Interface administrators, Administrators (Semantic MediaWiki), Curators (Semantic MediaWiki), Editors (Semantic MediaWiki), Suppressors, Administrators
7,785
edits
No edit summary |
No edit summary |
||
Line 12: | Line 12: | ||
*[[Recurrent layer]]: Recurrent neural networks (RNNs) employ layers with a memory mechanism that enables them to process sequences of data and recognize temporal dependencies. | *[[Recurrent layer]]: Recurrent neural networks (RNNs) employ layers with a memory mechanism that enables them to process sequences of data and recognize temporal dependencies. | ||
*[[Dense layer]] ([[Fully connected layer]]): Layers in which every neuron is connected to every neuron in the previous layer. | *[[Dense layer]] ([[Fully connected layer]]): Layers in which every neuron is connected to every neuron in the previous layer. | ||
*[[Pooling layer]]: | *[[Pooling layer]]: reduce the dimensions of the feature maps. | ||
===Pooling Layers=== | ===Pooling Layers=== | ||
Pooling layers are used to reduce the spatial dimension of input data by aggregating activations within a local neighborhood. The most popular type of pooling is max pooling, in which only the maximum activation from each neighborhood is retained and all others discarded. Pooling layers often work in conjunction with convolutional layers when performing image recognition tasks. | Pooling layers are used to reduce the spatial dimension of input data by aggregating activations within a local neighborhood. The most popular type of pooling is max pooling, in which only the maximum activation from each neighborhood is retained and all others discarded. Pooling layers often work in conjunction with convolutional layers when performing image recognition tasks. | ||
==How Layers Work in Neural Networks== | ==How Layers Work in Neural Networks== |