Pooling is a downsampling operation that reduces the spatial dimensions of feature maps while retaining important information. It operates on non-overlapping regions.
Key Characteristics:
β’ Parameter-free: No learnable weights
β’ Translation invariant: Small shifts don't change output much
β’ Dimensionality reduction: Reduces spatial size
β’ Feature selection: Keeps most important activations
β
Reduces computational complexity
β
Provides translation invariance
β
Controls overfitting
β
Increases receptive field
π― Types of Pooling
π₯ Max Pooling
Takes the maximum value from each region. Preserves the strongest activations and provides translation invariance.
π Average Pooling
Computes the average of all values in each region. Provides smoother downsampling with less information loss.
π Global Pooling
Pools over the entire feature map, reducing it to a single value per channel. Used before final classification layers.
ποΈ Adaptive Pooling
Automatically adjusts pooling to produce a fixed output size regardless of input dimensions.
Interactive Pooling Demonstration
Input Feature Map (4Γ4)
β
Max Pooling (2Γ2, stride=2)
Click "Play" or "Next" to see pooling step by step