Dense Connectivity
DenseNet takes connectivity to the extreme: each layer receives inputs from all preceding layers and feeds into all subsequent layers.
The Core Idea:
β’ If skip connections help, why not connect everything?
β’ Layer L receives feature maps from layers 0, 1, 2, ..., L-1
β’ Maximum information flow and feature reuse
β’ Dramatic parameter reduction while maintaining performance
πΈοΈ Dense Connectivity Pattern
Each layer concatenates feature maps from all previous layers as input, creating a densely connected network structure.
π± Growth Rate Concept
Each layer adds a fixed number of feature maps (growth rate k), leading to linear growth in channel dimensions.
Key Benefits
-
β‘ Parameter Efficiency
Fewer parameters than ResNet while achieving better performance
-
π Strong Gradient Flow
Direct connections to all previous layers eliminate vanishing gradients
-
β»οΈ Feature Reuse
All layers have access to all previous features, maximizing information usage
-
π‘ Implicit Deep Supervision
Each layer receives direct supervision signal through shorter paths
-
π¨ Feature Diversity
Encourages exploration of different feature representations