Avoid representational bottlenecks, especially early in the network.(簡單說就是feature map的大小要慢慢的減少。)網絡
Higher dimensional representations are easier to process locally within a network. Increasing the activations per tile in a convolutional network allows for more disentangled features. The resulting networks will train faster.(在網絡較深層應該利用更多的feature map,有利於容納更多的分解特徵。這樣能夠加速訓練)性能
Spatial aggregation can be done over lower dimensional embeddings without much or any loss in representational power.(也就是bottleneck layer的設計)lua
Balance the width and depth of the network.(Increasing both the width and the depth of the network can contribute to higher quality networks.同時增長網絡的深度和寬度)spa