23 Aug 2016 | Christian Szegedy, Sergey Ioffe, Vincent Vanhoucke, Alex Alemi
The paper "Inception-v4, Inception-ResNet and the Impact of Residual Connections on Learning" by Christian Szegedy, Sergey Ioffe, Vincent Vanhoucke, and Alex Alemi explores the integration of residual connections into the Inception architecture. The authors provide empirical evidence that training with residual connections significantly accelerates the training of Inception networks and can lead to slight improvements in performance compared to non-residual Inception networks. They also introduce new streamlined architectures, such as Inception-v4, which are more efficient and have a more uniform structure. These new architectures improve single-frame recognition performance on the ILSVRC 2012 classification task. The paper further demonstrates that proper activation scaling stabilizes the training of very wide residual Inception networks. Using an ensemble of three residual and one Inception-v4 models, the authors achieve a 3.08% top-5 error on the ImageNet classification challenge, setting a new state-of-the-art. The study also highlights the importance of scaling residuals to prevent training instabilities, especially when the number of filters exceeds 1000.The paper "Inception-v4, Inception-ResNet and the Impact of Residual Connections on Learning" by Christian Szegedy, Sergey Ioffe, Vincent Vanhoucke, and Alex Alemi explores the integration of residual connections into the Inception architecture. The authors provide empirical evidence that training with residual connections significantly accelerates the training of Inception networks and can lead to slight improvements in performance compared to non-residual Inception networks. They also introduce new streamlined architectures, such as Inception-v4, which are more efficient and have a more uniform structure. These new architectures improve single-frame recognition performance on the ILSVRC 2012 classification task. The paper further demonstrates that proper activation scaling stabilizes the training of very wide residual Inception networks. Using an ensemble of three residual and one Inception-v4 models, the authors achieve a 3.08% top-5 error on the ImageNet classification challenge, setting a new state-of-the-art. The study also highlights the importance of scaling residuals to prevent training instabilities, especially when the number of filters exceeds 1000.