4 Jan 2019 | Fisher Yu Dequan Wang Evan Shelhamer Trevor Darrell
The paper "Deep Layer Aggregation" by Fisher Yu, Dequan Wang, Evan Shelhamer, Trevor Darrell, and UC Berkeley explores the importance of aggregating layers and blocks in deep convolutional networks to improve visual recognition tasks. The authors argue that while deeper and wider networks can enhance accuracy, the effective aggregation of layers is crucial for better performance. They introduce two deep layer aggregation (DLA) structures: iterative deep aggregation (IDA) and hierarchical deep aggregation (HDA). IDA iteratively merges features from shallower to deeper layers, refining resolution and scale, while HDA uses a tree structure to hierarchically merge features across different depths and stages. These structures are designed to be compatible with various backbone networks and are evaluated across multiple tasks, including image classification, fine-grained recognition, semantic segmentation, and boundary detection. The results show that DLA improves performance, reduces parameter count, and enhances resolution compared to existing methods, achieving state-of-the-art results in several benchmarks.The paper "Deep Layer Aggregation" by Fisher Yu, Dequan Wang, Evan Shelhamer, Trevor Darrell, and UC Berkeley explores the importance of aggregating layers and blocks in deep convolutional networks to improve visual recognition tasks. The authors argue that while deeper and wider networks can enhance accuracy, the effective aggregation of layers is crucial for better performance. They introduce two deep layer aggregation (DLA) structures: iterative deep aggregation (IDA) and hierarchical deep aggregation (HDA). IDA iteratively merges features from shallower to deeper layers, refining resolution and scale, while HDA uses a tree structure to hierarchically merge features across different depths and stages. These structures are designed to be compatible with various backbone networks and are evaluated across multiple tasks, including image classification, fine-grained recognition, semantic segmentation, and boundary detection. The results show that DLA improves performance, reduces parameter count, and enhances resolution compared to existing methods, achieving state-of-the-art results in several benchmarks.