27 May 2015 | Mingsheng Long, Yue Cao, Jianmin Wang, Michael I. Jordan
The paper "Learning Transferable Features with Deep Adaptation Networks" by Mingsheng Long introduces a novel architecture called Deep Adaptation Network (DAN) to enhance the transferability of features learned by deep neural networks for domain adaptation tasks. The authors address the challenge that deep features, while powerful, become increasingly specific and less transferable as they move through the network, especially with increasing domain discrepancy. To mitigate this issue, DAN embeds the hidden representations of task-specific layers into a reproducing kernel Hilbert space, where mean embeddings of different domain distributions can be explicitly matched. An optimal multi-kernel selection method is used to further reduce domain discrepancy, and an unbiased estimate of kernel mean embedding is implemented for scalable training. The proposed architecture is trained by fine-tuning a pre-trained AlexNet model on ImageNet and is evaluated on standard domain adaptation benchmarks, showing state-of-the-art performance. The contributions of the paper include a novel deep neural network architecture for domain adaptation, multi-layer adaptation, and multi-kernel adaptation, which together enhance the transferability of features and improve domain adaptation performance.The paper "Learning Transferable Features with Deep Adaptation Networks" by Mingsheng Long introduces a novel architecture called Deep Adaptation Network (DAN) to enhance the transferability of features learned by deep neural networks for domain adaptation tasks. The authors address the challenge that deep features, while powerful, become increasingly specific and less transferable as they move through the network, especially with increasing domain discrepancy. To mitigate this issue, DAN embeds the hidden representations of task-specific layers into a reproducing kernel Hilbert space, where mean embeddings of different domain distributions can be explicitly matched. An optimal multi-kernel selection method is used to further reduce domain discrepancy, and an unbiased estimate of kernel mean embedding is implemented for scalable training. The proposed architecture is trained by fine-tuning a pre-trained AlexNet model on ImageNet and is evaluated on standard domain adaptation benchmarks, showing state-of-the-art performance. The contributions of the paper include a novel deep neural network architecture for domain adaptation, multi-layer adaptation, and multi-kernel adaptation, which together enhance the transferability of features and improve domain adaptation performance.