AMC: AutoML for Model Compression and Acceleration on Mobile Devices

AMC: AutoML for Model Compression and Acceleration on Mobile Devices

2018 | Yihui He, Ji Lin, Zhijian Liu, Hanrui Wang, Li-Jia Li, and Song Han
The MIT Open Access Articles paper titled "AMC: AutoML for Model Compression and Acceleration on Mobile Devices" introduces a novel approach to model compression and acceleration on mobile devices using reinforcement learning. The authors propose AutoML for Model Compression (AMC), which leverages reinforcement learning to automatically determine the optimal compression policy for each layer of a neural network. This learning-based approach outperforms traditional rule-based methods by achieving higher compression ratios, better accuracy preservation, and reducing the need for manual labor. AMC is designed to automatically find the compression policy for an arbitrary network, achieving better performance than human-designed, rule-based methods. The system processes a pre-trained network layer by layer, using a reinforcement learning agent (DDPG) to determine the sparsity ratio for each layer. The agent learns through trial and error, balancing the trade-off between accuracy loss and model shrinking. The system evaluates the accuracy of the pruned model without fine-tuning, which is an efficient approximation of the final accuracy. The paper presents two compression policy search protocols: resource-constrained compression and accuracy-guaranteed compression. Resource-constrained compression aims to achieve the best accuracy given the maximum hardware resources, while accuracy-guaranteed compression ensures the smallest model size with no loss of accuracy. The authors demonstrate the effectiveness of AMC on multiple neural networks, including VGG, ResNet, and MobileNet, achieving significant improvements in compression ratios and inference speed. The results show that AMC outperforms hand-crafted heuristic policies, achieving better performance on both classification and object detection tasks. The system is able to compress models with minimal loss of accuracy, and it is particularly effective on mobile devices, achieving significant speedups on Android phones and GPUs. The paper also highlights the generalization ability of AMC, showing that it can be applied to various tasks and models. In conclusion, AMC provides a powerful and efficient approach to model compression and acceleration on mobile devices, leveraging reinforcement learning to automatically determine the optimal compression policy. The system outperforms traditional methods in terms of compression ratio, accuracy preservation, and efficiency, making it a valuable tool for deploying deep neural networks on resource-constrained mobile devices.The MIT Open Access Articles paper titled "AMC: AutoML for Model Compression and Acceleration on Mobile Devices" introduces a novel approach to model compression and acceleration on mobile devices using reinforcement learning. The authors propose AutoML for Model Compression (AMC), which leverages reinforcement learning to automatically determine the optimal compression policy for each layer of a neural network. This learning-based approach outperforms traditional rule-based methods by achieving higher compression ratios, better accuracy preservation, and reducing the need for manual labor. AMC is designed to automatically find the compression policy for an arbitrary network, achieving better performance than human-designed, rule-based methods. The system processes a pre-trained network layer by layer, using a reinforcement learning agent (DDPG) to determine the sparsity ratio for each layer. The agent learns through trial and error, balancing the trade-off between accuracy loss and model shrinking. The system evaluates the accuracy of the pruned model without fine-tuning, which is an efficient approximation of the final accuracy. The paper presents two compression policy search protocols: resource-constrained compression and accuracy-guaranteed compression. Resource-constrained compression aims to achieve the best accuracy given the maximum hardware resources, while accuracy-guaranteed compression ensures the smallest model size with no loss of accuracy. The authors demonstrate the effectiveness of AMC on multiple neural networks, including VGG, ResNet, and MobileNet, achieving significant improvements in compression ratios and inference speed. The results show that AMC outperforms hand-crafted heuristic policies, achieving better performance on both classification and object detection tasks. The system is able to compress models with minimal loss of accuracy, and it is particularly effective on mobile devices, achieving significant speedups on Android phones and GPUs. The paper also highlights the generalization ability of AMC, showing that it can be applied to various tasks and models. In conclusion, AMC provides a powerful and efficient approach to model compression and acceleration on mobile devices, leveraging reinforcement learning to automatically determine the optimal compression policy. The system outperforms traditional methods in terms of compression ratio, accuracy preservation, and efficiency, making it a valuable tool for deploying deep neural networks on resource-constrained mobile devices.
Reach us at info@study.space