Granite Code Models: A Family of Open Foundation Models for Code Intelligence

Granite Code Models: A Family of Open Foundation Models for Code Intelligence

7 May 2024 | Mayank Mishra*, Matt Stallone*, Gaoyuan Zhang*, Yikang Shen, Aditya Prasad, Adriana Meza Soria, Michele Merler, Parameswaran Selvam, Saptha Surendran, Shivdeep Singh, Manish Sethi, Xuan-Hong Dang, Pengyuan Li, Kun-Lung Wu, Syed Zawad, Andrew Coleman, Matthew White, Mark Lewis, Raju Pavuluri, Yan Koyfman, Boris Lublinsky, Maximilien de Bayser, Ibrahim Abdelaziz, Kinjal Basu, Mayank Agarwal, Yi Zhou, Chris Johnson, Aanchal Goyal, Hima Patel, Yousaf Shah, Petros Zerfos, Heiko Ludwig, Asim Munawar, Maxwell Crouse, Pavan Kapanipathi, Shweta Salaria, Bob Calio, Sophia Wen, Seetharami Seelam, Brian Belgodere, Carlos Fonseca, Amith Singhee, Nirmit Desai, David D. Cox, Ruchir Puri†, Rameswar Panda†
The paper introduces the Granite series of decoder-only code models, designed to support enterprise software development across a wide range of coding tasks. The models, trained on 116 programming languages, range in size from 3 to 34 billion parameters. Evaluation on various benchmarks demonstrates that Granite Code models consistently achieve state-of-the-art performance among available open-source code LLMs. The models are optimized for enterprise software development workflows and perform well across tasks such as code generation, fixing, and explanation. The Granite Code models are released under an Apache 2.0 license for both research and commercial use. The paper details the data collection, model architecture, training methods, and evaluation results, highlighting the models' versatility and robustness in handling diverse coding tasks.The paper introduces the Granite series of decoder-only code models, designed to support enterprise software development across a wide range of coding tasks. The models, trained on 116 programming languages, range in size from 3 to 34 billion parameters. Evaluation on various benchmarks demonstrates that Granite Code models consistently achieve state-of-the-art performance among available open-source code LLMs. The models are optimized for enterprise software development workflows and perform well across tasks such as code generation, fixing, and explanation. The Granite Code models are released under an Apache 2.0 license for both research and commercial use. The paper details the data collection, model architecture, training methods, and evaluation results, highlighting the models' versatility and robustness in handling diverse coding tasks.
Reach us at info@study.space
Understanding Granite Code Models%3A A Family of Open Foundation Models for Code Intelligence