27 Jun 2024 | Ibrahim Abdelaziz, Kinjal Basu, Mayank Agarwal, Sadhana Kumaravel, Matthew Stallone, Rameswar Panda, Yara Rizk, GP Bhargav, Maxwell Crouse, Chulaka Gunasekara, Shajith Ikbal, Sachin Joshi, Hina Karanam, Vineet Kumar, Asim Munawar, Sumit Neelam, Dinesh Raghu, Udit Sharma, Adriana Meza Soria, Dheeraj Sreedhar, Praveen Venkateswaran, Merve Unuvar, David Cox, Salim Roukos, Luis Lastras, Pavan Kapanipathi
The paper introduces GRANITE-20B-FUNCTIONCALLING, an open-source function-calling model trained using multi-task learning on granular tasks such as nested function calling, function chaining, parallel functions, function name detection, parameter-value pair detection, next-best function, and response generation. The model is designed to enhance large language models (LLMs) by enabling them to interact with external tools and APIs to complete complex tasks. The evaluation of GRANITE-20B-FUNCTIONCALLING on multiple out-of-domain datasets, including the Berkeley Function Calling Leaderboard and academic benchmarks, shows that it performs well or better than other open and proprietary models, particularly in terms of generalizability and hallucination rates. The model's ability to handle granular tasks and its performance on various metrics make it a robust solution for function-calling applications.The paper introduces GRANITE-20B-FUNCTIONCALLING, an open-source function-calling model trained using multi-task learning on granular tasks such as nested function calling, function chaining, parallel functions, function name detection, parameter-value pair detection, next-best function, and response generation. The model is designed to enhance large language models (LLMs) by enabling them to interact with external tools and APIs to complete complex tasks. The evaluation of GRANITE-20B-FUNCTIONCALLING on multiple out-of-domain datasets, including the Berkeley Function Calling Leaderboard and academic benchmarks, shows that it performs well or better than other open and proprietary models, particularly in terms of generalizability and hallucination rates. The model's ability to handle granular tasks and its performance on various metrics make it a robust solution for function-calling applications.