May 14th, 2024 | Marco Polignano, Pierpaolo Basile, Giovanni Semeraro
The paper introduces LLaMAntino-3-ANITA-8B-Inst-DPO-ITA, a state-of-the-art Large Language Model (LLM) specifically fine-tuned for the Italian language using the Meta LLaMA-3 model. The model is fine-tuned on English and Italian datasets using Supervised Fine-tuning (SFT) and Dynamic Preference Optimization (DPO) to improve performance and align with user preferences. The use of QLoRA (Quantized Low-Rank Adaptation) reduces the computational load and memory requirements, making the model more efficient. The model is evaluated on various benchmarks, showing superior performance compared to existing models, both in Italian and English. The paper also discusses the application of the model in retrieval-augmented generation, topic modeling, sentiment analysis, and recommender systems, highlighting its versatility and potential in practical scenarios. The authors conclude by emphasizing the importance of ethical considerations and the responsible use of AI technology in language-specific LLMs.The paper introduces LLaMAntino-3-ANITA-8B-Inst-DPO-ITA, a state-of-the-art Large Language Model (LLM) specifically fine-tuned for the Italian language using the Meta LLaMA-3 model. The model is fine-tuned on English and Italian datasets using Supervised Fine-tuning (SFT) and Dynamic Preference Optimization (DPO) to improve performance and align with user preferences. The use of QLoRA (Quantized Low-Rank Adaptation) reduces the computational load and memory requirements, making the model more efficient. The model is evaluated on various benchmarks, showing superior performance compared to existing models, both in Italian and English. The paper also discusses the application of the model in retrieval-augmented generation, topic modeling, sentiment analysis, and recommender systems, highlighting its versatility and potential in practical scenarios. The authors conclude by emphasizing the importance of ethical considerations and the responsible use of AI technology in language-specific LLMs.