Papers
arxiv:2505.13772

Krikri: Advancing Open Large Language Models for Greek

Published on May 19
Authors:
,
,
,
,
,
,
,

Abstract

Llama-Krikri-8B, a multilingual LLM optimized for Greek, demonstrates significant improvements in understanding, generation, and code generation through extensive training and post-processing techniques.

AI-generated summary

We introduce Llama-Krikri-8B, a cutting-edge Large Language Model tailored for the Greek language, built on Meta's Llama 3.1-8B. Llama-Krikri-8B has been extensively trained on high-quality Greek data to ensure superior adaptation to linguistic nuances. With 8 billion parameters, it offers advanced capabilities while maintaining efficient computational performance. Llama-Krikri-8B supports both Modern Greek and English, and is also equipped to handle polytonic text and Ancient Greek. The chat version of Llama-Krikri-8B features a multi-stage post-training pipeline, utilizing both human and synthetic instruction and preference data, by applying techniques such as MAGPIE. In addition, for evaluation, we propose three novel public benchmarks for Greek. Our evaluation on existing as well as the proposed benchmarks shows notable improvements over comparable Greek and multilingual LLMs in both natural language understanding and generation as well as code generation.

Community

Paper author

You can try out Llama Krikri 8B Instruct here

Sign up or log in to comment

Models citing this paper 3

Datasets citing this paper 3

Spaces citing this paper 2

Collections including this paper 1