420
v1v2v3v4v5 (latest)

OpenLLM-Ro -- Technical Report on Open-source Romanian LLMs

Abstract

In recent years, Large Language Models (LLMs) have achieved almost human-like performance on various tasks. While some LLMs have been trained on multilingual data, most of the training data is in English. Hence, their performance in English greatly exceeds their performance in other languages. This document presents our approach to training and evaluating the first foundational and chat LLM specialized for Romanian.

View on arXiv
Comments on this paper