Tokenization efficiency of current foundational large language models for the Ukrainian language

Foundational large language models (LLMs) are deployed in multilingual environments across a range of general and narrow task domains. These models generate text token by token, making them slower and more computationally expensive for low-resource languages that are underrepresented in the tokenize...

Full description

Saved in:
Bibliographic Details
Main Authors: Daniil Maksymenko, Oleksii Turuta
Format: Article
Language:English
Published: Frontiers Media S.A. 2025-08-01
Series:Frontiers in Artificial Intelligence
Subjects:
Online Access:https://www.frontiersin.org/articles/10.3389/frai.2025.1538165/full
Tags: Add Tag
No Tags, Be the first to tag this record!