Multimodal AI

LightOnOCR: A 1B End-to-End Multilingual Vision-Language Model for State-of-the-Art OCR

SSaid TaghadouiniAAdrien CavaillèsBBaptiste Aubertin
arXiv ID
2601.14251
Published
January 20, 2026
Authors
3
Hugging Face Likes
17
Comments
2

Abstract

We present LightOnOCR-2-1B, a 1B-parameter end-to-end multilingual vision--language model that converts document images (e.g., PDFs) into clean, naturally ordered text without brittle OCR pipelines. Trained on a large-scale, high-quality distillation mix with strong coverage of scans, French documents, and scientific PDFs, LightOnOCR-2 achieves state-of-the-art results on OlmOCR-Bench while being 9times smaller and substantially faster than prior best-performing models. We further extend the output format to predict normalized bounding boxes for embedded images, introducing localization during pretraining via a resume strategy and refining it with RLVR using IoU-based rewards. Finally, we improve robustness with checkpoint averaging and task-arithmetic merging. We release model checkpoints under Apache 2.0, and publicly release the dataset and LightOnOCR-bbox-bench evaluation under their respective licenses.

Keywords

vision-language modelOCR pipelinedistillation mixpretrainingcheckpoint averagingtask-arithmetic merginglocalizationresume strategyRLVRIoU-based rewards

More in Multimodal AI

View all
LightOnOCR: A 1B End-to-End Multilingual Vision-Language Model for State-of-the-Art OCR | Paperchime