{{PAPER_TITLE}}

Abstract

Metadata plays a critical role in indexing, documenting, and analyzing scientific literature, yet extracting it accurately and efficiently remains a challenging task. Traditional approaches often rely on rule-based or task-specific models, which struggle to generalize across domains and schema variations. In this paper, we present MeXtract, a family of lightweight language models designed for metadata extraction from scientific papers. The models, ranging from 0.5B to 3B parameters, are built by fine-tuning Qwen 2.5 counterparts. In their size family, MeXtract achieves state-of-the-art performance on metadata extraction on the MOLE benchmark. To further support evaluation, we extend the MOLE benchmark to incorporate model-specific metadata, providing an out-of-domain challenging subset. Our experiments show that fine-tuning on a given schema not only yields high accuracy but also transfers effectively to unseen schemas, demonstrating the robustness and adaptability of our approach. We release all the code, datasets, and models openly for the research community..

🚀 Introduction

MeXtract is a set of light-weight models for metadata extraction from scientific papers. The models were created by finetuning 0.5B, 1.5B, and 3B versions of Qwen2.5. The SFT stage was followed by a preference optimization stage. We evaluate the models on the MOLE+ benchmark, and we achieve state-of-the-art results with repsect similar models in the literature.

📄 Example

Here is an example of extracted metadata from a given sample paper. Note this example is smilplified, the actual paper is more complex and contain more pages.

📚 Models

We finetune 3 different size models for metadata extraction. You can find the models on Hugging Face 🤗.

📊 Results

Our models achieve state of the art results on the MOLE+ benchmark compared to models similar in size.

Model ar en jp fr ru multi model Average
Falcon3 3B Instruct20.4616.3020.2917.8117.2316.1315.9617.74
Llama3.2 3B Instruct28.7725.1733.1427.7322.2122.5833.3727.57
Gemma 3 4B It44.8846.5048.4643.8546.0642.0556.0446.83
Qwen2.5 3B Instruct49.9956.7261.1357.0864.1052.0759.0557.16
MOLE 3B23.0350.8850.8350.0557.7243.3417.1741.86
Nuextract 2.0 4B44.6143.5743.8248.9647.7840.1449.9045.54
Nuextract 2.0 8B51.9358.9362.1158.4163.2138.2153.7055.21
MeXtract 0.5B65.9669.9573.7968.4272.0768.2032.4164.40
MeXtract 1.5B67.0673.7175.0871.5776.2871.8752.0569.66
MeXtract 3B70.8178.0278.3272.8777.5174.9260.1873.23

📝 Citation

If you find this work useful, please cite it as follows:

@misc{mextract,
        title={MeXtract: Light-Weight Metadata Extraction from Scientific Papers}, 
        author={Zaid Alyafeai and Maged S. Al-Shaibani and Bernard Ghanem},
        year={2025},
        eprint={2510.06889},
        archivePrefix={arXiv},
        primaryClass={cs.CL},
        url={https://arxiv.org/abs/2510.06889}, 
}
                

📑 References

1. Alyafeai, Zaid, et al. "Masader: Metadata sourcing for arabic text and speech data resources." arXiv preprint arXiv:2110.06744 (2021).