Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Pixtral 12B

About

We introduce Pixtral-12B, a 12--billion-parameter multimodal language model. Pixtral-12B is trained to understand both natural images and documents, achieving leading performance on various multimodal benchmarks, surpassing a number of larger models. Unlike many open-source models, Pixtral is also a cutting-edge text model for its size, and does not compromise on natural language performance to excel in multimodal tasks. Pixtral uses a new vision encoder trained from scratch, which allows it to ingest images at their natural resolution and aspect ratio. This gives users flexibility on the number of tokens used to process an image. Pixtral is also able to process any number of images in its long context window of 128K tokens. Pixtral 12B substanially outperforms other open models of similar sizes (Llama-3.2 11B \& Qwen-2-VL 7B). It also outperforms much larger open models like Llama-3.2 90B while being 7x smaller. We further contribute an open-source benchmark, MM-MT-Bench, for evaluating vision-language models in practical scenarios, and provide detailed analysis and code for standardized evaluation protocols for multimodal LLMs. Pixtral-12B is released under Apache 2.0 license.

Pravesh Agrawal, Szymon Antoniak, Emma Bou Hanna, Baptiste Bout, Devendra Chaplot, Jessica Chudnovsky, Diogo Costa, Baudouin De Monicault, Saurabh Garg, Theophile Gervet, Soham Ghosh, Am\'elie H\'eliou, Paul Jacob, Albert Q. Jiang, Kartik Khandelwal, Timoth\'ee Lacroix, Guillaume Lample, Diego Las Casas, Thibaut Lavril, Teven Le Scao, Andy Lo, William Marshall, Louis Martin, Arthur Mensch, Pavankumar Muddireddy, Valera Nemychnikova, Marie Pellat, Patrick Von Platen, Nikhil Raghuraman, Baptiste Rozi\`ere, Alexandre Sablayrolles, Lucile Saulnier, Romain Sauvestre, Wendy Shang, Roman Soletskyi, Lawrence Stewart, Pierre Stock, Joachim Studnia, Sandeep Subramanian, Sagar Vaze, Thomas Wang, Sophia Yang• 2024

Related benchmarks

TaskDatasetResultRank
Visual Question AnsweringVQA 2.0 (test-dev)
Accuracy80.2
337
Relation and Type ExtractionBPMN
Precision40.2
116
BPMN Element Type ClassificationBPMN Diagrams Dataset
Precision74.1
116
Name and Type ExtractionBPMN dataset
Precision28.2
116
Visual Question AnsweringChartQA (test)
Accuracy81.8
58
Name ExtractionBPMN Diagrams (test)
Precision71.6
58
Visual Question AnsweringAI2D (test)
Accuracy79
54
Visual CountingPixMo-Count (test)
Score51.7
50
Spatial Logical ReasoningSpatiaLQA
Rc48.1
42
Human Preference RankingHuman Evaluation Elo (test)
Elo Score1.02e+3
34
Showing 10 of 22 rows

Other info

Follow for update