FineVision:
Open Data Is All You Need

A new open dataset for data-centric training of Vision Language Models

Published

Sep 4, 2025

PDF

Table of Contents

huggingface.co/datasets/HuggingFaceM4/FineVision


Today, we release FineVision, a new multimodal dataset with 24 million samples. We created FineVision by collecting over 200 datasets containing 17M images, 89M question-answer turns, and 10B answer tokens, totaling 5TB of high-quality data. Additionally, we extensively processed all datasets to unify their format, clean them of duplicates and poor data, and rated all turns using 32B VLMs across 4 qualitative metrics with a score from 1-5 to enable the construction and study of individual training mixtures.

To enable everyone to construct state-of-the-art open Vision-Language Models (VLMs), we ran extensive ablations on FineVision, and compared it to publicly available alternatives. Models trained on FineVision lead in performance over 11 common benchmarks compared against every baseline, thanks to FineVision’s scale and diversity of data.

To use the dataset, simply load it with:

python
from datasets import load_dataset, get_dataset_config_names

# Get all subset names and load the first one
available_subsets = get_dataset_config_names('HuggingFaceM4/FineVision')
ds = load_dataset(
  'HuggingFaceM4/FineVision',
  name=available_subsets[0],
  split='train', streaming=True,
)

# Inspect the first sample
ds[0]

Why this dataset?

Even though open-weight Vision-Language Models are becoming ever more powerful, the accessibility of the training data used for these models is lagging behind. This data is often proprietary and inaccessible to the broader community. Projects like The Cauldron, LLaVa, and Cambrian aim to provide such datasets, but get quickly outpaced by the speed of the field and the emergence of novel applications for VLMs, like agentic tasks. For FineVision we set out to combine and unify existing available data sources to create a large and high-quality dataset. As a first step we need to collect and standardize the datasets.

How did we build FineVision?

FineVision was a giant act of data curation. We started by collecting publicly available datasets, and augmenting underrepresented categories. We then evaluated all datasets for duplicated data internally and benchmark contamination. This data is then cleaned and rated, before being added to the final mixture.

Data Collection

We manually collected over 200 image-text datasets from various publicly available sources and processed them to unify their formatting. On top of that, some datasets are not presented in chat form, so we converted them into question-answer pairs. In some cases, this goes as far as synthetically creating questions for all samples. Finally, we adressed underrepresented domains, such as GUI-oriented data. To fill this gap, we create and add a new dataset that was compiled from existing GUI datasets, after applying chat normalization and unifying the action space to convert their specific formats into a more general GUI action space.


FineVision Subsets - click to see more
Subset NameTotal ImagesTotal SamplesTotal TurnsTotal Question TokensTotal Answer TokensCategorySource
coco_colors118,287118,287118,2871,301,1576,376,672Captioning & Knowledge(1)
densefusion_1m1,058,7511,058,7511,058,75110,692,478263,718,217Captioning & Knowledge(2) (Converted)
face_emotion7977977978,7678,066Captioning & Knowledge(3)
google_landmarks299,993299,993842,1276,194,97810,202,980Captioning & Knowledge(4) (Converted)
image_textualization(filtered)99,57399,57399,573917,57719,374,090Captioning & Knowledge(5)
laion_gpt4v9,3019,3019,30193,9501,875,283Captioning & Knowledge(6)
localized_narratives199,998199,998199,9982,167,1798,021,473Captioning & Knowledge(7)
sharegpt4o57,28457,28457,284558,64736,555,323Captioning & Knowledge(8)
sharegpt4v(coco)50,01750,01750,017460,8939,825,387Captioning & Knowledge(9)
sharegpt4v(knowledge)1,9881,9881,98818,250293,850Captioning & Knowledge(9)
sharegpt4v(llava)29,98629,98629,986275,7836,175,899Captioning & Knowledge(9)
sharegpt4v(sam)8,9908,9908,99082,8741,668,797Captioning & Knowledge(9)
textcaps21,90621,90621,906240,966355,991Captioning & Knowledge(10)
chart2text26,96126,96130,215342,2152,670,580Chart & Table(11)
chartqa18,26518,26528,287625,569134,793Chart & Table(12)
CoSyn_400k_chart116,814116,8141,085,88217,617,59157,641,030Chart & Table(13)
CoSyn_400k_table46,51846,518416,5196,280,45523,335,054Chart & Table(13)
dvqa200,000200,0002,325,31644,603,3725,477,966Chart & Table(14)
figureqa100,000100,0001,327,36818,515,1532,654,736Chart & Table(15)
figureqa(mathv360k)17,58717,58717,587722,95997,404Chart & Table(16)
finqa5,2765,2766,2515,552,943224,015Chart & Table(17)
hitab2,5002,5007,782177,999335,013Chart & Table(18)
lrv_chart1,7761,7765,37276,477158,711Chart & Table(19)
mmc_instruct168,178168,178168,17850,008,82474,581,055Chart & Table(20)
multihiertt30,8757,6197,830218,840244,744Chart & Table(21)
plotqa157,070157,07020,249,479738,371,054118,122,387Chart & Table(22)
robut_sqa8,5148,51434,141368,9571,794,570Chart & Table(23)
robut_wikisql74,98974,98986,2021,454,9209,276,100Chart & Table(23)
robut_wtq38,24638,24644,096587,0406,415,830Chart & Table(23)
SynthChartNet500,000500,000500,0002,169,24067,392,223Chart & Table(24)
tabmwp22,72222,72223,021639,6391,883,243Chart & Table(25)
tabmwp(mathv360k)22,45222,45222,452963,498158,042Chart & Table(16)
tat_dqa2,4482,20713,251320,3561,177,852Chart & Table(26)
tat_qa2,1992,19913,215989,419254,790Chart & Table(27)
Unichart611,925611,9256,898,32496,702,288211,989,247Chart & Table(28) (Converted)
vistext9,9699,9699,96988,7701,191,127Chart & Table(29)
vqaonbd39,98639,9861,254,16536,066,8075,620,523Chart & Table(30)
alfworldgpt45,07345,07345,07317,864,0336,276,573General VQA(31)
allava_laion468,664468,664937,32818,654,303145,799,426General VQA(32)
allava_vflan177,078177,078387,87212,444,71155,305,642General VQA(32)
cambrian(filtered)_processed83,12383,12498,5341,410,3215,503,211General VQA(33)
chinesememe54,21254,21254,212538,93821,122,723General VQA(34)
cocoqa46,28746,28778,7361,136,238212,480General VQA(35)
CoSyn_400k_graphic26,96826,96826,9681,678,8628,235,679General VQA(13)
datik220,537222,385222,3852,234,054187,757,952General VQA(36)
datikz47,44147,97448,296441,04059,116,193General VQA(36)
drivelm90,0494,072161,0302,399,3621,431,417General VQA(37)
hateful_memes8,5008,5008,500128,37517,000General VQA(38)
iconqa27,30727,30729,841906,87772,492General VQA(39)
iconqa(mathv360k)22,58922,58922,589952,183134,029General VQA(16)
idk11,12311,12327,614235,262665,247General VQA(40)
indoor_qa3,3503,3503,35036,83219,700General VQA
LLaVA_Instruct_150K157,710157,710361,4054,412,60028,719,278General VQA(41)
llavar_gpt4_20k19,79019,79043,167546,7031,516,730General VQA(42)
lnqa302,780302,7801,520,94216,530,32319,027,663General VQA(43)
lrv_normal(filtered)10,48910,489155,2692,108,3213,134,247General VQA(44)
lvis_instruct4v222,711222,7111,050,62212,556,17343,726,782General VQA(45)
mimic_cgd141,87870,939141,8691,789,7404,304,380General VQA(46)
mmevol160,215160,215630,44116,203,12750,445,237General VQA(47)
mmra2,0481,0241,02472,52325,764General VQA(48)
nlvr2100,85250,42686,3734,629,641172,746General VQA(49)
sketchyvqa8,0008,0008,000182,1928,000General VQA(50)
spark3,9043,9046,24865,98273,973General VQA(51)
spatialsense10,44010,44017,498200,963418,883General VQA(52)
spot_the_diff17,1328,5669,52482,670209,630General VQA(53)
vision_flan(filtered)175,964175,964175,9649,983,7583,009,891General VQA(54)
visual7w14,36614,36669,8173,054,334209,451General VQA(55)
vizwiz(mathv360k)6,6046,6046,604197,14344,876General VQA(56)
vqav282,77282,772443,7575,722,4881,100,837General VQA(57)
vsr2,1572,1573,35479,5966,708General VQA(58)
websight10,00010,00010,000113,1145,237,381General VQA(59)
wildvision33333340550,16172,820General VQA(60)
yesbut4,3184,3184,31838,365157,229General VQA(61)
aguvis-stage-1458,957458,9573,831,66636,151,27293,546,182Grounding & Counting(62) (Converted)
groundui13,53113,53118,016200,094883,274Grounding & Counting(63)
objects365_qa1,742,2871,742,28712,329,259135,681,6802,146,619,635Grounding & Counting(64) (Converted)
oodvqa8,4888,4888,488227,0288,488Grounding & Counting(50)
tallyqa98,68098,680183,9862,674,306370,282Grounding & Counting(65)
clevr70,00070,000699,98919,277,8131,570,525Mathematics(66)
clevr_math70,00070,000556,0827,888,064580,324Mathematics(16)
clevr_math(mathv360k)5,2805,2805,280174,87927,536Mathematics(16)
CoSyn_400k_math66,71466,71466,714500,55428,631,388Mathematics(13)
geo170k(align)35,29735,29735,297336,1511,866,019Mathematics(67)
geo170k(qa)12,10112,10112,1011,254,8311,115,242Mathematics(67)
geo3k2,0912,0912,091130,2872,091Mathematics(68)
geometry3k(mathv360k)9,7249,7249,724541,90869,075Mathematics(16)
geomverse9,3039,3039,339662,7562,454,014Mathematics(69)
geoqa+(mathv360k)17,16217,16217,1621,449,094117,740Mathematics(70)
geos(mathv360k)49849849832,3943,509Mathematics(71)
intergps1,2801,2801,76097,7995,280Mathematics(72)
mavis_math_metagen87,34887,34887,3486,668,9205,486,485Mathematics(73)
mavis_math_rule_geo99,98699,98699,9868,211,07912,535,251Mathematics(73)
raven63,08142,00042,000584,84363,081Mathematics(74)
super_clevr(mathv360k)8,6428,6428,642307,43844,129Mathematics(16)
unigeo(mathv360k)11,94911,94911,9491,011,06981,781Mathematics(16)
art5,6035,6035,60356,573283,138Naive OCR(75)
captcha113,062113,062113,0621,469,548466,856Naive OCR
chrome_writting8,8258,8258,825150,025172,940Naive OCR(76)
cocotext16,16916,16916,169143,818177,111Naive OCR(77)
ctw24,29024,290180,6219,787,4851,653,254Naive OCR(78)
funsd1941943,87916,85629,996Naive OCR(79)
hme100k74,49274,49274,4921,117,3801,757,743Naive OCR(80)
hw_squad20,45720,45783,6821,071,534388,518Naive OCR(81)
iam5,6635,6635,66345,582130,794Naive OCR(82)
iiit5k1,9901,9901,99035,8204,259Naive OCR(83)
imgur5k5,9345,9345,93489,010288,054Naive OCR(84)
k12_printing256,636256,636256,63614,114,9807,465,001Naive OCR(19)
latex_handwritten39,58339,58339,583390,3431,874,733Naive OCR(85)
latexformulas552,340552,340552,3405,138,60343,094,747Naive OCR(86)
maptext2002007999,43470,813Naive OCR(87)
mathwriting-google300,000300,000300,0002,461,2705,954,806Naive OCR(88) (Converted)
memotion6,9916,9916,991194,718177,429Naive OCR(89)
orand_car_a1,9991,9991,99943,9789,035Naive OCR(90)
rendered_text10,00010,00010,00085,879244,183Naive OCR(91)
sroie33,61633,61633,616605,088243,240Naive OCR(92)
svrd4,3964,3964,39665,400834,514Naive OCR(93)
SynthCodeNet499,983499,983499,9832,000,683253,422,136Naive OCR(24)
synthdog500,000500,000500,0008,849,84848,010,145Naive OCR(94)
SynthFormulaNet499,997499,997499,9971,999,63151,215,097Naive OCR(24)
tal_ocr_eng256,646256,646256,6463,385,0127,465,207Naive OCR(95)
wordart19,0664,8044,80478,03254,263Naive OCR(96)
olmOCR-mix-0225-documents228,864228,864228,8582,197,147163,194,337Naive OCR(97) (Converted)
olmOCR-mix-0225-books15,19415,19415,194145,7507,962,779Naive OCR(97) (Converted)
a_okvqa54,60254,60254,6021,065,188360,990OCR QA(98)
aokvqa16,53916,53917,056743,458218,917OCR QA(98)
arxivqa100,000100,000100,0007,022,0016,422,269OCR QA(99)
bentham10,84310,84310,843103,042124,459OCR QA(81)
blockdiagramcomputerized5025025025,06734,453OCR QA(100)
blockdiagramhandwritten1,0291,0291,02911,44475,598OCR QA(100)
CoSyn_400k_diagram34,96334,963300,3573,356,84411,943,321OCR QA(13)
CoSyn_400k_document71,28271,282605,1736,216,51716,095,526OCR QA(13)
CoSyn_400k_music11,96911,96981,786792,1293,175,586OCR QA(13)
CoSyn_400k_nutrition6,9316,931112,0971,642,9363,687,254OCR QA(13)
diagram_image_to_text3003003003,63120,723OCR QA(101)
DoclingMatix2,465,2021,270,91110,626,898162,581,6602,996,338,775OCR QA(24)
docvqa10,18910,18939,463724,814275,510OCR QA(102)
est_vqa19,35819,35819,358286,343143,270OCR QA(103)
handwriting_forms1,4001,4001,40081,20041,490OCR QA(104)
infographic_vqa1,9824,39423,717392,45686,951OCR QA(105)
infographic_vqa_llava_format4,3942,11310,054174,35243,912OCR QA(105)
infographic(gpt4v)2,1131,9821,982275,4981,044,183OCR QA(105)
invoices_receipts3,0133,0133,01336,745771,948OCR QA(106)
mapqa37,41737,417483,4168,454,7225,657,339OCR QA(107)
mapqa(mathv360k)5,2255,2255,225168,39044,560OCR QA(16)
mmsoc_memotion6,9916,9916,991188,505421,250OCR QA(108)
ocrvqa165,746165,746801,57912,217,5644,801,833OCR QA(109)
pdfvqa8,5938,59395,0001,272,618939,948OCR QA(110)
screen2words15,73015,73015,743133,014120,781OCR QA(111)
screenqa80,76180,76180,761940,729826,795OCR QA(112)
slidevqa11,8681,91910,617333,065156,036OCR QA(113)
st_vqa17,24717,24723,121338,83798,892OCR QA(114)
sujet_finance9,8019,801107,0501,395,6241,925,361OCR QA(115)
textocr(gpt4v)25,06025,06025,060150,3602,436,974OCR QA(116)
textvqa21,95321,95334,602553,990141,882OCR QA(117)
ureader_cap91,21591,21591,2151,086,4841,435,964OCR QA(118)
ureader_ie17,32017,32017,320406,237128,229OCR QA(118)
ureader_kg_processed37,55037,55037,550352,9072,013,731OCR QA(118)
ureader_qa_processed252,953252,953252,9537,100,750930,617OCR QA(118)
visualmrc3,0273,02711,988139,751147,385OCR QA(119)
ai2d_merged4,8584,85812,325755,4551,319,140Science(120)
CoSyn_400k_chemical8,9428,94255,391634,8812,450,290Science(13)
CoSyn_400k_circuit10,47010,47067,939713,5752,637,618Science(13)
pathvqa32,63232,63232,632639,38585,168Science(121)
pmc_vqa(mathv360k)35,94835,94835,9481,889,167255,109Science(16)
scienceqa4,9764,9766,1491,081,22018,447Science(122)
scienceqa(nona_context)19,20819,20819,2081,624,58325,311Science(19)
tqa2,7492,74912,567395,956149,776Science(123)
visualwebinstruct(filtered)263,581263,581263,5818,341,54031,802,459Science(124)
vqarad3133131,79325,1816,003Science(125)
text_code_feedback066,383221,09619,349,05679,752,351Text-only(126)
text_codefeedback_filtered_instruction0156,525156,52527,684,17062,764,414Text-only(126)
text_infinitymath0101,380101,3809,158,132212,543Text-only(127)
text_mathinstruct0262,039262,03920,405,29544,145,362Text-only(128)
text_mathqa0394,996394,99623,552,03572,451,061Text-only(129)
text_mathstepdpo10k010,79510,795557,233989,312Text-only(130)
text_numinamath_cot0859,494859,49475,818,870387,758,581Text-only(131)
text_openhermes_2_501,001,5511,008,268142,376,960233,561,291Text-only(132)
text_openorca04,233,8534,233,8531,049,478,873468,042,176Text-only(133)
text_orcamath0200,035200,03512,691,01461,860,987Text-only(134)
text_pythoncode25k049,62649,6261,629,2864,945,892Text-only(135)
text_pythoncodealpaca018,61218,612655,1272,683,469Text-only(136)
text_ruozhiba01,4961,49669,795234,822Text-only(137)
text_theoremqa080080050,0653,468Text-only(138)
text_wizardlm_evol069,99969,9997,753,96321,955,856Text-only(139)
text_OpenMathInstruct-201,000,0001,000,00074,905,850413,132,418Text-only(140)
Totals17,372,29324,322,19388,928,3433,168,958,4179,459,677,828

Cleaning

After gathering all the sub-datasets, every turn is cleaned. We removed all individual turns whose combined question and answer length exceeds 8192 tokens. We resize big images to have a longest side of 2048 pixels while keeping the aspect ratio, and discard samples with corrupted images.

Rating

Finally, we rate every single turn in our dataset across 4 axes. For this, we used a LLM and VLM-as-a-judge pipeline (using Qwen3-32B and Qwen2.5VL-32B-Instruct), to rate every turn on a scale from 1-5 in these 4 categories:

This is the distribution of scores across the different filters for FineVision:

Filter12345
Formatting0.50.71.177.520.3
Relevance2.90.514.716.565.4
Visual Dependency11.020.42.624.241.8
Image Correspondence8.13.617.326.844.1

FineVision Base Dataset

We classify FineVision’s subsets into 9 categories: Captioning & Knowledge, Chart & Table, General VQA, Grounding & Counting, Mathematics, Naive OCR, OCR QA, Science and Text-only (Fig. 1).

There are multiple ways to count the data in a multimodal dataset. The most common are the number of samples and the number of images. Additionally, a single sample can consist of multiple question/answer pairs in the form of a multi-turn conversation. Similarly to text-only datasets, the number of answer tokens is also interesting, since these are the tokens the model is actually trained on. We count all these characteristics for FineVision and arrive at 17.3M images, 24.3M samples, 88.9M turns, and 9.5B answer tokens. Based on these 4 distributions, multiple different mixtures are possible. In conjunction with the provided ratings, we encourage the community to create their own mixtures and experiment with the data. For example, large categories could be downsamples, while high-quality data could be upsampled. After collecting and processing the data, we run multiple experiments and ablations to provide practical recommendations on how to train small, data-centric VLMs.


Figure 1: Distribution of Categories in FineVision by Answer Tokens, Number of Samples, Turns, and Images. While the distributions differ a bit with the different metrics, FineVision provides a good baseline mixture, especially when judging by the number of images in the individual categories. Samples from Chart & Table usually lend themselves well to multi-turn conversations, since multiple similar questions can be asked for a single Chart. Samples from OCR QA often have a lot of answer tokens, since they aim at detailed document understanding, which are rarely answered with a short sentence.

Experimental Setup

To ensure a fair comparison between different configurations, we use the same setup and evaluations for all of our ablations. This enables us to compare FineVision to other publicly available datasets as well as experiment with different intra-dataset configurations.

Model Architecture: nanoVLM

For all ablations and experiments, we train a 460M parameter VLM, since it provides a good trade-off between training time and model performance. We utilize the lightweight nanoVLM training framework with SmolLM2-360M-Instruct as the text backbone, and SigLIP2-Base-512 as the vision encoder. We experimented with a classic 2-stage training schedule where the first stage is used to train mainly the Modality Projection to align the Language and Image Embeddings, and the second stage is used to train the whole model. Interestingly, we did not observe any significant benefits from this additional first stage compared to training the whole model directly at our size and training duration, so we settled on a single-stage training for most ablations.

Baseline Datasets

We use three similar open source alternatives as baselines to compare our dataset to: The Cauldron, LLaVA-OneVision and Cambrian-7M.

NameImagesSamplesTurnsAnswer Tokens
Cauldron2.0M1.8M27.8M0.3B
LLaVa-Vision2.5M3.9M9.1M1.0B
Cambrian-7M5.4M7M12.2M0.8B
FineVision17.3M24.3M88.9M9.5B

Evaluations

We utilize lmms-eval during training to evaluate our ablations in a reproducible manner. We evaluate on a diverse set of 11 benchmarks: AI2D, ChartQA, DocVQA, InfoVQA, MME, MMMU, MMStar, OCRBench, ScienceQA, TextVQA and SEED-Bench. Since these benchmarks cover different topics and produce results on different scales, e.g. AI2D returns the accuracy of the exact matches (0-1), but MME returns a continuous score (0-2800), aggregating them is not trivial. In our ablations the relative performance between the different configurations matters, so to provide a robuts summary metric we determine the rank of each model compared to the others in every benchmark at every training step and average it over all the benchmarks. This way, we can judge where different configurations rank among each other over the course of training. To keep a sense of how big the absolute difference between models is, we also provide an average over all metrics and incorporate MME by normalizing it between 0 and 1.

Training Configuration

Each of our ablations trains the 460M model with a maximal image size of 1536x1536 pixel (without resizing smaller images) and a maximal input token length of 4096. This results in a maximum batch size of 2 for a single H100, which we adapt with 8 steps of gradient accumulation on each of the 32 GPUs for an effective batch size of 512. In all single stage configurations we train for 20k Steps on 32 H100s for approximately 20h while evaluating all 11 benchmarks every 1k Steps. If not specified otherwise, the “Baseline” in our intra dataset ablations refers to a training run on the full unfiltered and unchanged dataset. In this configuration, a full epoch of the unfiltered FineVision dataset takes 12k steps.

Experiments

While many interesting questions could be investigated, we mainly focus on the aspects of the training that are influenced by the data. Before we dive into the internal details of FineVision, let’s have a look at our performance against the baselines.

How does FineVision compare to other open datasets?

Here we see the first interesting trend: VLMs still benefit from training on a larger, more diverse dataset than what was available until today. FineVision doesn’t lead the race in the first few thousand training steps, after all, it does include new tasks such as pointing and agentic browsing, so it shouldn’t be better at first. But after seeing enough varied data, FineVision clearly shows the best performance across a wide set of benchmarks, which can be seen in its average ranking (Fig. 2). One epoch of FineVision in our setup takes 12k training steps, so we train for close to 2 epochs in these ablations. Looking at the average benchmark, we can see how the models saturate around different points: 18k steps for Cambrian, 12k for LLaVa and 7k for the Cauldron. In particular, over 11 different benchmarks, FineVision achieves an average improvement of 40.7% over the Cauldron, 12.1% over Cambrian, and 46.3% over LLaVa, which increases to 51.3%, 18.6% and 58.0% when comparing the deduplicated versions of the datasets. Additionally, FineVision includes data for tasks such as agentic browsing, and counting and pointing, which are not part of the other baselines.

Figure 2: Average Rank of Models trained on different open source datasets. FineVision shows both the highest average rank as well as the highest average over benchmarks.

How much test data is in publicly available datasets?

We investigate data leakage by finding images from test sets that appear in the dataset. For this, we constructed an image deduplication pipeline. We used this pipeline to compare all images in FineVision to all images of 66 image-text benchmarks from the lmms-eval framework.

For the comparison, we embed the images using the SSCD descriptor, and compute the cosine similarity between a given image in FineVision and all images from the test-set embeddings. Whenever a sample has a similarity higher than a threshold of 0.95 it is assumed to be a duplicate.

While our tests with various thresholds show that this is still flagging more false-positives than false-negatives, given the scale of data we have, we preferred to err on the side of caution.

Below is an example of a correctly identified Duplicate (“Photo”), a false-positive with a similarity score above 0.95 (“Chart”) and a false-negative with a similarity score below 0.95 (“Drawing”) (Fig. 3).

We open-source the deduplication pipeline here as well as the precomputed test-set embedding’s here.



Figure 3: Examples of the Deduplication Results.

We repeated this deduplication procedure on all the baselines to analyse how contaminated they are. We found that all baselines contain between 2-3% images from test benchmarks, and removing them results in a performance drop of 2.4-2.8%. Interestingly, we find that for some benchmarks the difference is negligible, while other benchmarks suffer significantly. For example, after deduplicating, ScienceQA falls by 14.49% on average while OCRBench only drops by 1.08%. This deduplications also shows that FineVision contains the smallest relative amount of duplicated data at 1%, and also suffers the smallest performance drop over all benchmarks after deduplication at just 1.45%.

NameSamplesContamination RatePerformance Drop
Cauldron1.8M3.05%2.39%
Llava-Vision3.9M2.15%2.72%
Cambrian-7M7.0M2.29%2.78%
FineVision24.3M1.02%1.45%

Additionally, we experimented with removing all found samples from all datasets to see if the outcome is different from Fig. 2, but we observe the same distribution (Fig. 4).


Figure 4: Average Rank of Models trained on different deduplicated open source datasets. Even after deduplicating all dataset, FineVision shows the best performance.

How diverse are the datasets?

Similarly to the size comparison, we also wanted to evaluate the datasets for diversity. Evaluating the diversity of a dataset is a field of study for itself, which we will not dive into here, rather we borrow techniques from computer vision and use the already computed SSCD embeddings as a proxy of visual diversity. To not rely on a subsample of the dataset in estimating the diversity, we analyse the covariance metric of the full embeddings. From this covariance matrix, we can calculate the eigenvalues for analysis. We get the effective rank of the covariance matrix, which measures how uniformly the variance is distributed across dimensions, as well as the participation ratio, which measures how many dimensions actively contribute to the overall variance. To obtain a single diversity score for the datasets, we normalize the effective rank and participation ratio with the embedding dimension and compute their geometric mean. We observe that FineVision is not only the biggest, but also the most diverse dataset. Additionally, you can also clearly see that more images do not necessarily result in more diversity, since LLaVa is substantially less diverse than the Cauldron, even with more images.

NameImagesEffective RankParticipation RatioDiversity
Cauldron2.0M324.05129.220.400
LLaVa-Vision2.5M267.8987.050.298
Cambrian-7M5.4M359.73152.700.458
FineVision17.3M359.22182.520.500

Should you merge multiple questions for the same image into a single multi turn conversation?

Since the training of a VLM already builds upon pretrained vision and language backbones, datasets are usually not completely unstructured, but follow an image+question and answer structure. Some works have shown that consolidating multiple questions for the same image into a multi-turn conversation where the image is shown only once improves model performance, reduces training budget, and reduces the datasets’ memory footprint. We therefore experiment with deduplicating every image in our dataset internally using the same SSCD descriptors, manually inspect the resulting clusters, and merge fitting samples into a multi-turn conversation.

When training with the same training budget, we find that both models perform very similarly (Fig. 5). Some benchmarks favor one image/several turns, while others favor one image/one turn. Given this, we decide to release the dataset without merging multiple questions for the same image, and open-source the pipeline in case users want to explore this further.


Figure 5: Average Ranking of Models trained with internally deduplicated / merged samples. No clear benefit in merging can be seen with respect to model performance.

Should you train on multilingual data if your language backbone was not?

There are some multilingual datasets in our mixture, but since our Language Backbone is only trained on English data, we experimented with removing all the multilingual, mainly Chinese, subsets. Our results show that there is a slight advantage in leaving the multilingual data, even if it was not part of the Language Backbone’s initial training. We believe this reinforces our hypothesis that more diversity in the dataset is generally preferable for VLM training. In our training setup with this configuration, one epoch over the whole non-deduplicated dataset equals ~12k steps, so the benefit of unseen languages only materializes after the first full epoch (Fig. 5).


Figure 6: Average Rank of Models trained with and without multilingual samples. Keeping samples in unseen langauges improves performance after the first epoch.

How can you assess the quality of the dataset?

The usual goal for every dataset, to collect samples with the highest quality possible, is quite an abstract endeavour in practice, especially for multimodal datasets. Additionally, different training stages usually have different qualitative and quantitative requirements. Finally, tuning the mixtures of different categories is also reliant on how much data with what quality is available. For image-text datasets, there are 3 different combinatorial ways to evaluate a sample: text-only, image-only, and image-text correspondence. The question persists, how do you actually measure the quality of a sample, especially if you have to do so in 3 different ways. We propose doing so by leveraging both a LLM and a VLM as a judge.

To try to quantify the quality of the training data and the effect it has on the model’s performance, we run extensive ablations on our generated ratings.


Figure 6: Average Rank of Models trained with samples that have all 4 ratings above a certain threshold. Keeping all samples results in the best performance.

Interestingly, both when only training on turns that have any of the 4 ratings under a certain threshold, as well as when training on turns where only a single rating at a time is used, we observe the same behaviour. Simply training on the most diverse data, that one containing all samples, outperforms in benchmarks (Fig. 6) (Fig. 7). This could mean multiple things. Firstly, we can see almost the same distribution in the ranks across all filters: from best to worst with an increase in the rating threshold. For example the visual dependency and the image correspondence rating both result in exactly the same distribution of rankings, corresponding to the natural order of options, 1 through 5. This could indicate that with a sufficiently large dataset that you train on long enough, it hurts more to remove samples, even if they were judged to be of low quality, than to train on them.

Additionally, the notion of quality in VLM datasets is inherently nuanced. Unlike LLMs, where pre-training often relies on massive web crawls, training a VLM is closer to the supervised fine-tuning (SFT) stage. We do not train on crawls of internet data, instead we train on individual samples of Image-Question and Answer pairs, and these datapoints are usually ‘curated rather than collected’. We also do not train on trillions of tokens, but on billions. This built-in curation provides a baseline level of quality from the start. FineVision follows this pattern: it brings together widely used VLM datasets along with a few new ones in low-resource domains. We could therefore be trying to measure and quantify noisy nuances in the quality of Image-Question-Answer Pairs, instead of using the fact that they are already curated SFT datasets as the measure for quality.

Alternatively, while we used state-of-the-art open source models to judge our datapoints, we still had to find a compromise between model quality and cost due to the raw required effort to rate every single turn of FineVision. The chosen models could simply not be powerful enough to recognize and judge the quality of samples. Even though our first proposal to judge the quality of multimodal data on a per-turn basis did not yield any improvement in model performance, we believe that this is still an exciting and important direction of research and hope the release of FineVision encourages the community to develop techniques for this at large scale.


Model Performance After Applying Individual Filters
Figure 7: Comparison across thresholds for all four filters individually: Formatting, Relevance, Visual Dependency, and Image-Question Correspondence. Keeping all samples results in the best average performance.

Should you train in multiple stages?

The standard training procedure of a VLM usually follows at least two stages. First, you train only the connecting module, potentially in addition the image encoder, and then you train the whole model in a second stage. Some work has even introduced an additional Stage 2.5 (141), where you train the full model on a smaller subset of higher quality data. To investigate this on small models, we experiment both with single, dual and triple stage training.


1 Stage vs 2 Stages

To evaluate if pre-training the Modality Projection and the Vision Encoder provides any benefits to the final model performance, we conduct this experiment at a higher image resolution of 2048px and train substantially longer. We can see that even for training longer, the general difference in model performance is quite small. Individual benchmarks, do show differences (ScienceQA drops by 5% but OCRBench improves by 5% in the two-stage setup) (Fig. 8), so the better setup is individual to the desired model capabilities. This also shows that evaluation (and through this also correctly training) a VLM is not straightforward tasks, since availible benchmarks are limited proxies for the underlying model performance.

Figure 8: Average Rank of a model trained for 60K steps in a single stage, and a model trained for the same 60k steps on top of pretraining the Modality Projection and Vision Encoder for 15k steps. The pre-training procedure is not depicted in this graph.

2 Stages vs 2.5 Stages

We also experiment if splitting the second stage results in any performance improvements.

We take the baseline, and continue training for another 20k steps, both with the unfiltered (>= 1) as well as filtered subsets of FineVision according to our ratings.

Figure 9: Average Rank if a model trained for an additional 20K steps on top of unfiltered training for 20K steps. Subselecting data for the final training steps does not yield a performance improvement with our quality measure. Only the 20k steps for the final stage are depicted here, the first 20k steps are the same for all variations.

As in the previous results, we observe that the best outcome is simply achieved by training on as much and as diverse data as possible (Fig. 9). Like before, this could also be due to the way we filter the data, and a different quality measure might yield different results.

Conclusion

We introduce FineVision, a new state of the art open dataset to train VLMs, that is both bigger and more diverse than previous open source datasets. We provide extensive analysis regarding size, diversity, contamination and data-centric model training, and hope we can empower both further research and the community with this.

  1. 1.
    hazal-karakus/mscoco-controlnet-canny-less-colors · Datasets at Hugging Face [Internet]. [cited 2025 Aug 28]. Available from: https://huggingface.co/datasets/hazal-karakus/mscoco-controlnet-canny-less-colors/viewer
  2. 2.
    Li X, Zhang F, Diao H, Wang Y, Wang X, Duan L. Densefusion-1m: Merging vision experts for comprehensive multimodal perception. Advances in Neural Information Processing Systems [Internet]. 2024 [cited 2025 Aug 28];37:18535–56. Available from: https://proceedings.neurips.cc/paper_files/paper/2024/hash/20ffc2b42c7de4a1960cfdadf305bbe2-Abstract-Datasets_and_Benchmarks_Track.html
  3. 3.
    Mollahosseini A, Hasani B, Mahoor MH. Affectnet: A database for facial expression, valence, and arousal computing in the wild. IEEE Transactions on Affective Computing [Internet]. 2017 [cited 2025 Sep 2];10(1):18–31. Available from: https://ieeexplore.ieee.org/abstract/document/8013713/
  4. 4.
    Weyand T, Araujo A, Cao B, Sim J. Google Landmarks Dataset v2 – A Large-Scale Benchmark for Instance-Level Recognition and Retrieval [Internet]. 2020. Available from: https://arxiv.org/abs/2004.01804
  5. 5.
    Pi R, Zhang J, Zhang J, Pan R, Chen Z, Zhang T. Image Textualization: An Automatic Framework for Creating Accurate and Detailed Image Descriptions [Internet]. arXiv; 2024 [cited 2025 Aug 28]. Available from: http://arxiv.org/abs/2406.07502
  6. 6.
    laion/gpt4v-dataset · Datasets at Hugging Face [Internet]. 2023 [cited 2025 Sep 2]. Available from: https://huggingface.co/datasets/laion/gpt4v-dataset
  7. 7.
    Pont-Tuset J, Uijlings J, Changpinyo S, Soricut R, Ferrari V. Connecting Vision and Language with Localized Narratives. In: Vedaldi A, Bischof H, Brox T, Frahm J-M, editors. Computer Vision – ECCV 2020 [Internet]. Cham: Springer International Publishing; 2020 [cited 2025 Aug 28]. p. 647–64. Available from: https://link.springer.com/10.1007/978-3-030-58558-7_38
  8. 8.
    ShareGPT-4o [Internet]. [cited 2025 Aug 28]. Available from: https://sharegpt4o.github.io/
  9. 9.
    Chen L, Li J, Dong X, Zhang P, He C, Wang J, et al. ShareGPT4V: Improving Large Multi-modal Models with Better Captions. In: Leonardis A, Ricci E, Roth S, Russakovsky O, Sattler T, Varol G, editors. Computer Vision – ECCV 2024 [Internet]. Cham: Springer Nature Switzerland; 2025 [cited 2025 Sep 2]. p. 370–87. Available from: https://link.springer.com/10.1007/978-3-031-72643-9_22
    back: 1, 2, 3, 4
  10. 10.
    Sidorov O, Hu R, Rohrbach M, Singh A. TextCaps: A Dataset for Image Captioning with Reading Comprehension. In: Vedaldi A, Bischof H, Brox T, Frahm J-M, editors. Computer Vision – ECCV 2020 [Internet]. Cham: Springer International Publishing; 2020 [cited 2025 Aug 28]. p. 742–58. Available from: https://link.springer.com/10.1007/978-3-030-58536-5_44
  11. 11.
    Kantharaj S, Leong RTK, Lin X, Masry A, Thakkar M, Hoque E, et al. Chart-to-Text: A Large-Scale Benchmark for Chart Summarization [Internet]. arXiv; 2022 [cited 2025 Aug 28]. Available from: http://arxiv.org/abs/2203.06486
  12. 12.
    Masry A, Long DX, Tan JQ, Joty S, Hoque E. ChartQA: A Benchmark for Question Answering about Charts with Visual and Logical Reasoning [Internet]. arXiv; 2022 [cited 2025 Aug 28]. Available from: http://arxiv.org/abs/2203.10244
  13. 13.
    Yang Y, Patel A, Deitke M, Gupta T, Weihs L, Head A, et al. Scaling Text-Rich Image Understanding via Code-Guided Synthetic Multimodal Data Generation [Internet]. arXiv; 2025 [cited 2025 Aug 28]. Available from: http://arxiv.org/abs/2502.14846
    back: 1, 2, 3, 4, 5, 6, 7, 8, 9, 10
  14. 14.
    Kafle K, Price B, Cohen S, Kanan C. Dvqa: Understanding data visualizations via question answering. In: Proceedings of the IEEE conference on computer vision and pattern recognition [Internet]. 2018 [cited 2025 Aug 28]. p. 5648–56. Available from: http://openaccess.thecvf.com/content_cvpr_2018/html/Kafle_DVQA_Understanding_Data_CVPR_2018_paper.html
  15. 15.
    Kahou SE, Michalski V, Atkinson A, Kadar A, Trischler A, Bengio Y. FigureQA: An Annotated Figure Dataset for Visual Reasoning [Internet]. arXiv; 2018 [cited 2025 Aug 28]. Available from: http://arxiv.org/abs/1710.07300
  16. 16.
    Shi W, Hu Z, Bin Y, Liu J, Yang Y, Ng S-K, et al. Math-LLaVA: Bootstrapping Mathematical Reasoning for Multimodal Large Language Models [Internet]. arXiv; 2024 [cited 2025 Sep 2]. Available from: http://arxiv.org/abs/2406.17294
    back: 1, 2, 3, 4, 5, 6, 7, 8, 9, 10
  17. 17.
    Chen Z, Chen W, Smiley C, Shah S, Borova I, Langdon D, et al. FinQA: A Dataset of Numerical Reasoning over Financial Data [Internet]. arXiv; 2022 [cited 2025 Aug 28]. Available from: http://arxiv.org/abs/2109.00122
  18. 18.
    Cheng Z, Dong H, Wang Z, Jia R, Guo J, Gao Y, et al. HiTab: A Hierarchical Table Dataset for Question Answering and Natural Language Generation [Internet]. arXiv; 2022 [cited 2025 Aug 28]. Available from: http://arxiv.org/abs/2108.06712
  19. 19.
    lmms-lab/LLaVA-OneVision-Data · Datasets at Hugging Face [Internet]. 2025 [cited 2025 Sep 2]. Available from: https://huggingface.co/datasets/lmms-lab/LLaVA-OneVision-Data
    back: 1, 2, 3
  20. 20.
    Liu F, Wang X, Yao W, Chen J, Song K, Cho S, et al. MMC: Advancing Multimodal Chart Understanding with Large-scale Instruction Tuning [Internet]. arXiv; 2024 [cited 2025 Aug 28]. Available from: http://arxiv.org/abs/2311.10774
  21. 21.
    Zhao Y, Li Y, Li C, Zhang R. MultiHiertt: Numerical Reasoning over Multi Hierarchical Tabular and Textual Data [Internet]. arXiv; 2022 [cited 2025 Aug 28]. Available from: http://arxiv.org/abs/2206.01347
  22. 22.
    Methani N, Ganguly P, Khapra MM, Kumar P. Plotqa: Reasoning over scientific plots. In: Proceedings of the ieee/cvf winter conference on applications of computer vision [Internet]. 2020 [cited 2025 Aug 28]. p. 1527–36. Available from: http://openaccess.thecvf.com/content_WACV_2020/html/Methani_PlotQA_Reasoning_over_Scientific_Plots_WACV_2020_paper.html
  23. 23.
    Zhao Y, Zhao C, Nan L, Qi Z, Zhang W, Tang X, et al. RobuT: A Systematic Study of Table QA Robustness Against Human-Annotated Adversarial Perturbations [Internet]. arXiv; 2023 [cited 2025 Aug 28]. Available from: http://arxiv.org/abs/2306.14321
    back: 1, 2, 3
  24. 24.
    Nassar A, Marafioti A, Omenetti M, Lysak M, Livathinos N, Auer C, et al. SmolDocling: An ultra-compact vision-language model for end-to-end multi-modal document conversion [Internet]. arXiv; 2025 [cited 2025 Sep 2]. Available from: http://arxiv.org/abs/2503.11576
    back: 1, 2, 3, 4
  25. 25.
    Lu P, Qiu L, Chang K-W, Wu YN, Zhu S-C, Rajpurohit T, et al. Dynamic Prompt Learning via Policy Gradient for Semi-structured Mathematical Reasoning [Internet]. arXiv; 2023 [cited 2025 Aug 28]. Available from: http://arxiv.org/abs/2209.14610
  26. 26.
    Zhu F, Lei W, Feng F, Wang C, Zhang H, Chua T-S. Towards Complex Document Understanding By Discrete Reasoning. In: Proceedings of the 30th ACM International Conference on Multimedia [Internet]. Lisboa Portugal: ACM; 2022 [cited 2025 Aug 28]. p. 4857–66. Available from: https://dl.acm.org/doi/10.1145/3503161.3548422
  27. 27.
    Zhu F, Lei W, Huang Y, Wang C, Zhang S, Lv J, et al. TAT-QA: A Question Answering Benchmark on a Hybrid of Tabular and Textual Content in Finance [Internet]. arXiv; 2021 [cited 2025 Aug 28]. Available from: http://arxiv.org/abs/2105.07624
  28. 28.
    Masry A, Kavehzadeh P, Do XL, Hoque E, Joty S. UniChart: A Universal Vision-language Pretrained Model for Chart Comprehension and Reasoning [Internet]. arXiv; 2023 [cited 2025 Aug 28]. Available from: http://arxiv.org/abs/2305.14761
  29. 29.
    Tang BJ, Boggust A, Satyanarayan A. VisText: A Benchmark for Semantically Rich Chart Captioning [Internet]. arXiv; 2023 [cited 2025 Aug 28]. Available from: http://arxiv.org/abs/2307.05356
  30. 30.
    jp1924/VQAonBD · Datasets at Hugging Face [Internet]. [cited 2025 Sep 2]. Available from: https://huggingface.co/datasets/jp1924/VQAonBD
  31. 31.
    Shridhar M, Yuan X, Côté M-A, Bisk Y, Trischler A, Hausknecht M. ALFWorld: Aligning Text and Embodied Environments for Interactive Learning [Internet]. arXiv; 2021 [cited 2025 Sep 2]. Available from: http://arxiv.org/abs/2010.03768
  32. 32.
    Chen GH, Chen S, Zhang R, Chen J, Wu X, Zhang Z, et al. ALLaVA: Harnessing GPT4V-Synthesized Data for Lite Vision-Language Models [Internet]. arXiv; 2024 [cited 2025 Aug 28]. Available from: http://arxiv.org/abs/2402.11684
    back: 1, 2
  33. 33.
    Tong P, Brown E, Wu P, Woo S, IYER AJV, Akula SC, et al. Cambrian-1: A fully open, vision-centric exploration of multimodal llms. Advances in Neural Information Processing Systems [Internet]. 2024 [cited 2025 Aug 28];37:87310–56. Available from: https://proceedings.neurips.cc/paper_files/paper/2024/hash/9ee3a664ccfeabc0da16ac6f1f1cfe59-Abstract-Conference.html
  34. 34.
    REILX/chinese-meme-description-dataset · Datasets at Hugging Face [Internet]. 2024 [cited 2025 Sep 2]. Available from: https://huggingface.co/datasets/REILX/chinese-meme-description-dataset
  35. 35.
    Ren M, Kiros R, Zemel R. Exploring models and data for image question answering. Advances in neural information processing systems [Internet]. 2015 [cited 2025 Aug 28];28. Available from: https://proceedings.neurips.cc/paper/2015/hash/831c2f88a604a07ca94314b56a4921b8-Abstract.html
  36. 36.
    Belouadi J, Lauscher A, Eger S. AutomaTikZ: Text-Guided Synthesis of Scientific Vector Graphics with TikZ [Internet]. arXiv; 2024 [cited 2025 Aug 28]. Available from: http://arxiv.org/abs/2310.00367
    back: 1, 2
  37. 37.
    Sima C, Renz K, Chitta K, Chen L, Zhang H, Xie C, et al. DriveLM: Driving with Graph Visual Question Answering. In: Leonardis A, Ricci E, Roth S, Russakovsky O, Sattler T, Varol G, editors. Computer Vision – ECCV 2024 [Internet]. Cham: Springer Nature Switzerland; 2025 [cited 2025 Aug 28]. p. 256–74. Available from: https://link.springer.com/10.1007/978-3-031-72943-0_15
  38. 38.
    Kiela D, Firooz H, Mohan A, Goswami V, Singh A, Ringshia P, et al. The hateful memes challenge: Detecting hate speech in multimodal memes. Advances in neural information processing systems [Internet]. 2020 [cited 2025 Aug 28];33:2611–24. Available from: https://proceedings.neurips.cc/paper/2020/hash/1b84c4cee2b8b3d823b30e2d604b1878-Abstract.html
  39. 39.
    Lu P, Qiu L, Chen J, Xia T, Zhao Y, Zhang W, et al. IconQA: A New Benchmark for Abstract Diagram Understanding and Visual Language Reasoning [Internet]. arXiv; 2022 [cited 2025 Aug 28]. Available from: http://arxiv.org/abs/2110.13214
  40. 40.
    Cha S, Lee J, Lee Y, Yang C. Visually Dehallucinative Instruction Generation: Know What You Don’t Know [Internet]. arXiv; 2024 [cited 2025 Aug 28]. Available from: http://arxiv.org/abs/2402.09717
  41. 41.
    Liu H, Li C, Wu Q, Lee YJ. Visual instruction tuning. Advances in neural information processing systems [Internet]. 2023 [cited 2025 Aug 28];36:34892–916. Available from: https://proceedings.neurips.cc/paper_files/paper/2023/hash/6dcf277ea32ce3288914faf369fe6de0-Abstract-Conference.html
  42. 42.
    Zhang Y, Zhang R, Gu J, Zhou Y, Lipka N, Yang D, et al. LLaVAR: Enhanced Visual Instruction Tuning for Text-Rich Image Understanding [Internet]. arXiv; 2024 [cited 2025 Aug 28]. Available from: http://arxiv.org/abs/2306.17107
  43. 43.
    Vik’s ML Research Blog [Internet]. [cited 2025 Aug 28]. Available from: https://vikhyat.net/posts/2024-08-17-lnqa.html
  44. 44.
    Liu F, Lin K, Li L, Wang J, Yacoob Y, Wang L. Mitigating Hallucination in Large Multi-Modal Models via Robust Instruction Tuning [Internet]. arXiv; 2024 [cited 2025 Aug 28]. Available from: http://arxiv.org/abs/2306.14565
  45. 45.
    Wang J, Meng L, Weng Z, He B, Wu Z, Jiang Y-G. To See is to Believe: Prompting GPT-4V for Better Visual Instruction Tuning [Internet]. arXiv; 2023 [cited 2025 Aug 28]. Available from: http://arxiv.org/abs/2311.07574
  46. 46.
    Li B, Zhang Y, Chen L, Wang J, Pu F, Yang J, et al. MIMIC-IT: Multi-Modal In-Context Instruction Tuning [Internet]. arXiv; 2023 [cited 2025 Sep 2]. Available from: http://arxiv.org/abs/2306.05425
  47. 47.
    Luo R, Zhang H, Chen L, Lin T-E, Liu X, Wu Y, et al. MMEvol: Empowering Multimodal Large Language Models with Evol-Instruct [Internet]. arXiv; 2024 [cited 2025 Aug 28]. Available from: http://arxiv.org/abs/2409.05840
  48. 48.
    Wu S, Zhu K, Bai Y, Liang Y, Li Y, Wu H, et al. MMRA: A Benchmark for Evaluating Multi-Granularity and Multi-Image Relational Association Capabilities in Large Visual Language Models [Internet]. arXiv; 2024 [cited 2025 Aug 28]. Available from: http://arxiv.org/abs/2407.17379
  49. 49.
    Suhr A, Zhou S, Zhang A, Zhang I, Bai H, Artzi Y. A Corpus for Reasoning About Natural Language Grounded in Photographs [Internet]. arXiv; 2019 [cited 2025 Aug 28]. Available from: http://arxiv.org/abs/1811.00491
  50. 50.
    Tu H, Cui C, Wang Z, Zhou Y, Zhao B, Han J, et al. How Many Unicorns Are in This Image? A Safety Evaluation Benchmark for Vision LLMs [Internet]. arXiv; 2023 [cited 2025 Aug 28]. Available from: http://arxiv.org/abs/2311.16101
    back: 1, 2
  51. 51.
    Yu Y, Chung S, Lee B-K, Ro YM. SPARK: Multi-Vision Sensor Perception and Reasoning Benchmark for Large-scale Vision-Language Models [Internet]. arXiv; 2024 [cited 2025 Aug 28]. Available from: http://arxiv.org/abs/2408.12114
  52. 52.
    Yang K, Russakovsky O, Deng J. Spatialsense: An adversarially crowdsourced benchmark for spatial relation recognition. In: Proceedings of the IEEE/CVF International Conference on Computer Vision [Internet]. 2019 [cited 2025 Aug 28]. p. 2051–60. Available from: http://openaccess.thecvf.com/content_ICCV_2019/html/Yang_SpatialSense_An_Adversarially_Crowdsourced_Benchmark_for_Spatial_Relation_Recognition_ICCV_2019_paper.html
  53. 53.
    Jhamtani H, Berg-Kirkpatrick T. Learning to Describe Differences Between Pairs of Similar Images [Internet]. arXiv; 2018 [cited 2025 Aug 28]. Available from: http://arxiv.org/abs/1808.10584
  54. 54.
    Xu Z, Feng C, Shao R, Ashby T, Shen Y, Jin D, et al. Vision-Flan: Scaling Human-Labeled Tasks in Visual Instruction Tuning [Internet]. arXiv; 2024 [cited 2025 Aug 28]. Available from: http://arxiv.org/abs/2402.11690
  55. 55.
    Zhu Y, Groth O, Bernstein M, Fei-Fei L. Visual7w: Grounded question answering in images. In: Proceedings of the IEEE conference on computer vision and pattern recognition [Internet]. 2016 [cited 2025 Aug 28]. p. 4995–5004. Available from: http://openaccess.thecvf.com/content_cvpr_2016/html/Zhu_Visual7W_Grounded_Question_CVPR_2016_paper.html
  56. 56.
    Gurari D, Li Q, Stangl AJ, Guo A, Lin C, Grauman K, et al. Vizwiz grand challenge: Answering visual questions from blind people. In: Proceedings of the IEEE conference on computer vision and pattern recognition [Internet]. 2018 [cited 2025 Aug 28]. p. 3608–17. Available from: http://openaccess.thecvf.com/content_cvpr_2018/html/Gurari_VizWiz_Grand_Challenge_CVPR_2018_paper.html
  57. 57.
    Goyal Y, Khot T, Summers-Stay D, Batra D, Parikh D. Making the v in vqa matter: Elevating the role of image understanding in visual question answering. In: Proceedings of the IEEE conference on computer vision and pattern recognition [Internet]. 2017 [cited 2025 Aug 28]. p. 6904–13. Available from: http://openaccess.thecvf.com/content_cvpr_2017/html/Goyal_Making_the_v_CVPR_2017_paper.html
  58. 58.
    Liu F, Emerson G, Collier N. Visual spatial reasoning. Transactions of the Association for Computational Linguistics [Internet]. 2023 [cited 2025 Aug 28];11:635–51. Available from: https://direct.mit.edu/tacl/article/doi/10.1162/tacl_a_00566/116470
  59. 59.
    Laurençon H, Tronchon L, Sanh V. Unlocking the conversion of Web Screenshots into HTML Code with the WebSight Dataset [Internet]. arXiv; 2024 [cited 2025 Aug 28]. Available from: http://arxiv.org/abs/2403.09029
  60. 60.
    Lu Y, Jiang D, Chen W, Wang WY, Choi Y, Lin BY. Wildvision: Evaluating vision-language models in the wild with human preferences. Advances in Neural Information Processing Systems [Internet]. 2024 [cited 2025 Aug 28];37:48224–55. Available from: https://proceedings.neurips.cc/paper_files/paper/2024/hash/563991b5c8b45fe75bea42db738223b2-Abstract-Datasets_and_Benchmarks_Track.html
  61. 61.
    Nandy A, Agarwal Y, Patwa A, Das MM, Bansal A, Raj A, et al. YesBut: A High-Quality Annotated Multimodal Dataset for evaluating Satire Comprehension capability of Vision-Language Models [Internet]. arXiv; 2024 [cited 2025 Aug 28]. Available from: http://arxiv.org/abs/2409.13592
  62. 62.
    Xu Y, Wang Z, Wang J, Lu D, Xie T, Saha A, et al. Aguvis: Unified Pure Vision Agents for Autonomous GUI Interaction [Internet]. arXiv; 2025 [cited 2025 Sep 2]. Available from: http://arxiv.org/abs/2412.04454
  63. 63.
    Zheng L, Huang Z, Xue Z, Wang X, An B, Yan S. AgentStudio: A Toolkit for Building General Virtual Agents [Internet]. arXiv; 2025 [cited 2025 Aug 28]. Available from: http://arxiv.org/abs/2403.17918
  64. 64.
    Shao S, Li Z, Zhang T, Peng C, Yu G, Zhang X, et al. Objects365: A Large-Scale, High-Quality Dataset for Object Detection. In: 2019 IEEE/CVF International Conference on Computer Vision (ICCV) [Internet]. Seoul, Korea (South): IEEE; 2019 [cited 2025 Sep 2]. p. 8429–38. Available from: https://ieeexplore.ieee.org/document/9009553/
  65. 65.
    Acharya M, Kafle K, Kanan C. Tallyqa: Answering complex counting questions. In: Proceedings of the AAAI conference on artificial intelligence [Internet]. 2019 [cited 2025 Aug 28]. p. 8076–84. Available from: https://ojs.aaai.org/index.php/AAAI/article/view/4815
  66. 66.
    Lindström AD, Abraham SS. CLEVR-Math: A Dataset for Compositional Language, Visual and Mathematical Reasoning [Internet]. arXiv; 2022 [cited 2025 Sep 2]. Available from: http://arxiv.org/abs/2208.05358
  67. 67.
    Gao J, Pi R, Zhang J, Ye J, Zhong W, Wang Y, et al. G-LLaVA: Solving Geometric Problem with Multi-Modal Large Language Model [Internet]. arXiv; 2025 [cited 2025 Aug 28]. Available from: http://arxiv.org/abs/2312.11370
    back: 1, 2
  68. 68.
    Lu P, Gong R, Jiang S, Qiu L, Huang S, Liang X, et al. Inter-GPS: Interpretable Geometry Problem Solving with Formal Language and Symbolic Reasoning [Internet]. arXiv; 2021 [cited 2025 Aug 28]. Available from: http://arxiv.org/abs/2105.04165
  69. 69.
    Kazemi M, Alvari H, Anand A, Wu J, Chen X, Soricut R. GeomVerse: A Systematic Evaluation of Large Models for Geometric Reasoning [Internet]. arXiv; 2023 [cited 2025 Aug 28]. Available from: http://arxiv.org/abs/2312.12241
  70. 70.
    Cao J, Xiao J. An augmented benchmark dataset for geometric question answering through dual parallel text encoding. In: Proceedings of the 29th international conference on computational linguistics [Internet]. 2022 [cited 2025 Aug 28]. p. 1511–20. Available from: https://aclanthology.org/2022.coling-1.130/
  71. 71.
    Seo M, Hajishirzi H, Farhadi A, Etzioni O, Malcolm C. Solving geometry problems: Combining text and diagram interpretation. In: Proceedings of the 2015 conference on empirical methods in natural language processing [Internet]. 2015 [cited 2025 Aug 28]. p. 1466–76. Available from: https://aclanthology.org/D15-1171.pdf
  72. 72.
    Lu P, Gong R, Jiang S, Qiu L, Huang S, Liang X, et al. Inter-GPS: Interpretable Geometry Problem Solving with Formal Language and Symbolic Reasoning [Internet]. arXiv; 2021 [cited 2025 Aug 28]. Available from: http://arxiv.org/abs/2105.04165
  73. 73.
    Zhang R, Wei X, Jiang D, Guo Z, Li S, Zhang Y, et al. MAVIS: Mathematical Visual Instruction Tuning with an Automatic Data Engine [Internet]. arXiv; 2024 [cited 2025 Aug 28]. Available from: http://arxiv.org/abs/2407.08739
    back: 1, 2
  74. 74.
    Zhang C, Gao F, Jia B, Zhu Y, Zhu S-C. Raven: A dataset for relational and analogical visual reasoning. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition [Internet]. 2019 [cited 2025 Aug 28]. p. 5317–27. Available from: http://openaccess.thecvf.com/content_CVPR_2019/html/Zhang_RAVEN_A_Dataset_for_Relational_and_Analogical_Visual_REasoNing_CVPR_2019_paper.html
  75. 75.
    Chng CK, Liu Y, Sun Y, Ng CC, Luo C, Ni Z, et al. Icdar2019 robust reading challenge on arbitrary-shaped text-rrc-art. In: 2019 International Conference on Document Analysis and Recognition (ICDAR) [Internet]. IEEE; 2019 [cited 2025 Sep 2]. p. 1571–6. Available from: https://ieeexplore.ieee.org/abstract/document/8978157/
  76. 76.
    Mouchere H, Viard-Gaudin C, Zanibbi R, Garain U, Kim DH, Kim JH. Icdar 2013 crohme: Third international competition on recognition of online handwritten mathematical expressions. In: 2013 12th International Conference on Document Analysis and Recognition [Internet]. IEEE; 2013 [cited 2025 Aug 28]. p. 1428–32. Available from: https://ieeexplore.ieee.org/abstract/document/6628849/
  77. 77.
    Veit A, Matera T, Neumann L, Matas J, Belongie S. COCO-Text: Dataset and Benchmark for Text Detection and Recognition in Natural Images [Internet]. arXiv; 2016 [cited 2025 Sep 2]. Available from: http://arxiv.org/abs/1601.07140
  78. 78.
    Yuan T-L, Zhu Z, Xu K, Li C-J, Mu T-J, Hu S-M. A Large Chinese Text Dataset in the Wild. J Comput Sci Technol [Internet]. 2019 May [cited 2025 Sep 2];34(3):509–21. Available from: http://link.springer.com/10.1007/s11390-019-1923-y
  79. 79.
    Jaume G, Ekenel HK, Thiran J-P. Funsd: A dataset for form understanding in noisy scanned documents. In: 2019 International Conference on Document Analysis and Recognition Workshops (ICDARW) [Internet]. IEEE; 2019 [cited 2025 Aug 28]. p. 1–6. Available from: https://ieeexplore.ieee.org/abstract/document/8892998/
  80. 80.
    Yuan Y, Liu X, Dikubab W, Liu H, Ji Z, Wu Z, et al. Syntax-aware network for handwritten mathematical expression recognition. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition [Internet]. 2022 [cited 2025 Aug 28]. p. 4553–62. Available from: http://openaccess.thecvf.com/content/CVPR2022/html/Yuan_Syntax-Aware_Network_for_Handwritten_Mathematical_Expression_Recognition_CVPR_2022_paper.html
  81. 81.
    Mathew M, Gomez L, Karatzas D, Jawahar CV. Asking questions on handwritten document collections. IJDAR [Internet]. 2021 Sep [cited 2025 Aug 28];24(3):235–49. Available from: https://link.springer.com/10.1007/s10032-021-00383-3
    back: 1, 2
  82. 82.
    Marti U-V, Bunke H. The IAM-database: an English sentence database for offline handwriting recognition. International Journal on Document Analysis and Recognition [Internet]. 2002 Nov 1 [cited 2025 Aug 28];5(1):39–46. Available from: http://link.springer.com/10.1007/s100320200071
  83. 83.
    Mishra A, Alahari K, Jawahar CV. Scene text recognition using higher order language priors. In: BMVC-British machine vision conference [Internet]. BMVA; 2012 [cited 2025 Aug 28]. Available from: https://inria.hal.science/hal-00818183/
  84. 84.
    Krishnan P, Kovvuri R, Pang G, Vassilev B, Hassner T. Textstylebrush: transfer of text aesthetics from a single example. IEEE Transactions on Pattern Analysis and Machine Intelligence [Internet]. 2023 [cited 2025 Aug 28];45(7):9122–34. Available from: https://ieeexplore.ieee.org/abstract/document/10027471/
  85. 85.
    Im2Latex [Internet]. [cited 2025 Sep 2]. Available from: https://sujayr91.github.io/Im2Latex/
  86. 86.
    OleehyO/latex-formulas · Datasets at Hugging Face [Internet]. 2024 [cited 2025 Aug 28]. Available from: https://huggingface.co/datasets/OleehyO/latex-formulas
  87. 87.
    Li Z, Lin Y, Chiang Y-Y, Weinman J, Tual S, Chazalon J, et al. ICDAR 2024 Competition on Historical Map Text Detection, Recognition, and Linking. In: Barney Smith EH, Liwicki M, Peng L, editors. Document Analysis and Recognition - ICDAR 2024 [Internet]. Cham: Springer Nature Switzerland; 2024 [cited 2025 Aug 28]. p. 363–80. Available from: https://link.springer.com/10.1007/978-3-031-70552-6_22
  88. 88.
    Gervais P, Fadeeva A, Maksai A. MathWriting: A Dataset For Handwritten Mathematical Expression Recognition. In: Proceedings of the 31st ACM SIGKDD Conference on Knowledge Discovery and Data Mining V2 [Internet]. Toronto ON Canada: ACM; 2025 [cited 2025 Sep 2]. p. 5459–69. Available from: https://dl.acm.org/doi/10.1145/3711896.3737436
  89. 89.
    Sharma C, Bhageria D, Scott W, PYKL S, Das A, Chakraborty T, et al. SemEval-2020 Task 8: Memotion Analysis – The Visuo-Lingual Metaphor! [Internet]. arXiv; 2020 [cited 2025 Aug 28]. Available from: http://arxiv.org/abs/2008.03781
  90. 90.
    Diem M, Fiel S, Kleber F, Sablatnig R, Saavedra JM, Contreras D, et al. ICFHR 2014 competition on handwritten digit string recognition in challenging datasets (HDSRC 2014). In: 2014 14th International Conference on Frontiers in Handwriting Recognition [Internet]. IEEE; 2014 [cited 2025 Aug 28]. p. 779–84. Available from: https://ieeexplore.ieee.org/abstract/document/6981115/
  91. 91.
    wendlerc/RenderedText · Datasets at Hugging Face [Internet]. 2024 [cited 2025 Aug 28]. Available from: https://huggingface.co/datasets/wendlerc/RenderedText
  92. 92.
    Huang Z, Chen K, He J, Bai X, Karatzas D, Lu S, et al. Icdar2019 competition on scanned receipt ocr and information extraction. In: 2019 International Conference on Document Analysis and Recognition (ICDAR) [Internet]. IEEE; 2019 [cited 2025 Aug 28]. p. 1516–20. Available from: https://ieeexplore.ieee.org/abstract/document/8977955/
  93. 93.
    Yu W, Zhang C, Cao H, Hua W, Li B, Chen H, et al. ICDAR 2023 Competition on Structured Text Extraction from Visually-Rich Document Images. In: Fink GA, Jain R, Kise K, Zanibbi R, editors. Document Analysis and Recognition - ICDAR 2023 [Internet]. Cham: Springer Nature Switzerland; 2023 [cited 2025 Sep 2]. p. 536–52. Available from: https://link.springer.com/10.1007/978-3-031-41679-8_32
  94. 94.
    Kim G, Hong T, Yim M, Nam J, Park J, Yim J, et al. OCR-Free Document Understanding Transformer. In: Avidan S, Brostow G, Cissé M, Farinella GM, Hassner T, editors. Computer Vision – ECCV 2022 [Internet]. Cham: Springer Nature Switzerland; 2022 [cited 2025 Aug 28]. p. 498–517. Available from: https://link.springer.com/10.1007/978-3-031-19815-1_29
  95. 95.
    https://ai.100tal.com/dataset [Internet]. [cited 2025 Sep 2]. Available from: https://ai.100tal.com/dataset
  96. 96.
    Xie X, Fu L, Zhang Z, Wang Z, Bai X. Toward Understanding WordArt: Corner-Guided Transformer for Scene Text Recognition. In: Avidan S, Brostow G, Cissé M, Farinella GM, Hassner T, editors. Computer Vision – ECCV 2022 [Internet]. Cham: Springer Nature Switzerland; 2022 [cited 2025 Aug 28]. p. 303–21. Available from: https://link.springer.com/10.1007/978-3-031-19815-1_18
  97. 97.
    Poznanski J, Rangapur A, Borchardt J, Dunkelberger J, Huff R, Lin D, et al. olmOCR: Unlocking Trillions of Tokens in PDFs with Vision Language Models [Internet]. arXiv; 2025 [cited 2025 Sep 2]. Available from: http://arxiv.org/abs/2502.18443
    back: 1, 2
  98. 98.
    Schwenk D, Khandelwal A, Clark C, Marino K, Mottaghi R. A-OKVQA: A Benchmark for Visual Question Answering Using World Knowledge. In: Avidan S, Brostow G, Cissé M, Farinella GM, Hassner T, editors. Computer Vision – ECCV 2022 [Internet]. Cham: Springer Nature Switzerland; 2022 [cited 2025 Aug 28]. p. 146–62. Available from: https://link.springer.com/10.1007/978-3-031-20074-8_9
    back: 1, 2
  99. 99.
    Li L, Wang Y, Xu R, Wang P, Feng X, Kong L, et al. Multimodal ArXiv: A Dataset for Improving Scientific Comprehension of Large Vision-Language Models [Internet]. arXiv; 2024 [cited 2025 Aug 28]. Available from: http://arxiv.org/abs/2403.00231
  100. 100.
    Bhushan S, Lee M. Block diagram-to-text: Understanding block diagram images by generating natural language descriptors. In: Findings of the Association for Computational Linguistics: AACL-IJCNLP 2022 [Internet]. 2022 [cited 2025 Sep 2]. p. 153–68. Available from: https://aclanthology.org/2022.findings-aacl.15/
    back: 1, 2
  101. 101.
    Kamizuru00/diagram_image_to_text · Datasets at Hugging Face [Internet]. 2024 [cited 2025 Aug 28]. Available from: https://huggingface.co/datasets/Kamizuru00/diagram_image_to_text
  102. 102.
    Mathew M, Karatzas D, Jawahar CV. Docvqa: A dataset for vqa on document images. In: Proceedings of the IEEE/CVF winter conference on applications of computer vision [Internet]. 2021 [cited 2025 Aug 28]. p. 2200–9. Available from: http://openaccess.thecvf.com/content/WACV2021/html/Mathew_DocVQA_A_Dataset_for_VQA_on_Document_Images_WACV_2021_paper.html
  103. 103.
    Wang X, Liu Y, Shen C, Ng CC, Luo C, Jin L, et al. On the general value of evidence, and bilingual scene-text visual question answering. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition [Internet]. 2020 [cited 2025 Aug 28]. p. 10126–35. Available from: http://openaccess.thecvf.com/content_CVPR_2020/html/Wang_On_the_General_Value_of_Evidence_and_Bilingual_Scene-Text_Visual_CVPR_2020_paper.html
  104. 104.
    ift/handwriting_forms · Datasets at Hugging Face [Internet]. [cited 2025 Sep 2]. Available from: https://huggingface.co/datasets/ift/handwriting_forms
  105. 105.
    Mathew M, Bagal V, Tito R, Karatzas D, Valveny E, Jawahar CV. Infographicvqa. In: Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision [Internet]. 2022 [cited 2025 Aug 28]. p. 1697–706. Available from: http://openaccess.thecvf.com/content/WACV2022/html/Mathew_InfographicVQA_WACV_2022_paper.html
    back: 1, 2, 3
  106. 106.
    mychen76/invoices-and-receipts_ocr_v1 · Datasets at Hugging Face [Internet]. 2025 [cited 2025 Sep 2]. Available from: https://huggingface.co/datasets/mychen76/invoices-and-receipts_ocr_v1
  107. 107.
    Chang S, Palzer D, Li J, Fosler-Lussier E, Xiao N. MapQA: A Dataset for Question Answering on Choropleth Maps [Internet]. arXiv; 2022 [cited 2025 Aug 28]. Available from: http://arxiv.org/abs/2211.08545
  108. 108.
    Sharma C, Paka W, Scott DB, Das A, Poria S, Chakraborty T, et al. Task report: Memotion analysis 1.0@ semeval 2020: The visuo-lingual metaphor. In: Proceedings of the 14th International Workshop on Semantic Evaluation (SemEval-2020), Barcelona, Spain, Sep Association for Computational Linguistics. 2020. p. 759–73.
  109. 109.
    Mishra A, Shekhar S, Singh AK, Chakraborty A. Ocr-vqa: Visual question answering by reading text in images. In: 2019 international conference on document analysis and recognition (ICDAR) [Internet]. IEEE; 2019 [cited 2025 Sep 1]. p. 947–52. Available from: https://ieeexplore.ieee.org/abstract/document/8978122/
  110. 110.
    Ding Y, Luo S, Chung H, Han SC. PDF-VQA: A New Dataset for Real-World VQA on PDF Documents. In: De Francisci Morales G, Perlich C, Ruchansky N, Kourtellis N, Baralis E, Bonchi F, editors. Machine Learning and Knowledge Discovery in Databases: Applied Data Science and Demo Track [Internet]. Cham: Springer Nature Switzerland; 2023 [cited 2025 Sep 1]. p. 585–601. Available from: https://link.springer.com/10.1007/978-3-031-43427-3_35
  111. 111.
    Wang B, Li G, Zhou X, Chen Z, Grossman T, Li Y. Screen2Words: Automatic Mobile UI Summarization with Multimodal Learning. In: The 34th Annual ACM Symposium on User Interface Software and Technology [Internet]. Virtual Event USA: ACM; 2021 [cited 2025 Sep 1]. p. 498–510. Available from: https://dl.acm.org/doi/10.1145/3472749.3474765
  112. 112.
    Hsiao Y-C, Zubach F, Baechler G, Sunkara S, Carbune V, Lin J, et al. ScreenQA: Large-Scale Question-Answer Pairs over Mobile App Screenshots [Internet]. arXiv; 2025 [cited 2025 Sep 1]. Available from: http://arxiv.org/abs/2209.08199
  113. 113.
    Tanaka R, Nishida K, Nishida K, Hasegawa T, Saito I, Saito K. Slidevqa: A dataset for document visual question answering on multiple images. In: Proceedings of the AAAI Conference on Artificial Intelligence [Internet]. 2023 [cited 2025 Sep 1]. p. 13636–45. Available from: https://ojs.aaai.org/index.php/AAAI/article/view/26598
  114. 114.
    Biten AF, Tito R, Mafla A, Gomez L, Rusinol M, Valveny E, et al. Scene text visual question answering. In: Proceedings of the IEEE/CVF international conference on computer vision [Internet]. 2019 [cited 2025 Sep 1]. p. 4291–301. Available from: http://openaccess.thecvf.com/content_ICCV_2019/html/Biten_Scene_Text_Visual_Question_Answering_ICCV_2019_paper.html
  115. 115.
    sujet-ai/Sujet-Finance-QA-Vision-100k · Datasets at Hugging Face [Internet]. 2024 [cited 2025 Sep 1]. Available from: https://huggingface.co/datasets/sujet-ai/Sujet-Finance-QA-Vision-100k
  116. 116.
    jimmycarter/textocr-gpt4v · Datasets at Hugging Face [Internet]. 2024 [cited 2025 Sep 1]. Available from: https://huggingface.co/datasets/jimmycarter/textocr-gpt4v
  117. 117.
    Singh A, Natarajan V, Shah M, Jiang Y, Chen X, Batra D, et al. Towards vqa models that can read. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition [Internet]. 2019 [cited 2025 Sep 1]. p. 8317–26. Available from: http://openaccess.thecvf.com/content_CVPR_2019/html/Singh_Towards_VQA_Models_That_Can_Read_CVPR_2019_paper.html
  118. 118.
    Ye J, Hu A, Xu H, Ye Q, Yan M, Xu G, et al. UReader: Universal OCR-free Visually-situated Language Understanding with Multimodal Large Language Model [Internet]. arXiv; 2023 [cited 2025 Sep 1]. Available from: http://arxiv.org/abs/2310.05126
    back: 1, 2, 3, 4
  119. 119.
    Tanaka R, Nishida K, Yoshida S. Visualmrc: Machine reading comprehension on document images. In: Proceedings of the AAAI Conference on Artificial Intelligence [Internet]. 2021 [cited 2025 Sep 1]. p. 13878–88. Available from: https://ojs.aaai.org/index.php/AAAI/article/view/17635
  120. 120.
    andito/ai2d-merged · Datasets at Hugging Face [Internet]. [cited 2025 Sep 2]. Available from: https://huggingface.co/datasets/andito/ai2d-merged
  121. 121.
    He X, Zhang Y, Mou L, Xing E, Xie P. PathVQA: 30000+ Questions for Medical Visual Question Answering [Internet]. arXiv; 2020 [cited 2025 Sep 1]. Available from: http://arxiv.org/abs/2003.10286
  122. 122.
    Lu P, Mishra S, Xia T, Qiu L, Chang K-W, Zhu S-C, et al. Learn to explain: Multimodal reasoning via thought chains for science question answering. Advances in Neural Information Processing Systems [Internet]. 2022 [cited 2025 Sep 1];35:2507–21. Available from: https://proceedings.neurips.cc/paper_files/paper/2022/hash/11332b6b6cf4485b84afadb1352d3a9a-Abstract-Conference.html
  123. 123.
    Kembhavi A, Seo M, Schwenk D, Choi J, Farhadi A, Hajishirzi H. Are you smarter than a sixth grader? textbook question answering for multimodal machine comprehension. In: Proceedings of the IEEE Conference on Computer Vision and Pattern recognition [Internet]. 2017 [cited 2025 Sep 1]. p. 4999–5007. Available from: http://openaccess.thecvf.com/content_cvpr_2017/html/Kembhavi_Are_You_Smarter_CVPR_2017_paper.html
  124. 124.
    Jia Y, Li J, Yue X, Li B, Nie P, Zou K, et al. VisualWebInstruct: Scaling up Multimodal Instruction Data through Web Search [Internet]. arXiv; 2025 [cited 2025 Sep 1]. Available from: http://arxiv.org/abs/2503.10582
  125. 125.
    Lau JJ, Gayen S, Ben Abacha A, Demner-Fushman D. A dataset of clinically generated visual questions and answers about radiology images. Scientific data [Internet]. 2018 [cited 2025 Sep 1];5(1):1–10. Available from: https://www.nature.com/articles/sdata2018251
  126. 126.
    Zheng T, Zhang G, Shen T, Liu X, Lin BY, Fu J, et al. OpenCodeInterpreter: Integrating Code Generation with Execution and Refinement [Internet]. arXiv; 2025 [cited 2025 Sep 1]. Available from: http://arxiv.org/abs/2402.14658
    back: 1, 2
  127. 127.
    Zhang B-W, Yan Y, Li L, Liu G. Infinity Math: A Scalable Instruction Tuning Dataset in Programmatic Mathematical Reasoning. In: Proceedings of the 33rd ACM International Conference on Information and Knowledge Management [Internet]. Boise ID USA: ACM; 2024 [cited 2025 Sep 1]. p. 5405–9. Available from: https://dl.acm.org/doi/10.1145/3627673.3679122
  128. 128.
    Yue X, Qu X, Zhang G, Fu Y, Huang W, Sun H, et al. MAmmoTH: Building Math Generalist Models through Hybrid Instruction Tuning [Internet]. arXiv; 2023 [cited 2025 Sep 1]. Available from: http://arxiv.org/abs/2309.05653
  129. 129.
    Amini A, Gabriel S, Lin P, Koncel-Kedziorski R, Choi Y, Hajishirzi H. MathQA: Towards Interpretable Math Word Problem Solving with Operation-Based Formalisms [Internet]. arXiv; 2019 [cited 2025 Sep 1]. Available from: http://arxiv.org/abs/1905.13319
  130. 130.
    Lai X, Tian Z, Chen Y, Yang S, Peng X, Jia J. Step-DPO: Step-wise Preference Optimization for Long-chain Reasoning of LLMs [Internet]. arXiv; 2024 [cited 2025 Sep 1]. Available from: http://arxiv.org/abs/2406.18629
  131. 131.
    AI-MO/NuminaMath-CoT · Datasets at Hugging Face [Internet]. 2025 [cited 2025 Sep 1]. Available from: https://huggingface.co/datasets/AI-MO/NuminaMath-CoT
  132. 132.
    teknium/OpenHermes-2.5 · Datasets at Hugging Face [Internet]. 2024 [cited 2025 Sep 1]. Available from: https://huggingface.co/datasets/teknium/OpenHermes-2.5
  133. 133.
    Open-Orca/OpenOrca · Datasets at Hugging Face [Internet]. 2024 [cited 2025 Sep 1]. Available from: https://huggingface.co/datasets/Open-Orca/OpenOrca
  134. 134.
    Mitra A, Khanpour H, Rosset C, Awadallah A. Orca-Math: Unlocking the potential of SLMs in Grade School Math [Internet]. arXiv; 2024 [cited 2025 Sep 1]. Available from: http://arxiv.org/abs/2402.14830
  135. 135.
    flytech/python-codes-25k · Datasets at Hugging Face [Internet]. 2024 [cited 2025 Sep 1]. Available from: https://huggingface.co/datasets/flytech/python-codes-25k
  136. 136.
    sahil2801/CodeAlpaca-20k · Datasets at Hugging Face [Internet]. 2023 [cited 2025 Sep 2]. Available from: https://huggingface.co/datasets/sahil2801/CodeAlpaca-20k
  137. 137.
    qywu/ruozhiba_en · Datasets at Hugging Face [Internet]. [cited 2025 Sep 1]. Available from: https://huggingface.co/datasets/qywu/ruozhiba_en
  138. 138.
    Chen W, Yin M, Ku M, Lu P, Wan Y, Ma X, et al. TheoremQA: A Theorem-driven Question Answering dataset [Internet]. arXiv; 2023 [cited 2025 Sep 1]. Available from: http://arxiv.org/abs/2305.12524
  139. 139.
    WizardLMTeam/WizardLM_evol_instruct_70k · Datasets at Hugging Face [Internet]. 2024 [cited 2025 Sep 1]. Available from: https://huggingface.co/datasets/WizardLMTeam/WizardLM_evol_instruct_70k
  140. 140.
    Toshniwal S, Du W, Moshkov I, Kisacanin B, Ayrapetyan A, Gitman I. OpenMathInstruct-2: Accelerating AI for Math with Massive Open-Source Instruction Data [Internet]. arXiv; 2024 [cited 2025 Sep 2]. Available from: http://arxiv.org/abs/2410.01560
  141. 141.
    Li Z, Chen G, Liu S, Wang S, VS V, Ji Y, et al. Eagle 2: Building Post-Training Data Strategies from Scratch for Frontier Vision-Language Models [Internet]. 2025. Available from: https://arxiv.org/abs/2501.14818