{ "config_general": { "model_name": "InternVL3-78B", "model_dtype": "float16", "model_size": 0 }, "results": { "CMMMU": { "艺术与设计": { "num": 88, "correct": 66, "accuracy": 75.0 }, "overall": { "num": 900, "correct": 476, "accuracy": 52.89 }, "商业": { "num": 126, "correct": 42, "accuracy": 33.33 }, "科学": { "num": 204, "correct": 97, "accuracy": 47.55 }, "健康与医学": { "num": 153, "correct": 89, "accuracy": 58.17 }, "人文社会科学": { "num": 85, "correct": 57, "accuracy": 67.06 }, "技术与工程": { "num": 244, "correct": 125, "accuracy": 51.23 }, "accuracy": 52.89, "acc_stderr": 0, "acc": 52.89 }, "MMMU": { "accuracy": 66.44, "subject_score": { "Accounting": 60.0, "Agriculture": 66.67, "Architecture": 50.0, "Art": 81.67, "Basic": 73.33, "Biology": 56.67, "Chemistry": 50.0, "Clinical": 70.0, "Computer": 60.0, "Design": 83.33, "Diagnostics": 50.0, "Economics": 73.33, "Electronics": 46.67, "Energy": 60.0, "Finance": 50.0, "Geography": 76.67, "History": 76.67, "Literature": 83.33, "Manage": 56.67, "Marketing": 90.0, "Materials": 46.67, "Math": 50.0, "Mechanical": 56.67, "Music": 56.67, "Pharmacy": 70.0, "Physics": 73.33, "Psychology": 80.0, "Public": 86.67, "Sociology": 76.67 }, "difficulty_score": { "Medium": 66.75, "Easy": 75.59, "Hard": 50.83 }, "acc_stderr": 0, "acc": 66.44 }, "MMMU_Pro_standard": { "accuracy": 49.65, "subject_score": { "History": 73.21, "Literature": 78.85, "Finance": 53.33, "Design": 68.33, "Sociology": 62.96, "Agriculture": 48.33, "Art": 69.81, "Accounting": 36.21, "Clinical_Medicine": 47.46, "Energy_and_Power": 22.41, "Pharmacy": 57.89, "Architecture_and_Engineering": 26.67, "Public_Health": 46.55, "Electronics": 60.0, "Physics": 46.67, "Art_Theory": 80.0, "Psychology": 50.0, "Manage": 42.0, "Economics": 42.37, "Biology": 47.46, "Mechanical_Engineering": 37.29, "Diagnostics_and_Laboratory_Medicine": 40.0, "Basic_Medical_Science": 48.08, "Computer_Science": 55.0, "Math": 43.33, "Music": 31.67, "Materials": 26.67, "Marketing": 55.93, "Chemistry": 45.0, "Geography": 55.77 }, "difficulty_score": { "Medium": 48.06, "Easy": 60.8, "Hard": 38.15 }, "acc_stderr": 0, "acc": 49.65 }, "MMMU_Pro_vision": { "accuracy": 39.13, "subject_score": { "Sociology": 50.0, "Literature": 57.69, "History": 39.29, "Design": 56.67, "Art": 54.72, "Clinical_Medicine": 33.9, "Architecture_and_Engineering": 21.67, "Public_Health": 31.03, "Agriculture": 28.33, "Art_Theory": 58.18, "Pharmacy": 42.11, "Psychology": 31.67, "Biology": 38.98, "Manage": 30.0, "Economics": 35.59, "Physics": 40.0, "Accounting": 48.28, "Diagnostics_and_Laboratory_Medicine": 31.67, "Mechanical_Engineering": 23.73, "Basic_Medical_Science": 46.15, "Finance": 46.67, "Electronics": 48.33, "Computer_Science": 40.0, "Math": 38.33, "Energy_and_Power": 25.86, "Music": 28.33, "Marketing": 40.68, "Materials": 26.67, "Chemistry": 41.67, "Geography": 44.23 }, "acc_stderr": 0, "acc": 39.13 }, "MmvetV2": { "accuracy": 69.0716, "capability_scores": { "ocr": 76.87500000000004, "math": 74.11764705882354, "spat": 66.75126903553297, "rec": 66.35922330097094, "know": 66.15384615384613, "gen": 69.3454545454546, "seq": 61.78571428571429 }, "capability_detail_scores": { "ocr_math": 57.27272727272727, "spat_ocr_math": 90.0, "rec_spat_ocr_math": 100.0, "rec_spat": 54.642857142857146, "spat_ocr": 80.0, "rec_spat_ocr": 75.0, "spat_ocr_know": 100.0, "rec_ocr": 75.0, "rec_spat_know": 39.0, "ocr": 86.875, "rec": 67.45762711864406, "rec_know": 53.84615384615385, "rec_know_gen": 69.89999999999998, "rec_ocr_know_gen": 77.6923076923077, "rec_spat_ocr_gen": 75.34883720930232, "spat_ocr_gen": 90.0, "spat_ocr_math_seq_gen": 40.0, "spat_ocr_rec_math_seq": 0.0, "rec_spat_gen": 50.0, "spat_ocr_math_gen": 50.0, "rec_seq_spat": 54.285714285714285, "rec_seq_spat_ocr": 60.0, "rec_spat_know_gen": 63.33333333333334, "rec_gen": 72.3529411764706, "rec_spat_ocr_know": 25.0, "spat_ocr_rec_gen_know": 65.0, "rec_ocr_math": 100.0, "rec_ocr_gen": 90.0, "rec_seq_ocr_gen": 80.0, "ocr_gen": 66.15384615384615, "rec_seq_gen": 60.0, "rec_seq": 65.0, "rec_seq_spat_gen": 68.75, "rec_seq_know": 0.0, "rec_seq_know_gen": 35.0, "spat_ocr_rec_seq_gen": 53.333333333333336, "ocr_rec_seq_gen_know": 95.0, "rec_math_know": 75.0, "rec_seq_ocr": 100.0 }, "acc_stderr": 0, "acc": 69.0716 }, "MathVerse": { "Text Dominant": { "accuracy": 54.31, "correct": 428, "total": 788 }, "Total": { "accuracy": 47.56, "correct": 1874, "total": 3940 }, "Vision Dominant": { "accuracy": 48.48, "correct": 382, "total": 788 }, "Vision Intensive": { "accuracy": 48.73, "correct": 384, "total": 788 }, "Text Lite": { "accuracy": 50.38, "correct": 397, "total": 788 }, "Vision Only": { "accuracy": 35.91, "correct": 283, "total": 788 }, "accuracy": 47.56, "acc_stderr": 0, "acc": 47.56 }, "Ocrlite": { "final_score": [ 1195, 1644 ], "accuracy": 72.689, "Key Information Extraction-Bookshelf": [ 31, 51, 0.608, { "Default": [ 31, 51, 0.608 ] } ], "Scene Text-centric VQA-diet_constraints": [ 61, 90, 0.678, { "Default": [ 61, 90, 0.678 ] } ], "Doc-oriented VQA-Control": [ 129, 189, 0.683, { "Default": [ 129, 189, 0.683 ] } ], "Doc-oriented VQA": [ 152, 204, 0.745, { "Default": [ 152, 204, 0.745 ] } ], "Scene Text-centric VQA-Fake_logo": [ 64, 119, 0.538, { "Default": [ 64, 119, 0.538 ] } ], "Handwritten Mathematical Expression Recognition": [ 1, 100, 0.01, { "Default": [ 1, 100, 0.01 ] } ], "Key Information Extraction": [ 184, 209, 0.88, { "Default": [ 184, 209, 0.88 ] } ], "Scene Text-centric VQA-Control": [ 166, 200, 0.83, { "Default": [ 166, 200, 0.83 ] } ], "Scene Text-centric VQA": [ 227, 282, 0.805, { "Default": [ 227, 282, 0.805 ] } ], "Artistic Text Recognition": [ 42, 50, 0.84, { "Default": [ 42, 50, 0.84 ] } ], "Irregular Text Recognition": [ 46, 50, 0.92, { "Default": [ 46, 50, 0.92 ] } ], "Non-Semantic Text Recognition": [ 42, 50, 0.84, { "Default": [ 42, 50, 0.84 ] } ], "Regular Text Recognition": [ 50, 50, 1.0, { "Default": [ 50, 50, 1.0 ] } ], "acc_stderr": 0, "acc": 72.689 }, "OcrliteZh": { "final_score": [ 153, 234 ], "accuracy": 65.385, "Docvqa": [ 4, 10, 0.4, { "Default": [ 4, 10, 0.4 ] } ], "Chartqa-human": [ 4, 10, 0.4, { "Default": [ 4, 10, 0.4 ] } ], "Chartqa-au": [ 6, 10, 0.6, { "Default": [ 6, 10, 0.6 ] } ], "infographic": [ 6, 10, 0.6, { "Default": [ 6, 10, 0.6 ] } ], "Key Information Extraction": [ 35, 45, 0.778, { "Default": [ 35, 45, 0.778 ] } ], "Scene Text-centric VQA": [ 27, 40, 0.675, { "Default": [ 27, 40, 0.675 ] } ], "Artistic Text Recognition": [ 5, 11, 0.455, { "Default": [ 5, 11, 0.455 ] } ], "IrRegular Text Recognition": [ 6, 11, 0.545, { "Default": [ 6, 11, 0.545 ] } ], "Non-semantic Text Recognition": [ 11, 12, 0.917, { "Default": [ 11, 12, 0.917 ] } ], "Regular Text Recognition": [ 10, 11, 0.909, { "Default": [ 10, 11, 0.909 ] } ], "Handwriting_CN": [ 16, 20, 0.8, { "Default": [ 16, 20, 0.8 ] } ], "Chinese Unlimited": [ 23, 44, 0.523, { "Default": [ 23, 44, 0.523 ] } ], "acc_stderr": 0, "acc": 65.385 }, "CharXiv": { "descriptive": { "Overall Score": 83.85, "By Question": { "Q1": 87.3, "Q2": 83.91, "Q3": 66.95, "Q4": 85.21, "Q5": 87.03, "Q6": 85.54, "Q7": 86.32, "Q8": 88.84, "Q9": 83.08, "Q10": 74.66, "Q11": 80.0, "Q12": 79.12, "Q13": 79.0, "Q14": 93.26, "Q15": 97.44, "Q16": 77.78, "Q17": 64.29, "Q18": 90.28, "Q19": 84.62 }, "By Category": { "Information Extraction": 83.27, "Enumeration": 89.35, "Pattern Recognition": 85.37, "Counting": 78.37, "Compositionality": 64.29 }, "By Subplot": { "1 Subplot": 87.05, "2-4 Subplots": 86.38, "5+ Subplots": 74.58 }, "By Subject": { "Computer Science": 84.13, "Economics": 83.88, "Electrical Engineering and Systems Science": 87.39, "Mathematics": 82.96, "Physics": 80.51, "Quantitative Biology": 81.55, "Quantitative Finance": 86.21, "Statistics": 84.73 }, "By Year": { "2020": 82.49, "2021": 81.42, "2022": 85.86, "2023": 85.79 }, "N_valid": 4000, "N_invalid": 5, "Question Type": "Descriptive" }, "reasoning": { "Overall Score": 41.4, "By Answer Type": { "Text-in-Chart": 45.91, "Text-in-General": 42.42, "Number-in-Chart": 43.53, "Number-in-General": 30.13 }, "By Source": { "GPT-Sourced": 53.8, "GPT-Inspired": 39.81, "Completely Human": 38.17 }, "By Subject": { "Computer Science": 38.1, "Economics": 46.38, "Electrical Engineering and Systems Science": 33.61, "Mathematics": 45.93, "Physics": 48.82, "Quantitative Biology": 46.83, "Quantitative Finance": 33.62, "Statistics": 35.4 }, "By Year": { "2020": 38.87, "2021": 45.98, "2022": 40.98, "2023": 39.52 }, "By Subplot": { "1 Subplot": 41.97, "2-4 Subplots": 41.01, "5+ Subplots": 41.1 }, "N_valid": 1000, "N_invalid": 0, "Question Type": "Reasoning" }, "accuracy": 62.62, "acc_stderr": 0, "acc": 62.62 }, "MathVision": { "accuracy": 39.97, "acc_stderr": 0, "acc": 39.97 }, "CII-Bench": { "accuracy": 67.32, "domain_score": { "CTC": 65.19, "Env.": 72.22, "Art": 61.03, "Society": 70.27, "Life": 68.4, "Politics": 70.83 }, "emotion_score": { "Positive": 70.09, "Negative": 64.91, "Neutral": 67.29 }, "acc_stderr": 0, "acc": 67.32 }, "Blink": { "accuracy": 61.13, "Art Style": 85.47, "Counting": 79.17, "Forensic Detection": 45.45, "Functional Correspondence": 37.69, "IQ Test": 27.33, "Jigsaw": 51.33, "Multi-view Reasoning": 54.14, "Object Localization": 53.28, "Relative Depth": 83.06, "Relative Reflectance": 32.09, "Semantic Correspondence": 52.52, "Spatial Relation": 91.61, "Visual Correspondence": 82.56, "Visual Similarity": 82.22, "acc_stderr": 0, "acc": 61.13 } } }