| { | |
| "@context": "https://schema.org", | |
| "@type": "Dataset", | |
| "name": "ALL Bench Leaderboard 2026", | |
| "alternateName": [ | |
| "ALL Bench", | |
| "ALLBench", | |
| "AI Benchmark Leaderboard 2026" | |
| ], | |
| "description": "The only AI benchmark leaderboard covering LLM, VLM, Agent, Image, Video, and Music generation in a single unified view. 91 models cross-verified across 6 modalities with confidence badges. Features composite 5-axis scoring, interactive comparison tools, and downloadable intelligence reports.", | |
| "url": "https://huggingface.co/spaces/FINAL-Bench/all-bench-leaderboard", | |
| "sameAs": [ | |
| "https://huggingface.co/datasets/FINAL-Bench/ALL-Bench-Leaderboard", | |
| "https://github.com/final-bench/ALL-Bench-Leaderboard" | |
| ], | |
| "license": "https://opensource.org/licenses/MIT", | |
| "version": "2.2.1", | |
| "datePublished": "2026-03-01", | |
| "dateModified": "2026-03-10", | |
| "creator": { | |
| "@type": "Organization", | |
| "name": "ALL Bench Team", | |
| "url": "https://huggingface.co/FINAL-Bench" | |
| }, | |
| "keywords": [ | |
| "AI benchmark", | |
| "LLM leaderboard", | |
| "GPT-5", | |
| "Claude", | |
| "Gemini", | |
| "VLM benchmark", | |
| "AI agent", | |
| "image generation", | |
| "video generation", | |
| "music generation", | |
| "MMLU-Pro", | |
| "GPQA", | |
| "ARC-AGI-2", | |
| "FINAL Bench", | |
| "metacognition", | |
| "multimodal AI", | |
| "AI evaluation", | |
| "benchmark comparison", | |
| "AI model ranking", | |
| "open source AI" | |
| ], | |
| "about": [ | |
| { | |
| "@type": "Thing", | |
| "name": "Large Language Model" | |
| }, | |
| { | |
| "@type": "Thing", | |
| "name": "Vision Language Model" | |
| }, | |
| { | |
| "@type": "Thing", | |
| "name": "AI Benchmark" | |
| }, | |
| { | |
| "@type": "Thing", | |
| "name": "Generative AI" | |
| }, | |
| { | |
| "@type": "Thing", | |
| "name": "Metacognition" | |
| } | |
| ], | |
| "measurementTechnique": "Cross-verified benchmark aggregation with 3-tier confidence system", | |
| "variableMeasured": [ | |
| { | |
| "@type": "PropertyValue", | |
| "name": "MMLU-Pro", | |
| "description": "57K expert-level multi-discipline questions" | |
| }, | |
| { | |
| "@type": "PropertyValue", | |
| "name": "GPQA Diamond", | |
| "description": "PhD-level expert questions in science" | |
| }, | |
| { | |
| "@type": "PropertyValue", | |
| "name": "AIME 2025", | |
| "description": "American Invitational Mathematics Examination" | |
| }, | |
| { | |
| "@type": "PropertyValue", | |
| "name": "HLE", | |
| "description": "Humanity's Last Exam — 2500 expert-sourced questions" | |
| }, | |
| { | |
| "@type": "PropertyValue", | |
| "name": "ARC-AGI-2", | |
| "description": "Abstract reasoning and novel pattern recognition" | |
| }, | |
| { | |
| "@type": "PropertyValue", | |
| "name": "FINAL Bench Metacognitive", | |
| "description": "AI self-correction ability measurement" | |
| }, | |
| { | |
| "@type": "PropertyValue", | |
| "name": "SWE-Pro", | |
| "description": "Software engineering benchmark by Scale AI" | |
| }, | |
| { | |
| "@type": "PropertyValue", | |
| "name": "IFEval", | |
| "description": "Instruction following evaluation" | |
| }, | |
| { | |
| "@type": "PropertyValue", | |
| "name": "LiveCodeBench", | |
| "description": "Continuously updated coding benchmark" | |
| }, | |
| { | |
| "@type": "PropertyValue", | |
| "name": "Union Eval S3", | |
| "description": "ALL Bench proprietary integrated benchmark, 100% JSON auto-graded" | |
| } | |
| ], | |
| "distribution": [ | |
| { | |
| "@type": "DataDownload", | |
| "encodingFormat": "application/json", | |
| "contentUrl": "https://huggingface.co/datasets/FINAL-Bench/ALL-Bench-Leaderboard/resolve/main/all_bench_leaderboard_v2.1.json", | |
| "name": "Unified JSON Dataset (75KB)" | |
| } | |
| ], | |
| "isPartOf": { | |
| "@type": "DataCatalog", | |
| "name": "Hugging Face Datasets", | |
| "url": "https://huggingface.co/datasets" | |
| }, | |
| "funder": { | |
| "@type": "Organization", | |
| "name": "FINAL Bench" | |
| } | |
| } |