Name | Variant | Published At | Dataset | Incoming Links | Outgoing Links |
AI2 Reasoning Challenge |
Benchmark
|
2018-03-14 | 1 | 0 | |
AMD |
Organization
|
1969-05-01 | 0 | 1 | |
Allen Institute for AI |
Organization
|
2014-01-01 | 1 | 1 | |
Antrophic |
Organization
|
2021-01-01 | 0 | 3 | |
Claude 3.7 Sonnet |
Model
|
2025-02-24 | 1 | 2 | |
Claude Opus 4 |
Model
|
2025-05-22 | 1 | 2 | |
Claude Sonnet 4 |
Model
|
2025-05-22 | 1 | 2 | |
CommonsenseQA |
Benchmark
|
2019-06-01 | 1 | 0 | |
DeepSeek-R1 |
Model
|
2025-01-20 | 0 | 2 | |
FineWeb |
Dataset
|
2024-05-31 | 1 | 6 | |
GPQA |
Benchmark
|
2023-11-20 | 0 | 0 | |
GPT-4.5 |
Model
|
2025-02-27 | 1 | 2 | |
GPT-5 |
Model
|
2025-08-07 | 1 | 3 | |
Gemini 2.0 |
Model
|
2024-12-11 | 1 | 2 | |
Gemini 2.5 |
Model
|
2025-03-25 | 1 | 2 | |
Organization
|
1998-09-04 | 0 | 2 | ||
Grok 3 |
Model
|
2025-02-19 | 1 | 1 | |
Grok 4 |
Model
|
2025-07-09 | 1 | 1 | |
HellaSwag |
Benchmark
|
2019-05-29 | 2 | 0 | |
Humanity's Last Exam |
Benchmark
|
2025-01-25 | 3 | 0 | |
Instella |
Model
|
2025-03-05 | 1 | 6 | |
MMLU |
Benchmark
|
2020-09-07 | 7 | 0 | |
MMLU Pro |
Benchmark
|
2025-06-03 | 6 | 0 | |
MMMU |
Benchmark
|
2023-11-27 | 8 | 0 | |
MMMU-Pro |
Benchmark
|
2024-09-05 | 5 | 0 | |
Mistral Small 3.1 |
Model
|
2025-03-17 | 0 | 4 | |
OpenAI |
Organization
|
2015-12-08 | 0 | 2 | |
OpenBookQA |
Benchmark
|
2018-09-08 | 2 | 1 | |
Physical Interaction: Question Answering |
Benchmark
|
2019-11-26 | 2 | 0 | |
Qwen2.5-Omni |
Model
|
2025-03-27 | 0 | 4 | |
WinoGrande |
Benchmark
|
2019-07-24 | 2 | 0 | |
xAI |
Organization
|
2023-03-09 | 0 | 2 |