🤖 LLM Benchmark Dashboard

Latest Open Source Models (April 2026) — Optimized for 64GB RAM

📊 Benchmark Explanations

MMLU (Massive Multitask Language Understanding)

Tests knowledge across 57 subjects including math, history, law, medicine, and more. Higher is better. 80%+ = excellent general knowledge, 70-80% = good, below 70% = average.

HumanEval

Measures code generation ability using 164 programming problems. Higher is better. 80%+ = excellent coder, 70-80% = good programming skills, below 70% = basic coding ability.

Total Models

0

CPU Compatible

0

New This Month

0

Best Overall

Gemma 4
Model
Size
RAM
VRAM
MMLU
HumanEval
Speed
Type
Compare

MMLU Score by Model

HumanEval Score by Model

RAM Requirements

Models by Organization

⚖️ Model Comparison

Select models from the table view to compare them here. Best values highlighted in green.

🙏 Open Source Credits