MMTU leaderboard

|πŸ€— Dataset |πŸ› οΈGitHub |πŸ†Leaderboard|πŸ“– Paper |

Tables and table-based use cases play a crucial role in many real-world applications, such as spreadsheets, databases, and computational notebooks, which traditionally require expert-level users like data engineers, analysts, and database administrators to operate. Although LLMs have shown remarkable progress in working with tables, comprehensive benchmarking of such capabilities remains limited, often narrowly focusing on tasks like NL-to-SQL and Table-QA, while overlooking the broader spectrum of real-world tasks that professional users face today.

We introduce MMTU, a large-scale benchmark with over 30K questions across 25 real-world table tasks, designed to comprehensively evaluate models ability to understand, reason, and manipulate real tables at the expert-level. These tasks are drawn from decades' worth of computer science research on tabular data, with a focus on complex table tasks faced by professional users. We show that MMTU require a combination of skills -- including table understanding, reasoning, and coding -- that remain challenging for today's frontier models, where even frontier reasoning models like OpenAI o4-mini and DeepSeek R1 score only around 60%, suggesting significant room for improvement. Our evaluation code is available at GitHub.

{
  • "headers": [
    • "Model",
    • "Model type",
    • "Model size",
    • "Overall",
    • "Data Source",
    • "Date"
    ],
  • "data": [
    • [
      • "o4-mini (2024-11-20)",
      • "Reasoning",
      • "unknown",
      • 0.639,
      • "MMTU",
      • "2025-06-10 00:00:00"
      ],
    • [
      • "Deepseek-R1",
      • "Reasoning",
      • "unknown",
      • 0.596,
      • "MMTU",
      • "2025-06-10 00:00:00"
      ],
    • [
      • "Deepseek-V3",
      • "Chat",
      • "unknown",
      • 0.517,
      • "MMTU",
      • "2025-06-10 00:00:00"
      ],
    • [
      • "GPT-4o (2024-11-20)",
      • "Chat",
      • "unknown",
      • 0.491,
      • "MMTU",
      • "2025-06-10 00:00:00"
      ],
    • [
      • "Llama-3.3-70B",
      • "Chat",
      • "70B",
      • 0.438,
      • "MMTU",
      • "2025-06-10 00:00:00"
      ],
    • [
      • "Mistral-Large-2411",
      • "Chat",
      • "123B",
      • 0.43,
      • "MMTU",
      • "2025-06-10 00:00:00"
      ],
    • [
      • "Mistral-Small-2503",
      • "Chat",
      • "70B",
      • 0.402,
      • "MMTU",
      • "2025-06-10 00:00:00"
      ],
    • [
      • "GPT-4o-mini (2024-07-18)",
      • "Chat",
      • "unknown",
      • 0.386,
      • "MMTU",
      • "2025-06-10 00:00:00"
      ],
    • [
      • "Llama-3.1-8B",
      • "Chat",
      • "8B",
      • 0.259,
      • "MMTU",
      • "2025-06-10 00:00:00"
      ]
    ],
  • "metadata": null
}