- 📅 Sep 18, 2025: EvaLearn was accepted to the NeurIPS 2025 main track with a high score of 5/5/5/5! 🎉
- 📅 Jul 15, 2025: We've released a new version! 🎉 Open-sourced complete Chinese rubrics, updated Chinese README documentation, and optimized evaluation scripts for improved efficiency and accuracy.
- 📅 Jun 5, 2025: EvaLearn is officially open-sourced! 🚀 We released this innovative benchmark for evaluating the learning capability and efficiency of large language models.
EvaLearn is a benchmark designed to evaluate large language models (LLMs) on their learning capability and efficiency. It contains 648 challenging problems across six task types, grouped into 182 sequences. Unlike traditional benchmarks that evaluate models in parallel, EvaLearn requires models to solve problems sequentially, allowing them to leverage experience from previous solutions.
The EvaLearn evaluation framework consists of:
- A streamlined sequential evaluation tool (
Evaluate/evaluate.py) that processes sequences of questions - A dataset of problem definitions (
Dataset/EvaLearn_Problem.json) - A dataset of sequence definitions (
Dataset/EvaLearn_Sequence.json) - A metrics evaluation tool (
Evaluate/evaluate_metric.py) for analyzing results
git clone https://github.com/YOUR_USERNAME/EvaLearn.git
cd EvaLearn
pip install -r requirements.txtRun the evaluation:
python Evaluate/evaluate.py --input Dataset/EvaLearn_Problem.json \
--seq Dataset/EvaLearn_Sequence.json \
--output results.json \
--workers 4 \
--client-api-key YOUR_CLIENT_API_KEY \
--judge-api-key YOUR_JUDGE_API_KEY| Argument | Description |
|---|---|
--input |
Path to the problem JSON file |
--seq |
Path to the sequence JSON file |
--output |
Path to save the evaluation results |
--workers |
Number of worker threads for parallel processing |
--no-check-empty |
Skip checking for empty responses |
--judge-api-key |
API key for the judge model |
--client-api-key |
API key for the client model |
--judge-model |
Model to use for judging (default: "gpt-4o-2024-11-20") |
--client-model |
Model to use for client responses (default: "gpt-4o-2024-11-20") |
--judge-api-base-url |
Custom base URL for judge API calls |
--client-api-base-url |
Custom base URL for client API calls |
- Checkpoint Recovery: Automatically resumes interrupted evaluations
- API Compatibility: Support for custom API endpoints
- Parallel Processing: Multi-threaded execution for faster processing
from Evaluate.evaluate import sequentialEval
sequentialEval(
input_json_path="Dataset/EvaLearn_Problem.json",
seq_json_path="Dataset/EvaLearn_Sequence.json",
output_json_path="results.json",
client_api_key="YOUR_CLIENT_API_KEY",
judge_api_key="YOUR_JUDGE_API_KEY"
)Use Evaluate/evaluate_metric.py to compute learning metrics from your results:
python Evaluate/evaluate_metric.py --results results.json --output report.json- Overall sequence accuracy
- Position-wise Accuracy
- Slope of fitted accuracy curve
- Average position of first correct solution
- Average number of consecutive correct solutions
- Post-warmup Accuracy
For detailed metric descriptions, please refer to the Section 2.3 of the paper.
Your results should be in a JSON file, where each item contains at least:sequence_id: Unique identifier for a sequence
position_in_sequence: Position (1-based) of the problem in the sequencetype: (Optional) Task type/categorygpt4judge: String containing a JSON with ananswer_scorefield
python Evaluate/evaluate_metric.py --results <results.json> [--problems 7] [--warmup 3] [--output <report.json>]--results: Path to your results JSON file (required)--problems: Number of problems per sequence (default: 7)--warmup: Number of initial problems to exclude for post-warmup accuracy (default: 3)--output: Path to save the report as JSON (default:report_<results.json>)
- Prints a summary of all metrics to the console, including:
- Overall metrics
- Position-wise accuracy
- Metrics by task type
- Saves a detailed report as a JSON file (if
--outputis specified).
python Evaluate/evaluate_metric.py --results my_eval_results.json --problems 7 --warmup 3 --output my_report.json- Logs are saved to
evaluation_metrics.logand also printed to the console.
Each problem in Dataset/EvaLearn_Problem.json has the following structure:
{
"id": 1,
"type": "Logical Reasoning",
"source": "LogicGame-crypto_puzzle",
"level": 1,
"prompt": ["The question text that will be presented to the model"],
"rubric_zh": "用于判断模型回答质量的中文评分标准",
"rubric_en": "English evaluation criteria used by the judge model",
"canonical_answer": "The expected correct answer"
}| Field | Description |
|---|---|
id |
Unique identifier for the problem |
type |
Category of the problem (e.g., "Logical Reasoning", "Mathematical Reasoning") |
source |
Origin of the problem |
level |
Difficulty level |
prompt |
The question text (can be a string or an array of strings) |
rubric_zh |
Chinese evaluation criteria used by the judge model |
rubric_en |
English evaluation criteria used by the judge model |
canonical_answer |
The expected correct answer |
Note: The results in our paper use the Chinese rubric, which was carefully annotated by our annotation team and is of high quality. The English version was translated using a large language model to help understand the meaning of the rubric. Therefore, we strongly recommend that everyone use the Chinese rubric for evaluation. We will also update it with a high-quality English rubric in the future.
Each sequence in Dataset/EvaLearn_Sequence.json has the following structure:
{
"sequence_id": 1,
"type": "Extraction",
"question_ids": [252, 258, 297, 263, 245, 273, 241]
}| Field | Description |
|---|---|
sequence_id |
Unique identifier for the sequence |
type |
Category of the sequence (e.g., "Extraction", "Logical Reasoning") |
question_ids |
Ordered list of problem IDs that form the sequence |
The main evaluation function that processes sequences of questions with checkpoint recovery and API flexibility.
sequentialEval(
input_json_path,
seq_json_path,
output_json_path,
worker_nums=None,
check_empty=True,
judge_api_key=None,
client_api_key=None,
judge_model="gpt-4o-2024-11-20",
client_model="gpt-4o-2024-11-20",
judge_api_base_url=None,
client_api_base_url=None
)Parameters:
input_json_path: Path to the problem JSON fileseq_json_path: Path to the sequence JSON fileoutput_json_path: Path to save evaluation resultsworker_nums: Number of worker threads (default: 5)check_empty: Whether to check and reprocess empty responses (default: True)judge_api_key: API key for judge modelclient_api_key: API key for client modeljudge_model: Model name for judging (default: "gpt-4o-2024-11-20")client_model: Model name for responses (default: "gpt-4o-2024-11-20")judge_api_base_url: Custom base URL for judge APIclient_api_base_url: Custom base URL for client API
Processes a sequence of questions with inference and judging.
Processes a batch of sequences in parallel with thread-safe result saving.
Identifies sequences with empty responses for reprocessing.
Handle JSON file loading and saving with error handling and backup mechanisms.
Generates formatted history prompts from previous interactions.
The script uses a CONFIG dictionary for various parameters:
max_retries: Maximum API call retries (default: 10)initial_delay: Initial retry delay in seconds (default: 1)max_delay: Maximum retry delay in seconds (default: 60)worker_nums: Default number of worker threads (default: 5)questions_per_sequence: Expected questions per sequence (default: 7)
This project is licensed under the Apache-2.0 License - see the LICENSE file for details.
Shihan Dou: [email protected]
Ming Zhang: [email protected]
We gratefully acknowledge the significant contributions made by the annotation teams at ByteDance, whose diligent work was essential to the success of this paper ❤️❤️. The core members of the annotation team include Di Cheng, Linhua Deng, Yanxi Fu, Yafei Qiao, Chaoqian Ren, Mei Su, Ying Wu, Baitong Yang, and Xingyu Zhu.
We also wish to express our sincere appreciation to an undisclosed third-party annotation company ❤️❤️ for their substantial support in data annotation. Finally, we would like to thank all individuals who participated in and supported this project for their valuable input.
@article{dou2025evalearn,
title={EvaLearn: Quantifying the Learning Capability and Efficiency of LLMs via Sequential Problem Solving},
author={Dou, Shihan and Zhang, Ming and Huang, Chenhao and Chen, Jiayi and Chen, Feng and Liu, Shichun and Liu, Yan and Liu, Chenxiao and Zhong, Cheng and Zhang, Zongzhang and others},
journal={arXiv preprint arXiv:2506.02672},
year={2025}
}
