Skip to content

smartliuhw/eval

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Intro

This repo is used for @smartliuhw thesis's model evaluation. The EleutherAI/lm-evaluation-harness is used as the basic framework.

The code was running on the RTX 4090 with 24G GPU-memory with accelarate package to enable data parallel.

How to use

Install the dependency

Enter the path which contanins the README file, then run the following command:

pip install -e .

Modify task configurations

All the task configurations are in this path. Enter it and modify the task's configuration you need.

An example is in nq_open_cot.yaml file, in which I customized the dataset path, task group, descriptions, input template and metrics. Also, the utils.py is modified to adapt to the special dataset. It is recommanded to save the dataset locally to save precious time.

Modify evaluation script

After customized the task you need, a shell script is needed to launch the evaluation. An example is in eval_test.sh file. Only a few params are needed to be changed.

If you have any question, feel free to ask me. And it's recommanded to read the framework's origin README to gain a better understanding about this framework.

About

personal code of evaluation

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages