Prompt templates for climate-specific tasks can be found under tasks
.
- Clone this repo
- Clone and install EleutherAI/lm-evaluation-harness (the big-refactor branch)
- The templates from this repo can be easily integrated with lm-evaluation-harness by either:
a. placing the tasks fromtasks
directory underlm_eval/tasks/
(in lm-evaluation-harness) and running evaluation using the following command:lm_eval lm-evaluation-harness/main.py \ --model hf \ --model_args pretrained=tiiuae/falcon-7b \ --tasks claim_binary \ --output_path /results/falcon-7b.jsonl \ --show_config --log_samples \ --num_fewshot 5
OR
b. by directly passing the paths to the tasks as command-line arguments using --include_path
. An example command is shown below:
lm_eval \
--model hf \
--model_args pretrained=tiiuae/falcon-7b \
--tasks claim_binary \
--output_path /results/falcon-7b.jsonl \
--show_config --log_samples \
--num_fewshot 5 --include_path <path/to/this/repo/tasks/exeter>
HuggingFace Link to Climate Evaluation Datasets | Paper Link
@misc{thulke2024climategpt,
title={ClimateGPT: Towards AI Synthesizing Interdisciplinary Research on Climate Change},
author={David Thulke and Yingbo Gao and Petrus Pelser and Rein Brune and Rricha Jalota and Floris Fok and Michael Ramos and Ian van Wyk and Abdallah Nasir and Hayden Goldstein and Taylor Tragemann and Katie Nguyen and Ariana Fowler and Andrew Stanco and Jon Gabriel and Jordan Taylor and Dean Moro and Evgenii Tsymbalov and Juliette de Waal and Evgeny Matusov and Mudar Yaghi and Mohammad Shihadah and Hermann Ney and Christian Dugast and Jonathan Dotan and Daniel Erasmus},
year={2024},
eprint={2401.09646},
archivePrefix={arXiv},
primaryClass={cs.LG}
}