C-Eval
A Multi-Level Multi-Discipline Chinese Evaluation
Suite for Foundation Models
(2023)
C-Eval is a comprehensive Chinese evaluation suite for foundation models. It consists of 13948 multi-choice questions spanning 52 diverse disciplines and four difficulty levels, as shown below. You may explore our dataset examples at Explore, or check our paper for more details.
Our data can be directly downloaded on Huggingface datasets. Please refer to our github instructions for how to read and use the data.
@inproceedings{huang2023ceval, title={C-Eval: A Multi-Level Multi-Discipline Chinese Evaluation Suite for Foundation Models}, author={Huang, Yuzhen and Bai, Yuzhuo and Zhu, Zhihao and Zhang, Junlei and Zhang, Jinghan and Su, Tangjun and Liu, Junteng and Lv, Chuancheng and Zhang, Yikai and Lei, Jiayi and Fu, Yao and Sun, Maosong and He, Junxian}, booktitle={Advances in Neural Information Processing Systems}, year={2023} }
Have any questions about C-Eval? Please contact us at ceval.benchmark@gmail.com or create an issue on Github. For potential collaboration, please contact junxianh@cse.ust.hk.