AI Model Evaluation

LLMEval3

LLMEval is a large-scale model evaluation benchmark launched by the NLP Laboratory of Fudan University.

Tags:

LLMEval is a large-scale model evaluation benchmark launched by the NLP Laboratory of Fudan University. The latest LLMEval-3 focuses on professional knowledge and ability evaluation, covering 13 subject categories and more than 50 secondary disciplines designated by the Ministry of Education, including philosophy, economics, law, education, literature, history, science, engineering, agriculture, medicine, military science, management, art, etc., with a total of about 200000 standard generative question answering questions.

data statistics

Relevant Navigation

No comments

No comments...