AI Model Evaluation
LLMEval3
LLMEval is a large-scale model evaluation benchmark launched by the NLP Laboratory of Fudan University.
Tags:AI Model EvaluationAI TestLLMEval is a large-scale model evaluation benchmark launched by the NLP Laboratory of Fudan University. The latest LLMEval-3 focuses on professional knowledge and ability evaluation, covering 13 subject categories and more than 50 secondary disciplines designated by the Ministry of Education, including philosophy, economics, law, education, literature, history, science, engineering, agriculture, medicine, military science, management, art, etc., with a total of about 200000 standard generative question answering questions.
data statistics
Relevant Navigation
No comments...