LLM-eval-survey
This resource provides an in-depth review of diverse evaluation methods for large language models (LLMs), covering aspects like natural language processing and reasoning abilities. It features academic papers and projects assessing the robustness, ethics, and trustworthiness of LLMs. Regular updates ensure the most recent insights, with an open invitation for contributions to further refine the survey.