Awesome-Efficient-LLM
Discover a curated list of cutting-edge research papers on improving the efficiency of Large Language Models (LLMs) through methods such as network pruning, knowledge distillation, and quantization. This resource provides insights into accelerating inference, optimizing architectures, and enhancing hardware performance, offering valuable information for both academic and industry professionals.