Project Icon

ChatGLM-Efficient-Tuning

Optimize Language Models Using Advanced Fine-Tuning Techniques

Product DescriptionThe project implements advanced fine-tuning techniques for the ChatGLM-6B model, including LoRA, P-Tuning V2, and Reinforcement Learning with Human Feedback (RLHF). It features a comprehensive Web UI for single GPU-based training, evaluation, and inference, highlighting its role in optimizing large language models. The repository supports various datasets like Stanford Alpaca, BELLE, and GPT-4 generated data, enhancing ChatGLM's adaptability to diverse datasets and tuning methods. Although the project is no longer actively maintained, it has significantly contributed to the efficient tuning of language models.
Project Details