#large language model

Logo of LangChain-ChatGLM-Webui
LangChain-ChatGLM-Webui
The LangChain-ChatGLM-Webui project provides a WebUI utilizing LangChain and ChatGLM-6B series models for applications grounded in local knowledge. Supporting multiple text file formats like txt, docx, md, and pdf, it includes models such as ChatGLM-6B and Belle for enhanced embedding functionalities. Designed for real-world AI model implementation, the project offers online accessibility through HuggingFace, ModelScope, and AIStudio. Compatible with Python 3.8.1+, it facilitates straightforward deployment. Continuous updates and community engagement ensure its dynamic advancement, inviting developer participation without exaggerated claims.
Logo of PIXIU
PIXIU
This initiative presents a structured framework for developing, fine-tuning, and evaluating Large Language Models (LLMs) aimed at the financial sector. It grants open access to financial LLMs, instructional tuning datasets, and comprehensive datasets across diverse tasks, promoting transparency and collaborative research efforts. The project emphasizes multi-task capabilities and accommodates multi-modal financial data, enhancing understanding and prediction within the field of financial NLP.
Logo of awesome-recommend-system-pretraining-papers
awesome-recommend-system-pretraining-papers
This paper list investigates the latest advancements in pretrained recommendation models, highlighting large language models and novel methods in sequence representation and user modeling. It provides a comprehensive overview with various datasets and studies, encouraging community collaboration through open contributions. Read up on significant studies presented at conferences like SIGIR, CIKM, and WSDM, and explore innovative techniques including graph pretraining and generative recommendations, under the guidance of Xiangyang Li from Peking University.
Logo of ChatGLM-6B
ChatGLM-6B
Explore a bilingual conversational AI model designed for deployment on consumer-grade hardware. Featuring 62 billion parameters and utilizing INT4 quantization, it allows local deployment with as little as 6GB memory. Tailored for Chinese question-answering, the model is enhanced through extensive bilingual training and reinforcement learning. The project supports academic research with open access to its weights and offers customization options for developers. Importantly, it stresses responsible use by encouraging compliance with open-source licenses and ethical standards.