Chinese-Mixtral
Discover Mixtral models tailored for Chinese language with enhanced architecture for effective long-text processing. The collection includes a base model amplified for Chinese and an Instruct model for interactive tasks. With native 32K context length, extendable to 128K, the models are ideal for tasks needing deep context, such as math reasoning and code generation. Offering open-source scripts for training and fine-tuning, users can easily adapt or develop custom models. Streamlined for integrations like transformers and llama.cpp, it facilitates quantification and deployment on local devices.