tevatron
Designed for scalable neural retrieval, this toolkit facilitates efficient model training and inference. It integrates parameter-efficient methods such as LoRA and advanced technologies like DeepSpeed and flash attention. Users can access and finetune top pre-trained models, including BGE-Embedding and Instruct-E5, via HuggingFace. Self-contained datasets support various tasks, ensuring efficient training on billion-scale LLMs with GPUs and TPUs. This makes it an excellent choice for researchers seeking to enhance retrieval systems using sophisticated techniques.