#RAG
quivr
Leveraging generative AI, this tool provides a robust RAG framework that is fast, efficient, and customizable. It is capable of integrating with various LLMs and supports multiple file types, facilitating enhanced workflows through internet search capabilities and additional tools. The quivr-core allows easy file ingestion and inquiry, designed to optimize focus on product development. This tool is ideal for those seeking innovative approaches to data retrieval and management.
trulens
TruLens provides tools for evaluating and tracking language model experiments, enabling developers to improve their applications by analyzing performance through detailed methods. Integrate feedback functions, explore the RAG Triad, and perform honest and helpful evaluations. Easily connect tools and logging in your development workflow, and iterate through app versions using an intuitive interface. Start by installing TruLens via PyPI and follow the quick guide to evaluate RAGs from scratch. Join the community for support and collaboration.
FlagEmbedding
Explore a comprehensive toolkit of embedding models aimed at enhancing retrieval-augmented language models. Includes models for inference, fine-tuning, and evaluation across multiple languages. Achieve high performance in text and image retrieval with models like OmniGen for image generation. Follow the latest innovations such as MemoRAG and lightweight rerankers for efficiency. Access community support for continuous updates. Improve your NLP projects with easy installation and integration via the FlagEmbedding platform.
chatbot
The MongoDB Chatbot Framework allows developers to build intelligent chatbots efficiently using MongoDB and Atlas Vector Search. It offers flexible customization, built-in enhancements, and advanced features like retrieval-augmented generation (RAG). Developers can quickly move from prototype to production, integrating various AI models and strategies. Extensive documentation and tutorials support the development of scalable, data-driven chatbots using Atlas Vector Search and ChatGPT API.
llm-zoomcamp
A free 10-week online course teaching the creation of AI systems with Language Learning Models (LLMs) for practical applications. No prior AI or ML experience is needed. Participants can learn hands-on with Docker, Python, OpenAI API, and Elasticsearch, covering topics like vector databases and AI evaluation metrics. The course features workshops, project examples, and community interactions via Slack and Telegram, making it perfect for programmers and tech enthusiasts.
Streamer-Sales
Streamer-Sales uses an AI-powered model for enhanced product sales via livestream presentations. It generates tailored narratives that highlight product features, improving purchase intent. Features include instant copywriting, accelerated inference, voice-to-text, digital human creation, and real-time interaction. Seamlessly integrates with ASR, TTS, and RAG technologies, optimizing sales for live or offline environments.
talk2arxiv
Talk2Arxiv is an open-source tool for enhancing interaction with academic papers through advanced PDF parsing and effective retrieval methods. It's built using cutting-edge algorithms for text analysis and robust frontend frameworks, ensuring an efficient user experience. Upcoming features aim to expand content extraction capabilities, further enhancing research access.
llm-books
This detailed guide explores AI application development with open source tools, focusing on large language models, LangChain basics, and practical implementations. It covers additional topics such as LlamaIndex, HuggingGPT, and LLMOps, alongside the latest updates in LLM application evaluation, RAG series, and major domestic API interpretations. An ideal resource for those interested in enhancing their AI knowledge and engaging with a collaborative learning community.
ChatGPTCLIBot
Explore the capabilities of running GPT models in the command line environment using extended context memory and customizable prompts. This CLI bot provides extensive memory through embeddings, supports custom documents for Q&A, and facilitates smooth operation with its customizable prompts and real-time streaming responses. Compatible with Windows, Linux, and macOS, it offers features like undo, reset, and management of chat history, catering to the needs of both LLM enthusiasts and professionals.
LARS
LARS is a locally executable application for Large Language Models (LLMs) that offers advanced citation features to enhance response accuracy. Utilizing Retrieval Augmented Generation (RAG) technology, it reduces AI inaccuracies by basing responses on user-uploaded documents, including detailed citations such as document names and page numbers. Supporting formats like PDFs and Word files, LARS provides a built-in document reader and customizable settings, making it ideal for a wide range of tasks.
casibase
Casibase is an open-source RAG knowledge database offering web UI and enterprise SSO, supporting AI models such as OpenAI, Azure, LLaMA, and Google Gemini. Its user-friendly interface and robust backend provide enhanced AI functionalities for businesses, facilitating integration and management. Discover its features through online demos or initiate setup via casibase.org. Connect with the community on Discord for support and collaboration.
codefuse-chatbot
Discover an AI-powered assistant designed to enhance the software development lifecycle. Developed by the Ant Group's CodeFuse team, this open-source project utilizes Multi-Agent coordination, offering tools and libraries for executing complex DevOps tasks. It supports customizable knowledge bases, facilitates code and document analysis, and is compatible with small-scale DevOps models. The solution enables innovative approaches to development and operational issues, with support for both offline model deployment and the OpenAI API.
BCEmbedding
BCEmbedding, developed by Youdao, offers a robust bilingual and crosslingual framework designed for efficient semantic vector generation and retrieval. Featuring both EmbeddingModel and RerankerModel, it is optimized for accurate search results and improved ranking. As a key component in Retrieval Augmented Generation (RAG), BCEmbedding supports languages such as Chinese, English, Japanese, and Korean, catering to sectors like education, law, and finance. It integrates effectively with LangChain and LlamaIndex platforms, delivering comprehensive embeddings and reranking solutions. The models have shown consistent performance across various linguistic settings and are validated by evaluations and Youdao's product use cases.
MaxKB
MaxKB is an open-source Q&A system using large language models and RAG to optimize enterprise knowledge, customer service, and educational interactions. The system allows for effortless document integration and intelligent responses with reduced errors. It is compatible with various local and global models and includes a robust workflow engine for complex AI tasks, ensuring seamless system integration without coding.
Play-with-LLMs
Explore effective methods for training and evaluating large language models through practical examples involving RAG, Agent, and Chain applications. Understand the use of Mistral-8x7b and Llama3-8b models with techniques such as CoT and ReAct Agents, transformers, and adaptations for specific languages like Chinese. The article offers comprehensive insights into pretraining, fine-tuning, and RLHF processes, supported by practical case studies. Ideal for those interested in model quantization and deployment on platforms such as Huggingface.
Advanced_RAG
Discover the advanced use of Retrieval-Augmented Generation (RAG) with the Langchain framework for Python, designed to enhance the understanding capabilities of Large Language Models (LLMs). This resource provides insights into essential components such as the Multi Query Retriever and discusses advanced techniques including Agentic RAG variants, which improve contextual knowledge and generate more accurate responses. Explore methods for query transformation, routing to data sources, and indexing within VectorDBs to refine retrieval processes. Understand various RAG systems, ranging from basic models to complex self-reflective and corrective agentic processes, aimed at achieving adaptability and precision in language generation.
cookbook
Mistral Cookbook offers practical examples and tutorials from users and partners that highlight the diverse applications of Mistral models, including model evaluations and embedding techniques. Contributions focus on clarity, originality, and are reproducible, adding value to the community. Accepts submissions in .md or .ipynb formats with Colab-compatible examples, fostering a collaborative learning environment.
Feedback Email: [email protected]