#llama.cpp

Logo of gguf-tools
gguf-tools
The gguf-tools library facilitates the manipulation and documentation of GGUF files, essential in the local machine learning landscape. The tools include utilities for displaying detailed GGUF information, analyzing tensor differences between models, and examining tensor weights. Although in active development, the project provides real-world applications, with experimental features still under refinement. While user-friendly and well-documented, its current limitations include missing quantization formats. Discover its potential applications in 'llama.cpp' and other machine learning projects.
Logo of CASALIOY
CASALIOY
Explore an effective toolkit for air-gapped LLMs featuring LangChain and Qdrant. It supports local data processing and query handling without internet requirement, facilitating diverse dataset ingestion and robust GUI interaction. Includes GPU support and model conversion from GGML for optimal use.
Logo of llama_ros
llama_ros
Discover how llama_ros enables the integration of llama.cpp's optimization features into ROS 2 projects. It supports GGUF-based LLMs, VLMs, real-time LoRA changes, and GBNF grammars, enhancing robotic applications. The repository includes detailed installation guides, Docker options, and usage examples. Enhance ROS 2 functionality with CUDA support and other tools offered by llama_ros, suitable for expanding project capabilities with LangChain and related demos.
Logo of llama-cpp-python
llama-cpp-python
Discover efficient Python bindings for llama.cpp, providing C API access and high-level APIs for text completion. Features include OpenAI-compatible servers, model variety, and integration with LangChain and LlamaIndex. The package is installable on Python 3.8+ with any C compiler, supporting hardware acceleration. Comprehensive documentation aids integration, making it ideal for both standard and complex AI applications.