#ControlNet
ArtLine
ArtLine employs advanced AI, featuring Fast.AI and PyTorch, to convert photos into detailed line art. Utilizing ControlNet and self-attention, it enhances facial features, yet seeks refinement in handling backgrounds and shadows. Explore the seamless creation of art through the Colab demo.
ComfyUI_UltimateSDUpscale
Improve image upscaling projects by integrating ComfyUI with custom nodes from the Ultimate Stable Diffusion Upscale script. This setup provides options like tiled sampling and customizable samplers, allowing precise control over image dimensions and tile sizes through inputs such as 'upscale_by' and 'force_uniform_tiles'. Ideal for professionals wanting precise image processing control.
EditAnything
Leverage Segment Anything and ControlNet for dynamic image editing, enabling features like cross-image merging and comprehensive sketch transformations. Includes support for text-guided segmentation and customizable DreamBooth editing. Recent UI and performance upgrades enhance user experience.
sd-webui-segment-anything
The sd-webui-segment-anything extension connects AUTOMATIC1111 Stable Diffusion WebUI, Mikubill ControlNet, and tools like segment anything and GroundingDINO to enhance inpainting and semantic segmentation. It facilitates automated image matting and LoRA/LyCORIS dataset creation, offering features such as mask expansion, API access, and layout generation. Providing adaptability through constant updates, it supports SAM-HQ, MobileSAM, and various models, while being a resourceful tool with comprehensive guides and community support.
SeargeSDXL
This update brings advanced image processing capabilities to ComfyUI by incorporating the SDXL 1.0 with base and refiner checkpoints. It integrates features like FreeU v2, Controlnet, and Multi-LoRA into a unified workflow. The installation is simplified with a user-friendly script, and comprehensive documentation is provided. Performance improvements include a 20% increase in processing speed and enhanced image quality, making it a valuable resource for generating high-resolution images with detailed customization.
roomGPT
Discover an open-source AI application to creatively redesign room images. Simply upload a photo to see innovative room variations created using ControlNet on Replicate. Effortlessly clone and run the app locally with straightforward instructions, utilizing Bytescale for image storage. Ideal for those exploring AI in interior design without the need for authentication or payments.
StableVideo
StableVideo utilizes text prompts for consistency-aware video editing via diffusion techniques, allowing precise manipulation and maintaining coherence across frames in an efficient VRAM environment. Easy installation and integration options on GitHub or Hugging Face allow experiments with pre-trained models or custom video training using NLA. Suitable for research and creative projects, the interactive interface supports dynamic video editing and mask region control. Comprehensive documentation is available for enhancing editing capabilities with StableVideo.
zero123plus
Zero123++ v1.2 enhances multi-view image synthesis from a single input, emphasizing 3D generation and improved handling of camera settings. With the addition of a ControlNet normal generator, it achieves better mask accuracy. The model is easy to use with tools such as torch and diffusers, making it efficient for VRAM. Available under Apache 2.0 and CC-BY-NC 4.0 licenses, it can be accessed for non-commercial purposes on Hugging Face. Discover multi-view synthesis with included scripts and demos.
ComfyUI-InstantID
This unofficial adaptation of InstantID for ComfyUI provides powerful tools including pose reference for enhanced ID creation. Version 2.0 enables improved model management through automatic downloads from Huggingface Hub and local storage. Users can choose from various styles, apply InsightFace models, and enjoy compatibility with diverse GPUs. Enhanced code efficiency and new functionalities refine the image generation process, ensuring precise styling and consistent performance. A detailed installation guide facilitates quick setup, supported by comprehensive testing for assured reliability.
x-flux
This repository provides fine-tuning scripts for the Flux model, utilizing LoRA and ControlNet technologies. With support for high-resolution output through DeepSpeed integration, it enables training of models like the IP-Adapter and various ControlNet versions at 1024x1024 resolution. Necessary tools include Python 3.10+, PyTorch 2.1+, and HuggingFace CLI for downloading models. Testing is supported through ComfyUI, Gradio, and CLI, with a low-memory mode available using Flux-dev-F8 on HuggingFace. Models are under the FLUX.1 Non-Commercial License.
krita-ai-diffusion
The Krita plugin leverages cutting-edge generative AI to improve image creation and editing precision. By enabling image generation from text and refinement of existing designs, it offers flexible control through sketches and depth maps. The project supports open-source models for customization and local execution, with cloud options available for rapid setup. Its features, such as inpainting, live painting, and upscaling, enhance creative processes while maintaining high-quality outputs.
Radiata
Radiata offers an optimized Stable Diffusion WebUI, using TensorRT for improved performance. It includes features like Stable Diffusion XL and ControlNet plugin compatibility, and supports Lora & Lycoris for varied uses. Installation is straightforward for Windows and Linux. Visit the official documentation for more on its features and setup.
sd-webui-controlnet
This WebUI extension integrates ControlNet with Stable Diffusion to improve image generation without needing to merge models. It supports ControlNet 1.1 and includes features like high-resolution fixes, inpainting, and multi-control modes. The interface is user-friendly, enabling advanced options like pixel-perfect mode and reference-only control, applicable to various creative processes. Regular updates provide ongoing enhancements.
ebsynth_utility
Discover the features of the AUTOMATIC1111 UI extension for streamlined video creation using img2img and Ebsynth, eliminating the need for additional tools like After Effects. It supports integration with Controlnet, offering enhanced video editing capabilities including masking, background blending, and auto-tagging through deep learning. The extension showcases examples of color correction, facial expression changes, and dynamic application of Lora. Detailed installation and usage instructions are available, facilitating efficient creation of high-quality videos.
DiffSynth-Studio
DiffSynth Studio is an efficient diffusion engine supporting models like CogVideoX, FLUX, and Stable Diffusion. It offers tools for text-to-video and image synthesis customized for high-resolution outputs. Recent updates add FLUX ControlNet enhancements and advanced video synthesis models. Features include toon shading and stylization, with access via Python and WebUI.
stable-diffusion-webui-colab
Discover varied WebUI options available on Google Colab, including DreamBooth and LoRA trainer. The repository supports ‘lite’, ‘stable’, and ‘nightly’ builds, each offering distinct features and updates. Access step-by-step installation guides and direct links to various diffusion models like cyberpunk anime and inpainting, ensuring efficient WebUI operation with frequent updates.
Depth-Anything
Depth Anything employs large-scale datasets comprising over 63.5 million images to improve monocular depth estimation. Recognized by CVPR 2024, this method enhances depth prediction capabilities in relative and metric contexts. The project integrates features such as an optimized depth-conditioned ControlNet and supports scene understanding. With the release of Depth Anything V2 and its integration with platforms like Hugging Face, the project offers accessible tools for enhancing depth perception technologies.
facechain
FaceChain is a cutting-edge framework for generating identity-preserved human portraits using advanced AI technology. It generates high-quality and diverse portraits quickly with just one photo, supporting multiple styles and seamless integration with tools such as ControlNet and LoRAs. The latest version enhances speed, image fidelity, and style retention, providing improved control and compatibility. Compatible across platforms including Python scripts, Gradio interface, and sd webui, FaceChain represents the forefront of image generation and personalization, with its methodologies and developments showcased at top conferences like NeurIPS and CVPR.
Real-Time-Latent-Consistency-Model
The Real-Time Latent Consistency Model uses advanced techniques to optimize the image-to-image transformation process. It pairs the Latent Consistency Model (LCM) with ControlNet for fast, efficient inference, leveraging pipelines like ControlNet Canny and LCM-LoRA. This model is compatible with CUDA and Python, and optimized for Apple silicon for enhanced performance. With Docker setup for effective resource management, it suits applications needing real-time rendering, offering various pipelines for image-to-image and text-to-image transformations, ensuring flexible, high-quality outcomes.
axodox-machinelearning
This project provides a complete C++ solution for Stable Diffusion image synthesis, eliminating the need for Python and enhancing deployment efficiency. It includes txt2img, img2img, inpainting functions, and ControlNet for guided generation. The library, optimized for DirectML, is aimed at real-time graphics and game developers, offering GPU-accelerated feature extraction. Prebuilt NuGet packages are available for seamless integration into Visual Studio C++ projects.
Feedback Email: [email protected]