#pytorch

Logo of Deep-reinforcement-learning-with-pytorch
Deep-reinforcement-learning-with-pytorch
This repository is a resource for deep reinforcement learning, including both classic and advanced algorithms implemented in PyTorch. It provides code to aid understanding and experimentation, featuring models like DQN, PPO, and A3C. Actively maintained, the project plans to expand with new algorithms. It supports environments such as CartPole and BipedalWalker and includes installation guides, dependencies, and links to academic papers.
Logo of pytorch-sentiment-neuron
pytorch-sentiment-neuron
This open-source project leverages PyTorch, CUDA, and Python 3.5 for sentiment analysis by generating and analyzing sentiments in reviews. It supports model implementation, visualization, and retraining with adjustable parameters including sequence length, batch size, and RNN setup, providing a flexible framework for developers to explore sentiment analysis.
Logo of fvcore
fvcore
fvcore is a compact library from FAIR's computer vision team, offering essential utilities for frameworks like Detectron2, PySlowFast, and ClassyVision. It includes PyTorch layers and tools for flop counting, parameter counting, BatchNorm statistics recalibration, and hyperparameter scheduling. Installation can be easily done through PyPI, Anaconda Cloud, GitHub, or a local clone. With support for Python 3.6 and above, it offers type-annotated, tested, and benchmarked components for reliable performance.
Logo of zero_nlp
zero_nlp
The project delivers a versatile framework for Chinese NLP tasks, utilizing PyTorch and Transformers. It includes comprehensive training and fine-tuning solutions for a variety of models such as text-to-vector and multimodal. Abundant open-source training data ensures easy setup with advanced processing methods suitable for large datasets. Models supported include GPT2, CLIP, and GPT-NeoX among others, offering multi-GPU training and deployment capabilities. Discover tutorials for model modification and explore a wide range of pretrained and custom models for diverse NLP needs.
Logo of flops-counter.pytorch
flops-counter.pytorch
The tool offers an accurate calculation of multiply-add operations and parameters in neural networks through PyTorch and ATEN backends. It provides detailed per-layer cost analysis, especially effective when using the ATEN backend for comprehensive support including transformer models. Key features include per-module statistics, detailed operation logs, and module exclusion options, accommodating complex research requirements. Supporting convolutional layers, activations, RNNs, and transformer architectures, the tool serves researchers and developers in assessing neural network complexities.
Logo of bayesian-flow-networks
bayesian-flow-networks
Discover Bayesian Flow Networks designed for effective modeling of continuous and discrete data. This project offers flexible loss functions, versatile probability models, and essential scripts for tasks such as training and testing. Experience practical experiments including MNIST and CIFAR-10, enhanced by PyTorch integration. It ensures reproducibility and offers integration options for advanced data analysis, benefiting researchers and practitioners in data-centric domains.
Logo of einops
einops
Einops offers a consistent API for tensor operations across platforms such as numpy, pytorch, and tensorflow, focusing on operations like rearrange, reduce, and repeat. Recent enhancements include framework support, functionality improvements, and features such as EinMix for linear layers. Its semantic clarity and uniform behavior facilitate intuitive tensor manipulation, independent of the framework.
Logo of rvc-tts-pipeline
rvc-tts-pipeline
The rvc-tts-pipeline facilitates high-quality TTS by replicating the original speaker's voice through TTS and RVC model integration. It allows the conversion of audio files from TTS systems like tortoise or vits into outputs that sound more authentic with trained RVC weights. PyTorch installation is a prerequisite for RVC's operation. As the package is still under development, some bugs may be present. The provided installation and usage instructions simplify the setup, enabling the use of the 'rvc_convert' function for conversions. Acknowledgment is given to RVC creators for their indispensable code.
Logo of FastDiff
FastDiff
FastDiff offers a PyTorch implementation of a quick conditional diffusion model for high-fidelity speech synthesis, with pretrained models and dataset support including LJSpeech, LibriTTS, and VCTK. It features multi-GPU support and guidance for text-to-speech synthesis using advanced methods like ProDiff and Tacotron. The project ensures ease of integration with well-documented instructions while emphasizing ethical standards for voice usage.
Logo of Tacotron-pytorch
Tacotron-pytorch
Discover the Pytorch implementation of the Tacotron model, a thorough end-to-end text-to-speech synthesis method. Utilizing the LJSpeech dataset, the project details steps from data preprocessing to audio synthesis. Aimed at researchers and developers in TTS technology, it allows hyperparameter adjustments to efficiently convert text to speech. Features include encoder, decoder, and post-processing networks essential for speech generation. The project is in early development stages, providing sample outputs and inviting community feedback for ongoing enhancement.
Logo of LyCORIS
LyCORIS
LyCORIS introduces efficient fine-tuning techniques for Stable Diffusion models, optimizing image generation without compromising model size or training speed. Incorporating methods like LoRA, LoHa, and DyLoRA, it ensures versatile and diverse output generation. It is compatible with platforms such as sd-webui, ComfyUI, and InvokeAI, facilitating easy integration. The project is backed by detailed documentation and dynamic community interaction. Recent updates include automatic detection of quantized layers and improved functional APIs, offering adaptable solutions for users.
Logo of glow
glow
Glow is a machine learning compiler designed for hardware accelerators, providing integration with high-level frameworks. It enhances code generation for neural network graphs and employs compiler optimizations. Glow's lowering phase supports diverse input operators and hardware targets. The project is continuously developed with industry partners and runs on macOS and Linux with support for a modern C++ compiler. Comprehensive documentation and examples are available for integration and testing.