recurrent-memory-transformer
The Recurrent Memory Transformer (RMT) enhances AI model performance by using memory-augmented segment-level transformers. Designed for Hugging Face models, it incorporates special tokens to optimize memory and sequence processing. Features include comprehensive training examples, gradient accumulation, and metric management. It supports tasks with extensive context requirements and is developed in partnership with DeepPavlov.ai, AIRI, and London Institute for Mathematical Sciences.