proger / mamba-cpu
☆11Updated 8 months ago
Related projects: ⓘ
- A fast RWKV Tokenizer written in Rust☆34Updated 2 weeks ago
- PyTorch video decoding☆47Updated last week
- Hugging Face Inference Toolkit used to serve transformers, sentence-transformers, and diffusers models.☆34Updated 3 weeks ago
- Github repo for Peifeng's internship project☆12Updated 10 months ago
- A toolkit enhances PyTorch with specialized functions for low-bit quantized neural networks.☆24Updated 2 months ago
- 完全依靠ChatGPT生成数据微调的西式翻译腔聊天风格中文大模型☆16Updated 5 months ago
- XVERSE-MoE-A36B: A multilingual large language model developed by XVERSE Technology Inc.☆31Updated last week
- A library for simplifying fine tuning with multi gpu setups in the Huggingface ecosystem.☆15Updated 3 months ago
- ☆40Updated 2 months ago
- Improving Text Embedding of Language Models Using Contrastive Fine-tuning☆54Updated last month
- First-principle implementations of various AI algorithms using a wide range of deep learning frameworks, accompanied by relevant research…☆24Updated this week
- Trying to deconstruct RWKV in understandable terms☆14Updated last year
- A public implementation of the ReLoRA pretraining method, built on Lightning-AI's Pytorch Lightning suite.☆33Updated 6 months ago
- Implementation of mamba with rust☆69Updated 6 months ago
- ☆42Updated 3 weeks ago
- Zeta implementation of a reusable and plug in and play feedforward from the paper "Exponentially Faster Language Modeling"☆15Updated last week
- Block Transformer: Global-to-Local Language Modeling for Fast Inference (Official Code)☆118Updated 2 weeks ago
- A byte-level decoder architecture that matches the performance of tokenized Transformers.☆57Updated 4 months ago
- ☆22Updated 3 months ago
- Implementation of the Mamba SSM with hf_integration.☆55Updated 3 weeks ago
- Evaluation of bm42 sparse indexing algorithm☆60Updated 2 months ago
- A transformer-based multimodal model for music.☆27Updated last month
- ☆50Updated 3 months ago
- Download full or partial git-lfs repos without temporarily using 2x disk space☆30Updated 11 months ago
- implementation of https://arxiv.org/pdf/2312.09299☆19Updated 2 months ago
- Fast LLM Training CodeBase With dynamic strategy choosing [Deepspeed+Megatron+FlashAttention+CudaFusionKernel+Compiler];☆32Updated 8 months ago
- Here we will test various linear attention designs.☆55Updated 4 months ago
- The simplest, fastest repository for training/finetuning medium-sized xLSTMs.☆38Updated 3 months ago
- Data preparation code for CrystalCoder 7B LLM☆42Updated 4 months ago
- QuIP quantization☆41Updated 6 months ago