catid / cuda_float_compressLinks
Python package for compressing floating-point PyTorch tensors
☆12Updated last year
Alternatives and similar repositories for cuda_float_compress
Users that are interested in cuda_float_compress are comparing it to the libraries listed below
Sorting:
- Latent Large Language Models☆18Updated last year
- ☆51Updated last year
- Official code for "SWARM Parallelism: Training Large Models Can Be Surprisingly Communication-Efficient"☆143Updated last year
- train with kittens!☆62Updated 11 months ago
- PCCL (Prime Collective Communications Library) implements fault tolerant collective communications over IP☆121Updated 2 weeks ago
- MPI Code Generation through Domain-Specific Language Models☆14Updated 10 months ago
- Tree Attention: Topology-aware Decoding for Long-Context Attention on GPU clusters☆129Updated 9 months ago
- Cerule - A Tiny Mighty Vision Model☆68Updated last year
- ☆46Updated last year
- Code repository for the paper - "AdANNS: A Framework for Adaptive Semantic Search"☆65Updated last year
- Fork of Flame repo for training of some new stuff in development☆17Updated 3 weeks ago
- Demonstration that finetuning RoPE model on larger sequences than the pre-trained model adapts the model context limit☆63Updated 2 years ago
- ☆13Updated 2 years ago
- Compression for Foundation Models☆35Updated 2 months ago
- ☆62Updated last year
- Train a SmolLM-style llm on fineweb-edu in JAX/Flax with an assortment of optimizers.☆19Updated 2 months ago
- ☆18Updated last year
- [ICML 2023] "Outline, Then Details: Syntactically Guided Coarse-To-Fine Code Generation", Wenqing Zheng, S P Sharan, Ajay Kumar Jaiswal, …☆41Updated last year
- Implementation of SelfExtend from the paper "LLM Maybe LongLM: Self-Extend LLM Context Window Without Tuning" from Pytorch and Zeta☆13Updated 10 months ago
- A tree-based prefix cache library that allows rapid creation of looms: hierarchal branching pathways of LLM generations.☆74Updated 7 months ago
- Demo of the unit_scaling library, showing how a model can be easily adapted to train in FP8.☆46Updated last year
- Make triton easier☆47Updated last year
- Matrix (Multi-Agent daTa geneRation Infra and eXperimentation framework) is a versatile engine for multi-agent conversational data genera…☆94Updated this week
- Linear Attention Sequence Parallelism (LASP)☆86Updated last year
- Token Omission Via Attention☆128Updated 11 months ago
- Lightweight toolkit package to train and fine-tune 1.58bit Language models☆88Updated 4 months ago
- An EXA-Scale repository of Multi-Modality AI resources from papers and models, to foundational libraries!☆40Updated last year
- CUDA and Triton implementations of Flash Attention with SoftmaxN.☆73Updated last year
- Advanced Ultra-Low Bitrate Compression Techniques for the LLaMA Family of LLMs☆110Updated last year
- ☆23Updated 9 months ago