daochenzha / dreamshardLinks
[NeurIPS 2022] DreamShard: Generalizable Embedding Table Placement for Recommender Systems
☆29Updated 2 years ago
Alternatives and similar repositories for dreamshard
Users that are interested in dreamshard are comparing it to the libraries listed below
Sorting:
- Repository for "GIST: Distributed training for large-scale graph convolutional networks"☆15Updated 2 years ago
- Can GPT-4 Perform Neural Architecture Search?☆87Updated 2 years ago
- JORA: JAX Tensor-Parallel LoRA Library (ACL 2024)☆36Updated last year
- Using FlexAttention to compute attention with different masking patterns☆45Updated last year
- Linear Attention Sequence Parallelism (LASP)☆87Updated last year
- Code for COMET: Cardinality Constrained Mixture of Experts with Trees and Local Search☆11Updated 2 years ago
- Lottery Ticket Adaptation☆40Updated 10 months ago
- ☆30Updated 3 weeks ago
- Official codebase for NeurIPS 2022 paper End-to-end Learning to Index and Search in Large Output Spaces☆12Updated 2 years ago
- ☆34Updated 10 months ago
- AutoMoE: Neural Architecture Search for Efficient Sparsely Activated Transformers☆47Updated 2 years ago
- The code implementation of MAGDi: Structured Distillation of Multi-Agent Interaction Graphs Improves Reasoning in Smaller Language Models…☆37Updated last year
- A testbed for agents and environments that can automatically improve models through data generation.☆27Updated 7 months ago
- Q-Probe: A Lightweight Approach to Reward Maximization for Language Models☆41Updated last year
- [KDD 2022] AutoShard: Automated Embedding Table Sharding for Recommender Systems☆22Updated 2 years ago
- Advantage Leftover Lunch Reinforcement Learning (A-LoL RL): Improving Language Models with Advantage-based Offline Policy Gradients☆26Updated last year
- Hyperparameter tuning via uncertainty modeling☆48Updated last year
- A RL env with procedurally generated symbolic reasoning data☆24Updated 2 weeks ago
- Revisiting Efficient Training Algorithms For Transformer-based Language Models (NeurIPS 2023)☆80Updated 2 years ago
- Learning from preferences is a common paradigm for fine-tuning language models. Yet, many algorithmic design decisions come into play. Ou…☆32Updated last year
- Implementation of TableFormer, Robust Transformer Modeling for Table-Text Encoding, in Pytorch☆39Updated 3 years ago
- Experimental scripts for researching data adaptive learning rate scheduling.☆22Updated last year
- [ICLR 2023] "Sparse MoE as the New Dropout: Scaling Dense and Self-Slimmable Transformers" by Tianlong Chen*, Zhenyu Zhang*, Ajay Jaiswal…☆55Updated 2 years ago
- 32 times longer context window than vanilla Transformers and up to 4 times longer than memory efficient Transformers.☆48Updated 2 years ago
- Understanding the correlation between different LLM benchmarks☆29Updated last year
- ☆33Updated 11 months ago
- ☆52Updated last year
- ☆10Updated last year
- Triton Implementation of HyperAttention Algorithm☆48Updated last year
- Minimum Description Length probing for neural network representations☆20Updated 8 months ago