deepshard / mixtral-8x7b-InferenceLinks
Eh, simple and works.
☆27Updated last year
Alternatives and similar repositories for mixtral-8x7b-Inference
Users that are interested in mixtral-8x7b-Inference are comparing it to the libraries listed below
Sorting:
- inference code for mixtral-8x7b-32kseqlen☆102Updated last year
- ☆88Updated last year
- Zeus LLM Trainer is a rewrite of Stanford Alpaca aiming to be the trainer for all Large Language Models☆69Updated 2 years ago
- Cerule - A Tiny Mighty Vision Model☆67Updated this week
- Simple GRPO scripts and configurations.☆59Updated 9 months ago
- Fast approximate inference on a single GPU with sparsity aware offloading☆38Updated last year
- an implementation of Self-Extend, to expand the context window via grouped attention☆118Updated last year
- Optimizing Causal LMs through GRPO with weighted reward functions and automated hyperparameter tuning using Optuna☆58Updated 3 weeks ago
- Track the progress of LLM context utilisation☆54Updated 6 months ago
- Simplex Random Feature attention, in PyTorch☆73Updated 2 years ago
- Scripts to create your own moe models using mlx☆90Updated last year
- Using multiple LLMs for ensemble Forecasting☆16Updated last year