deepshard / mixtral-8x7b-Inference
Eh, simple and works.
☆27Updated last year
Alternatives and similar repositories for mixtral-8x7b-Inference:
Users that are interested in mixtral-8x7b-Inference are comparing it to the libraries listed below
- inference code for mixtral-8x7b-32kseqlen☆99Updated last year
- Optimizing Causal LMs through GRPO with weighted reward functions and automated hyperparameter tuning using Optuna☆39Updated 2 months ago
- an implementation of Self-Extend, to expand the context window via grouped attention☆119Updated last year
- Scripts to create your own moe models using mlx☆89Updated last year
- Zeus LLM Trainer is a rewrite of Stanford Alpaca aiming to be the trainer for all Large Language Models☆69Updated last year
- ☆112Updated 4 months ago
- ☆48Updated last year
- Lego for GRPO☆27Updated 3 weeks ago
- Score LLM pretraining data with classifiers☆55Updated last year
- A repository of prompts and Python scripts for intelligent transformation of raw text into diverse formats.☆30Updated last year
- An example implementation of RLHF (or, more accurately, RLAIF) built on MLX and HuggingFace.☆25Updated 10 months ago
- an open source reproduction of NVIDIA's nGPT (Normalized Transformer with Representation Learning on the Hypersphere)☆96Updated last month
- Cerule - A Tiny Mighty Vision Model☆67Updated 7 months ago
- look how they massacred my boy☆63Updated 6 months ago
- ☆38Updated 9 months ago
- Simplex Random Feature attention, in PyTorch☆74Updated last year
- Chat Markup Language conversation library☆55Updated last year
- tiny_fnc_engine is a minimal python library that provides a flexible engine for calling functions extracted from a LLM.☆38Updated 7 months ago
- Just a bunch of benchmark logs for different LLMs☆119Updated 8 months ago
- Comprehensive analysis of difference in performance of QLora, Lora, and Full Finetunes.☆82Updated last year
- KMD is a collection of conversational exchanges between patients and doctors on various medical topics. It aims to capture the intricaci…☆24Updated last year
- Using multiple LLMs for ensemble Forecasting☆16Updated last year
- Simple GRPO scripts and configurations.☆58Updated 2 months ago
- smolLM with Entropix sampler on pytorch☆151Updated 5 months ago
- ☆38Updated last year
- LLM reads a paper and produce a working prototype☆52Updated 2 weeks ago
- ☆22Updated last year
- ☆66Updated 11 months ago
- ☆28Updated last year
- Fast approximate inference on a single GPU with sparsity aware offloading☆38Updated last year