AMD-AGI / InstellaLinks
Fully Open Language Models with Stellar Performance
☆248Updated last month
Alternatives and similar repositories for Instella
Users that are interested in Instella are comparing it to the libraries listed below
Sorting:
- Pivotal Token Search☆124Updated 2 months ago
- ☆189Updated last year
- Lightweight Inference server for OpenVINO☆211Updated this week
- Sparse Inferencing for transformer based LLMs☆197Updated last month
- Docs for GGUF quantization (unofficial)☆258Updated 2 months ago
- ☆417Updated 3 weeks ago
- Benchmark and optimize LLM inference across frameworks with ease☆41Updated last week
- Train, tune, and infer Bamba model☆132Updated 3 months ago
- Simple & Scalable Pretraining for Neural Architecture Research☆291Updated 3 weeks ago
- 1.58 Bit LLM on Apple Silicon using MLX☆223Updated last year
- Lightweight toolkit package to train and fine-tune 1.58bit Language models☆88Updated 4 months ago
- No-code CLI designed for accelerating ONNX workflows☆214Updated 3 months ago
- LocalScore is an open benchmark which helps you understand how well your computer can handle local AI tasks.☆58Updated 2 weeks ago
- ☆296Updated last month
- InferX is a Inference Function as a Service Platform☆133Updated this week
- Code to train and evaluate Neural Attention Memory Models to obtain universally-applicable memory systems for transformers.☆322Updated 10 months ago
- Local Qwen3 LLM inference. One easy-to-understand file of C source with no dependencies.☆110Updated 2 months ago
- ☆619Updated 3 weeks ago
- PyTorch implementation of models from the Zamba2 series.☆185Updated 7 months ago
- A companion toolkit to pico-train for quantifying, comparing, and visualizing how language models evolve during training.☆108Updated 5 months ago
- ☆334Updated this week
- LLM Inference on consumer devices☆124Updated 6 months ago
- Everything you need to know about LLM inference☆229Updated last week
- GRadient-INformed MoE☆264Updated 11 months ago
- ☆196Updated 4 months ago
- This is the documentation repository for SGLang. It is auto-generated from https://github.com/sgl-project/sglang/tree/main/docs.☆76Updated this week
- ☆231Updated 2 months ago
- Transformer GPU VRAM estimator☆66Updated last year
- Kolosal AI is an OpenSource and Lightweight alternative to LM Studio to run LLMs 100% offline on your device.☆297Updated 3 months ago
- ☆262Updated 2 months ago