AMD-AGI / InstellaLinks
Fully Open Language Models with Stellar Performance
☆312Updated last month
Alternatives and similar repositories for Instella
Users that are interested in Instella are comparing it to the libraries listed below
Sorting:
- ☆157Updated last month
- Pivotal Token Search☆142Updated 3 weeks ago
- SWE-Bench Pro: Can AI Agents Solve Long-Horizon Software Engineering Tasks?☆240Updated this week
- Code for Bolmo: Byteifying the Next Generation of Language Models☆112Updated 2 weeks ago
- Welcome to the official repository of SINQ! A novel, fast and high-quality quantization method designed to make any Large Language Model …☆585Updated 2 weeks ago
- This is the documentation repository for SGLang. It is auto-generated from https://github.com/sgl-project/sglang☆96Updated this week
- ☆191Updated last year
- Sparse Inferencing for transformer based LLMs☆216Updated 5 months ago
- Train, tune, and infer Bamba model☆137Updated 7 months ago
- Simple & Scalable Pretraining for Neural Architecture Research☆306Updated last month
- Reverse Engineering Gemma 3n: Google's New Edge-Optimized Language Model☆257Updated 7 months ago
- 1.58 Bit LLM on Apple Silicon using MLX☆237Updated last year
- PyTorch implementation of models from the Zamba2 series.☆185Updated 11 months ago
- Code to train and evaluate Neural Attention Memory Models to obtain universally-applicable memory systems for transformers.☆346Updated last year
- Tiny Model, Big Logic: Diversity-Driven Optimization Elicits Large-Model Reasoning Ability in VibeThinker-1.5B☆559Updated last month
- WeDLM: The fastest diffusion language model with standard causal attention and native KV cache compatibility, delivering real speedups ov…☆480Updated last week
- Everything you need to know about LLM inference☆257Updated last week
- ☆148Updated 5 months ago
- ☆1,257Updated last month
- Accelerating your LLM training to full speed! Made with ❤️ by ServiceNow Research☆276Updated this week
- No-code CLI designed for accelerating ONNX workflows☆221Updated 7 months ago
- ☆460Updated last month
- Benchmark and optimize LLM inference across frameworks with ease☆153Updated 3 months ago
- ☆236Updated last month
- All information and news with respect to Falcon-H1 series☆95Updated 3 months ago
- Developer Asset Hub for NVIDIA Nemotron — A one-stop resource for training recipes, usage cookbooks, and full end-to-end reference exampl…☆314Updated this week
- Verify Precision of all Kimi K2 API Vendor☆491Updated this week
- GRadient-INformed MoE☆264Updated last year
- Docs for GGUF quantization (unofficial)☆347Updated 5 months ago
- MLPerf Client is a benchmark for Windows, Linux and macOS, focusing on client form factors in ML inference scenarios.☆67Updated last month