AMD-AIG-AIMA / Instella
Fully Open Language Models with Stellar Performance
☆215Updated this week
Alternatives and similar repositories for Instella:
Users that are interested in Instella are comparing it to the libraries listed below
- Lightweight Inference server for OpenVINO☆144Updated this week
- A Pure Rust based LLM (Any LLM based MLLM such as Spark-TTS) Inference Engine, powering by Candle framework.☆86Updated last week
- Editor with LLM generation tree exploration☆65Updated last month
- ☆186Updated 7 months ago
- A minimalistic C++ Jinja templating engine for LLM chat templates☆128Updated this week
- Lightweight Llama 3 8B Inference Engine in CUDA C☆47Updated 2 weeks ago
- Run LLM Agents on Ryzen AI PCs in Minutes☆288Updated last week
- MockLLM, when you want it to do what you tell it to do!☆46Updated this week
- A fork of OpenBLAS with Armv8-A SVE (Scalable Vector Extension) support☆16Updated 4 years ago
- llama.cpp fork used by GPT4All☆54Updated last month
- 1.58 Bit LLM on Apple Silicon using MLX☆194Updated 10 months ago
- Kolosal AI is an OpenSource and Lightweight alternative to LM Studio to run LLMs 100% offline on your device.☆172Updated last week
- Turns devices into a scalable LLM platform☆127Updated this week
- ☆56Updated 8 months ago
- Run multiple resource-heavy Large Models (LM) on the same machine with limited amount of VRAM/other resources by exposing them on differe…☆55Updated last month
- TPI-LLM: Serving 70b-scale LLMs Efficiently on Low-resource Edge Devices☆171Updated 4 months ago
- AI Tensor Engine for ROCm☆150Updated this week
- A python package for serving LLM on OpenAI-compatible API endpoints with prompt caching using MLX.☆76Updated 3 months ago
- Moxin is a family of fully open-source and reproducible LLMs☆85Updated 3 weeks ago
- A companion toolkit to pico-train for quantifying, comparing, and visualizing how language models evolve during training.☆48Updated 2 weeks ago
- LLM Inference on consumer devices☆103Updated 2 weeks ago
- llama.cpp fork with additional SOTA quants and improved performance☆243Updated this week
- Modular, open source LLMOps stack that separates concerns: LiteLLM unifies LLM APIs, manages routing and cost controls, and ensures high-…☆92Updated last month
- cortex.llamacpp is a high-efficiency C++ inference engine for edge computing. It is a dynamic library that can be loaded by any server a…☆37Updated this week
- Granite 3.1 Language Models☆98Updated 3 months ago
- Schola is a plugin for enabling Reinforcement Learning (RL) in Unreal Engine. It provides tools to help developers create environments, d…☆34Updated this week
- Source code for Intel's Polite Guard NLP project☆29Updated this week
- Rust framework for LLM orchestration☆202Updated 8 months ago
- Testing LLM reasoning abilities with family relationship quizzes.☆62Updated 2 months ago
- See Through Your Models☆372Updated 3 weeks ago