Linaro / tinyBLAS
A fork of OpenBLAS with Armv8-A SVE (Scalable Vector Extension) support
☆15Updated 4 years ago
Alternatives and similar repositories for tinyBLAS:
Users that are interested in tinyBLAS are comparing it to the libraries listed below
- Large Model Proxy is designed to make it easy to run multiple resource-heavy Large Models (LM) on the same machine with limited amount of…☆49Updated 3 months ago
- ☆53Updated 5 months ago
- A faithful clone of Karpathy's llama2.c (one file inference, zero dependency) but fully functional with LLaMA 3 8B base and instruct mode…☆113Updated 6 months ago
- llama.cpp fork with additional SOTA quants and improved performance☆133Updated this week
- A library for incremental loading of large PyTorch checkpoints☆56Updated last year
- A minimalistic C++ Jinja templating engine for LLM chat templates☆104Updated this week
- iterate quickly with llama.cpp hot reloading. use the llama.cpp bindings with bun.sh☆47Updated last year
- GRDN.AI app for garden optimization☆70Updated 11 months ago
- Distributed Inference for mlx LLm☆79Updated 5 months ago
- General purpose GPU compute framework built on Vulkan to support 1000s of cross vendor graphics cards (AMD, Qualcomm, NVIDIA & friends). …☆44Updated 4 months ago
- 33B Chinese LLM, DPO QLORA, 100K context, AirLLM 70B inference with single 4GB GPU☆13Updated 8 months ago
- Resources regarding evML (edge verified machine learning)☆13Updated 3 weeks ago
- Lightweight Llama 3 8B Inference Engine in CUDA C☆43Updated last week
- Aana SDK is a powerful framework for building AI enabled multimodal applications.☆41Updated this week
- Inference Vision Transformer (ViT) in plain C/C++ with ggml☆31Updated last year
- Simple LLM inference server☆20Updated 7 months ago
- Mistral7B playing DOOM☆127Updated 6 months ago
- ☆27Updated 5 months ago
- Easy to use, High Performant Knowledge Distillation for LLMs☆40Updated 2 weeks ago
- A super simple web interface to perform blind tests on LLM outputs.☆27Updated 10 months ago
- This benchmark tests how well LLMs incorporate a set of 10 mandatory story elements (characters, objects, core concepts, attributes, moti…☆35Updated last week
- ☆65Updated 8 months ago
- A high-throughput and memory-efficient inference and serving engine for LLMs☆51Updated last year
- Local Startup Advisor Chatbot☆30Updated last year
- Testing LLM reasoning abilities with family relationship quizzes.☆57Updated this week
- Chat WebUI is an easy-to-use user interface for interacting with AI, and it comes with multiple useful built-in tools.☆15Updated 2 months ago
- GPU-targeted vendor-agnostic AI library for Windows, and Mistral model implementation.☆53Updated last year
- klmbr - a prompt pre-processing technique to break through the barrier of entropy while generating text with LLMs☆67Updated 4 months ago
- ☆53Updated 6 months ago
- A CLI to manage install and configure llama inference implemenation in multiple languages☆65Updated last year