groq / groqflowLinks
GroqFlow provides an automated tool flow for compiling machine learning and linear algebra workloads into Groq programs and executing those programs on GroqChip™ processors.
☆114Updated 3 months ago
Alternatives and similar repositories for groqflow
Users that are interested in groqflow are comparing it to the libraries listed below
Sorting:
- Machine Learning Agility (MLAgility) benchmark and benchmarking tools☆40Updated 3 months ago
- ☆89Updated last year
- Tutorial to get started with SkyPilot!☆57Updated last year
- A high-throughput and memory-efficient inference and serving engine for LLMs☆53Updated last year
- 1.58-bit LLaMa model☆83Updated last year
- vLLM: A high-throughput and memory-efficient inference and serving engine for LLMs☆93Updated this week
- AskIt: Unified programming interface for programming with LLMs (GPT-3.5, GPT-4, Gemini, Claude, Cohere, Llama 2)☆79Updated 10 months ago
- 1.58 Bit LLM on Apple Silicon using MLX☆225Updated last year
- ☆112Updated last year
- ☆36Updated last year
- Tenstorrent console based hardware information program☆57Updated this week
- ☆67Updated last year
- Parameter-Efficient Sparsity Crafting From Dense to Mixture-of-Experts for Instruction Tuning on General Tasks☆31Updated last year
- Transformer GPU VRAM estimator☆66Updated last year
- ☆116Updated 11 months ago
- look how they massacred my boy☆63Updated last year
- ☆29Updated last year
- A repository of projects and datasets under active development by Alignment Lab AI☆22Updated last year
- A tree-based prefix cache library that allows rapid creation of looms: hierarchal branching pathways of LLM generations.☆72Updated 9 months ago
- Fast parallel LLM inference for MLX☆232Updated last year
- 🏥 Health monitor for a Petals swarm☆39Updated last year
- LLM inference in C/C++☆102Updated 2 weeks ago
- inference code for mixtral-8x7b-32kseqlen☆102Updated last year
- This repository explains and provides examples for "concept anchoring" in GPT4.☆71Updated last year
- Deploy your autonomous agents to production grade environments with 99% Uptime Guarantee, Infinite Scalability, and self-healing.☆47Updated last month
- A guidance compatibility layer for llama-cpp-python☆36Updated 2 years ago
- LMQL implementation of tree of thoughts☆34Updated last year
- ☆124Updated last year
- Turing machines, Rule 110, and A::B reversal using Claude 3 Opus.☆58Updated last year
- ScalarLM - a unified training and inference stack☆93Updated this week