groq / groqflowLinks
GroqFlow provides an automated tool flow for compiling machine learning and linear algebra workloads into Groq programs and executing those programs on GroqChip™ processors.
☆114Updated 5 months ago
Alternatives and similar repositories for groqflow
Users that are interested in groqflow are comparing it to the libraries listed below
Sorting:
- Machine Learning Agility (MLAgility) benchmark and benchmarking tools☆40Updated 5 months ago
- vLLM: A high-throughput and memory-efficient inference and serving engine for LLMs☆93Updated this week
- Tutorial to get started with SkyPilot!☆58Updated last year
- ☆112Updated 2 years ago
- ☆90Updated last year
- A high-throughput and memory-efficient inference and serving engine for LLMs☆53Updated 2 years ago
- Repository of model demos using TT-Buda☆63Updated 9 months ago
- 🏥 Health monitor for a Petals swarm☆40Updated last year
- ☆122Updated last year
- AskIt: Unified programming interface for programming with LLMs (GPT-3.5, GPT-4, Gemini, Claude, Cohere, Llama 2)☆80Updated last year
- ScalarLM - a unified training and inference stack☆95Updated 2 months ago
- Parameter-Efficient Sparsity Crafting From Dense to Mixture-of-Experts for Instruction Tuning on General Tasks☆31Updated last year
- ☆125Updated last year
- AI Assistant running within your browser.☆78Updated last year
- ☆68Updated last year
- Official code for "SWARM Parallelism: Training Large Models Can Be Surprisingly Communication-Efficient"☆148Updated 2 years ago
- Chaiverse: the Chai AI developer platform.☆190Updated last year
- A repository of projects and datasets under active development by Alignment Lab AI☆22Updated 2 years ago
- ☆119Updated last year
- ☆64Updated last year
- 1.58 Bit LLM on Apple Silicon using MLX☆238Updated last year
- PCCL (Prime Collective Communications Library) implements fault tolerant collective communications over IP☆141Updated 4 months ago
- ☆36Updated last year
- Pipeline is an open source python SDK for building AI/ML workflows☆138Updated last year
- Inference examples☆65Updated 4 months ago
- Fast parallel LLM inference for MLX☆242Updated last year
- Transformer GPU VRAM estimator☆67Updated last year
- A Learning Journey: Micrograd in Mojo 🔥☆65Updated last year
- A tree-based prefix cache library that allows rapid creation of looms: hierarchal branching pathways of LLM generations.☆77Updated 11 months ago
- 1.58-bit LLaMa model☆83Updated last year