eliranwong / MultiAMDGPU_AIDev_Ubuntu
Multi AMD GPU Setup for AI Development on Ubuntu with ROCM
☆26Updated last week
Alternatives and similar repositories for MultiAMDGPU_AIDev_Ubuntu:
Users that are interested in MultiAMDGPU_AIDev_Ubuntu are comparing it to the libraries listed below
- Distributed Inference for mlx LLm☆87Updated 8 months ago
- ☆83Updated 3 months ago
- CursorCore: Assist Programming through Aligning Anything☆117Updated last month
- Service for testing out the new Qwen2.5 omni model☆17Updated last week
- llama.cpp fork with additional SOTA quants and improved performance☆231Updated this week
- Moxin is a family of fully open-source and reproducible LLMs☆85Updated 2 weeks ago
- automatically quant GGUF models☆164Updated this week
- ☆23Updated this week
- LM inference server implementation based on *.cpp.☆154Updated this week
- Lightweight Inference server for OpenVINO☆143Updated this week
- Lightweight Llama 3 8B Inference Engine in CUDA C☆47Updated last week
- LLM inference in C/C++☆67Updated last week
- Run multiple resource-heavy Large Models (LM) on the same machine with limited amount of VRAM/other resources by exposing them on differe…☆55Updated last month
- LLM as Interpreter for Natural Language Programming, Pseudo-code Programming and Flow Programming of AI Agents☆38Updated 8 months ago
- ☆17Updated last week
- Fast parallel LLM inference for MLX☆177Updated 8 months ago
- GenAI & agent toolkit for Apple Silicon Mac, implementing JSON schema-steered structured output (3SO) and tool-calling in Python. For mor…☆118Updated last month
- Minimal, clean code implementation of RAG with mlx using gguf model weights☆49Updated 11 months ago
- Something similar to Apple Intelligence?☆59Updated 9 months ago
- ☆37Updated 3 months ago
- LLM inference in C/C++☆19Updated last week
- Implementation of nougat that focuses on processing pdf locally.☆81Updated 2 months ago
- A toolkit for fine-tuning, inferencing, and evaluating GreenBitAI's LLMs.☆81Updated 3 weeks ago
- A python package for serving LLM on OpenAI-compatible API endpoints with prompt caching using MLX.☆76Updated 3 months ago
- Tool to download models from Huggingface Hub and convert them to GGML/GGUF for llama.cpp☆128Updated 6 months ago
- A pipeline parallel training script for LLMs.☆136Updated this week
- Dagger functions to import Hugging Face GGUF models into a local ollama instance and optionally push them to ollama.com.☆115Updated 10 months ago
- Fully-featured, beautiful web interface for vLLM - built with NextJS.☆118Updated last week
- ☆125Updated last week
- a simplified version of Google's Gemma model to be used for learning☆24Updated last year