kyuz0 / amd-strix-halo-vllm-toolboxesLinks
☆40Updated 2 months ago
Alternatives and similar repositories for amd-strix-halo-vllm-toolboxes
Users that are interested in amd-strix-halo-vllm-toolboxes are comparing it to the libraries listed below
Sorting:
- Ampere optimized llama.cpp☆27Updated 2 weeks ago
- No-code CLI designed for accelerating ONNX workflows☆216Updated 4 months ago
- vLLM: A high-throughput and memory-efficient inference and serving engine for LLMs☆92Updated last week
- ☆49Updated last month
- Route LLM requests to the best model for the task at hand.☆122Updated last week
- Welcome to the official repository of SINQ! A novel, fast and high-quality quantization method designed to make any Large Language Model …☆564Updated last week
- Inference engine for Intel devices. Serve LLMs, VLMs, Whisper, Kokoro-TTS, Embedding and Rerank models over OpenAI endpoints.☆236Updated last week
- Serving LLMs in the HF-Transformers format via a PyFlask API☆71Updated last year
- InferX: Inference as a Service Platform☆138Updated this week
- Run multiple resource-heavy Large Models (LM) on the same machine with limited amount of VRAM/other resources by exposing them on differe…☆82Updated last week
- For individual users, watsonx Code Assistant can access a local IBM Granite model☆37Updated 4 months ago
- ☆61Updated 4 months ago
- How to build an ACP compliant agent that uses MCP as well!☆11Updated 6 months ago
- This is the Mixture-of-Agents (MoA) concept, adapted from the original work by TogetherAI. My version is tailored for local model usage a…☆118Updated last year
- Building open version of OpenAI o1 via reasoning traces (Groq, ollama, Anthropic, Gemini, OpenAI, Azure supported) Demo: https://hugging…☆184Updated last year
- Use smol agents to do research and then update csv coumns with its findings.☆41Updated 9 months ago
- LLM Inference on consumer devices☆125Updated 7 months ago
- Practical and advanced guide to LLMOps. It provides a solid understanding of large language models’ general concepts, deployment techniqu…☆76Updated last year
- Use Codestral Mamba with Visual Studio Code and the Continue extension. A local LLM alternative to GitHub Copilot.☆29Updated last year
- A comprehensive platform for managing, testing, and leveraging Ollama AI models with advanced features for customization, workflow automa…☆46Updated 8 months ago
- ☆213Updated last week
- Intel® AI Assistant Builder☆117Updated last week
- FMS Model Optimizer is a framework for developing reduced precision neural network models.☆20Updated last week
- llmbasedos — Local-First OS Where Your AI Agents Wake Up and Work☆277Updated 2 months ago
- Offline LLM chatbot with personalized memory — works on CPU with multi-session memory support.☆22Updated 4 months ago
- ☆17Updated 2 months ago
- the rent a hal project for AI☆21Updated 2 months ago
- This repo contains documents of the OPEA project☆43Updated 2 months ago
- chrome & firefox extension to chat with webpages: local llms☆127Updated 10 months ago
- This is a Python package to add tool calling capabilities to newly released LLMs on LangChain's ChatOpenAI, AzureAIChatCompletionsModel a…☆121Updated 5 months ago