containers / ramalamaLinks
RamaLama is an open-source developer tool that simplifies the local serving of AI models from any source and facilitates their use for inference in production, all through the familiar language of containers.
☆1,723Updated this week
Alternatives and similar repositories for ramalama
Users that are interested in ramalama are comparing it to the libraries listed below
Sorting:
- Boot and upgrade via container images☆1,254Updated this week
- Work with LLMs on a local environment using containers☆227Updated this week
- Model swapping for llama.cpp (or any local OpenAPI compatible server)☆848Updated this week
- Generate Podman Quadlet files from a Podman command, compose file, or existing object☆878Updated 6 months ago
- the terminal client for Ollama☆1,905Updated this week
- LLM plugin providing access to models running on an Ollama server☆306Updated this week
- Support for bootable OS containers (bootc) and generating disk images☆436Updated this week
- Create microVMs from OCI images☆1,474Updated 4 months ago
- Local AI API Platform☆2,715Updated 2 weeks ago
- Dockerfile → Universal Wasm Binary☆877Updated last month
- VS Code extension for LLM-assisted code/text completion☆774Updated this week
- Connect home devices into a powerful cluster to accelerate LLM inference. More devices means faster inference.☆2,074Updated last month
- A powerful document AI question-answering tool that connects to your local Ollama models. Create, manage, and interact with RAG systems f…☆1,029Updated last week
- AI Inference Operator for Kubernetes. The easiest way to serve ML models in production. Supports VLMs, LLMs, embeddings, and speech-to-te…☆970Updated last week
- An external provider for Llama Stack allowing for the use of RamaLama for inference.☆15Updated this week
- Self-Hosted Plaform for Secure Execution of Untrusted User/AI Code☆1,710Updated this week
- Open source platform for AI Engineering: OpenTelemetry-native LLM Observability, GPU Monitoring, Guardrails, Evaluations, Prompt Manageme…☆1,585Updated this week
- llm-d is a Kubernetes-native high-performance distributed LLM inference framework☆987Updated this week
- WebAssembly binding for llama.cpp - Enabling on-browser LLM inference☆732Updated last month
- Effortlessly run LLM backends, APIs, frontends, and services with one command.☆1,702Updated 2 weeks ago
- A container for deploying bootable container images.☆209Updated this week
- CodeGate: Security, Workspaces and Multiplexing for AI Agentic Frameworks☆677Updated this week
- Taxonomy tree that will allow you to create models tuned with your data☆267Updated this week
- Artificial Neural Engine Machine Learning Library☆945Updated 2 weeks ago
- LSP-AI is an open-source language server that serves as a backend for AI-powered functionality, designed to assist and empower software e…☆2,780Updated 4 months ago
- Podman Desktop is the best free and open source tool to work with Containers and Kubernetes for developers. Get an intuitive and user-fri…☆5,874Updated this week
- Replace OpenAI with Llama.cpp Automagically.☆318Updated 11 months ago
- Podman Terminal UI☆787Updated this week
- What If Language Models Expertly Routed All Inference? WilmerAI allows prompts to be routed to specialized workflows based on the domain …☆684Updated last month
- Examples for building and running LLM services and applications locally with Podman☆158Updated last week