akx / ollama-dl
Download models from the Ollama library, without Ollama
☆72Updated 6 months ago
Alternatives and similar repositories for ollama-dl
Users that are interested in ollama-dl are comparing it to the libraries listed below
Sorting:
- A more memory-efficient rewrite of the HF transformers implementation of Llama for use with quantized weights.☆64Updated last year
- Deploy your GGML models to HuggingFace Spaces with Docker and gradio☆36Updated last year
- Extract structured data from local or remote LLM models☆42Updated 10 months ago
- Tool to download models from Huggingface Hub and convert them to GGML/GGUF for llama.cpp☆136Updated 2 weeks ago
- Pybind11 bindings for Whisper.cpp☆57Updated last week
- A lightweight proxy for filtering `<think>` tags from any OpenAI-compatible API endpoint. Designed for chain-of-thought language models t…☆37Updated 3 months ago
- A guidance compatibility layer for llama-cpp-python☆34Updated last year
- simple prompt script to convert hf/ggml files to gguf, and to quantize☆26Updated last year
- LLM inference in C/C++☆76Updated this week
- VSCode AI coding assistant powered by self-hosted llama.cpp endpoint.☆180Updated 3 months ago
- Dagger functions to import Hugging Face GGUF models into a local ollama instance and optionally push them to ollama.com.☆115Updated 11 months ago
- Something similar to Apple Intelligence?☆60Updated 10 months ago
- Local LLM inference & management server with built-in OpenAI API☆31Updated last year
- LLaMA Server combines the power of LLaMA C++ with the beauty of Chatbot UI.☆123Updated last year
- ☆24Updated 3 months ago
- Simple, Fast, Parallel Huggingface GGML model downloader written in python☆24Updated last year
- ☆87Updated 11 months ago
- an auto-sleeping and -waking framework around llama.cpp☆11Updated 3 months ago
- klmbr - a prompt pre-processing technique to break through the barrier of entropy while generating text with LLMs☆71Updated 7 months ago
- Text generation in Python, as easy as possible☆60Updated this week
- ☆156Updated 9 months ago
- Pressure testing the context window of open LLMs☆24Updated 8 months ago
- Review/Check GGUF files and estimate the memory usage and maximum tokens per second.☆161Updated this week
- Conduct in-depth research with AI-driven insights : DeepDive is a command-line tool that leverages web searches and AI models to generate…☆41Updated 8 months ago
- Embedding models from Jina AI☆59Updated last year
- For inferring and serving local LLMs using the MLX framework☆103Updated last year
- ☆69Updated this week
- Mistral-7B finetuned for function calling☆16Updated last year
- ☆202Updated 3 weeks ago
- A Python library to orchestrate LLMs in a neural network-inspired structure☆47Updated 7 months ago