bodaay / HuggingFaceModelDownloaderLinks
Simple go utility to download HuggingFace Models and Datasets
☆716Updated 9 months ago
Alternatives and similar repositories for HuggingFaceModelDownloader
Users that are interested in HuggingFaceModelDownloader are comparing it to the libraries listed below
Sorting:
- Web UI for ExLlamaV2☆506Updated 6 months ago
- The official API server for Exllama. OAI compatible, lightweight, and fast.☆1,031Updated last week
- Large-scale LLM inference engine☆1,524Updated this week
- An OpenAI API compatible API for chat with image input and questions about the images. aka Multimodal.☆259Updated 5 months ago
- Comparison of the output quality of quantization methods, using Llama 3, transformers, GGUF, EXL2.☆159Updated last year
- The llama-cpp-agent framework is a tool designed for easy interaction with Large Language Models (LLMs). Allowing users to chat with LLM …☆587Updated 6 months ago
- An extension for oobabooga/text-generation-webui that enables the LLM to search the web☆259Updated last week
- A multimodal, function calling powered LLM webui.☆216Updated 11 months ago
- An AI assistant beyond the chat box.☆328Updated last year
- Self-evaluating interview for AI coders☆594Updated 2 months ago
- Make abliterated models with transformers, easy and fast☆83Updated 4 months ago
- Dolphin System Messages☆345Updated 6 months ago
- An optimized quantization and inference library for running LLMs locally on modern consumer-class GPUs☆473Updated last week
- Pure C++ implementation of several models for real-time chatting on your computer (CPU & GPU)☆677Updated this week
- automatically quant GGUF models☆195Updated last week
- A proxy server for multiple ollama instances with Key security☆480Updated 3 weeks ago
- Convenience scripts to finetune (chat-)LLaMa3 and other models for any language☆312Updated last year
- Docker variants of oobabooga's text-generation-webui, including pre-built images.☆435Updated last month
- The RunPod worker template for serving our large language model endpoints. Powered by vLLM.☆355Updated this week
- VSCode AI coding assistant powered by self-hosted llama.cpp endpoint.☆183Updated 6 months ago
- A fast batching API to serve LLM models☆185Updated last year
- Download models from the Ollama library, without Ollama☆92Updated 9 months ago
- A more memory-efficient rewrite of the HF transformers implementation of Llama for use with quantized weights.☆2,894Updated last year
- A self-hosted github copilot guide using oobabooga webui☆162Updated last year
- Efficient visual programming for AI language models☆363Updated 3 months ago
- Falcon LLM ggml framework with CPU and GPU support☆247Updated last year
- A fast inference library for running LLMs locally on modern consumer-class GPUs☆4,279Updated last week
- Review/Check GGUF files and estimate the memory usage and maximum tokens per second.☆197Updated last week
- ☆120Updated 9 months ago
- Ingest files for retrieval augmented generation (RAG) with open-source Large Language Models (LLMs), all without 3rd parties or sensitive…☆687Updated last year