OpenAccess-AI-Collective / servereless-runpod-ggmlLinks
☆53Updated 2 years ago
Alternatives and similar repositories for servereless-runpod-ggml
Users that are interested in servereless-runpod-ggml are comparing it to the libraries listed below
Sorting:
- ☆50Updated 2 years ago
- Easily create LLM automation/agent workflows☆61Updated last year
- Client-side toolkit for using large language models, including where self-hosted☆113Updated last week
- Deploy your GGML models to HuggingFace Spaces with Docker and gradio☆37Updated 2 years ago
- Experimental sampler to make LLMs more creative☆31Updated 2 years ago
- Auto Data is a library designed for quick and effortless creation of datasets tailored for fine-tuning Large Language Models (LLMs).☆102Updated last year
- ☆134Updated last year
- A prompt/context management system☆170Updated 2 years ago
- Automated prompting and scoring framework to evaluate LLMs using updated human knowledge prompts☆108Updated 2 years ago
- Local character AI chatbot with chroma vector store memory and some scripts to process documents for Chroma☆34Updated last year
- Creates an Langchain Agent which uses the WebUI's API and Wikipedia to work☆73Updated 2 years ago
- TheBloke's Dockerfiles☆307Updated last year
- ☆116Updated 10 months ago
- AutoNL - Natural Language Automation tool☆86Updated last year
- ☆44Updated last year
- Dagger functions to import Hugging Face GGUF models into a local ollama instance and optionally push them to ollama.com.☆119Updated last year
- Generate Synthetic Data Using OpenAI, MistralAI or AnthropicAI☆221Updated last year
- auto fine tune of models with synthetic data☆75Updated last year
- ezlocalai is an easy to set up local artificial intelligence server with OpenAI Style Endpoints.☆88Updated last month
- 🔓 The open-source autonomous agent LLM initiative 🔓☆91Updated last year
- Gradio based tool to run opensource LLM models directly from Huggingface☆96Updated last year
- Generate visual podcasts about novels using open source models☆25Updated 2 years ago
- Locally running LLM with internet access☆97Updated 4 months ago
- Landmark Attention: Random-Access Infinite Context Length for Transformers QLoRA☆123Updated 2 years ago
- ☆47Updated last year
- Simple and fast server for GPTQ-quantized LLaMA inference☆24Updated 2 years ago
- Complex RAG backend☆29Updated last year
- Simple Graph Memory for AI applications☆89Updated 5 months ago
- Ollama models of NousResearch/Hermes-2-Pro-Mistral-7B-GGUF☆31Updated last year
- A Qt GUI for large language models☆44Updated last year