Leikoe / torch_to_ggmlLinks
convert a saved pytorch model to gguf and generate as much corresponding ggml c code as possible
☆15Updated 2 years ago
Alternatives and similar repositories for torch_to_ggml
Users that are interested in torch_to_ggml are comparing it to the libraries listed below
Sorting:
- AirLLM 70B inference with single 4GB GPU☆14Updated 6 months ago
- This public GitHub repository contains code for a fully self-hosted, on-premise transcription solution.☆53Updated last year
- Experimental sampler to make LLMs more creative☆31Updated 2 years ago
- Simple, Fast, Parallel Huggingface GGML model downloader written in python☆24Updated 2 years ago
- Embarrassingly Easy Fully Non-Autoregressive Zero-Shot TTS (E2 TTS) in MLX☆29Updated last year
- Yet Another (LLM) Web UI, made with Gemini☆12Updated last year
- Implementation of 'Vocos: Closing the gap between time-domain and Fourier-based neural vocoders for high-quality audio synthesis', in MLX☆23Updated last year
- Lightweight continuous batching OpenAI compatibility using HuggingFace Transformers include T5 and Whisper.☆29Updated 9 months ago
- Accepts a Hugging Face model URL, automatically downloads and quantizes it using Bits and Bytes.☆38Updated last year
- A repository to store helpful information and emerging insights in regard to LLMs☆21Updated 2 years ago
- Steer LLM outputs towards a certain topic/subject and enhance response capabilities using activation engineering by adding steering vecto…☆43Updated last year
- A simple speech-to-text and text-to-speech AI chatbot that can be run fully offline.☆46Updated last year
- Implementation of E2-TTS, "Embarrassingly Easy Fully Non-Autoregressive Zero-Shot TTS", in MLX☆20Updated last year
- Tcurtsni: Reverse Instruction Chat, ever wonder what your LLM wants to ask you?☆23Updated last year
- PyGPTPrompt: A CLI tool that manages context windows for AI models, facilitating user interaction and data ingestion for optimized long-t…☆30Updated last year
- Deploy your GGML models to HuggingFace Spaces with Docker and gradio☆38Updated 2 years ago
- run ollama & gguf easily with a single command☆52Updated last year
- Loader extension for tabbyAPI in SillyTavern☆26Updated 6 months ago
- A more memory-efficient rewrite of the HF transformers implementation of Llama for use with quantized weights.☆64Updated 2 years ago
- Python package wrapping llama.cpp for on-device LLM inference☆97Updated 2 months ago
- A Qt GUI for large language models☆45Updated 2 years ago
- ☆16Updated 2 years ago
- Adding a multi-text multi-speaker script (diffe) that is based on a script from asiff00 on issue 61 for Sesame: A Conversational Speech G…☆25Updated 9 months ago
- An OpenAI API compatible LLM inference server based on ExLlamaV2.☆25Updated last year
- Glyphs, acting as collaboratively defined symbols linking related concepts, add a layer of multidimensional semantic richness to user-AI …☆54Updated 10 months ago
- Local LLM inference & management server with built-in OpenAI API☆31Updated last year
- Create text chunks which end at natural stopping points without using a tokenizer☆26Updated last month
- A QT GUI for large language models☆38Updated 2 years ago
- Public reports detailing responses to sets of prompts by Large Language Models.☆32Updated 11 months ago
- Browser extension that lets you summarize and chat with any webpage using a local LLM of your choice.☆22Updated last year