ggml-org / ci
CI for ggml and related projects
☆28Updated this week
Alternatives and similar repositories for ci:
Users that are interested in ci are comparing it to the libraries listed below
- A super simple web interface to perform blind tests on LLM outputs.☆28Updated last year
- AirLLM 70B inference with single 4GB GPU☆12Updated 8 months ago
- Nexusflow function call, tool use, and agent benchmarks.☆19Updated 4 months ago
- GRDN.AI app for garden optimization☆70Updated last year
- llama.cpp gguf file parser for javascript☆34Updated 4 months ago
- GGML implementation of BERT model with Python bindings and quantization.☆56Updated last year
- ☆66Updated 10 months ago
- Public reports detailing responses to sets of prompts by Large Language Models.☆30Updated 3 months ago
- Port of Suno AI's Bark in C/C++ for fast inference☆52Updated last year
- Thin wrapper around GGML to make life easier☆24Updated this week
- Training hybrid models for dummies.☆20Updated 3 months ago
- Port of Facebook's LLaMA model in C/C++☆20Updated last year
- Course Project for COMP4471 on RWKV☆17Updated last year
- Local LLM inference & management server with built-in OpenAI API☆31Updated last year
- A python package for serving LLM on OpenAI-compatible API endpoints with prompt caching using MLX.☆77Updated 4 months ago
- Simple LLM inference server☆20Updated 10 months ago
- Unleash the full potential of exascale LLMs on consumer-class GPUs, proven by extensive benchmarks, with no long-term adjustments and min…☆26Updated 5 months ago
- Editor with LLM generation tree exploration☆66Updated 2 months ago
- Training a reward model for RLHF using RWKV.☆14Updated last year
- The official Python library for Formulaic☆16Updated 11 months ago
- Web browser version of StarCoder.cpp☆44Updated last year
- A library for simplifying fine tuning with multi gpu setups in the Huggingface ecosystem.☆16Updated 5 months ago
- Contains the model patches and the eval logs from the passing swe-bench-lite run.☆10Updated 9 months ago
- The hearth of The Pulsar App, fast, secure and shared inference with modern UI☆56Updated 4 months ago
- First token cutoff sampling inference example☆29Updated last year
- A python command-line tool to download & manage MLX AI models from Hugging Face.☆17Updated 7 months ago
- Yet Another (LLM) Web UI, made with Gemini☆11Updated 3 months ago
- ☆99Updated 7 months ago
- Generates grammer files from typescript for LLM generation☆37Updated last year
- Rust crate for some audio utilities☆22Updated last month