huggingface / api-inference-communityLinks
β169Updated 3 months ago
Alternatives and similar repositories for api-inference-community
Users that are interested in api-inference-community are comparing it to the libraries listed below
Sorting:
- Hugging Face's Zapier Integration π€β‘οΈβ48Updated 2 years ago
- The package used to build the documentation of our Hugging Face reposβ115Updated this week
- β198Updated last year
- Notus is a collection of fine-tuned LLMs using SFT, DPO, SFT+DPO, and/or any other RLHF techniques, while always keeping a data-first appβ¦β167Updated last year
- Exploring finetuning public checkpoints on filter 8K sequences on Pileβ114Updated 2 years ago
- manage histories of LLM applied applicationsβ87Updated last year
- **ARCHIVED** Filesystem interface to π€ Hubβ58Updated 2 years ago
- Use OpenAI with HuggingChat by emulating the text_generation_inference_serverβ44Updated last year
- [WIP] A π₯ interface for running code in the cloudβ85Updated 2 years ago
- QLoRA: Efficient Finetuning of Quantized LLMsβ78Updated last year
- Experiments with generating opensource language model assistantsβ97Updated 2 years ago
- Unofficial python bindings for the rust llm library. πβ€οΈπ¦β75Updated last year
- Convenient wrapper for fine-tuning and inference of Large Language Models (LLMs) with several quantization techniques (GTPQ, bitsandbytesβ¦β147Updated last year
- A voice-enabled chatbot application built using of π¦οΈπ LangChain, text-to-speech, and speech-to-text models from π€ Hugging Face, and β¦β191Updated last year
- Accelerated inference of π€ models using FuriosaAI NPU chips.β26Updated 11 months ago
- An easy-to-use LLMs quantization package with user-friendly apis, based on GPTQ algorithm.β37Updated last year
- β141Updated last year
- β123Updated 7 months ago
- Some notebooks about how we can fine-tune LLMβ22Updated 2 years ago
- β84Updated last year
- Pipeline for pulling and processing online language model pretraining data from the webβ178Updated last year
- HuggingChat like UI in Gradioβ70Updated 2 years ago
- Comprehensive analysis of difference in performance of QLora, Lora, and Full Finetunes.β81Updated last year
- DeepSpeed is a deep learning optimization library that makes distributed training easy, efficient, and effective.β166Updated last week
- Used for adaptive human in the loop evaluation of language and embedding models.β308Updated 2 years ago
- β67Updated 2 years ago
- An OpenAI Completions API compatible server for NLP transformers modelsβ65Updated last year
- Reimplementation of the task generation part from the Alpaca paperβ119Updated 2 years ago
- 4 bits quantization of SantaCoder using GPTQβ50Updated 2 years ago
- β17Updated 2 weeks ago