huggingface / api-inference-community
β168Updated last month
Alternatives and similar repositories for api-inference-community:
Users that are interested in api-inference-community are comparing it to the libraries listed below
- β199Updated last year
- [WIP] A π₯ interface for running code in the cloudβ86Updated 2 years ago
- The package used to build the documentation of our Hugging Face reposβ106Updated last week
- manage histories of LLM applied applicationsβ88Updated last year
- Google TPU optimizations for transformers modelsβ104Updated 2 months ago
- A Multilingual Dataset for Parsing Realistic Task-Oriented Dialogsβ114Updated 2 years ago
- An easy-to-use LLMs quantization package with user-friendly apis, based on GPTQ algorithm.β37Updated last year
- Some simple scripts that I use day-to-day when working with LLMs and Huggingface Hubβ158Updated last year
- Exploring finetuning public checkpoints on filter 8K sequences on Pileβ115Updated 2 years ago
- β129Updated 2 years ago
- Lightweight demos for finetuning LLMs. Powered by π€ transformers and open-source datasets.β73Updated 5 months ago
- Patch for MPT-7B which allows using and training a LoRAβ58Updated last year
- **ARCHIVED** Filesystem interface to π€ Hubβ58Updated last year
- Experiments with generating opensource language model assistantsβ97Updated last year
- Notus is a collection of fine-tuned LLMs using SFT, DPO, SFT+DPO, and/or any other RLHF techniques, while always keeping a data-first appβ¦β166Updated last year
- QLoRA: Efficient Finetuning of Quantized LLMsβ77Updated 11 months ago
- Repo for the Belebele dataset, a massively multilingual reading comprehension dataset.β321Updated 3 months ago
- Landmark Attention: Random-Access Infinite Context Length for Transformers QLoRAβ123Updated last year
- HuggingChat like UI in Gradioβ71Updated last year
- Unofficial python bindings for the rust llm library. πβ€οΈπ¦β75Updated last year
- Convenient wrapper for fine-tuning and inference of Large Language Models (LLMs) with several quantization techniques (GTPQ, bitsandbytesβ¦β147Updated last year
- LLM finetuningβ42Updated last year
- β84Updated last year
- QLoRA with Enhanced Multi GPU Supportβ36Updated last year
- Reimplementation of the task generation part from the Alpaca paperβ119Updated last year
- experiments with inference on llamaβ104Updated 9 months ago
- Comprehensive analysis of difference in performance of QLora, Lora, and Full Finetunes.β82Updated last year
- Tune MPTsβ84Updated last year
- Fully fine-tune large models like Mistral, Llama-2-13B, or Qwen-14B completely for freeβ230Updated 4 months ago
- inference code for mixtral-8x7b-32kseqlenβ99Updated last year