deep-diver / PingPong
manage histories of LLM applied applications
☆86Updated 10 months ago
Related projects: ⓘ
- Sakura-SOLAR-DPO: Merge, SFT, and DPO☆114Updated 8 months ago
- 1-Click is all you need.☆58Updated 4 months ago
- Evaluate your LLM apps, RAG pipeline, any generated text, and more!☆0Updated 4 months ago
- evolve llm training instruction, from english instruction to any language.☆108Updated last year
- Finetune Falcon, LLaMA, MPT, and RedPajama on consumer hardware using PEFT LoRA☆99Updated last month
- HuggingChat like UI in Gradio☆63Updated last year
- ☆37Updated last year
- ☆35Updated 5 months ago
- generate synthetic data for LLM fine-tuning in arbitrary situations within systematic way☆21Updated 6 months ago
- ☆31Updated 8 months ago
- QLoRA: Efficient Finetuning of Quantized LLMs☆74Updated 5 months ago
- Use OpenAI with HuggingChat by emulating the text_generation_inference_server☆45Updated last year
- Spherical Merge Pytorch/HF format Language Models with minimal feature loss.☆107Updated last year
- Low-Rank adapter extraction for fine-tuned transformers model☆154Updated 4 months ago
- This project breathes life into video characters by using AI to describe their personality and then chat with you as them.☆44Updated 6 months ago
- Experiments with generating opensource language model assistants☆97Updated last year
- ☆11Updated 5 months ago
- Convenient wrapper for fine-tuning and inference of Large Language Models (LLMs) with several quantization techniques (GTPQ, bitsandbytes…☆139Updated 11 months ago
- GPTQLoRA: Efficient Finetuning of Quantized LLMs with GPTQ☆96Updated last year
- ☆75Updated 3 weeks ago
- QLoRA with Enhanced Multi GPU Support☆36Updated last year
- An OpenAI Completions API compatible server for NLP transformers models☆54Updated 10 months ago
- Chat to LLaMa 2 that also provides responses with reference documents over vector database. Locally available model using GPTQ 4bit quant…☆30Updated 9 months ago
- ☆73Updated 8 months ago
- Anh - LAION's multilingual assistant datasets and models☆27Updated last year
- Official code for ACL 2023 (short, findings) paper "Recursion of Thought: A Divide and Conquer Approach to Multi-Context Reasoning with L…☆41Updated last year
- showing various ways to serve Keras based stable diffusion☆109Updated last year
- OSLO: Open Source for Large-scale Optimization☆172Updated last year
- RAFT, or Retrieval-Augmented Fine-Tuning, is a method comprising of a fine-tuning and a RAG-based retrieval phase. It is particularly sui…☆60Updated 2 weeks ago
- Alpaca-lora for huggingface implementation using Deepspeed and FullyShardedDataParallel☆23Updated last year