DreamGenX / DreamGenTrain
☆14Updated last year
Alternatives and similar repositories for DreamGenTrain:
Users that are interested in DreamGenTrain are comparing it to the libraries listed below
- Model REVOLVER, a human in the loop model mixing system.☆33Updated last year
- idea: https://github.com/nyxkrage/ebook-groupchat/☆86Updated 7 months ago
- Low-Rank adapter extraction for fine-tuned transformers models☆171Updated 10 months ago
- QLoRA: Efficient Finetuning of Quantized LLMs☆77Updated 11 months ago
- The one who calls upon functions - Function-Calling Language Model☆36Updated last year
- An unsupervised model merging algorithm for Transformers-based language models.☆104Updated 11 months ago
- run ollama & gguf easily with a single command☆50Updated 10 months ago
- This is our own implementation of 'Layer Selective Rank Reduction'☆233Updated 10 months ago
- Scripts to create your own moe models using mlx☆89Updated last year
- Little AI roleplay program☆56Updated last year
- ☆111Updated 3 months ago
- Convenient wrapper for fine-tuning and inference of Large Language Models (LLMs) with several quantization techniques (GTPQ, bitsandbytes…☆147Updated last year
- Let's create synthetic textbooks together :)☆74Updated last year
- Generate Synthetic Data Using OpenAI, MistralAI or AnthropicAI☆224Updated 11 months ago
- entropix style sampling + GUI☆25Updated 5 months ago
- Some simple scripts that I use day-to-day when working with LLMs and Huggingface Hub☆159Updated last year
- Full finetuning of large language models without large memory requirements☆93Updated last year
- GPT-2 small trained on phi-like data☆65Updated last year
- Experimental sampler to make LLMs more creative☆30Updated last year
- Text generation in Python, as easy as possible☆56Updated 2 weeks ago
- Traing PRO extension for oobabooga WebUI - recent dev version☆48Updated 2 months ago
- ☆152Updated 8 months ago
- Accepts a Hugging Face model URL, automatically downloads and quantizes it using Bits and Bytes.☆38Updated last year
- Image Diffusion block merging technique applied to transformers based Language Models.☆54Updated last year
- Landmark Attention: Random-Access Infinite Context Length for Transformers QLoRA☆123Updated last year
- A fast batching API to serve LLM models☆183Updated 11 months ago
- Client-side toolkit for using large language models, including where self-hosted☆107Updated 4 months ago
- Video+code lecture on building nanoGPT from scratch☆66Updated 9 months ago
- 5X faster 60% less memory QLoRA finetuning☆21Updated 10 months ago
- inference code for mixtral-8x7b-32kseqlen☆99Updated last year