pHaeusler / tinycatstoriesLinks
☆10Updated 2 years ago
Alternatives and similar repositories for tinycatstories
Users that are interested in tinycatstories are comparing it to the libraries listed below
Sorting:
- Multipack distributed sampler for fast padding-free training of LLMs☆199Updated last year
- Tune MPTs☆84Updated 2 years ago
- SparseGPT + GPTQ Compression of LLMs like LLaMa, OPT, Pythia☆41Updated 2 years ago
- Let us make Psychohistory (as in Asimov) a reality, and accessible to everyone. Useful for LLM grounding and games / fiction / business /…☆40Updated 2 years ago
- Landmark Attention: Random-Access Infinite Context Length for Transformers☆423Updated last year
- SAIL: Search Augmented Instruction Learning☆157Updated last month
- This project aims to make RWKV Accessible to everyone using a Hugging Face like interface, while keeping it close to the R and D RWKV bra…☆65Updated 2 years ago
- RWKV infctx trainer, for training arbitary context sizes, to 10k and beyond!☆147Updated last year
- Embed arbitrary modalities (images, audio, documents, etc) into large language models.☆187Updated last year
- Evaluating LLMs with Dynamic Data☆91Updated last month
- A bagel, with everything.☆324Updated last year
- RWKV in nanoGPT style☆192Updated last year
- Simple, hackable and fast implementation for training/finetuning medium-sized LLaMA-based models☆177Updated this week
- BABILong is a benchmark for LLM evaluation using the needle-in-a-haystack approach.☆210Updated 3 months ago
- Framework agnostic python runtime for RWKV models☆146Updated 2 years ago
- ☆96Updated 2 years ago
- Merge Transformers language models by use of gradient parameters.☆208Updated last year
- an implementation of Self-Extend, to expand the context window via grouped attention☆118Updated last year
- QLoRA: Efficient Finetuning of Quantized LLMs☆78Updated last year
- PyTorch implementation of Infini-Transformer from "Leave No Context Behind: Efficient Infinite Context Transformers with Infini-attention…☆291Updated last year
- MiniHF is an inference, human preference data collection, and fine-tuning tool for local language models. It is intended to help the user…☆180Updated last month
- ☆98Updated 2 years ago
- Spherical Merge Pytorch/HF format Language Models with minimal feature loss.☆136Updated last year
- Load multiple LoRA modules simultaneously and automatically switch the appropriate combination of LoRA modules to generate the best answe…☆158Updated last year
- TART: A plug-and-play Transformer module for task-agnostic reasoning☆200Updated 2 years ago
- ☆66Updated last year
- Tune LLaMa-7B on Alpaca Dataset using PEFT / LORA Based on @zphang's https://github.com/zphang/minimal-llama scripts.☆25Updated 2 years ago
- Fast modular code to create and train cutting edge LLMs☆68Updated last year
- Experiments with generating opensource language model assistants☆97Updated 2 years ago
- RWKV-7: Surpassing GPT☆94Updated 9 months ago