Jellyfish042 / uncheatable_eval
Evaluating LLMs with Dynamic Data
☆87Updated 3 weeks ago
Alternatives and similar repositories for uncheatable_eval
Users that are interested in uncheatable_eval are comparing it to the libraries listed below
Sorting:
- ☆121Updated this week
- RWKV-7: Surpassing GPT☆84Updated 5 months ago
- RWKV, in easy to read code☆72Updated last month
- Parameter-Efficient Sparsity Crafting From Dense to Mixture-of-Experts for Instruction Tuning on General Tasks☆143Updated 7 months ago
- Fast modular code to create and train cutting edge LLMs☆66Updated last year
- Official implementation for 'Extending LLMs’ Context Window with 100 Samples'☆77Updated last year
- RWKV infctx trainer, for training arbitary context sizes, to 10k and beyond!☆148Updated 9 months ago
- ☆18Updated 4 months ago
- Positional Skip-wise Training for Efficient Context Window Extension of LLMs to Extremely Length (ICLR 2024)☆205Updated 11 months ago
- Implementation of the paper: "Leave No Context Behind: Efficient Infinite Context Transformers with Infini-attention" from Google in pyTO…☆55Updated 3 weeks ago
- Code repository for the c-BTM paper☆106Updated last year
- ☆51Updated 6 months ago
- A large-scale RWKV v6, v7(World, ARWKV, PRWKV) inference. Capable of inference by combining multiple states(Pseudo MoE). Easy to deploy o…☆35Updated last week
- GPTQLoRA: Efficient Finetuning of Quantized LLMs with GPTQ☆103Updated last year
- Experiments on speculative sampling with Llama models☆126Updated last year
- Official github repo for the paper "Compression Represents Intelligence Linearly" [COLM 2024]☆134Updated 7 months ago
- My Implementation of Q-Sparse: All Large Language Models can be Fully Sparsely-Activated☆32Updated 9 months ago
- RWKV in nanoGPT style☆189Updated 11 months ago
- Multipack distributed sampler for fast padding-free training of LLMs☆189Updated 9 months ago
- QuIP quantization☆52Updated last year
- SparseGPT + GPTQ Compression of LLMs like LLaMa, OPT, Pythia☆41Updated 2 years ago
- A repository for research on medium sized language models.☆76Updated 11 months ago
- Layer-Condensed KV cache w/ 10 times larger batch size, fewer params and less computation. Dramatic speed up with better task performance…☆149Updated last month