fairydreaming / farel-bench
Testing LLM reasoning abilities with family relationship quizzes.
☆42Updated this week
Related projects ⓘ
Alternatives and complementary repositories for farel-bench
- Video+code lecture on building nanoGPT from scratch☆64Updated 5 months ago
- ☆118Updated 3 months ago
- ☆49Updated 8 months ago
- Full finetuning of large language models without large memory requirements☆93Updated 10 months ago
- Set of scripts to finetune LLMs☆36Updated 7 months ago
- The simplest, fastest repository for training/finetuning medium-sized xLSTMs.☆38Updated 5 months ago
- Cerule - A Tiny Mighty Vision Model☆67Updated 2 months ago
- An easy-to-understand framework for LLM samplers that rewind and revise generated tokens☆113Updated 3 weeks ago
- Gpu benchmark☆43Updated last month
- An introduction to LLM Sampling☆64Updated last week
- code for training & evaluating Contextual Document Embedding models☆117Updated this week
- ☆57Updated last week
- an implementation of Self-Extend, to expand the context window via grouped attention☆118Updated 10 months ago
- RWKV-7: Surpassing GPT☆44Updated this week
- ☆64Updated 5 months ago
- An efficent implementation of the method proposed in "The Era of 1-bit LLMs"☆154Updated last month
- Repo for "LoLCATs: On Low-Rank Linearizing of Large Language Models"☆177Updated last month
- A single repo with all scripts and utils to train / fine-tune the Mamba model with or without FIM☆50Updated 7 months ago
- Simple examples using Argilla tools to build AI☆40Updated this week
- ☆93Updated last month
- Fully fine-tune large models like Mistral, Llama-2-13B, or Qwen-14B completely for free☆221Updated 3 weeks ago
- Collection of autoregressive model implementation☆67Updated this week
- Demonstration that finetuning RoPE model on larger sequences than the pre-trained model adapts the model context limit☆63Updated last year
- ☆53Updated 5 months ago
- Low-Rank adapter extraction for fine-tuned transformers model☆162Updated 6 months ago
- look how they massacred my boy☆58Updated last month
- Train your own small bitnet model☆56Updated last month
- ☆57Updated 11 months ago
- Tree Attention: Topology-aware Decoding for Long-Context Attention on GPU clusters☆104Updated last month
- implementation of https://arxiv.org/pdf/2312.09299☆19Updated 4 months ago