fairydreaming / farel-bench
Testing LLM reasoning abilities with family relationship quizzes.
☆42Updated 2 weeks ago
Related projects ⓘ
Alternatives and complementary repositories for farel-bench
- An easy-to-understand framework for LLM samplers that rewind and revise generated tokens☆105Updated last week
- ☆49Updated 7 months ago
- Full finetuning of large language models without large memory requirements☆93Updated 10 months ago
- An introduction to LLM Sampling☆18Updated this week
- ☆116Updated 2 months ago
- Video+code lecture on building nanoGPT from scratch☆64Updated 4 months ago
- Cerule - A Tiny Mighty Vision Model☆67Updated 2 months ago
- look how they massacred my boy☆53Updated 3 weeks ago
- code for training & evaluating Contextual Document Embedding models☆92Updated this week
- an implementation of Self-Extend, to expand the context window via grouped attention☆118Updated 10 months ago
- Repo for "LoLCATs: On Low-Rank Linearizing of Large Language Models"☆171Updated 3 weeks ago
- Score LLM pretraining data with classifiers☆55Updated last year
- ☆52Updated 5 months ago
- The code repository for the CURLoRA research paper. Stable LLM continual fine-tuning and catastrophic forgetting mitigation.☆37Updated 2 months ago
- Set of scripts to finetune LLMs☆36Updated 7 months ago
- ☆93Updated 2 months ago
- inference code for mixtral-8x7b-32kseqlen☆98Updated 10 months ago
- The simplest, fastest repository for training/finetuning medium-sized xLSTMs.☆38Updated 5 months ago
- GRDN.AI app for garden optimization☆69Updated 9 months ago
- Comprehensive analysis of difference in performance of QLora, Lora, and Full Finetunes.☆81Updated last year
- Fast approximate inference on a single GPU with sparsity aware offloading☆38Updated 10 months ago
- Q-GaLore: Quantized GaLore with INT4 Projection and Layer-Adaptive Low-Rank Gradients.☆171Updated 3 months ago
- Fully fine-tune large models like Mistral, Llama-2-13B, or Qwen-14B completely for free☆219Updated last week
- Collection of autoregressive model implementation☆66Updated this week
- Low-Rank adapter extraction for fine-tuned transformers model☆162Updated 6 months ago
- Scripts to create your own moe models using mlx☆86Updated 8 months ago
- Simple examples using Argilla tools to build AI☆38Updated this week
- ☆55Updated 11 months ago
- An example implementation of RLHF (or, more accurately, RLAIF) built on MLX and HuggingFace.☆20Updated 4 months ago
- An efficent implementation of the method proposed in "The Era of 1-bit LLMs"☆154Updated 3 weeks ago