alxndrTL / ARC_LLMsLinks
Evaluating majors LLMs on the Abstraction and Reasoning Corpus
☆15Updated last year
Alternatives and similar repositories for ARC_LLMs
Users that are interested in ARC_LLMs are comparing it to the libraries listed below
Sorting:
- ☆56Updated 2 months ago
- Simple GRPO scripts and configurations.☆59Updated 5 months ago
- ARLC, a probabilistic abductive reasoner for solving Raven's progressive matrices.☆18Updated 2 months ago
- Jax like function transformation engine but micro, microjax☆33Updated 8 months ago
- Collection of autoregressive model implementation☆86Updated 2 months ago
- ☆81Updated last year
- my solution for Abstaction and reasoning challenge on kaggle☆10Updated last year
- My explorations into editing the knowledge and memories of an attention network☆35Updated 2 years ago
- LLM training in simple, raw C/CUDA☆15Updated 7 months ago
- HomebrewNLP in JAX flavour for maintable TPU-Training☆50Updated last year
- ☆53Updated last year
- Code for the examples presented in the talk "Training a Llama in your backyard: fine-tuning very large models on consumer hardware" given…☆14Updated last year
- JAX notebook showing how to LoRA + GPTQ arbitrary models☆10Updated last year
- ☆27Updated 10 months ago
- Official repository for the paper "Approximating Two-Layer Feedforward Networks for Efficient Transformers"☆38Updated last month
- ☆61Updated last year
- ☆45Updated last year
- A library for squeakily cleaning and filtering language datasets.☆47Updated 2 years ago
- ☆36Updated last year
- Explorations into the proposal from the paper "Grokfast, Accelerated Grokking by Amplifying Slow Gradients"☆101Updated 7 months ago
- ☆48Updated 8 months ago
- Simple repository for training small reasoning models☆33Updated 5 months ago
- FastFeedForward Networks☆20Updated last year
- PyTorch implementation for MRL☆19Updated last year
- ☆53Updated 8 months ago
- GoldFinch and other hybrid transformer components☆46Updated last year
- ☆18Updated last year
- ☆98Updated 6 months ago
- Large scale 4D parallelism pre-training for 🤗 transformers in Mixture of Experts *(still work in progress)*☆85Updated last year
- ☆23Updated 7 months ago