adamkarvonen / train_ChessGPT
A repository for training nanogpt-based Chess playing language models.
☆23Updated 9 months ago
Alternatives and similar repositories for train_ChessGPT:
Users that are interested in train_ChessGPT are comparing it to the libraries listed below
- A repo to evaluate various LLM's chess playing abilities.☆75Updated 9 months ago
- ☆37Updated 6 months ago
- look how they massacred my boy☆63Updated 3 months ago
- ☆60Updated last year
- ☆48Updated last year
- Simple Transformer in Jax☆130Updated 7 months ago
- Jax like function transformation engine but micro, microjax☆30Updated 3 months ago
- An example implementation of RLHF (or, more accurately, RLAIF) built on MLX and HuggingFace.☆22Updated 7 months ago
- This repository explains and provides examples for "concept anchoring" in GPT4.☆72Updated last year
- Chat Markup Language conversation library☆55Updated last year
- ☆80Updated 3 weeks ago
- Simplex Random Feature attention, in PyTorch☆72Updated last year
- Minimal, clean code implementation of RAG with mlx using gguf model weights☆46Updated 9 months ago
- Port of Andrej Karpathy's nanoGPT to Apple MLX framework.☆104Updated 11 months ago
- an implementation of Self-Extend, to expand the context window via grouped attention☆118Updated last year
- Scaling is a distributed training library and installable dependency designed to scale up neural networks, with a dedicated module for tr…☆53Updated 3 months ago
- Visualizing the internal board state of a GPT trained on chess PGN strings, and performing interventions on its internal board state and …☆199Updated 2 months ago
- ☆17Updated 4 months ago
- ☆18Updated last year
- Comprehensive analysis of difference in performance of QLora, Lora, and Full Finetunes.☆82Updated last year
- Cerule - A Tiny Mighty Vision Model☆67Updated 4 months ago
- ☆22Updated last year
- Karpathy's llama2.c transpiled to MLX for Apple Silicon☆15Updated last year
- MiniHF is an inference, human preference data collection, and fine-tuning tool for local language models. It is intended to help the user…☆163Updated this week
- run embeddings in MLX☆81Updated 4 months ago
- Full finetuning of large language models without large memory requirements☆93Updated last year
- Synthetic data derived by templating, few shot prompting, transformations on public domain corpora, and monte carlo tree search.☆30Updated last month
- σ-GPT: A New Approach to Autoregressive Models☆61Updated 5 months ago
- Public Inflection Benchmarks☆69Updated 10 months ago
- Alpha-Zero Connect Four NN trained via self play☆13Updated 4 months ago