sdascoli / boolformerLinks
☆164Updated last year
Alternatives and similar repositories for boolformer
Users that are interested in boolformer are comparing it to the libraries listed below
Sorting:
- Evaluation of neuro-symbolic engines☆41Updated last year
- ☆69Updated last year
- Code to reproduce "Transformers Can Do Arithmetic with the Right Embeddings", McLeish et al (NeurIPS 2024)☆198Updated last year
- Brain-Inspired Modular Training (BIMT), a method for making neural networks more modular and interpretable.☆174Updated 2 years ago
- Learning Universal Predictors☆81Updated last year
- ☆82Updated last year
- ☆105Updated last year
- Q-Probe: A Lightweight Approach to Reward Maximization for Language Models☆41Updated last year
- Repository for code used in the xVal paper☆147Updated last year
- Mixing Language Models with Self-Verification and Meta-Verification☆111Updated last year
- ☆238Updated last month
- ☆62Updated 2 years ago
- Explorations into the proposal from the paper "Grokfast, Accelerated Grokking by Amplifying Slow Gradients"☆103Updated last year
- Swarming algorithms like PSO, Ant Colony, Sakana, and more in PyTorch 😊☆136Updated this week
- ☆109Updated 5 months ago
- σ-GPT: A New Approach to Autoregressive Models☆70Updated last year
- Memoria is a human-inspired memory architecture for neural networks.☆82Updated last year
- Examining how large language models (LLMs) perform across various synthetic regression tasks when given (input, output) examples in their…☆160Updated 3 months ago
- Predicting the Future of AI with AI: High-quality link prediction in an exponentially growing knowledge network☆81Updated 2 years ago
- ☆56Updated last year
- ☆167Updated 2 years ago
- Understanding how features learned by neural networks evolve throughout training☆41Updated last year
- Google Research☆46Updated 3 years ago
- The GeoV model is a large langauge model designed by Georges Harik and uses Rotary Positional Embeddings with Relative distances (RoPER).…☆121Updated 2 years ago
- Official Repository of Pretraining Without Attention (BiGS), BiGS is the first model to achieve BERT-level transfer learning on the GLUE …☆116Updated last year
- ☆53Updated 2 years ago
- ☆31Updated 2 years ago
- Functional Benchmarks and the Reasoning Gap☆89Updated last year
- Repo for solving arc problems with an Neural Cellular Automata☆23Updated 8 months ago
- Entailment self-training☆25Updated 2 years ago