sdascoli / boolformerLinks
☆163Updated last year
Alternatives and similar repositories for boolformer
Users that are interested in boolformer are comparing it to the libraries listed below
Sorting:
- Evaluation of neuro-symbolic engines☆39Updated last year
- ☆69Updated last year
- Learning Universal Predictors☆79Updated last year
- Brain-Inspired Modular Training (BIMT), a method for making neural networks more modular and interpretable.☆173Updated 2 years ago
- Q-Probe: A Lightweight Approach to Reward Maximization for Language Models☆41Updated last year
- Memoria is a human-inspired memory architecture for neural networks.☆76Updated 11 months ago
- Google Research☆46Updated 2 years ago
- Automatic gradient descent☆213Updated 2 years ago
- Based on the tree of thoughts paper☆48Updated 2 years ago
- ☆81Updated last year
- ☆33Updated last year
- Code to reproduce "Transformers Can Do Arithmetic with the Right Embeddings", McLeish et al (NeurIPS 2024)☆193Updated last year
- ☆53Updated last year
- Swarming algorithms like PSO, Ant Colony, Sakana, and more in PyTorch 😊☆131Updated this week
- Explorations into the proposal from the paper "Grokfast, Accelerated Grokking by Amplifying Slow Gradients"☆102Updated 9 months ago
- ☆61Updated last year
- Repo for solving arc problems with an Neural Cellular Automata☆19Updated 4 months ago
- Official Repository of Pretraining Without Attention (BiGS), BiGS is the first model to achieve BERT-level transfer learning on the GLUE …☆114Updated last year
- σ-GPT: A New Approach to Autoregressive Models☆68Updated last year
- The GeoV model is a large langauge model designed by Georges Harik and uses Rotary Positional Embeddings with Relative distances (RoPER).…☆121Updated 2 years ago
- ☆101Updated 9 months ago
- ☆52Updated last year
- Functional local implementations of main model parallelism approaches☆96Updated 2 years ago
- Examining how large language models (LLMs) perform across various synthetic regression tasks when given (input, output) examples in their…☆155Updated last year
- Multibackend Graph Neural Networks in Keras 3☆25Updated last year
- ☆102Updated 2 months ago
- Entailment self-training☆25Updated 2 years ago
- ModuleFormer is a MoE-based architecture that includes two different types of experts: stick-breaking attention heads and feedforward exp…☆224Updated 3 weeks ago
- gzip Predicts Data-dependent Scaling Laws☆34Updated last year
- Extending Conformal Prediction to LLMs☆67Updated last year