sradc / pretraining-BERT
Pre-train BERT from scratch, with HuggingFace. Accompanies the blog post: sidsite.com/posts/bert-from-scratch
☆40Updated last year
Alternatives and similar repositories for pretraining-BERT:
Users that are interested in pretraining-BERT are comparing it to the libraries listed below
- ☆60Updated last year
- gzip Predicts Data-dependent Scaling Laws☆34Updated 8 months ago
- Jax like function transformation engine but micro, microjax☆30Updated 3 months ago
- ☆22Updated last year
- A place to store reusable transformer components of my own creation or found on the interwebs☆46Updated this week
- A lightweight PyTorch implementation of the Transformer-XL architecture proposed by Dai et al. (2019)☆37Updated 2 years ago
- Scripts to prep PC for development use after OS installs☆37Updated last week
- QLoRA for Masked Language Modeling☆21Updated last year
- ☆78Updated 10 months ago
- Gzip and nearest neighbors for text classification☆56Updated last year
- Your favourite classical machine learning algos on the GPU/TPU☆20Updated last month
- An introduction to LLM Sampling☆75Updated 2 months ago
- ☆47Updated 2 months ago
- BPE modification that implements removing of the intermediate tokens during tokenizer training.☆25Updated 2 months ago
- ☆92Updated last year
- PyTorch implementation for MRL☆18Updated 11 months ago
- a pipeline for using api calls to agnostically convert unstructured data into structured training data☆29Updated 4 months ago
- utilities for loading and running text embeddings with onnx☆44Updated 6 months ago
- Implementation of GateLoop Transformer in Pytorch and Jax☆87Updated 8 months ago
- ☆46Updated last year
- Deep learning library implemented from scratch in numpy. Mixtral, Mamba, LLaMA, GPT, ResNet, and other experiments.☆51Updated 10 months ago
- Collection of autoregressive model implementation☆81Updated last week
- ☆53Updated last year
- Official repository for the paper "Approximating Two-Layer Feedforward Networks for Efficient Transformers"☆36Updated last year
- ☆49Updated 11 months ago
- Comprehensive analysis of difference in performance of QLora, Lora, and Full Finetunes.☆82Updated last year
- Triton Implementation of HyperAttention Algorithm☆46Updated last year
- Supercharge huggingface transformers with model parallelism.☆76Updated 4 months ago
- Code for NeurIPS LLM Efficiency Challenge☆55Updated 10 months ago