fkodom / python-repo-templateLinks
Template repo for Python projects, especially those focusing on machine learning and/or deep learning.
β15Updated 6 months ago
Alternatives and similar repositories for python-repo-template
Users that are interested in python-repo-template are comparing it to the libraries listed below
Sorting:
- Implementation of the Llama architecture with RLHF + Q-learningβ168Updated 10 months ago
- A MAD laboratory to improve AI architecture designs π§ͺβ135Updated last year
- Collection of autoregressive model implementationβ85Updated 7 months ago
- Understand and test language model architectures on synthetic tasks.β246Updated 2 months ago
- NeurIPS Large Language Model Efficiency Challenge: 1 LLM + 1GPU + 1Dayβ258Updated 2 years ago
- The simplest, fastest repository for training/finetuning medium-sized GPTs.β180Updated 5 months ago
- A repository for log-time feedforward networksβ223Updated last year
- Deep learning library implemented from scratch in numpy. Mixtral, Mamba, LLaMA, GPT, ResNet, and other experiments.β54Updated last year
- Code repository for Black Mambaβ260Updated last year
- Annotated version of the Mamba paperβ492Updated last year
- FlashFFTConv: Efficient Convolutions for Long Sequences with Tensor Coresβ337Updated 11 months ago
- Code for exploring Based models from "Simple linear attention language models balance the recall-throughput tradeoff"β243Updated 6 months ago
- Prune transformer layersβ74Updated last year
- β83Updated 2 years ago
- Implementation of the conditionally routed attention in the CoLT5 architecture, in Pytorchβ231Updated last year
- β94Updated 2 years ago
- β178Updated last year
- Large scale 4D parallelism pre-training for π€ transformers in Mixture of Experts *(still work in progress)*β87Updated 2 years ago
- Just some miscellaneous utility functions / decorators / modules related to Pytorch and Accelerate to help speed up implementation of newβ¦β125Updated last year
- β91Updated last year
- nanoGPT-like codebase for LLM trainingβ113Updated last month
- Load compute kernels from the Hubβ347Updated last week
- An extension of the nanoGPT repository for training small MOE models.β218Updated 9 months ago
- some common Huggingface transformers in maximal update parametrization (Β΅P)β87Updated 3 years ago
- Implementation of ST-Moe, the latest incarnation of MoE after years of research at Brain, in Pytorchβ374Updated last year
- Implementation of Infini-Transformer in Pytorchβ113Updated 11 months ago
- Experiment of using Tangent to autodiff tritonβ81Updated last year
- Evaluating the Mamba architecture on the Othello gameβ48Updated last year
- β205Updated last week
- Various transformers for FSDP researchβ38Updated 3 years ago