msaroufim / vscode-pytorch-extensionLinks
☆12Updated 2 years ago
Alternatives and similar repositories for vscode-pytorch-extension
Users that are interested in vscode-pytorch-extension are comparing it to the libraries listed below
Sorting:
- ☆97Updated last week
- Functional local implementations of main model parallelism approaches☆96Updated 2 years ago
- BricksRL: A Platform for Democratizing Robotics and Reinforcement Learning Research and Education with LEGO☆62Updated 10 months ago
- Solve puzzles. Learn CUDA.☆64Updated last year
- seqax = sequence modeling + JAX☆165Updated 3 weeks ago
- ML/DL Math and Method notes☆63Updated last year
- A puzzle to learn about prompting☆132Updated 2 years ago
- Graph neural networks in JAX.☆67Updated last year
- Swarming algorithms like PSO, Ant Colony, Sakana, and more in PyTorch 😊☆129Updated 2 weeks ago
- Library for text-to-text regression, applicable to any input string representation and allows pretraining and fine-tuning over multiple r…☆124Updated this week
- Gradient Boosting Reinforcement Learning (GBRL)☆118Updated 3 weeks ago
- ☆45Updated last year
- Flax (Jax) implementation of DeepSeek-R1-Distill-Qwen-1.5B with weights ported from Hugging Face.☆22Updated 5 months ago
- Machine Learning eXperiment Utilities☆46Updated 2 weeks ago
- Cost aware hyperparameter tuning algorithm☆167Updated last year
- Official codebase for "Quantile Reward Policy Optimization: Alignment with Pointwise Regression and Exact Partition Functions" (Matrenok …☆24Updated last month
- Project 2 (Building Large Language Models) for Stanford CS324: Understanding and Developing Large Language Models (Winter 2022)☆105Updated 2 years ago
- ☆98Updated last week
- Clean RL implementation using MLX☆32Updated last year
- A FlashAttention implementation for JAX with support for efficient document mask computation and context parallelism.☆135Updated 4 months ago
- JAX implementation of the Llama 2 model☆219Updated last year
- A Jax-based library for building transformers, includes implementations of GPT, Gemma, LlaMa, Mixtral, Whisper, SWin, ViT and more.☆290Updated 11 months ago
- Intrinsic Motivation from Artificial Intelligence Feedback☆131Updated last year
- Large scale 4D parallelism pre-training for 🤗 transformers in Mixture of Experts *(still work in progress)*☆87Updated last year
- Minimal (400 LOC) implementation Maximum (multi-node, FSDP) GPT training☆130Updated last year
- A set of Python scripts that makes your experience on TPU better☆54Updated last year
- Learn online intrinsic rewards from LLM feedback☆43Updated 7 months ago
- ☆88Updated last year
- A reinforcement learning framework based on MLX.☆235Updated 5 months ago
- Simple repository for training small reasoning models☆32Updated 6 months ago