johnmarktaylor91 / torchlensLinks
Package for extracting and mapping the results of every single tensor operation in a PyTorch model in one line of code.
☆636Updated 4 months ago
Alternatives and similar repositories for torchlens
Users that are interested in torchlens are comparing it to the libraries listed below
Sorting:
- Tensors, for human consumption☆1,353Updated 2 weeks ago
- torchview: visualize pytorch models☆1,029Updated 8 months ago
- TensorDict is a pytorch dedicated tensor container.☆1,003Updated last week
- TensorHue is a Python library that allows you to visualize tensors right in your console, making understanding and debugging tensor conte…☆124Updated 11 months ago
- Interactively inspect module inputs, outputs, parameters, and gradients.☆354Updated last month
- Puzzles for exploring transformers☆384Updated 2 years ago
- CIFAR-10 speedruns: 94% in 2.6 seconds and 96% in 27 seconds☆352Updated 2 months ago
- ☆215Updated last year
- A tool to analyze and debug neural networks in pytorch. Use a GUI to traverse the computation graph and view the data from many different…☆298Updated last year
- Universal Notation for Tensor Operations in Python.☆464Updated 9 months ago
- For optimization algorithm research and development.☆558Updated 3 weeks ago
- ☆492Updated last year
- ☆793Updated 2 weeks ago
- Annotated version of the Mamba paper☆495Updated last year
- Speed up model training by fixing data loading.☆575Updated this week
- A library to inspect and extract intermediate layers of PyTorch models.☆476Updated 3 years ago
- Official Implementation of "ADOPT: Modified Adam Can Converge with Any β2 with the Optimal Rate"☆434Updated last year
- ☆289Updated last year
- Library for Jacobian descent with PyTorch. It enables the optimization of neural networks with multiple losses (e.g. multi-task learning)…☆298Updated this week
- Best practices & guides on how to write distributed pytorch training code☆575Updated 3 months ago
- Train to 94% on CIFAR-10 in <6.3 seconds on a single A100. Or ~95.79% in ~110 seconds (or less!)☆1,299Updated last year
- MLCommons Algorithmic Efficiency is a benchmark and competition measuring neural network training speedups due to algorithmic improvement…☆406Updated this week
- Type annotations and runtime checking for shape and dtype of JAX/NumPy/PyTorch/etc. arrays. https://docs.kidger.site/jaxtyping/☆1,723Updated last week
- Legible, Scalable, Reproducible Foundation Models with Named Tensors and Jax☆693Updated last week
- Exca - Execution and caching tool for python☆113Updated this week
- Deep learning for dummies. All the practical details and useful utilities that go into working with real models.☆829Updated 6 months ago
- Implementation of Diffusion Transformer (DiT) in JAX☆306Updated last year
- Reliable, minimal and scalable library for pretraining foundation and world models☆123Updated last week
- Code release for "Git Re-Basin: Merging Models modulo Permutation Symmetries"☆502Updated 2 years ago
- The boundary of neural network trainability is fractal☆221Updated last year