modular / mojo-gpu-puzzlesLinks
Learn GPU Programming in Mojo🔥 by Solving Puzzles
☆65Updated last week
Alternatives and similar repositories for mojo-gpu-puzzles
Users that are interested in mojo-gpu-puzzles are comparing it to the libraries listed below
Sorting:
- Dynamic Neural Networks and Function Transformations in Python + Mojo☆254Updated this week
- A Learning Journey: Micrograd in Mojo 🔥☆61Updated 8 months ago
- A working machine learning framework in pure Mojo 🔥☆129Updated last year
- ☆54Updated last year
- ☆28Updated 9 months ago
- A fast and compact Dict implementation in Mojo 🔥☆35Updated 11 months ago
- Write a fast kernel and run it on Discord. See how you compare against the best!☆46Updated this week
- Implementation of Karpathy's micrograd in Mojo☆73Updated last year
- port of Andrjey Karpathy's llm.c to Mojo☆352Updated 6 months ago
- Mojo🔥 language test runner plugin for pytest. (aka pytest-mojo)☆49Updated last month
- Accurate, Hardware Accelerated, Special Functions in Mojo 🔥☆35Updated 6 months ago
- GPU documentation for humans☆70Updated last week
- NuMojo is a library for numerical computing in Mojo 🔥 similar to numpy in Python.☆164Updated last week
- Contains supporting materials for developer relations blog posts, videos, and workshops☆45Updated last week
- A Mojo implementation of the Tiny Stable Diffusion model☆52Updated last year
- Tensor library with autograd using only Rust's standard library☆68Updated 11 months ago
- C API for MLX☆115Updated 2 months ago
- Machine Learning algorithms in pure Mojo 🔥☆36Updated 3 weeks ago
- 📖 Learn some mojo !☆132Updated 11 months ago
- A Machine Learning framework from scratch in Pure Mojo 🔥☆441Updated 5 months ago
- A user-friendly tool chain that enables the seamless execution of ONNX models using JAX as the backend.☆114Updated this week
- ☆24Updated 11 months ago
- A stand-alone implementation of several NumPy dtype extensions used in machine learning.☆276Updated 3 weeks ago
- SIMD quantization kernels☆71Updated last week
- ctypes wrappers for HIP, CUDA, and OpenCL☆130Updated 11 months ago
- PCCL (Prime Collective Communications Library) implements fault tolerant collective communications over IP☆95Updated last month
- ☆130Updated last year
- Small scale distributed training of sequential deep learning models, built on Numpy and MPI.☆134Updated last year
- ☆97Updated last week
- LLM training in simple, raw C/CUDA, migrated into Rust☆46Updated 3 months ago