soketlabs / coomLinks
A training framework for large-scale language models based on Megatron-Core, the COOM Training Framework is designed to efficiently handle extensive model training inspired by Deepseek's HAI-LLM optimizations.
☆24Updated last month
Alternatives and similar repositories for coom
Users that are interested in coom are comparing it to the libraries listed below
Sorting:
- A repository to unravel the language of GPUs, making their kernel conversations easy to understand☆195Updated 6 months ago
- A lightweight evaluation suite tailored specifically for assessing Indic LLMs across a diverse range of tasks☆38Updated last year
- rl from zero pretrain, can it be done? yes.☆282Updated 2 months ago
- A repository consisting of paper/architecture replications of classic/SOTA AI/ML papers in pytorch☆395Updated last month
- ☆46Updated 8 months ago
- "LLM from Zero to Hero: An End-to-End Large Language Model Journey from Data to Application!"☆141Updated last month
- Curated collection of community environments☆195Updated this week
- in this repository, i'm going to implement increasingly complex llm inference optimizations☆75Updated 6 months ago
- Following Karpathy with GPT-2 implementation and training, writing lots of comments cause I have memory of a goldfish☆172Updated last year
- everything i know about cuda and triton☆13Updated 10 months ago
- This repo has all the basic things you'll need in-order to understand complete vision transformer architecture and its various implementa…☆228Updated 11 months ago
- So, I trained a Llama a 130M architecture I coded from ground up to build a small instruct model from scratch. Trained on FineWeb dataset…☆16Updated 8 months ago
- This repository contains the code for dataset curation and finetuning of instruct variant of the Bilingual OpenHathi model. The resultin…☆23Updated last year
- small auto-grad engine inspired from Karpathy's micrograd and PyTorch☆277Updated last year
- ☆45Updated 6 months ago
- A zero-to-one guide on scaling modern transformers with n-dimensional parallelism.☆105Updated 2 months ago
- ☆89Updated 8 months ago
- an open source reproduction of NVIDIA's nGPT (Normalized Transformer with Representation Learning on the Hypersphere)☆108Updated 9 months ago
- Simple & Scalable Pretraining for Neural Architecture Research☆305Updated 2 weeks ago
- a tiny vectorstore implementation built with numpy.☆63Updated last year
- A curated list of resources for learning and exploring Triton, OpenAI's programming language for writing efficient GPU code.☆441Updated 9 months ago
- An interface library for RL post training with environments.☆848Updated this week
- ☆532Updated 4 months ago
- ☆225Updated 3 weeks ago
- ☆29Updated last year
- ☆213Updated last week
- Simple Transformer in Jax☆140Updated last year
- ⚖️ Awesome LLM Judges ⚖️☆146Updated 7 months ago
- ☆45Updated 5 months ago
- Learnings and programs related to CUDA☆428Updated 5 months ago