BottleCapAI / NoCap-TestLinks
Open Test for BottleCapAI
โ127Updated last month
Alternatives and similar repositories for NoCap-Test
Users that are interested in NoCap-Test are comparing it to the libraries listed below
Sorting:
- ๐งฑ Modula software packageโ315Updated 3 months ago
- โ285Updated last year
- Legible, Scalable, Reproducible Foundation Models with Named Tensors and Jaxโ685Updated 3 weeks ago
- Attention Kernels for Symmetric Power Transformersโ128Updated 2 months ago
- Extract full next-token probabilities via language model APIsโ248Updated last year
- The simplest, fastest repository for training/finetuning medium-sized GPTs.โ179Updated 5 months ago
- Draw more samplesโ196Updated last year
- Official JAX implementation of xLSTM including fast and efficient training and inference code. 7B model available at https://huggingface.โฆโ105Updated 11 months ago
- An interactive exploration of Transformer programming.โ270Updated 2 years ago
- Efficient optimizersโ277Updated last month
- Open-source framework for the research and development of foundation models.โ658Updated this week
- Our solution for the arc challenge 2024โ185Updated 5 months ago
- Puzzles for exploring transformersโ380Updated 2 years ago
- โ532Updated 4 months ago
- nanoGPT-like codebase for LLM trainingโ113Updated last month
- A character tokenizer for Hugging Face Transformersโ29Updated last year
- Fast bare-bones BPE for modern tokenizer trainingโ172Updated 5 months ago
- Accelerate, Optimize performance with streamlined training and serving options with JAX.โ325Updated last week
- Minimal yet performant LLM examples in pure JAXโ207Updated last week
- โ161Updated 4 months ago
- A Jax-based library for building transformers, includes implementations of GPT, Gemma, LlaMa, Mixtral, Whisper, SWin, ViT and more.โ297Updated last year
- Training API and CLIโ253Updated last week
- Understand and test language model architectures on synthetic tasks.โ245Updated 2 months ago
- The nnsight package enables interpreting and manipulating the internals of deep learned models.โ727Updated this week
- Minimalistic, extremely fast, and hackable researcher's toolbench for GPT models in 307 lines of code. Reaches <3.8 validation loss on wiโฆโ355Updated last year
- โ131Updated 11 months ago
- A MAD laboratory to improve AI architecture designs ๐งชโ135Updated 11 months ago
- โ460Updated last year
- Home for "How To Scale Your Model", a short blog-style textbook about scaling LLMs on TPUsโ724Updated 2 weeks ago
- Brain-Inspired Modular Training (BIMT), a method for making neural networks more modular and interpretable.โ174Updated 2 years ago