Sentdex / Lambda-Cloud
Helpers and such for working with Lambda Cloud
☆51Updated last year
Alternatives and similar repositories for Lambda-Cloud:
Users that are interested in Lambda-Cloud are comparing it to the libraries listed below
- Comprehensive analysis of difference in performance of QLora, Lora, and Full Finetunes.☆82Updated last year
- Minimal example scripts of the Hugging Face Trainer, focused on staying under 150 lines☆197Updated 9 months ago
- ☆92Updated last year
- Simple embedding -> text model trained on a small subset of Wikipedia sentences.☆153Updated last year
- ☆143Updated last year
- ☆48Updated last year
- Functional local implementations of main model parallelism approaches☆95Updated last year
- Simple Transformer in Jax☆136Updated 7 months ago
- ☆22Updated last year
- Fully fine-tune large models like Mistral, Llama-2-13B, or Qwen-14B completely for free☆230Updated 3 months ago
- Highly commented implementations of Transformers in PyTorch☆132Updated last year
- A puzzle to learn about prompting☆124Updated last year
- ☆60Updated last year
- Full finetuning of large language models without large memory requirements☆93Updated last year
- Just large language models. Hackable, with as little abstraction as possible. Done for my own purposes, feel free to rip.☆44Updated last year
- A miniture AI training framework for PyTorch☆39Updated 2 weeks ago
- The history files when recording human interaction while solving ARC tasks☆97Updated this week
- [Added T5 support to TRLX] A repo for distributed training of language models with Reinforcement Learning via Human Feedback (RLHF)☆47Updated 2 years ago
- git extension for {collaborative, communal, continual} model development☆207Updated 3 months ago
- A miniature version of Modal☆19Updated 8 months ago
- An introduction to LLM Sampling☆75Updated 2 months ago
- Drop in replacement for OpenAI, but with Open models.☆154Updated last year
- Exploring finetuning public checkpoints on filter 8K sequences on Pile☆115Updated last year
- ☆94Updated last year
- Minimalistic, extremely fast, and hackable researcher's toolbench for GPT models in 307 lines of code. Reaches <3.8 validation loss on wi…☆341Updated 6 months ago
- ☆65Updated 2 years ago
- ☆412Updated last year
- Run GGML models with Kubernetes.☆174Updated last year
- an implementation of Self-Extend, to expand the context window via grouped attention☆118Updated last year
- ☆199Updated last year