CohleM / lilLMLinks
A little(lil) Language Model (LM). A tiny reproduction of LLaMA 3's model architecture.
☆55Updated 9 months ago
Alternatives and similar repositories for lilLM
Users that are interested in lilLM are comparing it to the libraries listed below
Sorting:
- ☆112Updated 7 months ago
- Video+code lecture on building nanoGPT from scratch☆68Updated last year
- AI management tool☆119Updated last year
- Testing LLM reasoning abilities with family relationship quizzes.☆63Updated last year
- ☆159Updated 9 months ago
- ☆242Updated 4 months ago
- ☆109Updated 5 months ago
- A compact LLM pretrained in 9 days by using high quality data☆339Updated 10 months ago
- Transplants vocabulary between language models, enabling the creation of draft models for speculative decoding WITHOUT retraining.☆49Updated 3 months ago
- ☆137Updated last year
- An easy-to-understand framework for LLM samplers that rewind and revise generated tokens☆150Updated last month
- Docs for GGUF quantization (unofficial)☆366Updated 6 months ago
- ☆135Updated 9 months ago
- Local Qwen3 LLM inference. One easy-to-understand file of C source with no dependencies.☆157Updated 7 months ago
- 1.58-bit LLaMa model☆82Updated last year
- Q-GaLore: Quantized GaLore with INT4 Projection and Layer-Adaptive Low-Rank Gradients.☆201Updated last year
- Fully fine-tune large models like Mistral, Llama-2-13B, or Qwen-14B completely for free☆232Updated last year
- Train your own small bitnet model☆77Updated last year
- Low-Rank adapter extraction for fine-tuned transformers models☆180Updated last year
- Dataset Crafting w/ RAG/Wikipedia ground truth and Efficient Fine-Tuning Using MLX and Unsloth. Includes configurable dataset annotation …☆193Updated last year
- Comparison of the output quality of quantization methods, using Llama 3, transformers, GGUF, EXL2.☆165Updated last year
- REAP: Router-weighted Expert Activation Pruning for SMoE compression☆232Updated 2 months ago
- ☆207Updated last year
- A stable, fast and easy-to-use inference library with a focus on a sync-to-async API☆47Updated last year
- ☆141Updated 5 months ago
- klmbr - a prompt pre-processing technique to break through the barrier of entropy while generating text with LLMs☆86Updated last year
- Easy to use, High Performant Knowledge Distillation for LLMs☆97Updated 9 months ago
- ☆304Updated 3 months ago
- Enhancing LLMs with LoRA☆206Updated 3 months ago
- Gemma 2 optimized for your local machine.☆378Updated last year