huggingface / local-gemma
Gemma 2 optimized for your local machine.
☆360Updated 6 months ago
Alternatives and similar repositories for local-gemma:
Users that are interested in local-gemma are comparing it to the libraries listed below
- Fully fine-tune large models like Mistral, Llama-2-13B, or Qwen-14B completely for free☆230Updated 3 months ago
- ☆152Updated 7 months ago
- A Lightweight Library for AI Observability☆233Updated this week
- ☆205Updated last week
- ☆679Updated 2 weeks ago
- ☆198Updated 8 months ago
- Fast parallel LLM inference for MLX☆163Updated 7 months ago
- ☆446Updated 10 months ago
- FastMLX is a high performance production ready API to host MLX models.☆260Updated 2 months ago
- This project showcases an LLMOps pipeline that fine-tunes a small-size LLM model to prepare for the outage of the service LLM.☆295Updated this week
- ☆609Updated 2 months ago
- ☆806Updated 5 months ago
- ☆111Updated 2 months ago
- GRadient-INformed MoE☆261Updated 4 months ago
- MLX-VLM is a package for inference and fine-tuning of Vision Language Models (VLMs) on your Mac using MLX.☆845Updated 2 weeks ago
- Q-GaLore: Quantized GaLore with INT4 Projection and Layer-Adaptive Low-Rank Gradients.☆191Updated 7 months ago
- Banishing LLM Hallucinations Requires Rethinking Generalization☆270Updated 7 months ago
- Automated Identification of Redundant Layer Blocks for Pruning in Large Language Models☆217Updated 9 months ago
- Accelerate your Hugging Face Transformers 7.6-9x. Native to Hugging Face and PyTorch.☆688Updated 5 months ago
- An efficent implementation of the method proposed in "The Era of 1-bit LLMs"☆154Updated 4 months ago
- ☆113Updated 4 months ago
- Phi-3.5 for Mac: Locally-run Vision and Language Models for Apple Silicon☆260Updated 5 months ago
- ☆172Updated 6 months ago
- ☆251Updated 2 months ago
- Generate Synthetic Data Using OpenAI, MistralAI or AnthropicAI☆222Updated 9 months ago
- Official implementation of Half-Quadratic Quantization (HQQ)☆748Updated this week
- Evaluate and Enhance Your LLM Deployments for Real-World Inference Needs☆191Updated this week
- ☆207Updated 7 months ago
- Start a server from the MLX library.☆173Updated 6 months ago
- Maybe the new state of the art vision model? we'll see 🤷♂️☆160Updated last year