namtranase / gemma-cpp-pythonLinks
A Python wrapper for gemma.cpp
β51Updated last year
Alternatives and similar repositories for gemma-cpp-python
Users that are interested in gemma-cpp-python are comparing it to the libraries listed below
Sorting:
- Maybe the new state of the art vision model? we'll see π€·ββοΈβ163Updated last year
- Fully fine-tune large models like Mistral, Llama-2-13B, or Qwen-14B completely for freeβ231Updated 7 months ago
- Friendly Terminal Assistant for Developersβ17Updated last year
- LLaVA server (llama.cpp).β179Updated last year
- An efficent implementation of the method proposed in "The Era of 1-bit LLMs"β153Updated 7 months ago
- Python bindings for ggmlβ141Updated 9 months ago
- Video+code lecture on building nanoGPT from scratchβ67Updated 11 months ago
- This is our own implementation of 'Layer Selective Rank Reduction'β238Updated last year
- β114Updated 5 months ago
- Embed arbitrary modalities (images, audio, documents, etc) into large language models.β184Updated last year
- Tune MPTsβ84Updated last year
- Google TPU optimizations for transformers modelsβ112Updated 4 months ago
- Train your own small bitnet modelβ71Updated 7 months ago
- β66Updated last year
- Generate Synthetic Data Using OpenAI, MistralAI or AnthropicAIβ222Updated last year
- β130Updated 9 months ago
- Testing and evaluating the capabilities of Vision-Language models (PaliGemma) in performing computer vision tasks such as object detectioβ¦β80Updated last year
- QLoRA: Efficient Finetuning of Quantized LLMsβ78Updated last year
- Q-GaLore: Quantized GaLore with INT4 Projection and Layer-Adaptive Low-Rank Gradients.β199Updated 10 months ago
- run paligemma in real timeβ131Updated last year
- A Gradio component that can be used to annotate images with bounding boxes.β52Updated 3 months ago
- β157Updated 10 months ago
- β49Updated last year
- Low-Rank adapter extraction for fine-tuned transformers modelsβ171Updated last year
- Convenient wrapper for fine-tuning and inference of Large Language Models (LLMs) with several quantization techniques (GTPQ, bitsandbytesβ¦β147Updated last year
- β70Updated last month
- A more memory-efficient rewrite of the HF transformers implementation of Llama for use with quantized weights.β63Updated last year
- Full finetuning of large language models without large memory requirementsβ93Updated last year
- An OpenAI Completions API compatible server for NLP transformers modelsβ65Updated last year
- Cerule - A Tiny Mighty Vision Modelβ66Updated 9 months ago