JackZeng0208 / llama.cpp-android-tutorialLinks
llama.cpp tutorial on Android phone
☆143Updated 8 months ago
Alternatives and similar repositories for llama.cpp-android-tutorial
Users that are interested in llama.cpp-android-tutorial are comparing it to the libraries listed below
Sorting:
- A mobile Implementation of llama.cpp☆324Updated last year
- ☆281Updated last month
- IRIS is an android app for interfacing with GGUF / llama.cpp models locally.☆261Updated 11 months ago
- A mobile Implementation of llama.cpp☆26Updated 2 years ago
- This is a simple shell script to install the alpaca llama 7B model on termux for Android phones. All credits goes to the original develop…☆64Updated 2 years ago
- Inference Llama 2 in one file of pure C☆43Updated 2 years ago
- 使用Android手机的CPU推理stable diffusion☆157Updated 2 years ago
- A set of bash scripts to automate deployment of GGML/GGUF models [default: RWKV] with the use of KoboldCpp on Android - Termux☆44Updated last year
- automatically quant GGUF models☆219Updated 3 weeks ago
- React Native binding of llama.cpp☆45Updated last month
- LLaMA Server combines the power of LLaMA C++ with the beauty of Chatbot UI.☆130Updated 2 years ago
- ☆66Updated last year
- Testing LLM reasoning abilities with family relationship quizzes.☆63Updated 11 months ago
- ☆127Updated last year
- Tool to download models from Huggingface Hub and convert them to GGML/GGUF for llama.cpp☆167Updated 8 months ago
- Port of Facebook's LLaMA model in C/C++☆105Updated last month
- High-speed and easy-use LLM serving framework for local deployment☆140Updated 5 months ago
- Docker compose to run vLLM on Windows☆113Updated 2 years ago
- After my server ui improvements were successfully merged, consider this repo a playground for experimenting, tinkering and hacking around…☆53Updated last year
- A minimal Android demo app for Kokoro-TTS☆39Updated 11 months ago
- Run SD1.x/2.x/3.x, SDXL, and FLUX.1 on your phone device☆78Updated 5 months ago
- OpenGPT 4o is a free alternative to OpenAI GPT 4o☆211Updated last year
- Cortex.Tensorrt-LLM is a C++ inference library that can be loaded by any server at runtime. It submodules NVIDIA’s TensorRT-LLM for GPU a…☆42Updated last year
- An OpenAI API compatible API for chat with image input and questions about the images. aka Multimodal.☆267Updated 10 months ago
- Running any GGUF SLMs/LLMs locally, on-device in Android☆639Updated last week
- Comparison of the output quality of quantization methods, using Llama 3, transformers, GGUF, EXL2.☆165Updated last year
- LLM inference in C/C++☆21Updated 9 months ago
- This reference can be used with any existing OpenAI integrated apps to run with TRT-LLM inference locally on GeForce GPU on Windows inste…☆127Updated last year
- ☆108Updated 4 months ago
- Instructions for installing Open Interpreter on your Android device.☆236Updated last year