Load and run Llama from safetensors files in C
☆15Oct 24, 2024Updated last year
Alternatives and similar repositories for llama_st
Users that are interested in llama_st are comparing it to the libraries listed below. We may earn a commission when you buy through links labeled 'Ad' on this page.
Sorting:
- Local Qwen3 LLM inference. One easy-to-understand file of C source with no dependencies.☆165Jul 5, 2025Updated 8 months ago
- Accelerate multihead attention transformer model using HLS for FPGA☆11Dec 7, 2023Updated 2 years ago
- ☆14Mar 22, 2024Updated 2 years ago
- A local-first LLM development studio. Build, test, and customize inference workflows with your own models — no cloud, totally local.☆17May 21, 2025Updated 10 months ago
- Optimizing the Deployment of Tiny Transformers on Low-Power MCUs☆33Sep 2, 2024Updated last year
- Managed hosting for WordPress and PHP on Cloudways • AdManaged hosting with the flexibility to host WordPress, Magento, Laravel, or PHP apps, on multiple cloud providers. Cloudways by DigitalOcean.
- ☆64Jul 10, 2025Updated 8 months ago
- JavaScript bindings for the ggml-js library☆45Nov 10, 2025Updated 4 months ago
- An educational Rust project for exporting and running inference on Qwen3 LLM family☆40Aug 3, 2025Updated 7 months ago
- run ollama & gguf easily with a single command☆52May 15, 2024Updated last year
- Implementation of IceFormer: Accelerated Inference with Long-Sequence Transformers on CPUs (ICLR 2024).☆25Feb 22, 2026Updated last month
- A simple interface for using Ollama with LangChain's RAGChain☆30Mar 5, 2024Updated 2 years ago
- Llama2 inference in one TypeScript file☆20May 29, 2025Updated 9 months ago
- An efficient spatial accelerator enabling hybrid sparse attention mechanisms for long sequences☆32Mar 7, 2024Updated 2 years ago
- An experimental desktop client for using Claude Desktop's MCP with Novelcrafter codices.☆10Dec 3, 2024Updated last year
- Virtual machines for every use case on DigitalOcean • AdGet dependable uptime with 99.99% SLA, simple security tools, and predictable monthly pricing with DigitalOcean's virtual machines, called Droplets.
- Collection of PureBasic Headers and Libraries I made over the years.☆17Aug 4, 2023Updated 2 years ago
- LLM inference in C/C++☆23Oct 4, 2024Updated last year
- ☆34Nov 9, 2025Updated 4 months ago
- Adapted version of llama3.np (NumPy) to a CuPy implementation for the Llama 3 model.☆34May 16, 2024Updated last year
- ☆38Oct 21, 2025Updated 5 months ago
- An API for VoiceCraft.☆25Jun 27, 2024Updated last year
- My collection of dotfiles☆14Mar 16, 2026Updated last week
- Step by step explanation/tutorial of llama2.c☆226Oct 9, 2023Updated 2 years ago
- Tiny evaluation of leading LLMs on competitive programming problems☆14Nov 28, 2024Updated last year
- 1-Click AI Models by DigitalOcean Gradient • AdDeploy popular AI models on DigitalOcean Gradient GPU virtual machines with just a single click and start building anything your business needs.
- Distributed LLM Inference for Apple Silicon Clusters☆65Jan 6, 2026Updated 2 months ago
- Training framework for Large Behavioral Models☆27Sep 17, 2025Updated 6 months ago
- A straightforward method to reduce your LLM inference API costs and token usage.☆22May 18, 2025Updated 10 months ago
- Enemies for your LLM☆35Jan 20, 2026Updated 2 months ago
- Experience the power of AI with this free AI voice generator demo. Utilizing Deepgram and Groq, we transform text into voice seamlessly. …☆37Jun 12, 2024Updated last year
- A Beginner's Guide to Monetizing Your Python AI Chatbot☆16Apr 22, 2025Updated 11 months ago
- PyTorch code for "ADEM-VL: Adaptive and Embedded Fusion for Efficient Vision-Language Tuning"☆21Oct 28, 2024Updated last year
- ☆43Aug 2, 2025Updated 7 months ago
- HippocampAI — Autonomous Memory Engine for LLM Agents☆62Feb 13, 2026Updated last month
- DigitalOcean Gradient AI Platform • AdBuild production-ready AI agents using customizable tools or access multiple LLMs through a single endpoint. Create custom knowledge bases or connect external data.
- Single-file, pure CUDA C implementation for running inference on Qwen3 0.6B GGUF. No Dependencies.☆23Nov 26, 2025Updated 4 months ago
- ☆43Sep 15, 2025Updated 6 months ago
- Note about running ollama 🦙☆36May 2, 2024Updated last year
- A comprehensive hands-on project for learning GPU programming with CUDA and HIP, covering fundamental concepts through advanced optimizat…☆35Nov 20, 2025Updated 4 months ago
- implement GPT-OSS 20B & 120B C++ inference from scratch on AMD GPUs☆170Oct 25, 2025Updated 5 months ago
- L2E llama2.c on Commodore C-64☆18Feb 22, 2025Updated last year
- Explore, Install, Innovate — in 1 Click.☆167Feb 7, 2026Updated last month