okuvshynov / cubestat
Horizon chart for CPU/GPU/Neural Engine utilization monitoring on Apple M1/M2 and nVidia GPUs on Linux
☆25Updated 4 months ago
Alternatives and similar repositories for cubestat:
Users that are interested in cubestat are comparing it to the libraries listed below
- A super simple web interface to perform blind tests on LLM outputs.☆27Updated 11 months ago
- First token cutoff sampling inference example☆29Updated last year
- llm plugin for Cerebras fast inference API☆23Updated last month
- Benchmarks comparing PyTorch and MLX on Apple Silicon GPUs☆75Updated 7 months ago
- 🛠 Self-hosted, fast, and consistent remote configuration for apps.☆14Updated 2 years ago
- Visualize expert firing frequencies across sentences in the Mixtral MoE model☆17Updated last year
- LLM plugin providing access to the LLM documentation☆17Updated 2 weeks ago
- A framework for collecting a large human-sourced chain-of-thoughts dataset☆19Updated 7 months ago
- Benchmark that evaluates LLMs using 436 NYT Connections puzzles☆35Updated this week
- Implementation of nougat that focuses on processing pdf locally.☆79Updated last month
- Finetune your embeddings in-browser☆32Updated 10 months ago
- A library for incremental loading of large PyTorch checkpoints☆56Updated 2 years ago
- A Python framework for building AI agent systems with robust task management in the form of a graph execution engine, inference capabilit…☆21Updated last week
- Create embeddings for LLM using the Nomic API☆22Updated 3 months ago
- asynchronous/distributed speculative evaluation for llama3☆37Updated 6 months ago
- A minimalistic C++ Jinja templating engine for LLM chat templates☆123Updated last week
- Exploration of Vector database Index for fast approximate nearest neighbour search.☆20Updated 7 months ago
- GGUF implementation in C as a library and a tools CLI program☆258Updated last month
- Columnar database on SSD NVMe☆13Updated 3 years ago
- A fork of llama3.c used to do some R&D on inferencing☆19Updated 2 months ago
- Shared personal notes created while working with the Apple MLX machine learning framework☆21Updated 8 months ago
- Command line tool for Deep Infra cloud ML inference service☆29Updated 8 months ago
- ☆52Updated 10 months ago
- ☆12Updated 5 months ago
- ☆15Updated last year
- Roberta Question Answering using MLX.☆24Updated last year
- Lightweight Llama 3 8B Inference Engine in CUDA C☆46Updated 2 weeks ago
- Run Llama 2 using MLX on macOS☆33Updated last year