ModelCloud / Device-SMI
Self-contained Python lib with zero-dependencies that give you a unified device properties for gpu, cpu, and npu. No more calling separate tools such as nvidia-smi or /proc/cpuinfo and parsing it yourself.
☆9Updated this week
Alternatives and similar repositories for Device-SMI:
Users that are interested in Device-SMI are comparing it to the libraries listed below
- AirLLM 70B inference with single 4GB GPU☆12Updated 5 months ago
- A collection of notebooks for the Hugging Face blog series (https://huggingface.co/blog).☆42Updated 5 months ago
- implementation of https://arxiv.org/pdf/2312.09299☆20Updated 6 months ago
- Public reports detailing responses to sets of prompts by Large Language Models.☆28Updated last week
- ☆16Updated 3 weeks ago
- Unleash the full potential of exascale LLMs on consumer-class GPUs, proven by extensive benchmarks, with no long-term adjustments and min…☆23Updated last month
- Experiments with BitNet inference on CPU☆52Updated 9 months ago
- ☆16Updated 11 months ago
- Framework for synthetic data generation with AI feedback☆26Updated 2 weeks ago
- The simplest, fastest repository for training/finetuning medium-sized xLSTMs.☆38Updated 7 months ago
- Modified Beam Search with periodical restart☆12Updated 3 months ago
- Experimental sampler to make LLMs more creative☆30Updated last year
- Training hybrid models for dummies.☆16Updated 3 weeks ago
- alternative way to calculating self attention☆18Updated 7 months ago
- Web Interface for Vision Language Models Including InternVLM2☆16Updated 5 months ago
- Apps that run on modal.com☆12Updated 7 months ago
- An open source replication of the stawberry method that leverages Monte Carlo Search with PPO and or DPO☆26Updated this week
- A swarm of LLM agents that will help you test, document, and productionize your code!☆13Updated last week
- Lightweight continuous batching OpenAI compatibility using HuggingFace Transformers include T5 and Whisper.☆18Updated last month
- A list of language models with permissive licenses such as MIT or Apache 2.0☆24Updated 2 months ago
- High level library for batched embeddings generation, blazingly-fast web-based RAG and quantized indexes processing ⚡☆62Updated 2 months ago
- new optimizer☆19Updated 5 months ago
- ☆12Updated 3 months ago
- ☆23Updated last month
- Easily convert HuggingFace models to GGUF-format for llama.cpp☆21Updated 5 months ago
- Uses a Gradio interface to stream coding related responses from local and cloud based large language models. Pulls context from GitHub Re…☆19Updated 4 months ago
- Prompt Jinja2 templates for LLMs☆27Updated last week
- ☆24Updated last year
- Parameter-Efficient Sparsity Crafting From Dense to Mixture-of-Experts for Instruction Tuning on General Tasks☆31Updated 7 months ago
- Yet Another (LLM) Web UI, made with Gemini☆11Updated 2 weeks ago