tugot17 / tokenomicsLinks
Estimate the throughput of OAI compatible servers
β21Updated 3 weeks ago
Alternatives and similar repositories for tokenomics
Users that are interested in tokenomics are comparing it to the libraries listed below
Sorting:
- β159Updated 11 months ago
- Where GPUs get cooked π©βπ³π₯β319Updated 2 months ago
- β170Updated last year
- The Batched API provides a flexible and efficient way to process multiple requests in a batch, with a primary focus on dynamic batching oβ¦β153Updated 4 months ago
- A Lightweight Library for AI Observabilityβ251Updated 9 months ago
- Simple UI for debugging correlations of text embeddingsβ301Updated 6 months ago
- β233Updated 5 months ago
- Atropos is a Language Model Reinforcement Learning Environments framework for collecting and evaluating LLM trajectories through diverse β¦β756Updated this week
- smol models are fun tooβ92Updated last year
- β267Updated this week
- Modded vLLM to run pipeline parallelism over public networksβ40Updated 6 months ago
- Minimal example scripts of the Hugging Face Trainer, focused on staying under 150 linesβ196Updated last year
- Training-Ready RL Environments + Evalsβ182Updated this week
- High-Performance Engine for Multi-Vector Searchβ185Updated 2 weeks ago
- Google TPU optimizations for transformers modelsβ122Updated 10 months ago
- Inference server benchmarking toolβ128Updated last month
- Module, Model, and Tensor Serialization/Deserializationβ276Updated 3 months ago
- Fault tolerance for PyTorch (HSDP, LocalSGD, DiLoCo, Streaming DiLoCo)β454Updated 2 weeks ago
- ShellSage saves sysadminsβ sanity by solving shell script snafus super swiftlyβ380Updated 2 weeks ago
- A tool to configure, launch and manage your machine learning experiments.β208Updated this week
- Real-time terminal monitor for InfiniBand networks - htop for high-speed interconnectsβ48Updated 2 months ago
- β78Updated last year
- β54Updated last year
- Fine-tune an LLM to perform batch inference and online serving.β114Updated 6 months ago
- Manage scalable open LLM inference endpoints in Slurm clustersβ277Updated last year
- Manage ML configuration with pydanticβ16Updated 6 months ago
- Load compute kernels from the Hubβ335Updated this week
- A repository to unravel the language of GPUs, making their kernel conversations easy to understandβ196Updated 5 months ago
- Faster structured generationβ260Updated 3 weeks ago
- β212Updated last week