replicate / cog
Containers for machine learning
β8,611Updated this week
Alternatives and similar repositories for cog
Users that are interested in cog are comparing it to the libraries listed below
Sorting:
- πΈ Run LLMs at home, BitTorrent-style. Fine-tuning and inference up to 10x faster than offloadingβ9,633Updated 8 months ago
- Large Language Model Text Generation Inferenceβ10,145Updated this week
- Run any open-source LLMs, such as DeepSeek and Llama, as OpenAI compatible API endpoint in the cloud.β11,275Updated this week
- A minimal PyTorch re-implementation of the OpenAI GPT (Generative Pretrained Transformer) trainingβ21,927Updated 9 months ago
- Build and share delightful machine learning apps, all in Python. π Star to support our work!β38,130Updated this week
- tiktoken is a fast BPE tokeniser for use with OpenAI's models.β14,523Updated 2 months ago
- AITemplate is a Python framework which renders neural network into high performance CUDA/HIP C++ code. Specialized for FP16 TensorCore (Nβ¦β4,635Updated last month
- Official Implementation of OCR-free Document Understanding Transformer (Donut) and Synthetic Document Generator (SynthDoG), ECCV 2022β6,237Updated 10 months ago
- Inference Llama 2 in one file of pure Cβ18,399Updated 9 months ago
- SkyPilot: Run AI and batch jobs on any infra (Kubernetes or 16+ clouds). Get unified execution, cost savings, and high GPU availability vβ¦β8,124Updated this week
- The standard data-centric AI package for data quality and machine learning with messy, real-world data and labels.β10,545Updated last month
- The easiest way to serve AI apps and models - Build Model Inference APIs, Job queues, LLM apps, Multi-model pipelines, and more!β7,712Updated this week
- A guidance language for controlling large language models.β20,215Updated this week
- π A list of open LLMs available for commercial use.β12,016Updated 3 months ago
- CodeGen is a family of open-source model for program synthesis. Trained on TPU-v4. Competitive with OpenAI Codex.β5,081Updated 3 months ago
- DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.β38,481Updated this week
- OpenAssistant is a chat-based assistant that understands tasks, can interact with third-party systems, and retrieve information dynamicalβ¦β37,345Updated 9 months ago
- Simple, safe way to store and distribute tensorsβ3,268Updated 2 weeks ago
- A Bulletproof Way to Generate Structured JSON from Language Modelsβ4,724Updated last year
- Gorilla: Training and Evaluating LLMs for Function Calls (Tool Calls)β12,092Updated this week
- Aim π« β An easy-to-use & supercharged open-source experiment tracker.β5,587Updated this week
- An implementation of model parallel autoregressive transformers on GPUs, based on the Megatron and DeepSpeed librariesβ7,184Updated last week
- Hackable and optimized Transformers building blocks, supporting a composable construction.β9,487Updated last week
- A collection of libraries to optimise AI model performancesβ8,371Updated 10 months ago
- π₯ Fast State-of-the-Art Tokenizers optimized for Research and Productionβ9,684Updated last month
- Code and documentation to train Stanford's Alpaca models, and generate the data.β30,001Updated 10 months ago
- QLoRA: Efficient Finetuning of Quantized LLMsβ10,439Updated 11 months ago
- A scalable generative AI framework built for researchers and developers working on Large Language Models, Multimodal, and Speech AI (Autoβ¦β14,448Updated this week
- π‘ All-in-one open-source AI framework for semantic search, LLM orchestration and language model workflowsβ10,936Updated last week
- An unnecessarily tiny implementation of GPT-2 in NumPy.β3,355Updated 2 years ago