huggingface / transformers.js-benchmarking
β15Updated 2 weeks ago
Alternatives and similar repositories for transformers.js-benchmarking
Users that are interested in transformers.js-benchmarking are comparing it to the libraries listed below
Sorting:
- β‘Delightful WebNN resources, curated list of awesome things around WebNN ecosystem.πβ56Updated 3 weeks ago
- Website with current metrics on the fastest AI models.β41Updated 6 months ago
- Retrieve large (GBs) AI binary model files from cloud, cache locally as sharded blobs to load faster on 2nd page load, returns stored filβ¦β21Updated 3 weeks ago
- Train text generation model with JavaScript.β15Updated 10 months ago
- Proof of concept for running moshi/hibiki using webrtcβ18Updated 2 months ago
- Browse, search, and visualize ONNX models.β25Updated last week
- Node.js binding of llama.cppβ10Updated 2 weeks ago
- GPU accelerated client-side embeddings for vector search, RAG etc.β66Updated last year
- Profile your CoreML models directly from Python πβ27Updated 7 months ago
- LLama implementations benchmarking frameworkβ12Updated last year
- A JavaScript implementation of Llama 3 using node-mlx.β72Updated 9 months ago
- A website to show WebGPU infoβ68Updated last week
- β42Updated last month
- π§ͺ Model Loader APIβ33Updated last year
- Inference Llama 2 in one file of pure JavaScript(HTML)β33Updated 10 months ago
- Rust crate for some audio utilitiesβ23Updated 2 months ago
- Thin wrapper around GGML to make life easierβ29Updated this week
- Code for training & inference with FLAN family of modelsβ17Updated last year
- JavaScript bindings for the ggml-js libraryβ43Updated last month
- MLX binary vectors and associated algorithms.β14Updated 2 months ago
- GGML implementation of BERT model with Python bindings and quantization.β26Updated last year
- β26Updated 5 months ago
- β12Updated 10 months ago
- Add local LLMs to your Web or Electron apps! Powered by Rust + WebGPUβ102Updated last year
- Using headless Chrome on server side environments for true client side browser emulation with NVIDIA T4 GPUs for Web AI model testing or β¦β73Updated last year
- β55Updated last week
- llama.cpp fork used by GPT4Allβ55Updated 2 months ago
- Proof of concept for a generative AI application framework powered by WebAssembly and Extismβ14Updated last year
- ONNX Runtime Web benchmark toolβ8Updated last year
- A python command-line tool to download & manage MLX AI models from Hugging Face.β17Updated 8 months ago