uygarkurt / BERT-PyTorchLinks
☆17Updated 5 months ago
Alternatives and similar repositories for BERT-PyTorch
Users that are interested in BERT-PyTorch are comparing it to the libraries listed below
Sorting:
- LLM_library is a comprehensive repository serves as a one-stop resource hands-on code, insightful summaries.☆69Updated last year
- Set of scripts to finetune LLMs☆37Updated last year
- Notebook and Scripts that showcase running quantized diffusion models on consumer GPUs☆38Updated 8 months ago
- Using open source LLMs to build synthetic datasets for direct preference optimization☆64Updated last year
- ☆39Updated last month
- Playground for Transformers☆51Updated last year
- A single repo with all scripts and utils to train / fine-tune the Mamba model with or without FIM☆55Updated last year
- A Python wrapper around HuggingFace's TGI (text-generation-inference) and TEI (text-embedding-inference) servers.☆33Updated last month
- A byte-level decoder architecture that matches the performance of tokenized Transformers.☆63Updated last year
- Fine-tuning large language models (LLMs) is crucial for enhancing performance across domain-specific task applications. This comprehensiv…☆12Updated 9 months ago
- A framework for few-shot evaluation of autoregressive language models.☆16Updated last year
- ☆41Updated 2 months ago
- Distributed training (multi-node) of a Transformer model☆72Updated last year
- Complete implementation of Llama2 with/without KV cache & inference 🚀☆47Updated last year
- Lightweight demos for finetuning LLMs. Powered by 🤗 transformers and open-source datasets.☆77Updated 8 months ago
- nanogpt turned into a chat model☆68Updated last year
- Official repo for the paper PHUDGE: Phi-3 as Scalable Judge. Evaluate your LLMs with or without custom rubric, reference answer, absolute…☆49Updated 11 months ago
- Tutorial for how to build BERT from scratch☆94Updated last year
- Collection of resources for RL and Reasoning☆25Updated 4 months ago
- Quantization of LLMs and benchmarking.☆10Updated last year
- BERT explained from scratch☆14Updated last year
- Supplementary material for "Understanding Parameter-Efficient Finetuning of Large Language Models: From Prefix Tuning to Adapters"☆46Updated 2 years ago
- Trully flash implementation of DeBERTa disentangled attention mechanism.☆58Updated last month
- Fine-Tuning Llama3-8B LLM in a multi-GPU environment using DeepSpeed☆18Updated last year
- Fine-tuning Open-Source LLMs for Adaptive Machine Translation☆80Updated last month
- Advanced Retrieval-Augmented Generation (RAG) through practical notebooks, using the power of the Langchain, OpenAI GPTs ,META LLAMA3 , A…☆69Updated last year
- minimal scripts for 24GB VRAM GPUs. training, inference, whatever☆40Updated last week
- The simplest, fastest repository for training/finetuning medium-sized xLSTMs.☆41Updated last year
- A set of scripts and notebooks on LLM finetunning and dataset creation☆110Updated 9 months ago
- This is the code that went into our practical dive using mamba as information extraction☆53Updated last year