uygarkurt / BERT-PyTorchLinks
β17Updated last year
Alternatives and similar repositories for BERT-PyTorch
Users that are interested in BERT-PyTorch are comparing it to the libraries listed below
Sorting:
- Lightweight demos for finetuning LLMs. Powered by π€ transformers and open-source datasets.β77Updated last year
- LLM_library is a comprehensive repository serves as a one-stop resource hands-on code, insightful summaries.β69Updated 2 years ago
- Complete implementation of Llama2 with/without KV cache & inference πβ49Updated last year
- Tutorial for how to build BERT from scratchβ101Updated last year
- A set of scripts and notebooks on LLM finetunning and dataset creationβ115Updated last year
- Model, Code & Data for the EMNLP'23 paper "Making Large Language Models Better Data Creators"β137Updated 2 years ago
- Unofficial implementation of https://arxiv.org/pdf/2407.14679β53Updated last year
- Fine-tuning Open-Source LLMs for Adaptive Machine Translationβ90Updated 6 months ago
- Set of scripts to finetune LLMsβ37Updated last year
- minimal scripts for 24GB VRAM GPUs. training, inference, whateverβ50Updated last month
- Using open source LLMs to build synthetic datasets for direct preference optimizationβ72Updated last year
- Fine-Tuning Llama3-8B LLM in a multi-GPU environment using DeepSpeedβ19Updated last year
- Prune transformer layersβ74Updated last year
- This repository contains an implementation of the LLaMA 2 (Large Language Model Meta AI) model, a Generative Pretrained Transformer (GPT)β¦β74Updated 2 years ago
- Building a 2.3M-parameter LLM from scratch with LLaMA 1 architecture.β197Updated last year
- β78Updated 2 years ago
- Official repo for the paper PHUDGE: Phi-3 as Scalable Judge. Evaluate your LLMs with or without custom rubric, reference answer, absoluteβ¦β51Updated last year
- LORA: Low-Rank Adaptation of Large Language Models implemented using PyTorchβ122Updated 2 years ago
- Composition of Multimodal Language Models From Scratchβ15Updated last year
- A Python wrapper around HuggingFace's TGI (text-generation-inference) and TEI (text-embedding-inference) servers.β32Updated 4 months ago
- β16Updated last year
- A pipeline for LLM knowledge distillationβ112Updated 10 months ago
- EvolKit is an innovative framework designed to automatically enhance the complexity of instructions used for fine-tuning Large Language Mβ¦β249Updated last year
- Combining ViT and GPT-2 for image captioning. Trained on MS-COCO. The model was implemented mostly from scratch.β48Updated 2 years ago
- Fine-tune ModernBERT with custom tokenizers, curriculum learning, and next-gen optimizers.β74Updated 2 weeks ago
- β105Updated 10 months ago
- Distributed training (multi-node) of a Transformer modelβ92Updated last year
- A framework for few-shot evaluation of language models.β35Updated 10 months ago
- Okapi: Instruction-tuned Large Language Models in Multiple Languages with Reinforcement Learning from Human Feedbackβ96Updated 2 years ago
- Spherical Merge Pytorch/HF format Language Models with minimal feature loss.β143Updated 2 years ago