alinourian / Fine-tuning-Mistral-7b-QA
Fine tuning Mistral-7b with PEFT(Parameter Efficient Fine-Tuning) and LoRA(Low-Rank Adaptation) on Puffin Dataset(multi-turn conversations between GPT-4 and real humans)
☆12Updated 11 months ago
Related projects ⓘ
Alternatives and complementary repositories for Fine-tuning-Mistral-7b-QA
- ☆41Updated 2 weeks ago
- Code and data for "StructLM: Towards Building Generalist Models for Structured Knowledge Grounding" (COLM 2024)☆68Updated last month
- Small and Efficient Mathematical Reasoning LLMs☆71Updated 9 months ago
- Repository containing the SPIN experiments on the DIBT 10k ranked prompts☆23Updated 8 months ago
- The code implementation of MAGDi: Structured Distillation of Multi-Agent Interaction Graphs Improves Reasoning in Smaller Language Models…☆30Updated 9 months ago
- A public implementation of the ReLoRA pretraining method, built on Lightning-AI's Pytorch Lightning suite.☆33Updated 8 months ago
- Using open source LLMs to build synthetic datasets for direct preference optimization☆40Updated 8 months ago
- This is a new metric that can be used to evaluate faithfulness of text generated by LLMs. The work behind this repository can be found he…☆31Updated last year
- ☆14Updated 2 weeks ago
- The first dense retrieval model that can be prompted like an LM☆63Updated 2 months ago
- Explore the use of DSPy for extracting features from PDFs 🔎☆33Updated 8 months ago
- Script for processing OpenAI's PRM800K process supervision dataset into an Alpaca-style instruction-response format☆27Updated last year
- ☆41Updated last month
- ReBase: Training Task Experts through Retrieval Based Distillation☆27Updated 4 months ago
- minimal LLM scripts for 24GB VRAM GPUs. training, inference, whatever☆33Updated this week
- ☆59Updated last month
- ☆22Updated 2 months ago
- Aioli: A unified optimization framework for language model data mixing☆13Updated last week
- SWIM-IR is a Synthetic Wikipedia-based Multilingual Information Retrieval training set with 28 million query-passage pairs spanning 33 la…☆44Updated last year
- distill chatGPT coding ability into small model (1b)☆24Updated last year
- A repository for research on medium sized language models.☆74Updated 5 months ago
- Anchored Preference Optimization and Contrastive Revisions: Addressing Underspecification in Alignment☆46Updated 2 months ago
- Codebase accompanying the Summary of a Haystack paper.☆72Updated 2 months ago
- ☆33Updated 6 months ago
- From GaLore to WeLore: How Low-Rank Weights Non-uniformly Emerge from Low-Rank Gradients. Ajay Jaiswal, Lu Yin, Zhenyu Zhang, Shiwei Liu,…☆43Updated 4 months ago
- Lightweight demos for finetuning LLMs. Powered by 🤗 transformers and open-source datasets.☆64Updated last month
- Simple Model Similarities Analysis☆20Updated 9 months ago
- ☆48Updated last month
- ☆32Updated last year