fshnkarimi / Fine-tuning-an-LLM-using-LoRA
π Text Classification with LoRA (Low-Rank Adaptation) of Language Models - Efficiently fine-tune large language models for text classification tasks using the Stanford Sentiment Treebank (SST-2) dataset and the LoRA technique.
β48Updated last year
Alternatives and similar repositories for Fine-tuning-an-LLM-using-LoRA:
Users that are interested in Fine-tuning-an-LLM-using-LoRA are comparing it to the libraries listed below
- Implementation and evaluation of multimodal RAG with text and image inputs for industrial applicationsβ45Updated 5 months ago
- MathPrompter Implementation: This repository hosts an implementation based on the 'MathPrompter: Mathematical Reasoning Using Large Languβ¦β13Updated 2 weeks ago
- This repository contains a custom implementation of the BERT model, fine-tuned for specific tasks, along with an implementation of Low Raβ¦β70Updated last year
- Gemma2(9B), Llama3-8B-Finetune-and-RAG, code base for sample, implemented in Kaggle platformβ21Updated 2 months ago
- Truth Forest: Toward Multi-Scale Truthfulness in Large Language Models through Intervention without Tuningβ46Updated last year
- β24Updated 7 months ago
- β45Updated 3 weeks ago
- Source code of the paper: RetrievalQA: Assessing Adaptive Retrieval-Augmented Generation for Short-form Open-Domain Question Answering [Fβ¦β62Updated 10 months ago
- β28Updated last year
- Finance specialized RAG System for the ACM-ICAIF '24 Competition.β33Updated 4 months ago
- Fine-Tuning LLM and embedding modelsβ27Updated last year
- Scripts for fine-tuning Llama2 via SFT and DPO.β197Updated last year
- Official repo for the paper PHUDGE: Phi-3 as Scalable Judge. Evaluate your LLMs with or without custom rubric, reference answer, absoluteβ¦β49Updated 9 months ago
- Codes and datasets for the paper Measuring and Enhancing Trustworthiness of LLMs in RAG through Grounded Attributions and Learning to Refβ¦β48Updated last month
- β74Updated 3 months ago
- β29Updated 5 months ago
- An overview of GRPO & DeepSeek-R1 Training with Open Source GRPO Model Fine Tuningβ31Updated 2 months ago
- Testing speed and accuracy of RAG with, and without Cross Encoder Reranker.β48Updated last year
- Code for "Democratizing Reasoning Ability: Tailored Learning from Large Language Model", EMNLP 2023β34Updated last year
- A simplified implementation for experimenting with Reinforcement Learning (RL) on GSM8K, inspired by RLVR and Deepseek R1. This repositorβ¦β78Updated 2 months ago
- minimal GRPO implementation from scratchβ85Updated last month
- a curated list of the role of small models in the LLM eraβ100Updated 7 months ago
- β16Updated 6 months ago
- RL significantly the reasoning capability of Qwen2.5-1.5B-Instructβ28Updated 2 months ago
- β47Updated 11 months ago
- This is the code of MMOA-RAG.β50Updated last month
- β27Updated last year
- LangChain, Llama2-Chat, and zero- and few-shot prompting are used to generate synthetic datasets for IR and RAG system evaluationβ36Updated last year
- Official repository for RAGViz: Diagnose and Visualize Retrieval-Augmented Generation [EMNLP 2024]β82Updated 3 months ago
- Qwen2 VL Fine Tuning using Llama Factoryβ20Updated 7 months ago