HAE-RAE / haerae-evaluation-toolkitLinks
The most modern LLM evaluation toolkit
โ70Updated last week
Alternatives and similar repositories for haerae-evaluation-toolkit
Users that are interested in haerae-evaluation-toolkit are comparing it to the libraries listed below
Sorting:
- Benchmark in Korean Contextโ137Updated 2 years ago
- [KO-Platy๐ฅฎ] Korean-Open-platypus๋ฅผ ํ์ฉํ์ฌ llama-2-ko๋ฅผ fine-tuningํ KO-platypus modelโ75Updated 2 months ago
- Forked repo from https://github.com/EleutherAI/lm-evaluation-harness/commit/1f66adcโ80Updated last year
- โ110Updated 3 months ago
- Official repository for KoMT-Bench built by LG AI Researchโ70Updated last year
- huggingface์ ์๋ ํ๊ตญ์ด ๋ฐ์ดํฐ ์ธํธโ30Updated last year
- LLM ๋ชจ๋ธ์ ์ธ๊ตญ์ด ํ ํฐ ์์ฑ์ ๋ง๋ ์ฝ๋ ๊ตฌํโ81Updated 3 months ago
- โ68Updated last year
- IA3๋ฐฉ์์ผ๋ก KoAlpaca๋ฅผ fine tuningํ ํ๊ตญ์ด LLM๋ชจ๋ธโ69Updated 2 years ago
- KURE: ๊ณ ๋ ค๋ํ๊ต์์ ๊ฐ๋ฐํ, ํ๊ตญ์ด ๊ฒ์์ ํนํ๋ ์๋ฒ ๋ฉ ๋ชจ๋ธโ193Updated 2 months ago
- ํ๊ตญ์ด ์ธ์ด๋ชจ๋ธ ๋ค๋ถ์ผ ์ฌ๊ณ ๋ ฅ ๋ฒค์น๋งํฌโ198Updated last year
- Korean Sentence Embedding Model Performance Benchmark for RAGโ48Updated 9 months ago
- This repository aims to develop CoT Steering based on CoT without Prompting. It focuses on enhancing the modelโs latent reasoning capabilโฆโ114Updated 4 months ago
- Liner LLM Meetup archiveโ71Updated last year
- Official datasets and pytorch implementation repository of SQuARe and KoSBi (ACL 2023)โ248Updated 2 years ago
- ํ๊ตญ์ด ์๋ฃ ๋ถ์ผ ํนํ ์ฑ๋ด ํ๋ก์ ํธโ32Updated last year
- bpe based korean t5 model for text-to-text unified frameworkโ63Updated last year
- Kor-IR: Korean Information Retrieval Benchmarkโ88Updated last year
- ๊ตฌ๊ธ์์ ๋ฐํํ Chain-of-Thought Reasoning without Prompting์ ์ฝ๋๋ก ๊ตฌํํ ๋ ํฌ์ ๋๋ค.โ67Updated last year
- ํ๊ตญ์ด ์ฌ๋ฆฌ ์๋ด ๋ฐ์ดํฐ์ โ80Updated 2 years ago
- AutoRAG example about benchmarking Korean embeddings.โ41Updated last year
- [Google Meet] MLLM Arxiv Casual Talkโ52Updated 2 years ago
- ํ๊ตญ์ด LLM ๋ฆฌ๋๋ณด๋ ๋ฐ ๋ชจ๋ธ ์ฑ๋ฅ/์์ ์ฑ ๊ด๋ฆฌโ22Updated 2 years ago
- 42dot LLM consists of a pre-trained language model, 42dot LLM-PLM, and a fine-tuned model, 42dot LLM-SFT, which is trained to respond to โฆโ130Updated last year
- SKT A.X LLM 4.0โ142Updated 4 months ago
- Curation note of NLP datasetsโ99Updated 2 years ago
- โ40Updated last year
- ํ๊ตญ์ด ๋ฒค์น๋งํฌ ํ๊ฐ ์ฝ๋ ํตํฉ๋ณธ(?)โ20Updated last year
- ์์ฒด ๊ตฌ์ถํ ํ๊ตญ์ด ํ๊ฐ ๋ฐ์ดํฐ์ ์ ์ด์ฉํ ํ๊ตญ์ด ๋ชจ๋ธ ํ๊ฐโ31Updated last year
- Train GEMMA on TPU/GPU! (Codebase for training Gemma-Ko Series)โ48Updated last year