Beomi / transformers-language-modeling
Train ๐คtransformers with DeepSpeed: ZeRO-2, ZeRO-3
โ23Updated 3 years ago
Alternatives and similar repositories for transformers-language-modeling:
Users that are interested in transformers-language-modeling are comparing it to the libraries listed below
- Difference-based Contrastive Learning for Korean Sentence Embeddingsโ24Updated last year
- ์ธ์ด๋ชจ๋ธ์ ํ์ตํ๊ธฐ ์ํ ๊ณต๊ฐ ํ๊ตญ์ด instruction dataset๋ค์ ๋ชจ์๋์์ต๋๋ค.โ19Updated last year
- Beyond LM: How can language model go forward in the future?โ15Updated last year
- Hate speech detection corpus in Korean, shared with EMNLP 2023 paperโ13Updated 9 months ago
- KLUE Benchmark 1st place (2021.12) solutions. (RE, MRC, NLI, STS, TC)โ25Updated 2 years ago
- Google ๊ณต์ Rouge Implementation์ ํ๊ตญ์ด์์ ์ฌ์ฉํ ์ ์๋๋ก ์ฒ๋ฆฌโ14Updated last year
- [Findings of NAACL2022] A Dog Is Passing Over The Jet? A Text-Generation Dataset for Korean Commonsense Reasoning and Evaluationโ12Updated 2 years ago
- โ26Updated 4 years ago
- Megatron LM 11B on Huggingface Transformersโ27Updated 3 years ago
- Abstractive summarization using Bert2Bert framework.โ31Updated 4 years ago
- Don't Judge a Language Model by Its Last Layer: Contrastive Learning with Layer-Wise Attention Poolingโ9Updated 2 years ago
- โ10Updated 3 months ago
- Korean Nested Named Entity Corpusโ18Updated last year
- Korean Named Entity Corpusโ25Updated last year
- Official code and dataset repository of KoBBQ (TACL 2024)โ14Updated 9 months ago
- baikal.ai's pre-trained BERT models: descriptions and sample codesโ12Updated 3 years ago
- 2019 ๊ตญ์ด๊ฒฝ์ง๋ํ ํ๊ตญ์ด ์์กด๊ตฌ๋ฌธ ๋ถ์ ๋์(๋ฌธ์ฒด๋ถ ์ฅ๊ด์)โ16Updated 2 years ago
- Korean Commonsense Knowledge Graphโ14Updated 2 years ago
- โ14Updated 3 years ago
- ๋ชจ๋์ ๋ง๋ญ์น ๋ฐ์ดํฐ๋ฅผ ๋ถ์์ ํธ๋ฆฌํ ํํ๋ก ๋ณํํ๋ ๊ธฐ๋ฅ์ ์ ๊ณตํฉ๋๋ค.โ11Updated 2 years ago
- This is project for korean auto spacingโ12Updated 4 years ago
- KoRean based ELECTRA pre-trained models (KR-ELECTRA) for Tensorflow and PyTorchโ14Updated 3 years ago
- A Framework aims to wisely initialize unseen subword embeddings in PLMs for efficient large-scale continued pretrainingโ14Updated last year
- Korean Abstract Meaning Representation (AMR) Corpusโ10Updated 2 years ago
- ํ๊ตญ์ด ์ดํ ์๋ฏธ ๋ถ์ ๋ชจ๋ธโ20Updated 2 years ago
- โ19Updated 2 years ago
- This repository forked from parlAI. Korean Wizard of Wikipedia task was added to this repo. This repository is going to be moved after EMโฆโ16Updated 2 years ago
- Calculating Expected Time for training LLM.โ38Updated last year
- Machine Generated Captions for Best Artworksโ22Updated 2 years ago
- ๋งค์ฃผ ๋ชฉ์์ผ, 20:00 ๋ชจ์โ16Updated 4 years ago