Mryangkaitong / big_dataloaderLinks
☆35Updated 3 years ago
Alternatives and similar repositories for big_dataloader
Users that are interested in big_dataloader are comparing it to the libraries listed below
Sorting:
- pytorch分布式训练,支持多机多卡,单机多卡。☆41Updated 3 years ago
- NTK scaled version of ALiBi position encoding in Transformer.☆68Updated last year
- ICLR2023 - Tailoring Language Generation Models under Total Variation Distance☆21Updated 2 years ago
- Python下shuffle几百G文件☆33Updated 3 years ago
- This repository open-sources our GEC system submitted by THU KELab (sz) in the CCL2023-CLTC Track 1: Multidimensional Chinese Learner Tex…☆15Updated last year
- R-Drop方法在中文任务上的简单实验☆91Updated 3 years ago
- 科大讯飞低资源多语种文本翻译挑战赛获奖方案☆28Updated last year
- 基于Gated Attention Unit的Transformer模型(尝鲜版)☆98Updated 2 years ago
- Correcting Chinese Spelling Errors with Phonetic Pre-training 非官方实现☆40Updated 3 years ago
- 百川Dynamic NTK-ALiBi的代码实现:无需微调即可推理更长文本☆47Updated last year
- Lion and Adam optimization comparison☆61Updated 2 years ago
- 论文模型复现☆42Updated 2 years ago
- 我的数据竞赛方案总结☆68Updated last year
- A wide variety of research projects developed by the SpokenNLP team of Speech Lab, Alibaba Group.☆117Updated last month
- A pre-trained model with multi-exit transformer architecture.☆54Updated 2 years ago
- This is a personal reimplementation of Google's Infini-transformer, utilizing a small 2b model. The project includes both model and train…☆57Updated last year
- 记录NLP、CV、搜索、推荐等AI岗位最新情况。☆29Updated 2 years ago
- Apply the Circular to the Pretraining Model☆37Updated 3 years ago
- 基于python的BM25文本匹配算法实现☆32Updated 3 years ago
- Ongoing research training transformer language models at scale, including: BERT & GPT-2☆19Updated last year
- Code for "SLIM: Explicit Slot-Intent Mapping with BERT for Joint Multi-Intent Detection and Slot Filling"☆17Updated 2 years ago
- dpo算法实现☆38Updated last year
- ☆53Updated 3 years ago
- ChatGPT相关资源汇总☆55Updated 2 years ago
- Ongoing research training transformer language models at scale, including: BERT & GPT-2☆69Updated last year
- FLASHQuad_pytorch☆67Updated 3 years ago
- The code of paper "Learning to Break the Loop: Analyzing and Mitigating Repetitions for Neural Text Generation" published at NeurIPS 202…☆46Updated 2 years ago
- A paper list of pre-trained language models (PLMs).☆81Updated 3 years ago
- The complete training code of the open-source high-performance Llama model, including the full process from pre-training to RLHF.☆66Updated 2 years ago
- Clustering and Ranking: Diversity-preserved Instruction Selection through Expert-aligned Quality Estimation☆81Updated 7 months ago