euirim / goodwikiLinks
Package and scripts used to build a dataset of Wikipedia articles in Markdown.
☆20Updated 2 years ago
Alternatives and similar repositories for goodwiki
Users that are interested in goodwiki are comparing it to the libraries listed below
Sorting:
- ☆50Updated last year
- ☆17Updated last year
- Demonstration that finetuning RoPE model on larger sequences than the pre-trained model adapts the model context limit☆63Updated 2 years ago
- Multi-Domain Expert Learning☆67Updated 2 years ago
- ☆48Updated last year
- QLoRA with Enhanced Multi GPU Support☆37Updated 2 years ago
- A toolkit implementing advanced methods to transfer models and model knowledge across tokenizers.☆62Updated 7 months ago
- Exploring finetuning public checkpoints on filter 8K sequences on Pile☆116Updated 2 years ago
- Using open source LLMs to build synthetic datasets for direct preference optimization☆72Updated last year
- ☆20Updated 2 years ago
- Experiments for efforts to train a new and improved t5☆76Updated last year
- Code for NeurIPS LLM Efficiency Challenge☆60Updated last year
- minimal pytorch implementation of bm25 (with sparse tensors)☆104Updated 3 months ago
- Experiments with generating opensource language model assistants☆97Updated 2 years ago
- ☆32Updated 2 years ago
- Small and Efficient Mathematical Reasoning LLMs☆73Updated 2 years ago
- Vocabulary Trimming (VT) is a model compression technique, which reduces a multilingual LM vocabulary to a target language by deleting ir…☆61Updated last year
- Implementation of "LM-Infinite: Simple On-the-Fly Length Generalization for Large Language Models"☆40Updated last year
- Spherical Merge Pytorch/HF format Language Models with minimal feature loss.☆144Updated 2 years ago
- Experiments on speculative sampling with Llama models☆128Updated 2 years ago
- ☆38Updated last year
- A library for squeakily cleaning and filtering language datasets.☆49Updated 2 years ago
- GPTQLoRA: Efficient Finetuning of Quantized LLMs with GPTQ☆101Updated 2 years ago
- QLoRA: Efficient Finetuning of Quantized LLMs☆79Updated last year
- Merge LLM that are split in to parts☆27Updated 6 months ago
- Repo for training MLMs, CLMs, or T5-type models on the OLM pretraining data, but it should work with any hugging face text dataset.☆96Updated 3 years ago
- ☆53Updated 2 years ago
- ☆59Updated last year
- [TMLR 2026] When Attention Collapses: How Degenerate Layers in LLMs Enable Smaller, Stronger Models☆122Updated last year
- experiments with inference on llama☆103Updated last year