Spico197 / random-luck
Automatically select the best random seed based on ancient Chinese I Ching. Good luck and best wishes !
☆44Updated 3 years ago
Alternatives and similar repositories for random-luck:
Users that are interested in random-luck are comparing it to the libraries listed below
- Feeling confused about super alignment? Here is a reading list☆42Updated last year
- [NeurIPS'24] Weak-to-Strong Search: Align Large Language Models via Searching over Small Language Models☆54Updated 2 months ago
- 服务器 GPU 监控程序,当 GPU 属性满足预设条件时通过微信发送提示消息☆29Updated 3 years ago
- The source code of "Merging Experts into One: Improving Computational Efficiency of Mixture of Experts (EMNLP 2023)":☆35Updated 10 months ago
- The code and data for the paper JiuZhang3.0☆40Updated 8 months ago
- Token level visualization tools for large language models☆74Updated last month
- The implementation of paper "LLM Critics Help Catch Bugs in Mathematics: Towards a Better Mathematical Verifier with Natural Language Fee…☆38Updated 6 months ago
- [NeurIPS 2024] A Novel Rank-Based Metric for Evaluating Large Language Models☆40Updated 3 months ago
- [ICLR'24 spotlight] Tool-Augmented Reward Modeling☆44Updated last month
- Code and Data Repo for [NeurIPS 2024] Paper "Embedding Trajectory for Out-of-Distribution Detection in Mathematical Reasoning"☆22Updated 8 months ago
- Plug-and-Play Document Modules for Pre-trained Models☆25Updated last year
- 😎 A simple and easy-to-use toolkit for GPU scheduling.☆42Updated 3 years ago
- Code for preprint "Metadata Conditioning Accelerates Language Model Pre-training (MeCo)"☆34Updated last month
- Ladder Side-Tuning在CLUE上的简单尝试☆19Updated 2 years ago
- [ACL 2023] Solving Math Word Problems via Cooperative Reasoning induced Language Models (LLMs + MCTS + Self-Improvement)☆48Updated last year
- 基于Gated Attention Unit的Transformer模型(尝鲜版)☆97Updated last year
- This is a personal reimplementation of Google's Infini-transformer, utilizing a small 2b model. The project includes both model and train…☆56Updated 10 months ago
- [EMNLP 2023] Knowledge Rumination for Pre-trained Language Models☆16Updated last year
- Official completion of “Training on the Benchmark Is Not All You Need”.☆29Updated last month
- Research without Re-search: Maximal Update Parametrization Yields Accurate Loss Prediction across Scales☆31Updated last year
- Policies of scientific publisher and conferences towards large language model (LLM), such as ChatGPT☆73Updated 2 years ago
- On Memorization of Large Language Models in Logical Reasoning☆39Updated 3 months ago
- ☆61Updated 8 months ago
- Use the tokenizer in parallel to achieve superior acceleration☆15Updated 11 months ago
- The this is the official implementation of "DAPE: Data-Adaptive Positional Encoding for Length Extrapolation"☆35Updated 4 months ago
- ☆30Updated 5 months ago
- ☆45Updated 8 months ago
- 1.4B sLLM for Chinese and English - HammerLLM🔨☆44Updated 10 months ago
- ⚡Research papers about leveraging the capabilities of language models⚡☆52Updated last year