aksh555 / deciphering_cot
[EMNLP 2024 Findings] Code for deciphering CoT using shift ciphers
☆11Updated this week
Related projects ⓘ
Alternatives and complementary repositories for deciphering_cot
- Implementation of SelfExtend from the paper "LLM Maybe LongLM: Self-Extend LLM Context Window Without Tuning" from Pytorch and Zeta☆13Updated last week
- An example implementation of RLHF (or, more accurately, RLAIF) built on MLX and HuggingFace.☆21Updated 5 months ago
- Latent Large Language Models☆16Updated 2 months ago
- Public reports detailing responses to sets of prompts by Large Language Models.☆26Updated last year
- Trying to deconstruct RWKV in understandable terms☆14Updated last year
- Code for RATIONALYST: Pre-training Process-Supervision for Improving Reasoning https://arxiv.org/pdf/2410.01044☆30Updated last month
- ☆25Updated 2 months ago
- A library for simplifying fine tuning with multi gpu setups in the Huggingface ecosystem.☆15Updated 3 weeks ago
- [WIP] Transformer to embed Danbooru labelsets☆13Updated 7 months ago
- Training hybrid models for dummies.☆15Updated 3 weeks ago
- Binary vector search example using Unum's USearch engine and pre-computed Wikipedia embeddings from Co:here and MixedBread☆19Updated 7 months ago
- Learning to Retrieve by Trying - Source code for Grounding by Trying: LLMs with Reinforcement Learning-Enhanced Retrieval☆24Updated 3 weeks ago
- Fullstack chatbot application☆11Updated 3 months ago
- ☆36Updated 3 months ago
- Zeus LLM Trainer is a rewrite of Stanford Alpaca aiming to be the trainer for all Large Language Models☆69Updated last year
- ☆42Updated 4 months ago
- A framework for evaluating the effectiveness of chain-of-thought reasoning in language models.☆12Updated last month
- ☆20Updated 9 months ago
- GPT-4 Level Conversational QA Trained In a Few Hours☆55Updated 3 months ago
- LMQL implementation of tree of thoughts☆33Updated 9 months ago
- One Line To Build Zero-Data Classifiers in Minutes☆33Updated last month
- ☆41Updated 2 weeks ago
- Demonstration that finetuning RoPE model on larger sequences than the pre-trained model adapts the model context limit☆63Updated last year
- ☆12Updated 3 weeks ago
- Implementation of the paper: "AssistantBench: Can Web Agents Solve Realistic and Time-Consuming Tasks?"☆41Updated last month
- ☆22Updated 3 months ago
- Run GreenBitAI's Quantized LLMs on Apple Devices with MLX☆15Updated this week
- ☆38Updated 8 months ago
- Github repo for Peifeng's internship project☆12Updated last year
- Code and data for the paper "Why think step by step? Reasoning emerges from the locality of experience"☆48Updated last year