idiap / dialog2flow
Dialog2Flow: convert your dialogs to flows. This repository accompanies the paper "Dialog2Flow: Pre-training Soft-Contrastive Sentence Embeddings for Automatic Dialog Flow Extraction", accepted to the EMNLP 2024 main conference
☆14Updated last week
Alternatives and similar repositories for dialog2flow
Users that are interested in dialog2flow are comparing it to the libraries listed below
Sorting:
- ☆10Updated 8 months ago
- A collection of various LLM sampling methods implemented in pure Pytorch☆24Updated 5 months ago
- Entailment self-training☆25Updated last year
- Explorations into adversarial losses on top of autoregressive loss for language modeling☆36Updated 2 months ago
- Code for paper "Do Language Models Have Beliefs? Methods for Detecting, Updating, and Visualizing Model Beliefs"☆28Updated 3 years ago
- Implementations of growing and pruning in neural networks☆22Updated last year
- Implementation of Metaformer, but in an autoregressive manner☆24Updated 2 years ago
- Why Do We Need Weight Decay in Modern Deep Learning? [NeurIPS 2024]☆66Updated 7 months ago
- Official Code Repository for the paper "Key-value memory in the brain"☆25Updated 2 months ago
- ☆32Updated last year
- ☆10Updated last year
- Official repository of "LiNeS: Post-training Layer Scaling Prevents Forgetting and Enhances Model Merging"☆26Updated 6 months ago
- User-friendly implementation of the Mixture-of-Sparse-Attention (MoSA). MoSA selects distinct tokens for each head with expert choice rou…☆13Updated 2 weeks ago
- Official code for the paper "Attention as a Hypernetwork"☆33Updated 10 months ago
- Code of "Deep invariant networks with differentiable augmentation layers"☆18Updated 2 years ago
- Exploring an idea where one forgets about efficiency and carries out attention across each edge of the nodes (tokens)☆51Updated last month
- https://footprints.baulab.info☆17Updated 7 months ago
- Towards Understanding the Mixture-of-Experts Layer in Deep Learning☆29Updated last year
- Implementation of the Mamba SSM with hf_integration.☆56Updated 8 months ago
- We study toy models of skill learning.☆26Updated 3 months ago
- Code and pretrained models for the paper: "MatMamba: A Matryoshka State Space Model"☆59Updated 5 months ago
- This is official project in our paper: Is Bigger and Deeper Always Better? Probing LLaMA Across Scales and Layers☆30Updated last year
- Goldfish: Monolingual language models for 350 languages.☆17Updated 8 months ago
- Personal implementation of ASIF by Antonio Norelli☆25Updated 11 months ago
- Code for 'Emergent Analogical Reasoning in Large Language Models'☆49Updated last year
- Official Code for MIMETIC^2☆12Updated 5 months ago
- Hrrformer: A Neuro-symbolic Self-attention Model (ICML23)☆54Updated last year
- My explorations into editing the knowledge and memories of an attention network☆34Updated 2 years ago
- A byte-level decoder architecture that matches the performance of tokenized Transformers.☆63Updated last year
- ☆17Updated last year