File size: 2,303 Bytes
74cd113 8234588 fbb22f1 8234588 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 |
---
license: apache-2.0
datasets:
- Nuo97/Dolphin-DPO
language:
- zh
metrics:
- bleu
pipeline_tag: conversational
---
# COMEDY: COmpressive Memory-Enhanced Dialogue sYstems framework.
Github: https://github.com/nuochenpku/COMEDY
<br>
<div align="center">
<img src="comedy.png" width="40%" title="Introduction Figure">
</div>
### Task: Long-Term Conversation Dialogue Generation
Different from previous retrieval-based methods, COMEDY doesn't rely on any **retrieval module or database**.
Instead, COMEDY adopts a groundbreaking ''**One-for-All**'' approach, utilizing a single, unified model to manage the entire process from memory generation, compression to final response generation for long-term memory dialogue generation.
- COMEDY firstly involves distilling session-specific memory from past dialogues, encompassing fine-grained session summaries, including event recaps, and detailed user and bot portraits;
- In a break from traditional systems, COMEDY eschews the use of a memory database for storing these insights. Instead, it reprocesses and condenses memories from all past interactions, forming a *Compressive Memory*: The first part is the **concise events** that have occurred throughout all the conversations, creating a historical narrative that the system can draw upon. The second and third parts consist of a **detailed user profile** and the **dynamic relationship changes** between the user and chatbot across sessions, both derived from past conversational events.
- Finally, COMEDY skillfully integrates this compressive memory into ongoing conversations, enabling contextually memory-enhanced interactions.
### Training Dataset
**Dolphin**, the biggest Chinese long-term conversation dataset, from actual online user-chatbot interactions.
This dataset contains three tasks:
**Session-Level Memory Summarization**;
**Memory Compression**;
**Memory-Grounded Response Generation**,
comprising an extensive collection of 100k samples.
Dolphin is available at [**Dolphin**](https://huggingface.co/datasets/Nuo97/Dolphin-DPO)
### Training Strategy
Our training strategies include two stages: Mixed-task training and DPO Alignment.
<br>
<div align="center">
<img src="training_strategy.png" width="90%" title="Introduction Figure">
</div>
|