Search engine for discovering works of Art, research articles, and books related to Art and Culture
ShareThis
Javascript must be enabled to continue!

Deep Dialog Act Recognition using Multiple Token, Segment, and Context Information Representations

View through CrossRef
Automatic dialog act recognition is a task that has been widely explored over the years. In recent works, most approaches to the task explored different deep neural network architectures to combine the representations of the words in a segment and generate a segment representation that provides cues for intention. In this study, we explore means to generate more informative segment representations, not only by exploring different network architectures, but also by considering different token representations, not only at the word level, but also at the character and functional levels. At the word level, in addition to the commonly used uncontextualized embeddings, we explore the use of contextualized representations, which are able to provide information concerning word sense and segment structure. Character-level tokenization is important to capture intention-related morphological aspects that cannot be captured at the word level. Finally, the functional level provides an abstraction from words, which shifts the focus to the structure of the segment. Additionally, we explore approaches to enrich the segment representation with context information from the history of the dialog, both in terms of the classifications of the surrounding segments and the turn-taking history. This kind of information has already been proved important for the disambiguation of dialog acts in previous studies. Nevertheless, we are able to capture additional information by considering a summary of the dialog history and a wider turn-taking context. By combining the best approaches at each step, we achieve performance results that surpass the previous state-of-the-art on generic dialog act recognition on both the Switchboard Dialog Act Corpus (SwDA) and the ICSI Meeting Recorder Dialog Act Corpus (MRDA), which are two of the most widely explored corpora for the task. Furthermore, by considering both past and future context, similarly to what happens in an annotation scenario, our approach achieves a performance similar to that of a human annotator on SwDA and surpasses it on MRDA.
Title: Deep Dialog Act Recognition using Multiple Token, Segment, and Context Information Representations
Description:
Automatic dialog act recognition is a task that has been widely explored over the years.
In recent works, most approaches to the task explored different deep neural network architectures to combine the representations of the words in a segment and generate a segment representation that provides cues for intention.
In this study, we explore means to generate more informative segment representations, not only by exploring different network architectures, but also by considering different token representations, not only at the word level, but also at the character and functional levels.
At the word level, in addition to the commonly used uncontextualized embeddings, we explore the use of contextualized representations, which are able to provide information concerning word sense and segment structure.
Character-level tokenization is important to capture intention-related morphological aspects that cannot be captured at the word level.
Finally, the functional level provides an abstraction from words, which shifts the focus to the structure of the segment.
Additionally, we explore approaches to enrich the segment representation with context information from the history of the dialog, both in terms of the classifications of the surrounding segments and the turn-taking history.
This kind of information has already been proved important for the disambiguation of dialog acts in previous studies.
Nevertheless, we are able to capture additional information by considering a summary of the dialog history and a wider turn-taking context.
By combining the best approaches at each step, we achieve performance results that surpass the previous state-of-the-art on generic dialog act recognition on both the Switchboard Dialog Act Corpus (SwDA) and the ICSI Meeting Recorder Dialog Act Corpus (MRDA), which are two of the most widely explored corpora for the task.
Furthermore, by considering both past and future context, similarly to what happens in an annotation scenario, our approach achieves a performance similar to that of a human annotator on SwDA and surpasses it on MRDA.

Related Results

Enhancing Non-Formal Learning Certificate Classification with Text Augmentation: A Comparison of Character, Token, and Semantic Approaches
Enhancing Non-Formal Learning Certificate Classification with Text Augmentation: A Comparison of Character, Token, and Semantic Approaches
Aim/Purpose: The purpose of this paper is to address the gap in the recognition of prior learning (RPL) by automating the classification of non-formal learning certificates using d...
ANALISIS KALIMAT DIALOG DALAM NASKAH DRAMA �PADA SUATU HARI� KARYA ARIFIN C. NOER
ANALISIS KALIMAT DIALOG DALAM NASKAH DRAMA �PADA SUATU HARI� KARYA ARIFIN C. NOER
Judul penelitian ini adalah Analisis Kalimat Dialog dalam Naskah Drama �Pada Suatu Hari� Karya Arifin C. Noer (Sebuah Analisis Pragmatik tentang Deiksis dan Praanggapan). Rumusan m...
SPIRITUAL-HUMANISME HASAN ASKARI DAN DIALOG INTERRELIGI DI INDONESIA
SPIRITUAL-HUMANISME HASAN ASKARI DAN DIALOG INTERRELIGI DI INDONESIA
<p><strong>Abstrak</strong></p><p>Paper inimembahasspiritual-humanisme dariHasanAskari sebagai dialog interreligi alternatif. Data dikumpulkan melalui...
Meta-Representations as Representations of Processes
Meta-Representations as Representations of Processes
In this study, we explore how the notion of meta-representations in Higher-Order Theories (HOT) of consciousness can be implemented in computational models. HOT suggests that consc...
Advancing Transformer Efficiency with Token Pruning
Advancing Transformer Efficiency with Token Pruning
Transformer-based models have revolutionized natural language processing (NLP), achieving state-of-the-art performance across a wide range of tasks. However, their high computation...
HiRED: Attention-Guided Token Dropping for Efficient Inference of High-Resolution Vision-Language Models
HiRED: Attention-Guided Token Dropping for Efficient Inference of High-Resolution Vision-Language Models
High-resolution Vision-Language Models (VLMs) are widely used in multimodal tasks to enhance accuracy by preserving detailed image information. However, these models often generate...
Depth-aware salient object segmentation
Depth-aware salient object segmentation
Object segmentation is an important task which is widely employed in many computer vision applications such as object detection, tracking, recognition, and ret...

Back to Top