Advanced Multi-Graph Architectures for Natural Language Understanding

Publication Type:
Thesis
Issue Date:
2023
Full metadata record
Natural language processing (NLP) is a challenge while important subfield of artificial intelligence (AI). And natural language understanding (NLU) is a crucial component of NLP. In recent years, graph structures and graph neural networks are widely utilized in NLU. In single task scenarios, graph structures can represent the sentence or document in graphs (e.g. syntax graph of a sentence). And sometimes only one graph cannot represent sufficient information, therefore multiple graph structures are proposed to sufficiently represent the relations and dependencies. In multi-task scenarios, the interactions among the multiple tasks can be represented in a multi-task graph. And different kinds of graph neural networks have been proposed to achieve information aggregation on the graphs. In this thesis, the multi-graph structures for single tasks, the multi-task graph structures for multi-task learning, and the graph neural networks working on them are collectively referred to as multi-graph architectures. In this thesis, we design multi-graph architectures based on interconnected graphs to represent diverse linguistic dimensions and multi-task interactions, including syntactic, semantic, and contextual information. Each graph encapsulates specific linguistic features, fostering a more comprehensive understanding of language nuances. The fusion of these graphs enables the model to capture intricate relationships and dependencies among words, concepts, sentences and different tasks, contributing to a more robust and context-aware NLU system. This thesis explores the design principles, implementation details, and experimental results of multi-graph architectures applied to various NLU tasks, such as aspect sentiment classification, joint dialog sentiment classification and act recognition, and joint multiple intent detection and slot filling. Comparative analyses against state-of-the-art models demonstrate the efficacy of the proposed multi-graph architectures in handling ambiguous language constructs and improving overall NLU performance in both single-task and multi-task scenarios.
Please use this identifier to cite or link to this item: