Open Collections will undergo maintenance on Thursday, July 24th, 2025. The site will not be available from 8:00 AM - 9:00 AM PST and performance may be impacted from 9:00 AM - 12:00 PM PST.
- Library Home /
- Search Collections /
- Open Collections /
- Browse Collections /
- UBC Theses and Dissertations /
- Discourse-guided text-generation from knowledge graphs...
Open Collections
UBC Theses and Dissertations
UBC Theses and Dissertations
Discourse-guided text-generation from knowledge graphs and image scene graphs Ivanova, Inna
Abstract
This thesis introduces a discourse-guided approach for generating text from semi-structured data - knowledge graphs and image scene graphs. We provide a novel architecture that integrates discourse planning as an intermediary structuring step, with the objective of enhancing coherence, readability, and the overall quality of generated content. The proposed method reorders input graph nodes into a coherent discourse sequence prior to decoding, utilizing both Pointer Networks and Large Language Models (LLMs) to represent discourse structures. Our experiments focus on two distinct datasets - Agenda (scientific abstracts) and Visual Genome (image captioning) - illustrating that explicit discourse planning consistently enhances performance across standard natural language generation metrics and improves output quality as evaluated by both human and LLM-based assessments. This thesis presents a generalizable approach for integrating discourse structure into neural text generation systems and emphasizes the potential of large language models as both planners and evaluators in natural language generation tasks.
Item Metadata
Title |
Discourse-guided text-generation from knowledge graphs and image scene graphs
|
Creator | |
Supervisor | |
Publisher |
University of British Columbia
|
Date Issued |
2025
|
Description |
This thesis introduces a discourse-guided approach for generating text from semi-structured data - knowledge graphs and image scene graphs. We provide a novel architecture that integrates discourse planning as an intermediary structuring step, with the objective of enhancing coherence, readability, and the overall quality of generated content. The proposed method reorders input graph nodes into a coherent discourse sequence prior to decoding, utilizing both Pointer Networks and Large Language Models (LLMs) to represent discourse structures. Our experiments focus on two distinct datasets - Agenda (scientific abstracts) and Visual Genome (image captioning) - illustrating that explicit discourse planning consistently enhances performance across standard natural language generation metrics and improves output quality as evaluated by both human and LLM-based assessments. This thesis presents a generalizable approach for integrating discourse structure into neural text generation systems and emphasizes the potential of large language models as both planners and evaluators in natural language generation tasks.
|
Genre | |
Type | |
Language |
eng
|
Date Available |
2025-07-04
|
Provider |
Vancouver : University of British Columbia Library
|
Rights |
Attribution-NonCommercial-NoDerivatives 4.0 International
|
DOI |
10.14288/1.0449275
|
URI | |
Degree | |
Program | |
Affiliation | |
Degree Grantor |
University of British Columbia
|
Graduation Date |
2025-11
|
Campus | |
Scholarly Level |
Graduate
|
Rights URI | |
Aggregated Source Repository |
DSpace
|
Item Media
Item Citations and Data
Rights
Attribution-NonCommercial-NoDerivatives 4.0 International