Login (DCU Staff Only)
Login (DCU Staff Only)

DORAS | DCU Research Repository

Explore open access research and scholarly works from DCU

Advanced Search

Generation of Semantically Consistent Text and Its Evaluation

Huidrom, Rudali orcid logoORCID: 0000-0003-0630-3603 (2025) Generation of Semantically Consistent Text and Its Evaluation. PhD thesis, Dublin City University.

Abstract
Generating semantically consistent text remains a crucial challenge in Natural Language Processing (NLP), especially in input-controlled settings such as data-to-text generation. Despite advances in neural language generation, current models often produce outputs that contain semantic inconsistencies with the input, including omissions, hallucinations, and distortions. These issues stem from the black-box, non-deterministic nature of neural models, which limits their interpretability and controllability. This thesis investigates how semantic consistency, the property that all information conveyed by the output is entailed by the input, can be more reliably evaluated and ultimately improved in generated text. Focusing on data-to-text generation, the thesis proposes that advancing semantic controllability requires a combination of principled formalisation of semantic errors and robust evaluation methods. First, it synthesises prior work on error annotation to develop a task-agnostic consensus taxonomy for semantic errors, distinguishing omission, addition, and substitution errors. Second, it applies this taxonomy in human evaluation studies, including span-based semantic error annotation to better understand the nature and reliability of human judgements. Third, it assesses the capacity of automatic evaluation methods, including lexical, embedding-based, and prompt-based approaches, to reflect human ratings of semantic consistency. In this context, the thesis explores LLM-as-judge setups for scoring system outputs and uses LLMs to sanity-check repeated human evaluation experiments, investigating their generalisability and reliability. Together, these contributions provide new tools insights, approaches and results for the fine-grained assessment of semantic consistency and lay the groundwork for more interpretable, reliable, and scalable evaluation frameworks in neural language generation.
Metadata
Item Type:Thesis (PhD)
Date of Award:2025
Refereed:No
Supervisor(s):Belz, Anya
Subjects:Computer Science > Artificial intelligence
Computer Science > Computational linguistics
DCU Faculties and Centres:DCU Faculties and Schools > Faculty of Engineering and Computing > School of Computing
Research Institutes and Centres > ADAPT
Use License:This item is licensed under a Creative Commons Attribution-NonCommercial-No Derivative Works 4.0 License. View License
Funders:Faculty of Engineering and Computing
ID Code:31476
Deposited On:21 Nov 2025 12:07 by Rudali Huidrom . Last Modified 21 Nov 2025 12:07
Documents

Full text available as:

[thumbnail of PhD Thesis]
Preview
PDF (PhD Thesis) - Requires a PDF viewer such as GSview, Xpdf or Adobe Acrobat Reader
Creative Commons: Attribution-Noncommercial-No Derivative Works 4.0
5MB
Downloads

Downloads

Downloads per month over past year

Archive Staff Only: edit this record