- Library Home /
- Search Collections /
- Open Collections /
- Browse Collections /
- UBC Theses and Dissertations /
- Combating disinformation in the age of generative AI...
Open Collections
UBC Theses and Dissertations
UBC Theses and Dissertations
Combating disinformation in the age of generative AI : from watermarking LLMs to persuasion analysis of memes Dabiriaghdam, Amirhossein
Abstract
In the era of generative Artificial Intelligence (AI), Large Language Models (LLMs) and Vision-Language Models (VLMs), such as ChatGPT, have become double-edged swords that have revolutionized content creation but also worsened the spread of disinformation. These technologies offer impressive abilities but also raise two questions: 1. Who created this textual content, AI or human? 2. What is the underlying intent of this multimodal content? This thesis addresses each question from distinct yet complementary perspectives to advance disinformation mitigation. First, from the “authenticity” aspect: we investigate how we can identify machine-generated text algorithmically, and introduce SimMark, a semantic-based sentence-level watermarking algorithm that injects visually imperceptible yet statistically detectable patterns into the LLM-generated text based on inter-sentence semantic similarity. SimMark is model-agnostic and operates without needing access to the models’ internals. We carried out extensive experiments and found that SimMark achieves high detection rates even under various adversarial conditions such as paraphrasing attacks, while preserving text quality and fluency. Second, from the “intent” aspect: we explore the persuasive messages conveyed through internet memes—the amalgamation of textual and visual elements—to spread disinformation and influence public opinion. A core challenge in this task lies in effectively understanding the meaning expressed through the metaphorical combination of text and images in memes. We propose a novel multimodal pipeline to tackle this problem by adding an intermediate captioning step using VLMs. These detailed captions, combined with the original meme text and visual features, then feed into a multimodal, hierarchical classification model. Experimental results show our approach effectively detects persuasion techniques across multilingual and multimodal data. Together, these contributions offer tools for tracking AI-generated text and analyzing the persuasive tactics in disinformation campaigns in this new age of generative artificial intelligence.
Item Metadata
Title |
Combating disinformation in the age of generative AI : from watermarking LLMs to persuasion analysis of memes
|
Creator | |
Supervisor | |
Publisher |
University of British Columbia
|
Date Issued |
2025
|
Description |
In the era of generative Artificial Intelligence (AI), Large Language Models (LLMs) and Vision-Language Models (VLMs), such as ChatGPT, have become double-edged swords that have revolutionized content creation but also worsened the spread of disinformation. These technologies offer impressive abilities but also raise two questions: 1. Who created this textual content, AI or human? 2. What is the underlying intent of this multimodal content? This thesis addresses each question from distinct yet complementary perspectives to advance disinformation mitigation.
First, from the “authenticity” aspect: we investigate how we can identify machine-generated text algorithmically, and introduce SimMark, a semantic-based sentence-level watermarking algorithm that injects visually imperceptible yet statistically detectable patterns into the LLM-generated text based on inter-sentence semantic similarity. SimMark is model-agnostic and operates without needing access to the models’ internals. We carried out extensive experiments and found that SimMark achieves high detection rates even under various adversarial conditions such as paraphrasing attacks, while preserving text quality and fluency.
Second, from the “intent” aspect: we explore the persuasive messages conveyed through internet memes—the amalgamation of textual and visual elements—to spread disinformation and influence public opinion. A core challenge in this task lies in effectively understanding the meaning expressed through the metaphorical combination of text and images in memes. We propose a novel multimodal pipeline to tackle this problem by adding an intermediate captioning step using VLMs. These detailed captions, combined with the original meme text and visual features, then feed into a multimodal, hierarchical classification model. Experimental results show our approach effectively detects persuasion techniques across multilingual and multimodal data.
Together, these contributions offer tools for tracking AI-generated text and analyzing the persuasive tactics in disinformation campaigns in this new age of generative artificial intelligence.
|
Genre | |
Type | |
Language |
eng
|
Date Available |
2025-08-26
|
Provider |
Vancouver : University of British Columbia Library
|
Rights |
Attribution-NonCommercial-NoDerivatives 4.0 International
|
DOI |
10.14288/1.0449874
|
URI | |
Degree (Theses) | |
Program (Theses) | |
Affiliation | |
Degree Grantor |
University of British Columbia
|
Graduation Date |
2025-11
|
Campus | |
Scholarly Level |
Graduate
|
Rights URI | |
Aggregated Source Repository |
DSpace
|
Item Media
Item Citations and Data
Rights
Attribution-NonCommercial-NoDerivatives 4.0 International