View a PDF of the paper titled Unmasking the Imposters: How Censorship and Domain Adaptation Affect the Detection of Machine-Generated Tweets, by Bryan E. Tuck and Rakesh M. Verma
Abstract:The rapid development of large language models (LLMs) has significantly improved the generation of fluent and convincing text, raising concerns about their potential misuse on social media platforms. We present a comprehensive methodology for creating nine Twitter datasets to examine the generative capabilities of four prominent LLMs: Llama 3, Mistral, Qwen2, and GPT4o. These datasets encompass four censored and five uncensored model configurations, including 7B and 8B parameter base-instruction models of the three open-source LLMs. Additionally, we perform a data quality analysis to assess the characteristics of textual outputs from human, “censored,” and “uncensored” models, employing semantic meaning, lexical richness, structural patterns, content characteristics, and detector performance metrics to identify differences and similarities. Our evaluation demonstrates that “uncensored” models significantly undermine the effectiveness of automated detection methods. This study addresses a critical gap by exploring smaller open-source models and the ramifications of “uncensoring,” providing valuable insights into how domain adaptation and content moderation strategies influence both the detectability and structural characteristics of machine-generated text.
Submission history
From: Bryan Tuck [view email]
[v1]
Tue, 25 Jun 2024 22:49:17 UTC (2,660 KB)
[v2]
Tue, 17 Sep 2024 21:29:13 UTC (3,472 KB)
[v3]
Wed, 15 Jan 2025 22:20:15 UTC (3,589 KB)
Source link
lol