Can AI-Generated Persuasion Be Detected? Persuaficial Benchmark and AI vs. Human Linguistic Differences¶
Conference: ACL 2026 arXiv: 2601.04925 Code: https://github.com/ArkadiusDS/Persuaficial Area: Robotics & Embodied AI Keywords: Persuasion Detection, AI-Generated Text, Multilingual Benchmark, Linguistic Difference Analysis, Controllable Generation
TL;DR¶
Persuaficial is a high-quality multilingual benchmark covering six languages for AI-generated persuasive text. Systematic evaluation reveals that subtle AI persuasion is harder to detect than human persuasion (F1 drops ~20%), while intensified persuasion is paradoxically easier to detect.
Method¶
Key Designs¶
-
Four Controllable Generation Strategies: Paraphrasing (semantic equivalence), subtle rewriting (more covert), intensified rewriting (enhanced persuasion), and open-ended generation. Each simulates different real-world AI persuasion abuse scenarios.
-
Multilingual Multi-Source Construction: Three human persuasion datasets × four LLMs × six languages.
-
196-Dimensional Linguistic Feature Analysis: Using StyloMetrix for interpretable fine-grained analysis.
Key Experimental Results¶
| Strategy | F1 | Change vs Human |
|---|---|---|
| Human | 0.740 | — |
| Subtle Rewriting | 0.403 | ↓46% |
| Intensified Rewriting | 0.815 | ↑10% |
| Open-Ended | 0.896 | ↑21% |
Highlights & Insights¶
- "More subtle = harder to detect, more intense = easier to detect" is intuitive but first quantitatively verified here
- First systematic study of AI persuasion vs human persuasion detectability differences
Rating¶
- Novelty: ⭐⭐⭐⭐
- Experimental Thoroughness: ⭐⭐⭐⭐⭐
- Writing Quality: ⭐⭐⭐⭐
- Value: ⭐⭐⭐⭐