Abstract
Liars and truth-tellers often communicate differently. When tasked with writing deceptive text, humans use prior knowledge and experiences to intentionally deceive their target. Automatically generated text via large language models; LLMs, on the other hand, mirror training instances that are most likely written by humans. In the case of content like reviews, automatically generated language is inherently deceptive because the system is not grounded in material world experiences. In this paper, we characterize differences between (a) truthful text written by humans, (b) intentionally deceptive text written by humans, and (c) inherently deceptive text written by state-of-the-art language models (ChatGPT). We examined the expression of thirteen psychologically grounded and fundamental human traits (e.g., personality and empathy) across truthful and deceptive hotel reviews, finding that texts written by humans were more diverse (had more variation) in their expressions of personality than texts written by ChatGPT. Across all human traits we found that truthful and deceptive human language was easier to distinguish from machine generated language. Building on these differences, we trained a classifier using only the thirteen human traits to automatically discriminate between truthful and deceptive language, with a classification AUC of up to 0.966. Thus, despite the fact that large language models mirror text written by genuine (and truthful) humans, their lack of diversity in human traits makes them easier to identify. These results suggest that psychologically grounded human traits offer a robust feature set unaffected by the “human-ness” of LLM language, and further suggest that AI and humans are behaviorally different when communicating about experiences.
| Original language | English |
|---|---|
| Pages (from-to) | 23-37 |
| Number of pages | 15 |
| Journal | CEUR Workshop Proceedings |
| Volume | 3477 |
| State | Published - 2023 |
| Event | 1st International Workshop on Implicit Author Characterization from Texts for Search and Retrieval, IACT 2023 - Taipei, Taiwan, Province of China Duration: Jul 27 2023 → … |
Keywords
- AI Language
- ChatGPT
- Deception
- Large Language Models
- Personality
- Text analysis
Fingerprint
Dive into the research topics of '“I Slept Like a Baby”: Using Human Traits To Characterize Deceptive ChatGPT and Human Text'. Together they form a unique fingerprint.Cite this
- APA
- Author
- BIBTEX
- Harvard
- Standard
- RIS
- Vancouver