SAFETYLIT WEEKLY UPDATE

We compile citations and summaries of about 400 new articles every week.
RSS Feed

HELP: Tutorials | FAQ
CONTACT US: Contact info

Search Results

Journal Article

Citation

Wang S, Kim KJ. Cyberpsychol. Behav. Soc. Netw. 2023; ePub(ePub): ePub.

Copyright

(Copyright © 2023, Mary Ann Liebert Publishers)

DOI

10.1089/cyber.2022.0158

PMID

37140448

Abstract

Artificial intelligence (AI) has been increasingly integrated into content moderation to detect and remove hate speech on social media. An online experiment (Nā€‰=ā€‰478) was conducted to examine how moderation agents (AI vs. human vs. human-AI collaboration) and removal explanations (with vs. without) affect users' perceptions and acceptance of removal decisions for hate speech targeting social groups with certain characteristics, such as religion or sexual orientation. The results showed that individuals exhibit consistent levels of perceived trustworthiness and acceptance of removal decisions regardless of the type of moderation agent. When explanations for the content takedown were provided, removal decisions made jointly by humans and AI were perceived as more trustworthy than the same decisions made by humans alone, which increased users' willingness to accept the verdict. However, this moderated mediation effect was only significant when Muslims, not homosexuals, were the target of hate speech.


Language: en

Keywords

social media; artificial intelligence; hate speech; content moderation; transparency

NEW SEARCH


All SafetyLit records are available for automatic download to Zotero & Mendeley
Print