TY - JOUR
T1 - Detecting Harmful Content on Online Platforms
T2 - What Platforms Need vs. Where Research Efforts Go
AU - Arora, Arnav
AU - Nakov, Preslav
AU - Hardalov, Momchil
AU - Sarwar, Sheikh Muhammad
AU - Nayak, Vibha
AU - Dinkov, Yoan
AU - Zlatkova, Dimitrina
AU - Dent, Kyle
AU - Bhatawdekar, Ameya
AU - Bouchard, Guillaume
AU - Augenstein, Isabelle
N1 - Publisher Copyright:
© 2023 Copyright held by the owner/author(s). Publication rights licensed to ACM.
PY - 2023
Y1 - 2023
N2 - The proliferation of harmful content on online platforms is a major societal problem, which comes in many different forms, including hate speech, offensive language, bullying and harassment, misinformation, spam, violence, graphic content, sexual abuse, self-harm, and many others. Online platforms seek to moderate such content to limit societal harm, to comply with legislation, and to create a more inclusive environment for their users. Researchers have developed different methods for automatically detecting harmful content, often focusing on specific sub-problems or on narrow communities, as what is considered harmful often depends on the platform and on the context. We argue that there is currently a dichotomy between what types of harmful content online platforms seek to curb, and what research efforts there are to automatically detect such content. We thus survey existing methods as well as content moderation policies by online platforms in this light and suggest directions for future work.
AB - The proliferation of harmful content on online platforms is a major societal problem, which comes in many different forms, including hate speech, offensive language, bullying and harassment, misinformation, spam, violence, graphic content, sexual abuse, self-harm, and many others. Online platforms seek to moderate such content to limit societal harm, to comply with legislation, and to create a more inclusive environment for their users. Researchers have developed different methods for automatically detecting harmful content, often focusing on specific sub-problems or on narrow communities, as what is considered harmful often depends on the platform and on the context. We argue that there is currently a dichotomy between what types of harmful content online platforms seek to curb, and what research efforts there are to automatically detect such content. We thus survey existing methods as well as content moderation policies by online platforms in this light and suggest directions for future work.
KW - Additional Key Words and PhrasesOnline harms
KW - bullying and harassment
KW - content moderation
KW - graphic content
KW - hate speech
KW - misinformation
KW - offensive language
KW - self-harm
KW - sexual abuse
KW - spam
KW - violence
U2 - 10.1145/3603399
DO - 10.1145/3603399
M3 - Journal article
AN - SCOPUS:85176785424
VL - 56
SP - 1
EP - 17
JO - ACM Computing Surveys
JF - ACM Computing Surveys
SN - 0360-0300
IS - 3
M1 - 72
ER -