Abstract:Recent years have seen a strong uptick in both the prevalence and real-world consequences of false information spread through online platforms. At the same time, encrypted messaging systems such as WhatsApp, Signal, and Telegram, are rapidly gaining popularity as users seek increased privacy in their digital lives. The challenge we address is how to combat the viral spread of misinformation without compromising privacy. Our FACTS system tracks user complaints on messages obliviously, only revealing the message… Show more
“…In these designs, code inspection and continued use should demonstrate the correct functionality of the system; there is no remote server that needs periodic inspection or auditing. A privacy improvement can also be achieved even for partial client privacy, by reducing the amount of information that is sent to the server [185] or by performing detection only after a threshold of problematic content was detected [33,231].…”
Section: Transparency Methods In the Literaturementioning
confidence: 99%
“…Message franking [62,98,151,164,167,173,183,222,359,376] (total: 10) Reveal source, traceback, or popular messages [173,231,285,360] (total: 4) Other user reporting [26,86,128,192,207,214,237,245,248,377,384] (total: 11)…”
Section: Corporate Network Monitoringmentioning
confidence: 99%
“…We did not see any other examples of such methods being applied, and we would expect them to be inappropriate for most problem contexts. For misinformation, two works specifically attempted to detect or limit "viral" disinformation [118,231] rather than general misinformation. We hope to see more research and innovation on this topic both in and out of the encrypted setting.…”
Section: Problem Context Affects Choice Of Detection and Response Mec...mentioning
confidence: 99%
“…In a different take on user reporting, Liu et al [231] described an approach for revealing messages once they reached a specific threshold of reports globally across all users, by constructing a "collaborative counting bloom filter. " Their goal was to reveal misinformation that was especially "viral" and thus by definition reached a large number of users; after this the server would be free to take action on those specific images (e.g.…”
Section: Responses Unique To End-to-end Encryptionmentioning
Popular messaging applications now enable end-to-end-encryption (E2EE) by default, and E2EE data storage is becoming common. These important advances for security and privacy create new content moderation challenges for online services, because services can no longer directly access plaintext content. While ongoing public policy debates about E2EE and content moderation in the United States and European Union emphasize child sexual abuse material and misinformation in messaging and storage, we identify and synthesize a wealth of scholarship that goes far beyond those topics. We bridge literature that is diverse in both content moderation subject matter, such as malware, spam, hate speech, terrorist content, and enterprise policy compliance, as well as intended deployments, including not only privacy-preserving content moderation for messaging, email, and cloud storage, but also private introspection of encrypted web traffic by middleboxes. In this work, we systematize the study of content moderation in E2EE settings. We set out a process pipeline for content moderation, drawing on a broad interdisciplinary literature that is not specific to E2EE. We examine cryptography and policy design choices at all stages of this pipeline, and we suggest areas of future research to fill gaps in literature and better understand possible paths forward.
“…In these designs, code inspection and continued use should demonstrate the correct functionality of the system; there is no remote server that needs periodic inspection or auditing. A privacy improvement can also be achieved even for partial client privacy, by reducing the amount of information that is sent to the server [185] or by performing detection only after a threshold of problematic content was detected [33,231].…”
Section: Transparency Methods In the Literaturementioning
confidence: 99%
“…Message franking [62,98,151,164,167,173,183,222,359,376] (total: 10) Reveal source, traceback, or popular messages [173,231,285,360] (total: 4) Other user reporting [26,86,128,192,207,214,237,245,248,377,384] (total: 11)…”
Section: Corporate Network Monitoringmentioning
confidence: 99%
“…We did not see any other examples of such methods being applied, and we would expect them to be inappropriate for most problem contexts. For misinformation, two works specifically attempted to detect or limit "viral" disinformation [118,231] rather than general misinformation. We hope to see more research and innovation on this topic both in and out of the encrypted setting.…”
Section: Problem Context Affects Choice Of Detection and Response Mec...mentioning
confidence: 99%
“…In a different take on user reporting, Liu et al [231] described an approach for revealing messages once they reached a specific threshold of reports globally across all users, by constructing a "collaborative counting bloom filter. " Their goal was to reveal misinformation that was especially "viral" and thus by definition reached a large number of users; after this the server would be free to take action on those specific images (e.g.…”
Section: Responses Unique To End-to-end Encryptionmentioning
Popular messaging applications now enable end-to-end-encryption (E2EE) by default, and E2EE data storage is becoming common. These important advances for security and privacy create new content moderation challenges for online services, because services can no longer directly access plaintext content. While ongoing public policy debates about E2EE and content moderation in the United States and European Union emphasize child sexual abuse material and misinformation in messaging and storage, we identify and synthesize a wealth of scholarship that goes far beyond those topics. We bridge literature that is diverse in both content moderation subject matter, such as malware, spam, hate speech, terrorist content, and enterprise policy compliance, as well as intended deployments, including not only privacy-preserving content moderation for messaging, email, and cloud storage, but also private introspection of encrypted web traffic by middleboxes. In this work, we systematize the study of content moderation in E2EE settings. We set out a process pipeline for content moderation, drawing on a broad interdisciplinary literature that is not specific to E2EE. We examine cryptography and policy design choices at all stages of this pipeline, and we suggest areas of future research to fill gaps in literature and better understand possible paths forward.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.