The provision of toxic content and misinformation is a frequent phenomenon in current social media with specific impact and risks for younger users. We report on efforts taken in the project Courage to mitigate and overcome these threats through dedicated educational technology inspired by psychological and pedagogical approaches. The aim is to empower adolescents to confidently interact with and utilize social media and to increase their awareness and resilience. For this purpose, we have adopted approaches from the field of Intelligent Tutoring Systems, namely the provision of a virtual learning companion (VLC). The technical system is a browser-based environment that allows for combining a controllable social media space with a VLC as a plugin. This environment is backed by an API that bundles Machine Learning and Natural Language Processing algorithms for detecting and classifying different types of risks. The pedagogical scenarios that are supported by this technical environment and approach range from chat-based dialogues to more complex narrative scripts.
Pre-training of neural networks has recently revolutionized the field of Natural Language Processing (NLP) and has before demonstrated its effectiveness in computer vision. At the same time, advances around the detection of fake news were mainly driven by the context-based paradigm, where different types of signals (e.g. from social media) form graph-like structures that hold contextual information apart from the news article to classify. We propose to merge these two developments by applying pretraining of Graph Neural Networks (GNNs) in the domain of context-based fake news detection. Our experiments provide an evaluation of different pre-training strategies for graph-based misinformation detection and demonstrate that transfer learning does currently not lead to significant improvements over training a model from scratch in the domain. We argue that a major current issue is the lack of suitable large-scale resources that can be used for pre-training.
The rise of deep learning methods has transformed the research area of natural language processing beyond recognition. New benchmark performances are reported on a daily basis ranging from machine translation to question-answering. Yet, some of the unsolved practical research questions are not in the spotlight and this includes, for example, issues arising at the interface between spoken and written language processing.We identify sentence boundary detection and speaker change detection applied to automatically transcribed texts as two NLP problems that have not yet received much attention but are nevertheless of practical relevance. We frame both problems as binary tagging tasks that can be addressed by fine-tuning a transformer model and we report promising results.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.