Social media have undoubtedly changed our ways of living. Their presence concerns an increasing number of users (over 4,74 billion) and pervasively expands in the most diverse areas of human life. Marketing, education, news, data, and sociality are just a few of the many areas in which social media play now a central role. Recently, some attention toward the link between social media and political participation has emerged. Works in the field of artificial intelligence have already pointed out that there is a close link between the use of machine learning algorithms in social media and possible epistemic isolation, which could lead to political radicalization. The idea supporting this paper is that artificial intelligence for social media can actively put users’ deliberative capacity at risk and foster political extremism. To prove these claims, I proceed along two lines of inquiry. First, I focus on filter bubbles, namely the result of selections made by algorithms that recommend contents that meet users’ expectations and opinions. To analyze this phenomenon, I refer to the Deweyan model of experience. Second, I connect the filter bubbles problem to the Deweyan idea of deliberative and participatory democracy and Nussbaum’s concept of political compassion. The purpose of this paper is to provide a philosophical foundation that can both (1) effectively serve as a method for analyzing machine learning algorithms and their potential problems in relation to political extremism, and (2) be adopted as a standard to counter the danger of extremism associated with social media experience.