No Cover Image

Journal article 940 views 127 downloads

Recommender systems and the amplification of extremist content

Joe Whittaker Orcid Logo, Sean Looney, Alastair Reed Orcid Logo, Fabio Votta

Internet Policy Review, Volume: 10, Issue: 2

Swansea University Authors: Joe Whittaker Orcid Logo, Sean Looney, Alastair Reed Orcid Logo

  • policyreview-2021-2-1565.pdf

    PDF | Version of Record

    Released under the terms of a Creative Commons Attribution 3.0 Germany

    Download (483.51KB)

Check full text

DOI (Published version): 10.14763/2021.2.1565

Abstract

Policymakers have recently expressed concerns over the role of recommendation algorithms and their role in forming “filter bubbles.” This is a particularly prescient concern in the context of extremist content online; these algorithms may promote extremist content at the expense of more moderate voi...

Full description

Published in: Internet Policy Review
ISSN: 2197-6775
Published: Internet Policy Review, Alexander von Humboldt Institute for Internet and Society 2021
Online Access: Check full text

URI: https://cronfa.swan.ac.uk/Record/cronfa57054
Tags: Add Tag
No Tags, Be the first to tag this record!
Abstract: Policymakers have recently expressed concerns over the role of recommendation algorithms and their role in forming “filter bubbles.” This is a particularly prescient concern in the context of extremist content online; these algorithms may promote extremist content at the expense of more moderate voices. In this article, we make two contributions to this debate. Firstly, we provide a novel empirical analysis of three platforms’ recommendation systems when interacting with far-right content. We find that one platform – YouTube – does amplify extreme and fringe content, while two – Reddit and Gab – do not. Secondly, we contextualise these findings into the regulatory debate. There are currently few policy instruments for dealing with algorithmic amplification, and those that do exist largely focus on transparency. We argue that policymakers have yet to fully understand the problems inherent in “de-amplifying” legal, borderline content and argue that a co-regulatory approach may offer a route towards tackling many of these challenges.
Keywords: Filter bubble, Online radicalisation, Algorithms, Extremism, Regulation
College: Faculty of Humanities and Social Sciences
Funders: Global Research Network on Terrorism & Technology
Issue: 2