Facebook users tend almost entirely to click on links that they agree with, meaning that their news feeds can become an echo chamber, according to new research.
Most users of the site have friends with a broad range of political views. But they tend only to click on posts by those friends that they agree with politically, according to new research published in Science, and so the News Feed tends to show them the same kind of content.
Researchers call that self-sustaining effect the “filter bubble”. As technology companies learn more about their users, they show them the kind of results that they have clicked on in the past — which has the effect of showing people things that they already agreed with and wanted to see.
The new study was conducted by Facebook’s in-house scientists, and aimed to establish whether the site’s algorithm was creating a filter bubble, and whether that led to political polarisation. Facebook’s algorithm decides what users see on their news feed based on a range of criteria, including how often a link has been clicked by other users and whether it is the kind of thing that users have engaged with in the past.
To establish whether people tended to engage with people they disagreed with, Facebook researchers mapped out the site’s users according to which parties they’d indicated that they support. Over 10 million Facebook users were unknowingly mapped on a five point scale according to whether they were Conservative or Liberal.
The site then analysed news content to decide whether the organisation posting it tended to be conservative or liberal. Researchers calculated that by looking at the affiliations of the people who liked them — ranking the New York Times as slightly liberal but Fox News as slightly conservative, for instance.
Researchers then analysed when stories were “cross-cutting” — stories from conservative sources that were seen by liberal Facebook users, or vice versa. By analysing the two sets of data together, researchers could work out how often people see stories that they weren’t expected to agree with.
The Facebook news feed does tend to work as an echo chamber, the researchers found. Users were about 1 per cent less likely to see stories that they didn’t agree with, they said — less of a bias than some critics of the News Feed had suggested, but still enough to be significant.
COMMENT RULES: Comments that are judged to be defamatory, abusive or in bad taste are not acceptable and contributors who consistently fall below certain criteria will be permanently blacklisted. The moderator will not enter into debate with individual contributors and the moderator’s decision is final. It is Belfast Telegraph policy to close comments on court cases, tribunals and active legal investigations. We may also close comments on articles which are being targeted for abuse. Problems with commenting? email@example.com