Facebook has been quietly experimenting with reducing the amount of political content it puts in usersā news feeds. The move is a tacit acknowledgment that the way the companyās algorithms work can be a problem.
The heart of the matter is the distinction between provoking a response and providing content people want. Social media algorithms ā the rules their computers follow in deciding the content that you see ā rely heavily on peopleās behavior to make these decisions. In particular, they watch for content that people respond to or āengageā with by liking, commenting and sharing.
As a computer scientist who studies the ways large numbers of people interact using technology, I understand the logic of using the wisdom of the crowds in these algorithms. I also see substantial pitfalls in how the social media companies do so in practice.
[…]