YouTube’s tweaks to recommend fewer conspiracy videos seem to be working

[ad_1]

The team trained its computers by having them analyze hundreds of conspiracy and non-conspiracy videos, helping them “learn” how to differentiate between the two. Once they were trained, the computers analyzed eight million recommended videos over the course of one year. Unfortunately, they were unable to classify the videos by actually watching them — instead they used transcripts, metadata (titles, descriptions and tags), the top 200 comments of each video and the perceived intent of those comments. Depending on the contents of each of these sources, the videos would be classified as either conspiratorial or not.

The researchers considered conspiracy theory videos as those which cover secret plots by those in power, ideas that are contrary to scientific consensus and views that are not backed by evidence or are unfalsifiable. These include both harmful conspiracies like those that state that mourning Sandy Hook family members are actually “crisis actors,” as well as the relatively innocuous ones, such as the assertion that the Mothman caused the collapse of the Silver Bridge in Point Pleasant, West Virginia.

The results show that there has been a dramatic reduction in the amount of conspiracy theory videos that YouTube serves to its users. However, the study did not use an actual user account during the analysis. Rather, all content was viewed in a “logged out” state. This may have had a major impact on results, as watch history is an important factor in how YouTube determines which videos to suggest to a logged-in user.

Diving into the content of the videos shows that YouTube may be selective in which conspiratorial subjects it suppresses. The rate at which climate change denial videos are being suggested hasn’t changed much, while the aforementioned Sandy Hook videos have been noticeably reduced. While leaving the judgement over what constitutes a harmful conspiracy to YouTube’s employees could be a problem, one could argue that suppressing conspiracy theory videos across the board could also be an issue. Non-harmful conspiracies can be fun — the strange antics of Andrew W.K. come to mind — and, after all, several conspiracy theories, such as MKUltra and Project Sunshine, turned out to be true.

On a broader level, these conspiracy theory videos are a problem because YouTube is becoming an increasingly popular source for news and facts — the platform has two billion active users every month — as well as its monopoly on the longform social video market. As such an active platform, bad actors are more frequently leveraging YouTube as a forum for social engineering, disinformation and misinformation.

[ad_2]

Source link