Can Wikipedia-like citations on YouTube curb misinformation?
While Google has long been synonymous with search, people are increasingly seeking information directly through video platforms such as YouTube. Videos can be dense with information: text, audio, and image after image. Yet each of these layers presents a potential source of error or deceit. And when people search for videos directly on a site […]
While Google has long been synonymous with search, people are increasingly seeking information directly through video platforms such as YouTube. Videos can be dense with information: text, audio, and image after image. Yet each of these layers presents a potential source of error or deceit. And when people search for videos directly on a site like YouTube, sussing out which videos are credible sources can be tricky.
Credit: Hughes et al./CHI 2024
While Google has long been synonymous with search, people are increasingly seeking information directly through video platforms such as YouTube. Videos can be dense with information: text, audio, and image after image. Yet each of these layers presents a potential source of error or deceit. And when people search for videos directly on a site like YouTube, sussing out which videos are credible sources can be tricky.
To help people vet videos, University of Washington researchers created and tested Viblio, a browser extension that lets viewers and creators add Wikipedia-like citations to YouTube videos. The prototype offers users an alternate timeline, studded with notes and links to sources that support, refute or expand on the information presented in the video. Those links also appear in a list view, like the “References” section at the end of Wikipedia articles. In tests, 12 participants found the tool useful for gauging the credibility of videos on topics ranging from biology to political news to COVID-19 vaccines.
The team will present its findings May 14 in Honolulu at the ACM CHI Conference on Human Factors in Computing Systems. Viblio is not available to the public.
“We wanted to come up with a method to encourage people watching videos to do what’s called ‘lateral reading,’ which is that you go look at other places on the web to establish whether something is credible or true, as opposed to diving deep into the thing itself,” said senior author Amy X. Zhang, an assistant professor in the Paul G. Allen School of Computer Science & Engineering. “In previous research, I’d worked with the people at X’s Community Notes and with Wikipedia and seen that crowdsourcing citations and judgments can be a useful way to call out misinformation on platforms.”
To inform Viblio’s design, the team studied how 12 participants — mostly college students under 30 — gauged the credibility of YouTube videos when searching for them on the platform and while watching them. All said familiarity with the video’s source and the name of the channel were important. But many cited signs of a video’s potentially faulty credibility: the quality of the video, the user’s degree of interest in it, its ranking in search results, its length and the number of views or subscribers.
The team also found that in one case a participant misinterpreted a YouTube information panel as an endorsement of the video from the Centers for Disease Control and Prevention. But these panels are actually links to supplemental information that the site attaches to videos on “topics prone to misinformation.”
“The trouble is that a lot of YouTube videos, especially more educational ones, don’t offer a great way for people to prove they’re presenting good information,” said Emelia Hughes, a doctoral student at University of Notre Dame who completed this research as a UW undergraduate student in the Information School. “I’ve stumbled across a couple of YouTubers who were coming up with their own ways to cite sources within videos. There’s also not a great way to fight bad information. People can report a whole video, but that’s a pretty extreme measure when someone makes one or two mistakes.”
The researchers designed Viblio so users can better understand videos’ content while also avoiding things like users misinterpreting the additional information. To add a citation, users click a button on the extension. They can then add a link, select the timespan their citation references and add optional comments. They can also select the type of citation, which marks it with a colored dot in the timeline: “refutes the video clip’s claim” (red), “supports the video clip’s claim” (green) or “provides further explanation” (blue dot).
To test the system, the team had the study participants use Viblio for two weeks on a range of videos, including clips from Good Morning America, Fox News and ASAPScience. Participants could add citations as well as watch videos with other participants’ citations. For many, the added citations changed their opinion of certain videos’ credibility. But the participants also highlighted potential difficulties with deploying Viblio at a larger scale, such as the conflicts that arise in highly political videos or those on controversial topics that don’t fall into true-false binaries.
“What happens when people with different value systems add conflicting citations?” said co-author Tanu Mitra, a UW assistant professor in the Information School. “We of course have the issue with bad actors potentially adding misinformation and incorrect citations, but even when the users are acting in good faith, but have conflicting options, whose citation should be prioritized? Or should we be showing both conflicting citations? These are big challenges at scale.”
The researchers highlight a few areas for further study, such as expanding Viblio to other video platforms such as TikTok or Instagram; studying its useability at a greater scale to see whether users are motivated enough to continue adding citations; and exploring ways to create citations for videos that don’t get as much traffic and thus have fewer citations.
“Once we get past this initial question of how to add citations to videos, then the community vetting question remains very challenging,” Zhang said. “It can work. At X, Community Notes is working on ways to prevent people from ‘gaming’ voting by looking at whether someone always takes the same political side. And Wikipedia has standards for what should be considered a good citation. So it’s possible. It just takes resources.”
Additional co-authors on the paper include Renee Wang, who completed this work as an undergraduate at the UW and is now at Microsoft; Prerna Juneja, who completed this research as a UW doctoral student in the iSchool and is now an assistant professor at Seattle University; and Tony Li, who completed this work as a UW graduate student in human centered design and engineering and is now a doctoral student at University of California San Diego. This research was funded by the WikiCred Grants Initiative.
For more information, contact Zhang at axz@cs.uw.edu, Hughes at ehughes8@nd.edu, and Mitra at tmitra@uw.edu.
DOI
10.1145/3613904.3642490
Article Title
Viblio: Introducing Credibility Signals and Citations to Video-Sharing Platforms
Article Publication Date
14-May-2024
What's Your Reaction?