AI Disclosure Labels: Potentially More Harmful Than Helpful

Sissa Medialab

The growing use of AI-generated scientific and science-related content, especially on social media, raises important concerns: these texts may contain false or highly persuasive information that is difficult for users to detect, potentially shaping public opinion and decision-making.

Several jurisdictions and platforms are moving toward clearer disclosure of AI-generated or AI-synthesised content to protect the public. However, a new study published in JCOM warns that these labels may have the opposite effect of what regulators intend, decreasing the credibility of true scientific information while increasing that of false claims.

The Risks of AI-Generated Scientific Content

AI-generated content can be misleading for at least two reasons. First, language models may "hallucinate," producing plausible but factually incorrect statements. Second, users can deliberately prompt AI systems to create false yet credible messages. For this reason, several countries have introduced transparency obligations requiring online content generated or synthesized by AI to be clearly labeled.

In their new study, Teng Lin, a PhD candidate at the School of Journalism and Communication, University of Chinese Academy of Social Sciences (UCASS), Beijing, and Yiqing Zhang, a Master's student at the same school, set out to test whether these disclosure labels actually achieve their stated goal of protecting the public from misinformation.

The Experimental Study

"We focused on science-related information shared on social media," explains Teng.

The experimental study involved 433 participants recruited online through the Credamo platform between March and May 2024. The researchers created four types of social media posts: correct information with or without an AI label, and misinformation with or without an AI label. The texts were adapted using GPT-4 from items published by China's Science Rumour Debunking Platform, creating both accurate and misleading Weibo-style versions, and were then independently checked by the researchers. Participants were asked to rate the perceived credibility of each post on a scale from 1 to 5. The researchers also measured participants' negative attitudes toward AI and their level of involvement with the topic.

A Paradoxical Effect

The results revealed a counterintuitive pattern. "Our most important finding is what we call a 'truth-falsity crossover effect,'" says Teng. "The same AI label pushes credibility in opposite directions depending on whether the information is true or false: it reduces the credibility of true messages and increases the credibility of false ones." He adds that this does not necessarily mean the effect would be identical across all platforms or formats, but in their experimental setting the pattern was clear.

In this context, AI disclosure does not help people distinguish between true and false information. Instead, it appears to redistribute credibility in a paradoxical way.

Teng and Zhang also found that individual attitudes toward AI play a role. Participants who held more negative views of AI penalized correct information even more strongly when it was labeled as AI-generated. However, even among those with negative attitudes, the credibility boost observed for misinformation did not disappear entirely; it was only partially reduced, and this attenuation was topic-dependent, as it weakened in one topic but was not eliminated overall.

This suggests that so-called "algorithm aversion" does not lead to a uniform rejection of AI-generated content, but rather to a more complex and asymmetric reaction.

The Need for Careful Policy Design

Research like this highlights the need for careful testing before implementing regulatory interventions, as well-intended transparency measures may produce unintended consequences.

"In our paper we put forward some recommendations, although they need further research to be validated," Teng explains. "One proposal is to implement a dual-labeling approach. Instead of simply stating that the content is AI-generated, the label could also include a disclaimer indicating that the information has not been independently verified, or add a risk warning." In short, simply informing audiences that a text was generated by AI may not be sufficient on its own.

"Another recommendation is to adopt a graded or categorical labeling system," Teng adds. "Different types of scientific information carry different levels of risk. For example, medical or health-related information may require a stronger warning, while information about new technologies may involve lower risk. So we suggest using different levels of disclosure depending on the type and risk level of the content."

The paper Visible Sources and Invisible Risks: Exploring the Impact of AI Disclosure on Perceived Credibility of AI-Generated Content by Teng Lin and Yiqing Zhang is published in the Journal of Science Communcation (JCAP)

/Public Release. This material from the originating organization/author(s) might be of the point-in-time nature, and edited for clarity, style and length. Mirage.News does not take institutional positions or sides, and all views, positions, and conclusions expressed herein are solely those of the author(s).View in full here.