Skip to content
Human or AI?

Evaluating Labels on AI-Generated Social Media Content

March 2025

Human or AI?: 
Safeguarding youth privacy in the age of generative artificial intelligence

Authors

Angus Lockhart

Angus Lockhart
Senior Policy Analyst

Christelle Tessono
Policy and Research Assistant


Contributors

  • Nina Rafeek Dow
    Zaynab Choudhry
    Suzanne Bowness

Funder

This work is funded by The Canadian Digital Media Research Network


Tags


Share

BOLD IDEA
The current labelling approach by social media platforms isn’t working. More effective methods must be implemented to help improve trust and transparency online. 


It’s official—when it comes to online content, seeing is no longer believing. Rapid advancements in artificial intelligence (AI) technology has made it possible to create hyper-realistic synthetic images, videos, and audio, commonly called “deepfakes” when created for malicious purposes. And as the content-generating tools inevitably grow in sophistication, fake content will become virtually indistinguishable from real content. This has serious implications for misinformation, trust in the media, and even our democracy.

As such, navigating online platforms and information is becoming more complex, leaving users with greater challenges than ever before in determining whether what they are seeing is real or synthetic.

Because of this, Canadian residents want tools and strategies that help them confidently evaluate the authenticity of the content they encounter. Research from the Dais’ Survey of Online Harms in Canada 2024 found that 82 percent of Canadian residents want online platforms to label synthetic media and deepfakes so they know what they are looking at.

In response to the proliferation of AI-generated content, major platforms have begun rolling out strategies designed to label or otherwise highlight this content. But are they effective? Without a clear understanding of the effectiveness of these measures, it’s difficult to know whether they are helping to build public trust or simply window-dressing.

To investigate this question, we conducted a novel survey experiment examining Canadian residents’ impressions of AI-generated content in a mock social media environment (Facebook), applying a variety of labelling approaches.

This report offers an overview of the existing literature on AI-generated content, provides a scan of the current regulatory landscape surrounding AI-generated content on social media platforms, and shares the results of our survey on Canadian residents’ experience with synthetic and deepfake media online.

  • Small AI-generated content labels have no meaningful effect on user trust or sharing behaviour. This makes these labels functionally ineffective.
  • Full-screen labels are most effective. Only a full-screen label blocking AI content until manually removed significantly reduces exposure and improves perceptions of effective labelling. However, no social media platforms use this label type.
  • Nearly half (47 percent) of Canadian residents see deepfakes on at least a weekly basis—double the figure from our previous study: Survey of Online Harms in Canada.
  • One in five Canadian residents see synthetic media or deepfakes multiple times a day. Seventy percent encounter deepfakes at least a few times a month.
  • Canadian residents over 60 do not have a solid grasp of deepfakes. Older Canadian residents are more likely to be unsure if content is real or synthetic, suggesting difficulties in identifying AI-generated material. While most Canadian residents are at least somewhat familiar with deepfakes, younger individuals (16-29) and men of any age are significantly more aware than older demographics.
  • Exposure to deepfakes is highest among TikTok users (78 percent) and Instagram users (76 percent). Daily exposure was highest with YouTube users. This finding highlights social media as a key vector for synthetic content.
  • Canada’s governance approach remains largely voluntary and has stalled. Despite the growing presence of generative AI, legislative efforts—Canada’s Artificial Intelligence and Data Act (Bill C-27) and Online Harms Act (Bill C-63)— that would apply to AI-generated content on social platforms have stalled due to the prorogation of Canada’s parliament.

Introduction


Distinguishing human-generated from AI-generated content will only become increasingly difficult. According to our Survey of Online Harms in Canada, social media is growing as a news source, and these platforms are home to the majority of deepfake content. How can Canadians navigate this new reality? 

 This report examines the growing presence of generative AI (genAI) content in online spaces and evaluates the effectiveness of various labelling techniques implemented by social media platforms. We begin with a background on genAI, outlining its technological evolution and the role of social media platforms in shaping the digital information landscape.

Next, we explore how generative AI is used to create and disseminate content, highlighting the ease with which synthetic media can be produced and shared online. From there, we discuss factors influencing user trust in social media content, drawing on research that demonstrates how content presentation and source attribution shape public perceptions.

We then provide an overview of content moderation strategies used by the social media platforms, assessing their approaches to mitigating misinformation and disinformation, including AI-generated content labelling. We examine the effectiveness of current labelling techniques and compare the labelling practices of major social media platforms.

Following this, we review the regulatory landscape for genAI governance in Canada, identifying gaps in statutory oversight, especially compared with jurisdictions such as the European Union, which has introduced clear obligations for AI-generated content disclosure.

To test the impact of different AI-labelling techniques, we conducted an experiment simulating a social media environment, assessing whether labels influence user perception, engagement, and trust in AI-generated content, followed by a survey. The survey was conducted online using the Leger panel with 2,472 residents of Canada aged 16 and older. The survey was conducted from September 20, 2024 to November 21, 2024, in English and French.

The survey findings provide insights into Canadians’ familiarity with deepfakes, their exposure to synthetic content, and their reactions to various labelling methods.

Our findings conclude that the light-touch labelling approach currently employed by many platforms is largely ineffective. While Canadians want transparency around AI-generated content, subtle disclaimers do little to change user behaviour. More assertive interventions—such as content overlays that require user action to view AI-generated material—may be more effective in reducing engagement and improving transparency. Given the absence of meaningful AI governance in Canada, platforms remain largely unaccountable, reinforcing the need for stronger policy interventions to ensure that digital spaces remain trustworthy.

Background


Generative AI is defined as the set of AI systems that “create new content in response to prompts, based on their training data”.1 This content can include , but not be limited to, text, video, audio, images, and code, for example. Although OpenAI’s ChatGPT raised the technology’s public profile when it launched in 2022, generative AI tools are not a recent development. Instead, they stem from decades of research and innovation at the intersections of computer science, linguistics, and statistics.2 The first types of generative technologies were language models. 

Developed in the early 1950s, language models consisted of rule-based systems that relied on linguistic rules to process language. In the following decades, statistical models were developed using probabilistic methods that determined the likelihood of a word sequence in a given context (e.g., determining the likelihood that the word “rainy” could follow the word “day”).3 In the early 2000s, advancements in computing infrastructure allowed researchers to develop systems that could analyze large sets of various types of data. By the mid-2010s, decades of foundational theoretical research, combined with powerful computing infrastructure, allowed the creation of more sophisticated computational techniques to analyze significantly large data sets. The growing accessibility of consumer electronics (e.g., phones, laptops, and tablets), increased quality of internet connection, the digitization of many aspects of our professional lives, and the rise of social media platforms. These factors have all converged to allow the creation of generative AI tools like ChatGPT, Perplexity AI, and Midjourney.     

The arrival of generative AI coincides with an information ecosystem that is dominated by a small number of powerful social media platforms. Information about current events domestically and internationally is delivered through algorithmically-driven methods. These algorithms rapidly share information and prioritize user engagement to drive ad revenue. This has led to misinformation (the unknowing spread of false information) and disinformation (the intentional spread of false information to serve political objectives, cause harm, generate revenue) to undermine the integrity of information ecosystems.4 Moreover, extremist content has proliferated on social media as well, which has created unhealthy digital environments. GenAI tools make it easier for users to develop their own content, which opens the door to malicious uses that can further spread misinformation. 

Considering these issues, various jurisdictions are grappling with how to develop the necessary regulatory oversight mechanisms to ensure that social media platforms are held accountable. In the following section, we outline the scholarship dedicated to understanding the integrity of information circulating on social media platforms, and discuss how information is created, disseminated, perceived, and subsequently moderated in the context of generative AI.

With regards to content creation, there are various computational techniques that can be adopted to generate text, images, audio, and videos. These include simple editing software such as Photoshop, meme generators,5 and now genAI tools like tools like Dall-E and Stable Diffusion. In light of the popularity of genAI, traditional editing software companies have now added AI capabilities to their products (e.g. Adobe).

While editing images and other media content is not a recent phenomenon, genAI tools increase the average user’s ability to generate original (fake) content, or modify existing content, due to the ease of use, accessibility and ability of these tools to create realistic content easily, quickly, and to a high level of realism. According to Sensity AI, a company that tracks malicious use of generative AI through deepfake detection software, by 2024 there were nearly 3,000 dedicated tools for face swap, lip sync, face reenactment and AI avatars available. Additionally, there are over 10,200 tools for image generation and 11,018 for voice generation and cloning.6

Scholars have studied how information appears on social media feeds. In recent years, particular attention has been paid to how automated systems are used by platforms to create engaging feeds for users. For instance, researchers have identified that “recommender systems” are used by platforms like Facebook, X, and TikTok, to match content with user preferences and help users discover new interests.7 The use of recommender systems has been found to influence user behaviour to optimize user engagement.8 They have garnered a lot of scholarly attention as researchers have also investigated the harms that emerge from them. For instance, research has demonstrated that recommender systems can exacerbate the spread of false news online,lead users to extremist content, enable cyberbullying, and promote toxic content that triggers mental health issues.9 As a result of these harms, scholars and technologists have worked towards building technical interventions that centralize harm reduction and fairness in the design features of recommender systems.10 

Researchers have focused on understanding the factors that build a sense of trust in information among social media users. Notably, researchers have found that how information is presented to viewers influences their perception of its accuracy. For example, Ahmed (2021) conducted a US-based survey and found that captioning fake videos on Instagram decreased the proportion of participants who perceived the content to be accurate by 18 percent and reduced their likelihood of sharing the content by 15 percent.11 Yet, to date, there have been few studies attempting to understand how the public perceives AI-generated information.12 

There is also a growing body of scholarship analyzing the impact of media “provenance” — a term for the creator or source originating the content—on user perception of trust and accuracy. Notably, Feng et al. (2023), presented a mock Twitter-like feed to 595 participants based in the United States and United Kingdom. Participants were asked to rate the trust and accuracy of the content they saw on the feed containing a clickable pop-up box with various types of details about the source of the information. They found that providing source information through the pop-up boxes lowered the participant’s trust in deceptive, or fake, content shown to them.13

To address issues of misinformation and disinformation, social media platforms began to establish trust and safety teams that developed a wide range of content moderation policies and practices. As part of their arsenal, platforms moderated content manually through data workers and through automated methods powered by algorithmic systems.14 Data workers will review content and flag that it as harmful or as going against the platform’s guidelines. Moreover, this manual moderation is also used to train algorithmic systems to identify and subsequently take down content. However, research has demonstrated that platforms do not moderate content effectively.15 For instance, in a study that assessed the efficacy of Facebook, YouTube, and TikTok policies for political advertising, the authors found that platforms unevenly applied their policies across their respective platforms.16

In recent years, platforms have adopted a wide range of technical tools in an attempt to create healthier information ecosystems. These include, but are not limited to, labelling posts (e.g. using tags, identifying that it is sponsored/monetized content, political, etc.), inserting community notes, and adding descriptions, and/or account verification for public figures. These tools are meant to help a user decide whether to trust the source of information, or identify misleading information.

While the creation of trust and safety teams was a promising start in the effort to moderate content, platforms such as X (formerly Twitter) and Meta have significantly scaled back these initiatives. 17 With the increased use and popularity of content generated with AI systems on social media, and the reduction in content moderation, malicious actors may see this as an opportunity to increase misinformation and disinformation. 

Researchers have sought to assess the efficacy of labelling techniques. For instance, a study by Lewis et al. (2022) found an increase in users’ ability to detect deepfakes when labels were provided. However, despite detection moving from 11 percent to 22 percent with the aid of labels, the majority of participants were unable to identify a deepfake video.18 In another study of social media users in Italy, Iacobucci et al. (2021) found that describing the threat of deepfakes to participants, along with providing a definition of a deepfake ahead of viewing manipulated video increased their correct identification of fake content by 28 percent and decreased sharing intention by 57 percent.19

In Epstein et al. (2023), the authors attempt to discern the public’s understanding of nine different labelling terms.20 The authors recruited thousands of participants from the United States, Mexico, Brazil, India, and China. The researchers found that the labelling terms vary in their effectiveness at accurately representing the moderation goals. That is, for content that is intended to mislead users, labelling terms “deepfake” and “manipulated” were consistently able to represent the content’s misleading nature to users. Whereas labels such as “AI-generated,” “Generated with an AI Tool” and “AI manipulated” were perceived as useful for content that is made with AI, however, it did not push users to consider whether the content was truthful or misleading.

In Gruzd, Mai & Soares (2024), the authors used ModSimulator, an open-source tool that creates an interface similar to Facebook’s, to understand the effectiveness of two types of warning labels: footnotes and the “blur” filter. In terms of content, they created posts with false claims about the Russia-Ukraine war and presented the experiment to 1,500 participants. The researchers found that both warning levels were effective at decreasing user engagement with the false posts, with no significant difference in effectiveness.21 In Lu and Yuan (2024), the researchers conducted two experiments where they had 2,808 US participants watch deepfake videos of politicians expressing opposing views on climate change, with some participants viewing the video with AI disclaimers and others without the disclaimers. They found that AI disclaimers significantly enhanced viewer’s recognition of the content as deepfake parodies.22

In light of the increased use of generative AI tools to produce social media content, efforts have increased in recent years to identify synthetic images and videos. Methods such as watermarking and labelling posts have been introduced by platforms. Some technology companies have been collaborating in self-governance efforts. For instance, the Coalition for Content Provenance and Authenticity (C2PA) is a project that seeks to develop global technical standards for content provenance. The C2PA has many members, including corporations such as Google, Intel, Sony, Meta, Microsoft, Amazon, Adobe, and OpenAI.23

In Table 1, we outline the various approaches that popular platforms have adopted to label AI-generated content on their platform.

PlatformLabelling approachRestrictionsPenalty
TikTokContent creators must label content, not platform automated.*AI-generated content containing the likeness (visual or audio) of a real or fictional person are not allowed (even with the AI-generated content label), this includes: private figures (adults or minors who are not public figures), and public figures (e.g., adults with important public roles such as celebrities, business leaders, or a governance official).**Undisclosed penalty.
X
(formerly Twitter)
Does not use labels. Community guidelines may be used to flag content instead.***N/A.N/A.
Meta
(Facebook, Messenger Instagram, Threads)
Meta platforms use automated detection software and users can label their content as AI-generated.****Meta requires an AI label when content has photorealistic video or realistic-sounding audio that was digitally created, modified or altered, including with AI.

Meta does not require a label for images that have been created or modified with AI. However, these images may still require a label if the systems detect that they were AI-generated or if they were modified using AI.
Undisclosed penalty.
YouTubeRequire content creators to label when they’ve created altered or synthetic content that is realistic.*****Labels required for content that is photorealistic.Creators who consistently choose not to use labels may be subject to content removal, suspension from the YouTube partner program, or other penalties. 
BlueskyDoes not label generative AI content.******N/A.N/A.

*24 **25 ***26 ****27 *****28 ******29

Canada has faced a series of challenges in passing AI legislation. Most significantly, despite introducing the Artificial Intelligence and Data Act (AIDA) part of Bill C-27 in June 2022, the Act was unable to move past committee consideration before parliament was prorogued in January 2025.30 As a result, Canada does not have a federal statutory framework governing AI systems. However, there is a patchwork of strategies, regulations, codes, standards, and directives developed by public-sector departments and agencies for governing their respective uses of these technologies.31 Policy initiatives that cover the public sector tend to come from the Treasury Board Secretariat (TBS), while private-sector initiatives come from the department for Innovation, Science and Economic Development (ISED).

While most AI governance initiatives tend to focus on all AI systems, there is also a dedicated effort in focusing on generative AI systems. For instance, in September 2023, ISED launched the Voluntary Code of Conduct on the Responsible Development and Management of Advanced Generative AI Systems. Intended for the private sector, the Code outlines a series of principles for companies developing and deploying generative AI systems, regarding safety, transparency, fairness, and human oversight. They were initially developed as an interim measure while awaiting formal regulations through the proposed Artificial Intelligence and Data Act (AIDA) part of Bill C-27; however, with the Bill failing to pass prior to the recent prorogation of parliament, the future remains uncertain for private-sector governance of generative AI systems.32 

There is also a Guide on the Use of generative Artificial Intelligence developed by the Treasury Board Secretariat, but that applies only to how government institutions use generative systems. The Guide consists of principles and best practices that public servants should follow to responsibly use generative AI technologies.33 Through the Standards Council of Canada, the federal government is in the process of developing AI and data governance standards.34 The Digital Governance Council has also developed voluntary industry standards for AI.35

Moreover, the federal government’s efforts at regulating social media platforms have not only been hampered by prorogation, but they also do not focus on content moderation practices. In particular, Bill C-63: Online Harms Act, was tabled in February 2024 by the Minister of Justice. It sought to establish a regulatory framework for social media platforms by requiring them to comply with a series of duties and disclosure practices. The Bill sought to regulate seven categories of harm: content that sexually victimizes a child or revictimizes a survivor, intimate content communicated without consent, content used to bully a child, content that induces a child to harm themselves, content that incites hatred, content that incites violence, and content that incites violent extremism or terrorism.36 While important to consider child safety and violent extremism as part of the scope of online harms, other types of harms brought on by misinformation and disinformation need to be considered. Furthermore, the proposed approach would not compel platforms to engage in content moderation practices such as labelling and fact-checking.

In short, when it comes to genAI, the federal government has developed a variety of voluntary governance mechanisms in lieu of legislative statutes. While proposed laws such as Bill C-27 and Bill C-63 had the potential to address the increased use of genAI and its harmful effects, the limitations of the bills, and now the prorogation of parliament, have cut short those efforts.

Compare this to Europe, where there has already been significant action taken to govern the use of AI—specifically the EU’s Artificial Intelligence Act (AIA).

Having come into force in August 2024, the AIA seeks to govern AI systems by establishing sets of responsibilities and obligations to developers based on four categories of risk: 1) unacceptable risk, high risk, limited risk, and minimal risk (see Figure 1 for examples of systems in scope for each risk category). General purpose AI systems (GPAI) are also part of the scope of the Act and have separate requirements and obligations for developers.

As part of the scope of the AIA, companies that make AI systems that can be used to generate deepfakes and synthetic media must ensure that their tools label outputs as generated with AI. This may include techniques such as watermarks, or any other approach prescribed by regulation.37 Additionally, the providers must establish mechanisms that prevent the misuse of their tools, as well as measures to address any harms that may arise from their misuse. In light of the recency of AIA’s adoption, it is hard to assess the effectiveness of these measures—only time will tell. However, the requirements demanded by this Bill represent a significant step towards the labelling of AI-generated content online.

The EU’s Digital Services Act (DSA) seeks to protect fundamental human rights and promote a safe, fair, and open online ecosystem for users. It does so by regulating app stores, social networks, content sharing, online travel and accommodation platforms.38 Compared to the AIA, generative AI systems are not explicitly accounted for in the DSA. For instance, search engines must follow a set of obligations and produce documentation about the mitigation measures that they’re adopting. However, genAI systems may be used as search engines by users (i.e. retrieve information from the internet) which blurs the line between a standalone genAI system and traditional search engines.39 This distinction is difficult to draw because tools like ChatGPT can also be used to generate images, audio and video, instead of generating text that was developed by retrieving information online.40

As a result of the lack of specificity about generative AI in the DSA, regulators in charge of enforcing it will need to develop regulations that address this gap. For instance, regulators have already compelled search engines (Bing and Google) and online platforms (Meta, Snapchat, TikTok, YouTube, X) to provide information about the mitigation measures they are using to address the risks associated with generative AI.41

Experiment: Testing the Impact of AI-Labelling 


This study used a browser extension developed by Who Targets Me in order to simulate the Facebook feed of participants. Upon the completion of a short pre-survey eligibility assessment, participants were prompted to download a browser extension that collects the first 50 posts from participants’ Facebook feeds. The browser extension was only available for desktop computers and only through the Google Chrome browser.

Participants were then randomly assigned to three groups, with one group seeing the AI-generated posts without a label, and the two others seeing the posts each with a different style of label (either a small disclaimer or a full warning screen). Figure 3 shows the different labelling conditions. Participants were asked to scroll through to the bottom and interact with their feed as normal, and were subsequently provided with a post-exercise survey where they were asked questions about the manipulated posts, as well as additional questions regarding their social media use and knowledge of AI.

Figure 2
AI-Generated content that was injected into survey respondents’ Facebook feed.

Post A

Figure 3
Examples of the different AI content labeling methods.

No Label

This survey was conducted online with 2,472 residents of Canada aged 16 and older from September 20, 2024 to November 21, 2024, in English and French. A random sample of panellists from Leger’s research panel was invited to complete the survey, with response quotas set by region, language, age, and gender to ensure the sample reflected Canada’s population. Due to the nature of the survey experiment, respondents were required to have a Facebook account and to be using Google Chrome on a desktop computer.

Margins of error are typically not applicable to surveys conducted using an online panel such as this one. The data were weighted according to census data to ensure that the sample matched Canada’s population according to age, gender and region. The detailed survey methodology, including a list of questions, is available in the appendix.

Survey Findings


In addition to the survey experiment on the impact of various labelling approaches, we also tested Canadian residents’ familiarity with deepfakes and who is most exposed to them. This included both their self-reported familiarity with deepfakes, but also how well they could define deepfakes when asked to. We also benchmarked exposure to deepfakes specifically against exposure to misinformation more generally seen online.

Most people are at least somewhat familiar with deepfakes—only 18 percent said they were not at all familiar. Unsurprisingly, there are huge differences across age groups. Among those 16 to 29, 35 percent are very familiar with deepfakes while among those over 60, that number is only 12 percent. Men of all ages are also more familiar, as are people with a university degree.


Those who describe themselves as more politically centrist are less likely to say they are very familiar with deepfakes—only 18 percent of those who place themselves between 4 and 6 on a scale of 1 to 9 from politically left to right say they are very familiar with deepfakes. This is lower than 26 percent of people on the right who are very familiar, and 30 percent of people on the left.

However, despite claiming to be familiar with deepfakes, Canadian residents describe the concept of deepfakes or synthetic images in a wide variety of ways.

Most respondents who said they were familiar with deepfakes could come up with a definition that was at least partially correct - touching on at least one element of what constitutes a deepfake. A total of 73 percent were able to at least in part correctly describe deepfakes—although that includes three percent who gave a very specific example of deepfakes rather than defining them. However, 13 percent of respondents who thought they were familiar with the concept actually ended up describing misinformation generally. Out of all respondents—including those who said they don’t know what a deepfake is, 45 percent of Canadian residents were able to define deepfakes correctly.

In defining deepfakes, 32 percent of respondents mentioned anything related to artificial intelligence—either as a core part of deepfakes, or simply one avenue to create them—while many others simply described any media that had been altered or manufactured. While this suggests a close tie between deepfakes and AI technologies, it is clear that not all Canadians immediately make that connection.

Canadian residents are exposed to a significant amount of fake content online—ranging from misinformation to deepfakes across a huge variety of topic areas. Our survey suggests that Canadian residents reported seeing what they believed to be deepfakes less frequently than what they believed to be fake news, but about as often as they see fake news they believe to be true. One in five Canadian residents say they see synthetic media or deepfaked content online a few times a day, 47 percent of Canadians see deepfakes on at least a weekly basis, and 70 percent report seeing it at least a few times a month. Comparatively, 17 percent of Canadian residents see fake information about the news or current events they believed to be true at the time, but later find out was fake multiple times a day online and 67 percent see this type of content at least a few times a month.

Exposure to deepfakes is largely consistent with age—we find that just under half of Canadian residents across age groups report seeing synthetic media or deepfakes multiple times a week. However, older residents (60 and above) are more likely to be unsure how often they see deepfakes, potentially signalling that older residents have a harder time identifying whether or not content they see online is synthetic or real.

In general, those who use the internet more often are both more familiar with and more exposed to deepfakes. Only 61 percent of all respondents said they were at least somewhat familiar with the concept of deepfakes while 72 percent of those who use YouTube were familiar and 80 percent of those who use X were familiar. Exposure to deepfakes is highest on TikTok and Instagram, with 78 percent of TikTok users and 76 percent of Instagram users saying they see deepfakes at least a few times a month.

Label Testing Experiment


All respondents to the survey were shown both injected posts in their feed; however, not all respondents were able to recall seeing them.
Because we asked respondents to browse their re-constituted Facebook feed as they normally would, many spent limited time on the page, getting through the content quickly without closely engaging with every post.

In general, men and those on the right of the political spectrum were more likely to have seen Post A, while that relationship disappears for Post B (see Figure 11). Regarding Post B, women were slightly more likely to remember seeing the post, with no significant differences across political ideologies.

Among those who did recall seeing the injected AI-generated content, there were differences in both who believed it and who was likely to share it. Across both posts, those on the right of the political spectrum were more likely to say the information was accurate and also that they would share the content—however, this likely due to the nature of the content rather than anything specific related to the content being AI-generated.


Across both posts, older respondents were more likely to say the information presented was accurate, however, younger respondents were more likely to say they would share it. This is particularly pronounced for Post A, which dealt with the House of Commons, where the youngest group was twice as likely as the oldest group to say they would share the content, even though they were significantly less likely to say the content was accurate (18 points less likely).

Each respondent was randomly placed into one of three treatment groups: one with no label, one with a small label at the top of the post, and one with a full label blocking the post until a user decided to remove it. We found that labelling does matter, but only to limit the share of respondents who actually saw the posts, with no significant impact on how users respond to the posts if they do see them.

Across both posts, users were less likely to see the post if it was blocked by the full label, while the small disclaimer text had no significant impact on visibility of the posts. For Post A, this was a modest decrease in visibility from 46 percent with no label to 39 percent saying they saw the post with the label (see Figure 12). The effect was larger for Post B with only 43 percent having seen it with the full label blocking it, compared to 67 percent of respondents with no labelling saying they saw it. In both cases, the group that saw only a small disclaimer label were closer to the control group with no label than the full labelling group—although the difference is much larger for Post B.

However, when it comes to both credibility and the likelihood of sharing the content, the labelling had no significant impact. Those who saw either labelling condition were equally likely to say they felt both posts were accurate or inaccurate, and similarly, they were equally likely to say they would share the content.

The biggest differences came when respondents were asked to assess the effectiveness of the labelling method they were shown at informing them the content was AI-generated. We asked this of all respondents - even those who had not been shown any label on the AI-generated content. Those who saw either the small labels or no labels at all had a similar reaction - in both cases 39% of respondents felt the label was effective while 37% said that the no label condition was ineffective and 39% said the small disclaimer was ineffective (see Figure 15). 

However, for the full label condition, 60% of respondents said that the labelling was effective while only 21% said it was ineffective. This includes 36% who said it was definitely effective - far higher than the share for either other condition tested. So while respondents reacted exactly the same to no labelling and the small disclaimers, they felt that the screen was significantly more effective at identifying content as AI-generated.

Taken together, we see that the small disclaimer identifying that content is AI-generated is functionally no different from leaving content entirely unlabelled. Users find it no more helpful than no label, and it leaves them equally trusting and likely to share the AI-generated content. Adding a screen on top of AI content, however, leaves Canadian residents less likely to see the AI-generated content, and more likely to feel it was effectively labelled.

Conclusion


Canadians want to know when they are looking at AI-generated content, but the current approach by platforms has been haphazard—often implementing the lowest effort option without appropriately considering how effective that solution would be. YouTube, TikTok, and Meta platforms already allow users to flag their content as AI-generated and then leave a small label to indicate that to readers. Others like BlueSky and X (formerly Twitter) have no labels for AI-generated content. However, no platforms to date offer the full screening method for labelling AI that we tested in this project.

This research suggests that users are no more satisfied with the small label approach than they would be with no labelling approach at all. These small labels do not limit the reach of AI-generated content, they do not reduce the likelihood of this content being believed, and they don’t reduce the likelihood of the content being shared.

If platforms want to implement an AI content labelling regime that would actually limit the reach of AI-generated content, or leave Canadian residents feeling that AI-generated content was being labelled at all, a more effective approach would be to cover up the AI-generated content until users explicitly click to view it. 

This leaves open the question of how platforms should best identify AI-generated content on their platforms. While some have started implementing schemes to identify it automatically, many platforms still rely on users self-declaring that the content they are posting is AI-generated. This has little impact when those uploading content are doing so with the intent of deceiving viewers.

When content is uploaded to deliberately deceive viewers, self-regulation by platforms proves largely ineffective. Despite surveys showing that Canadians frequently encounter synthetic and deepfake content on social media—and many Canadian residents saying they want to be informed about its presence—platforms have chosen not to implement transparent labelling that would help users identify such content.

This situation is compounded by the absence of a statutory governance regime in Canada. If this country implemented statues similar to the ones in the European Union, that would mandate clearer disclosure practices. Without legal requirements compelling platforms to adopt more effective transparency measures, their current self-governance practices continue to fall short of addressing the concerns of Canadian users. This regulatory gap underscores the need for policy intervention to ensure digital platforms better protect consumers and promote informed decision making.

1

“AI Language Models: Technological, Socio-Economic and Policy Considerations,” OECD Digital Economy Papers, Vol. 352, April 13, 2023, https://doi.org/10.1787/13d38f92-en.

2

Arvind Narayanan and Sayash Kapoor,  AI Snake Oil : What Artificial Intelligence Can Do, What It Can’t, and How to Tell the Difference, (Princeton: Princeton University Press, 2024). 

3

Muhammad Usman Hadi, Qasem Al Tashi, Rizwan Qureshi, Abbas Shah, Amgad Muneer, Muhammad Irfan, Anas Zafar, et al., “A Survey on Large Language Models: Applications, Challenges, Limitations, and Practical Usage,” Preprint (2023), https://doi.org/10.36227/techrxiv.23589741.v1.

4

André Côté and Joe Masoodi, Understanding and Responding to Disinformation Targeting Canadian Companies, The Dais, May 30, 2024, https://dais.ca/reports/understanding-and-responding-to-disinformation-targeting-canadian-companies/.

5

Memes are humorous images, videos, texts, and audio used to reference comical or relatable experiences through social media platforms. Meme generators are digital tools that allow people to create custom memes through their pre-designed templates of popular meme formats.

6

Sensity AI, “The State of Deepfakes 2024,” https://5865987.fs1.hubspotusercontent-na1.net/hubfs/5865987/SODF%202024.pdf. 

7

Eli Lucherini, Matthew Sun, Amy Winecoff and Arvind Narayanan, “T-RECS: A Simulation Tool to Study the Societal Impact of Recommender Systems,” arXiv, July 28, 2021, http://arxiv.org/abs/2107.08959

8

Dietmar Jannach, and Gediminas Adomavicius, “Recommendations with a Purpose,” In Proceedings of the 10th ACM Conference on Recommender Systems, RecSys ’16, New York, NY: Association for Computing Machinery, 2016: 7–10, https://doi.org/10.1145/2959100.2959186.

9

Cristiano Lima-Strong, “A Whistleblower’s Power: Key Takeaways from the Facebook Papers,” Washington Post, October 26, 2021, https://www.washingtonpost.com/technology/2021/10/25/what-are-the-facebook-papers/; Hugues Sampasa-Kanyinga and Hayley A. Hamilton, “Use of Social Networking Sites and Risk of Cyberbullying Victimization: A Population-level Study of Adolescents,” Cyberpsychology, Behavior, and Social Networking 18, no. 12 (2015): 704-710, https://doi.org/10.1089/cyber.2015.0145

10

Hossein A. Rahmani, Mohammadmehdi Naghiaei, and Yashar Deldjoo, “A Personalized Framework for Consumer and Producer Group Fairness Optimization in Recommender Systems,” ACM Transactions on Recommender Systems 2, no. 3 (June 5, 2024): 19:1-19:24, https://dl.acm.org/doi/10.1145/3651167; Jonathan Stray, Alon Halevy, Parisa Assar, Dylan Hadfield-Menell, Craig Boutilier, Amar Ashar, Chloe Bakalar, et al. “Building Human Values into Recommender Systems: An Interdisciplinary Synthesis,” ACM Transactions on Recommender Systems 2, no. 3 (June 5, 2024): 20:1-20:57, https://dl.acm.org/doi/10.1145/3632297.  

11

Saifuddin Ahmed, “Fooled by the Fakes: Cognitive Differences in Perceived Claim Accuracy and Sharing Intention of Non-political Deepfakes,” Personality and Individual Differences, 182 (2021), https://doi.org/10.1016/j.paid.2021.111074.

12

Benjamin Toff and Felix M. Simon, 2023. “‘Or They Could Just Not Use It?’: The Paradox of AI Disclosure for Audience Trust in News,” SocArXiv Papers, December 2023, https://doi.org/10.31235/osf.io/mdvak.

13

K.J. Kevin Feng, Nick Ritchie, Pia Blumenthal, Andy Parsons and Amy X. Zhang, “Examining the Impact of Provenance-Enabled Media on Trust and Accuracy Perceptions,” Proceedings of the ACM on Human-Computer Interaction 7 (CSCW2 2023): 270:1-42, https://doi.org/10.1145/3610061.

14

Tarleton Gillespie, Custodians of the Internet: Platforms, Content Moderation, and the Hidden Decisions That Shape Social Media, (New Haven: Yale University Press, 2018); Mary L. Gray and Siddharth Suri, Ghost Work: How to Stop Silicon Valley from Building a New Global Underclass, (Boston: Houghton Mifflin Harcourt, 2019).

15

Tarleton Gillespie, Custodians of the Internet: Platforms, Content Moderation, and the Hidden Decisions That Shape Social Media. (New Haven: Yale University Press, 2018).

16

Orestis Papakyriakopoulos, Christelle Tessono, Arvind Narayanan, and Mihir Kshirsagar, “How Algorithms Shape the Distribution of Political Advertising: Case Studies of Facebook, Google, and TikTok,” In Proceedings of the 2022 AAAI/ACM Conference on AI, Ethics, and Society (2022): 532–46. 

17

Rod McGuirk, “X Corp. Has Slashed 30% of Trust and Safety Staff, an Australian Online Safety Watchdog Says,” AP News, January 10, 2024, https://apnews.com/article/x-corp-musk-australia-staff-safety-bc4772369cab1fe8dd975132fd8d61ed; Lauren Feiner, “The Fallout of Meta’s Content Moderation Overhaul,” The Verge, February 27, 2025, https://www.theverge.com/24339131/meta-content-moderation-fact-check-zuckerberg-texas.

18

Andrew Lewis, Patrick Vu, Raymond Duch and Areeq Chowdhury, “Do Content Warnings Help People Spot a Deepfake? Evidence from Two Experiments,” 2022,  https://doi.org/10.31219/osf.io/v4bf6.

19

Serena Iacobucci, Roberta De Cicco, Francesca Michetti, Riccardo Palumbo, and Stefano Pagliaro, “Deepfakes Unmasked: The Effects of Information Priming and Bullshit Receptivity on Deepfake Recognition and Sharing Intention,” Cyberpsychology, Behavior, and Social Networking 24, vol 3 (2021): 194–202, https://doi.org/10.1089/cyber.2020.0149

20

Ziv Epstein, Mengying Cathy Fang, Antonio Alonso Arechar and David Rand, “What Label Should Be Applied to Content Produced by Generative AI?” PsyArXiv (2023), https://doi.org/10.31234/osf.io/v4mfz.

21

Anatoliy Gruzd,  Philip Mai and Felipe B. Soares, “To Share or Not to Share: Randomized Controlled Study of Misinformation Warning Labels on Social Media,” In Disinformation in Open Online Media, edited by Mike Preuss, Agata Leszkiewicz, Jean-Christopher Boucher, Ofer Fridman, and Lucas Stampe, 46–69. (Springer, Cham, 2024), https://doi.org/10.1007/978-3-031-71210-4_4.

22

Hang Lu and Shupei Yuan, “‘I Know It’s a Deepfake’: The Role of AI Disclaimers and Comprehension in the Processing of Deepfake Parodies,” Journal of Communication 74 no. 5 (2024): 359–73, https://doi.org/10.1093/joc/jqae022.

23

“Membership – C2PA,” n.d., Coalition for Content Provenance and Authenticity, Accessed January 16, 2025, https://c2pa.org/membership/.

24

“New Labels for Disclosing AI-generated Content,” TikTok, September 19, 2023, https://newsroom.tiktok.com/en-us/new-labels-for-disclosing-ai-generated-content; “About AI-generated Content,” TikTok, n.d., Accessed March 7, 2025, https://support.tiktok.com/en/using-tiktok/creating-videos/ai-generated-content; “Partnering With Our Industry to Advance AI Transparency and Literacy,” TikTok, May 9, 2024, https://newsroom.tiktok.com/en-ca/partnering-with-our-industry-to-advance-ai-transparency-and-literacy-ca. 

25

“New Labels for Disclosing AI-generated Content,” TikTok, September 19, 2023, https://newsroom.tiktok.com/en-us/new-labels-for-disclosing-ai-generated-content; “About AI-generated Content,” TikTok, n.d., Accessed March 7, 2025

26

 “Platform Use Guidelines,” X Help Center, n.d., Accessed March 7, 2025, https://help.x.com/en/rules-and-policies#platform-use-guidelines

27

“How to Identify AI Content on Meta Products,” Meta Help Center, n.d., Accessed January 24, 2025. https://www.meta.com/help/artificial-intelligence/how-ai-generated-content-is-identified-and-labeled-on-meta/.

28

Jennifer Flannery O’Connor and Emily Moxley, “Our Approach to Responsible AI Innovation,” YouTube Official Blog, November 14, 2023,  https://blog.youtube/inside-youtube/our-approach-to-responsible-ai-innovation/.

29

“Bluesky’s Stackable Approach to Moderation,” Bluesky, March 12, 2024, https://bsky.social/about/blog/03-12-2024-stackable-moderation.

30

Bill C-27, An Act to enact the Consumer Privacy Protection Act,the Personal Information and Data Protection Tribunal Act and the Artificial Intelligence and Data Act and to make consequential and related amendments to other Acts, 1st Sess, 44th Parl, 2022. 

31

Blair Attard-Frost, Ana Brandusescu and Kelly Lyons, “The Governance of Artificial Intelligence in Canada: Findings and Opportunities from a Review of 84 AI Governance Initiatives,” Government Information Quarterly 41, no. 2 (2024): 101929, https://doi.org/10.1016/j.giq.2024.101929.

32

“Voluntary Code of Conduct on the Responsible Development and Management of Advanced Generative AI Systems,” Innovation, Science and Economic Development Canada, September 2024, https://ised-isde.canada.ca/site/ised/en/voluntary-code-conduct-responsible-development-and-management-advanced-generative-ai-systems

33

“Guide on the Use of Generative Artificial Intelligence,” Treasury Board Secretariat of Canada, October 15, 2024, https://www.canada.ca/en/government/system/digital-government/digital-government-innovations/responsible-use-ai/guide-use-generative-ai.html. 

34

“AI and Data Governance Standardization Collaborative,” Standards Council of Canada, May 3, 2023, https://scc-ccn.ca/areas-work/digital-technology/ai-and-data-governance-standardization-collaborative.

35

“Standards in Automated Decision Systems (AI),” n.d., Digital Governance Council (blog), Accessed January 24, 2025, https://dgc-cgn.org/standards/find-a-standard/standards-in-automated-decision-systems-ai/.

36

Bill C-63, An Act to enact the Online Harms Act, to amend the Criminal Code, the Canadian Human Rights Act and An Act respecting the mandatory reporting of Internet child pornography by persons who provide an Internet service and to make consequential and related amendments to other Acts,” 1st Sess, 44th Parl, 2024, Section 2, https://www.parl.ca/DocumentViewer/en/44-1/bill/C-63/first-reading.

37

Article 134 in European Parliament and Council, Regulation (EU) 2024/1689 of 13 June 2024 Laying Down Harmonised Rules on Artificial Intelligence and Amending Various Regulations and Directives (Artificial Intelligence Act), Official Journal of the European Union L-1689, 2024, https://eur-lex.europa.eu/legal-content/EN/TXT/PDF/?uri=OJ:L_202401689.

38

“The Digital Services Act,” European Commission, Accessed January 17, 2025, https://commission.europa.eu/strategy-and-policy/priorities-2019-2024/europe-fit-digital-age/digital-services-act_en

39

Jordi Calvet-Bademunt and Joan Barata, “The Digital Services Act Meets the AI Act: Bridging Platform and AI Governance | TechPolicy.Press,” Tech Policy.Press., May 29, 2024, https://techpolicy.press/the-digital-services-act-meets-the-ai-act-bridging-platform-and-ai-governance.

40

Florence G’sell, “Regulating under Uncertainty: Governance Options for Generative AI,” 2024,  https://doi.org/10.2139/ssrn.4918704.

41

“Commission Sends Requests for Information on Generative AI Risks to 6 Very Large Online Platforms and 2 Very Large Online Search Engines under the Digital Services Act | Shaping Europe’s Digital Future” (press release), European Commission, March 14, 2024, https://digital-strategy.ec.europa.eu/en/news/commission-sends-requests-information-generative-ai-risks-6-very-large-online-platforms-and-2-very.