IGF 2025 WS #281 Can I trust this? Provenance in the age of generative AI.

    Organizer 1: Civil Society, Asia-Pacific Group
    Organizer 2: Private Sector, Western European and Others Group (WEOG)
    Organizer 3: Private Sector, Latin American and Caribbean Group (GRULAC)
    Speaker 1: Farzaneh Badii, Civil Society, Asia-Pacific Group
    Speaker 2: Paree Zarolia, Technical Community, Western European and Others Group (WEOG)
    Speaker 3: Gisella Lomax, Intergovernmental Organization, Intergovernmental Organization
    Speaker 4: Sanderson Zeve, Civil Society, Western European and Others Group (WEOG)
    Speaker 5: Len Manriquez, Private Sector, Asia-Pacific Group
    Format
    Roundtable
    Duration (minutes): 90
    Format description: Forget panels that bore you to tears with PowerPoint presentations. Our 90-minute roundtable is your opportunity to actively shape the future of information literacy and AI provenance. Unlike typical panels, this format ensures your voice—whether as an end user, policymaker, technologist, or advocate—is heard and contributes to real solutions. We'll begin with essential insights and real-world challenges, illustrated with interactive exercises. We’ll then collaborate on technical, policy, and ethical strategies. You'll leave with practical frameworks and clear next steps, empowering you to find information to evaluate content trustworthiness, or implement frameworks for providing provenance information in AI governance. Join us to transform expertise into actionable solutions.
    Policy Question(s)
    How can policymakers strike a balance between promoting transparency in content labeling without inadvertently increasing trust in unlabeled content? What strategies can be implemented to ensure that assertive provenance techniques, such as watermarking, are widely adopted by various stakeholders, including content creators, platforms, and developers, without stifling innovation in AI and creative industries? How can governments and educational institutions collaborate with industry leaders and civil society to improve information literacy skills and empower users to critically evaluate online content, especially AI-generated content?
    What will participants gain from attending this session? Participants will learn about approaches to information literacy that focus on making more contextual and provenance information about content available to users. This includes assertive provenance tools such as content watermarking standards, and inferred context tools, such as the About This Image and About This Result features on Google Search. Our experts will share perspectives on the technical, policy, and ethical dimensions of provenance, offering insights into its role in combating misinformation, ensuring transparency, and fostering trust in AI systems. Attendees will discuss the opportunities and challenges for implementing different provenance tools across different platforms, regulatory and business environments. The session will provide practical takeaways, including emerging policy frameworks, industry best practices, and potential governance models to enhance AI accountability and empower users to make decisions about content trustworthiness. Participants will leave with a clearer roadmap for integrating holistic provenance approaches into AI governance discussions at national and global levels.
    Description:

    As tools powered by generative AI become more accessible and widespread, debate around the trustworthiness of content – especially synthetic images, video, and voice – have become more acute. While concerns around misinformation and disinformation are not new, generative AI capabilities do bring a new dimension to the conversation, which will require careful consideration and discussion. Google’s Paree Zarolia will present a paper titled “Determining trustworthiness through context and provenance.” This paper focuses on situating the issues of synthetic content within the broader context of information literacy and information quality, for which a deep and rich evidence-base already exists. The question of “Is this AI-generated?” is not equivalent to “Is this trustworthy?” Though these two questions can overlap, they do not always – and additional contextual information is often needed to make an accurate assessment of a piece of content’s trustworthiness. This paper also outlines the benefits and drawbacks of both assertive and inferred provenance, both of which will be required as part of a holistic approach to empower users to make sound decisions on content they encounter in the information ecosystem. Assertive provenance, which features heavily in current policy debates, focuses on providing labels, watermarks, and/or metadata that can indicate whether a piece of content has been AI-generated or not. Inferred provenance meanwhile focuses on empowering users with contextual information that can help them determine where a piece of content came from, what claims are being made about it, and who is responsible for those claims. The goal of our session is to show that by taking a holistic approach to addressing provenance in the age of generative AI, we can help mitigate the risks associated with fake content, give users tools to navigate the digital landscape, preserve trust in digital media, and uphold the integrity of information ecosystems.
    Expected Outcomes
    It’s difficult to know where the conversation with the audience will take us but we hope some of the following may be outcomes from the session: *A clearer understanding of how governments, businesses, civil society view the need for provenance with generative AI. *Recommendations for educating the public on assertive and inferred provenance tools and how to evaluate content trustworthiness. *Identification of concrete opportunities for cross-stakeholder partnership to increase availability of provenance information and information literacy skills.
    Hybrid Format: Using Zoom will allow both onsite and online participants to see and hear each other. We will ask all participants, both in person and remote to be logged in so we can manage the question queue in a neutral manner. Our onsite and online moderators will be in constant communication to ensure that we can facilitate questions and comments from both onsite and online participants. We will urge our speakers to use clear and concise language, avoid technical jargon, and provide context for all information discussed during the session to ensure that both onsite and online participants can follow along and understand the content. Finally, we plan to reserve 40 minutes for audience interaction and will explore the use of the quick show of hands feature in Zoom to ask questions and get feedback from both onsite and online participants in real-time.