An internet protocol called C2PA adds a “nutrition label” to images, video, and audio.

since C2PA relies on creators to opt in, the protocol doesn’t really address the problem of bad actors using AI-generated content. And it’s not yet clear just how helpful the provision of metadata will be when it comes to media fluency of the public. Provenance labels do not necessarily mention whether the content is true or accurate.

Interesting approach, but I can’t help but feel the actual utility is fairly limited. For example, I could see it being useful for large corporate creative studios that have contractual / union agreements that govern AI content usage.

If they’re using enterprise tools that build in C2PA, it’d give them a metadata audit trail showing exactly when and where AI was used.

That’s completely useless in the context where AI content flagging is most useful though. As the quote says, this provenance data is applied at the point of creation, and in a world where there are open source branches of generation models, there’s no way to ensure provenance tagging is built in.

This technology is most needed to combat AI powered misinformation campaigns, when that is the use case this is least able to address.

Create a post

A nice place to discuss rumors, happenings, innovations, and challenges in the technology sphere. We also welcome discussions on the intersections of technology and society. If it’s technological news or discussion of technology, it probably belongs here.

Remember the overriding ethos on Beehaw: Be(e) Nice. Each user you encounter here is a person, and should be treated with kindness (even if they’re wrong, or use a Linux distro you don’t like). Personal attacks will not be tolerated.

Subcommunities on Beehaw:


This community’s icon was made by Aaron Schneider, under the CC-BY-NC-SA 4.0 license.

  • 1 user online
  • 64 users / day
  • 174 users / week
  • 621 users / month
  • 2.31K users / 6 months
  • 1 subscriber
  • 3.28K Posts
  • 67K Comments
  • Modlog