Current:Home > MarketsAI expert says Princess Kate photo scandal shows our "sense of shared reality" being eroded -Trailblazer Capital Learning
AI expert says Princess Kate photo scandal shows our "sense of shared reality" being eroded
View
Date:2025-04-27 16:01:09
London — The European Parliament passed the world's first comprehensive law regulating the use of artificial intelligence on Wednesday, as controversy swirled around an edited photo of Catherine, the Princess of Wales, that experts say illustrates how even the awareness of new AI technologies is affecting society.
"The reaction to this image, if it were released before, pre-the big AI boom we've seen over the last few years, probably would be: 'This is a really bad job with editing or Photoshop,'" Henry Ajder, an expert on AI and deepfakes, told CBS News. "But because of the conversation about Kate Middleton being absent in the public eye and the kind of conspiratorial thinking that that's encouraged, when that combines with this new, broader awareness of AI-generated images… the conversation is very, very different."
Princess Kate, as she's most often known, admitted to "editing" the photo of herself and her three children that was posted to her official social media accounts on Sunday. Neither she nor Kensington Palace provided any details of what she had altered on the photo, but one royal watcher told CBS News it could have been a composite image created from a number of photographs.
Ajder said AI technology, and the rapid increase in public awareness of what it can do, means people's "sense of shared reality, I think, is being eroded further or more quickly than it was before."
Countering this, he said, will require work on the part of companies and individuals.
What's in the EU's new AI Act?
The European Union's new AI Act takes a risk-based approach to the technology. For lower risk AI systems such as spam filters, companies can choose to follow voluntary codes of conduct.
For technologies considered higher risk, where AI is involved in electricity networks or medical devices, for instance, there will be tougher requirements under the new law. Some uses of AI, such as police scanning people's faces using AI technology while they're in public places, will be outright banned apart from in exceptional circumstances.
The EU says the law, which is expected to come into effect by early summer, "will guarantee the safety and fundamental rights of people and businesses when it comes to AI."
Losing "our trust in content"?
Millions of people view dozens of images every day on their smartphones and other devices. Especially on small screens, it can be very difficult to detect inconsistencies that might indicate tampering or the use of AI, if it's possible to detect them at all.
"It shows our vulnerability towards content and towards how we make up our realities," Ramak Molavi Vasse'i, a digital rights lawyer and senior researcher at the Mozilla Foundation, told CBS News. "If we cannot trust what we see, this is really bad. Not only do we have, already, a decrease in trust in institutions. We have a decrease in trust and media, we have a decrease in trust, even for big tech… and for politicians. So this part is really bad for democracies and can be destabilizing."
Vasse'i co-authored a recent report looking at the effectiveness of different methods of marking and detecting whether a piece of content has been generated using AI. She said there were a number of possible solutions, including educating consumers and technologists and watermarking and labeling images, but none of them are perfect.
"I fear that the speed in which the development happens is too quick. We cannot grasp and really govern and control the technology that is kind of, not creating the problem in the first place, but accelerating the speed and distributing the problem," Vasse'i told CBS News.
"I think that we have to rethink the whole informational ecosystem that we have," she said. "Societies are built on trust on a private level, on a democratic level. We need to recreate our trust in content."
How can I know what I'm looking at is real?
Ajder said that, beyond the wider aim of working toward ways to bake transparency around AI into our technologies and information ecosystems, it's difficult on the individual level to tell whether AI has been used to change or create a piece of media.
That, he said, makes it vitally important for media consumers to identify sources that have clear quality standards.
"In this landscape where there is increasing distrust and dismissal of this kind of legacy media, this is a time when actually traditional media is your friend, or at least it is more likely to be your friend than getting your news from random people tweeting out stuff or, you know, Tiktok videos where you've got some guy in his bedroom giving you analysis of why this video is fake," Adjer said. "This is where trained, rigorous investigative journalism will be better resourced, and it will be more reliable in general."
He said tips about how to identify AI in imagery, such as watching to see how many times someone blinks in a video, can quickly become outdated as technologies are developing at lightning speed.
His advice: "Try to recognize the limitations of your own knowledge and your own ability. I think some humility around information is important in general right now."
- In:
- Deepfake
- Artificial Intelligence
- AI
- Catherine Princess of Wales
Haley Ott is the CBS News Digital international reporter, based in the CBS News London bureau.
Twitter InstagramveryGood! (35449)
Related
- Brianna LaPaglia Reveals The Meaning Behind Her "Chickenfry" Nickname
- Zendaya Aces With 4th Head-Turning Look for Met Gala 2024 After-Party
- How to Grow Long, Strong Natural Nails At Home, According To A Nail Artist
- 'Dreams do come true': Man wins $837K lottery prize after sister dreams he'd find gold
- Highlights from Trump’s interview with Time magazine
- White coated candy shipped nationwide recalled over salmonella contamination concerns
- Boeing’s first astronaut launch is off until late next week to replace a bad rocket valve
- Aaron Hernandez's fiancée responds to jokes made about late NFL player at Tom Brady's roast: Such a cruel world
- Taylor Swift makes surprise visit to Kansas City children’s hospital
- Most FTX customers to get all their money back less than 2 years after catastrophic crypto collapse
Ranking
- Taylor Swift makes surprise visit to Kansas City children’s hospital
- Mother of Australian surfers killed in Mexico gives moving tribute to sons at a beach in San Diego
- Bucks' Patrick Beverley: 'I was absolutely wrong' for throwing basketball at Pacers fans
- Keep Up With Kendall Jenner's 2 Jaw-Dropping Met Gala After-Party Looks
- Nevada attorney general revives 2020 fake electors case
- Future of MLB’s Tampa Bay Rays to come into focus with key meetings on $1.3B stadium project
- Connecticut House passes plan to spend remaining COVID funds, forgoing changes to state budget
- Watch live: USA TODAY discusses highlights from May 7 Apple event, 'Let Loose'
Recommendation
Intel's stock did something it hasn't done since 2022
Taylor Swift bill is signed into Minnesota law, boosting protections for online ticket buyers
Macklemore defends college protesters in pro-Palestine song, slams Biden: 'I'm not voting for you'
Chicago Tribune, other major newspapers accuse artificial intelligence companies of stealing content
A South Texas lawmaker’s 15
Police investigating shooting outside Drake’s mansion that left security guard wounded
Americans are reluctantly spending $500 a year tipping, a new study says.
Cardi B Unveils the Unbelievable Dress She Almost Wore to the 2024 Met Gala