- February 12, 2024
- AI in Business
On Tuesday, Meta announced a collaboration with other technology companies to develop standards that will enable advanced detection and labeling of AI-generated images shared by a large user base.
The Silicon Valley-based tech giant expects to be ready to roll out a system on its platforms — Facebook, Instagram, and Threads — to accurately identify and tag AI-generated visuals within months. With upcoming elections in various countries that account for half of the world’s population, platforms like Meta feel an urgency to monitor AI-generated content due to concerns about the increased spread of misinformation by malicious actors. “This technology needs further development and it won’t cover everything, but it’s a start,” Nick Clegg, the company’s head of international affairs, told AFP in an interview.
Since December, Meta has marked images captured by its AI tools with visible and hidden indicators. However, Meta is looking to expand these efforts by partnering with outside companies to increase user awareness of such content, Clegg shared. Meta mentioned in one of their blog updates that they are looking to establish universal technical standards with industry peers that would signal when AI has had a hand in creating a piece of content. These efforts will involve engagement with organizations that Meta has previously worked with on AI recommendations. These partners include industry leaders such as OpenAI, Google, Microsoft, and Midjourney.
But as Clegg pointed out, while there is some progress in embedding “signals” into AI-generated images, the practice of tagging AI-generated audio or video has not progressed as quickly in the industry. While acknowledging that invisible tagging won’t completely eradicate the threat of fraudulent images, Clegg believes it will significantly reduce the distribution of such content as far as current technology allows.
Meanwhile, Meta encourages users to be skeptical of online content, assessing the reliability of sources and scrutinizing details that may seem contrived. In particular, high-ranking individuals and women have been affected by the realistic but false manipulations known as “deep fakes”. A notable case involved fake nude images of mega-pop star Taylor Swift that went viral on the platform formerly known as Twitter.
The development of AI tools capable of generating content has raised concerns about possible abuse, such as using ChatGPT for political upheaval through disinformation or duplicate AI. Just last month, OpenAI announced a ban on the use of its technology by political figures or organizations. Meta insists that advertisers be transparent about any AI involvement in the creation or editing of both visual and audio content in political ads.
- Innovative Ai Technology Predicts Data Trends To Improve Storage Efficiency
- OpenAI Introduces Sora, An Ai Tools That Turns Text Prompts Into Instant Video
- The Meta Calls For Standardized Labeling Of AI-Generated Visual Content
- Reinforcement Learning with OpenAI Gym
- Amazon Introduces Rufus, a New AI-Powered Shopping Assistant in Its Mobile App
- Scientists Create Algorithm to Analyze Users’ Eye Movements on Screens
- Using AI to Enhance Creative Expression in Art Therapy
- Theano Overview and Guide
- Anthropology Researchers Have Discovered That AI Models Can Be Taught To Cheat
- Google Introduces “Circle to Search” – Revolutionary On-The-Go Search Feature for Android Devices
Get regular updates on data science, artificial intelligence, machine