This article has been reviewed according to Science X's editorial process and policies. Editors have highlighted the following attributes while ensuring the content's credibility:

fact-checked

trusted source

written by researcher(s)

proofread

TV network was slammed for 'AI editing' a Victorian MP's dress. How can news media use AI responsibly?

adobe Photoshop
Credit: Unsplash/CC0 Public Domain

Earlier this week, Channel Nine published an altered image of Victorian MP Georgie Purcell that showed her in a midriff-exposing tank top. The outfit was actually a dress.

Purcell chastised the channel for the and accused it of being sexist. Nine apologized for the edit and blamed it on an (AI) tool in Adobe Photoshop.

Generative AI has become increasingly prevalent over the past six months, as popular image editing and design tools like Photoshop and Canva have started integrating AI features into their programs.

But what are they capable of, exactly? Can they be blamed for doctored images? As these tools become more widespread, learning more about them and their dangers—alongside opportunities—is increasingly important.

What happened with the photo of Purcell?

Typically, making AI-generated or AI-augmented images involves "prompting"—using text commands to describe what you want to see or edit.

But late last year, Photoshop unveiled a new feature, generative fill. Among its options is an "expand" tool that can add content to images, even without text prompts.

For example, to expand an image beyond its original borders, a user can simply extend the canvas and Photoshop will "imagine" content that could go beyond the frame. This ability is powered by Firefly, Adobe's own generative AI tool.

Nine resized the image to better fit its television composition but, in doing so, also generated new parts of the image that weren't there originally.

The source material—and if it's cropped—are of critical importance here.

In the above example where the frame of the photo stops around Purcell's hips, Photoshop just extends the dress as might be expected. But if you use generative expand with a more tightly cropped or composed photo, Photoshop has to "imagine" more of what is going on in the image, with variable results.

Is it legal to alter someone's image like this? It's ultimately up to the courts to decide. It depends on the jurisdiction and, among other aspects, the risk of reputational harm. If a party can argue that publication of an altered image has caused or could cause them "serious harm," they might have a defamation case.

How else is generative AI being used?

Generative fill is just one way news organizations are using AI. Some are also using it to make or publish images, including photorealistic ones, depicting current events. An example of this is the ongoing Israel-Hamas conflict.

Others use it in place of stock photography or to create illustrations for hard-to-visualize topics, like AI itself.

Many adhere to institutional or industry-wide codes of conduct, such as the Journalist Code of Ethics from the Media, Entertainment & Arts Alliance of Australia. This states journalists should "present pictures and sound which are true and accurate" and disclose "any manipulation likely to mislead."

Some outlets do not use AI-generated or augmented images at all, or only when reporting on such images if they go viral.

Newsrooms can also benefit from generative AI tools. An example includes uploading a spreadsheet to a service like ChatGPT-4 and receiving suggestions on how to visualize the data. Or using it to help create a three-dimensional model that illustrates how a process works or how an event unfolded.

What safeguards should media have for responsible generative AI use?

I've spent the last year interviewing photo editors and people in related roles about how they use generative AI and what policies they have in place to do so safely.

I've learned that some bar their staff from using AI to generate any content. Others allow it only for non-realistic illustrations, such as using AI to create a bitcoin symbol or illustrate a story about finance.

News outlets, according to editors I spoke to, want to be transparent with their audiences about the content they create and how it is edited.

In 2019, Adobe started the Content Authenticity Initiative, which now includes major media organizations, image libraries and multimedia companies. This has led to the rollout of content credentials, a digital history of what equipment was used to make an image and what edits have been done to it.

This has been touted as a way to be more transparent with AI-generated or augmented content. But content credentials are not widely used yet. Besides, audiences shouldn't outsource their critical thinking to a third party.

In addition to transparency, news editors I spoke to were sensitive to AI potentially displacing human labor. Many outlets strive to use only AI generators that have been trained with proprietary content. This is because of the ongoing cases in jurisdictions around the world over AI training data and whether resulting generations breach copyright.

Lastly, news editors said they are aware of the potential for bias in AI generations, given the unrepresentative data AI models are trained on.

This year, the World Economic Forum has named AI-fueled misinformation and disinformation as the world's greatest short-term risk. It placed this above even disasters like extreme weather events, inflation and armed conflict.

Because of this risk and the elections happening in the United States and around the world this year, engaging in healthy skepticism about what you see online is a must.

As is being thoughtful about where you get your news and information from. Doing so makes you better equipped to participate in a democracy, and less likely to fall for scams.

Provided by The Conversation

This article is republished from The Conversation under a Creative Commons license. Read the original article.The Conversation

Citation: TV network was slammed for 'AI editing' a Victorian MP's dress. How can news media use AI responsibly? (2024, February 1) retrieved 27 April 2024 from https://phys.org/news/2024-02-tv-network-slammed-ai-victorian.html
This document is subject to copyright. Apart from any fair dealing for the purpose of private study or research, no part may be reproduced without the written permission. The content is provided for information purposes only.

Explore further

Stock photo giant Getty releases AI image generator

18 shares

Feedback to editors