Embracing AI: How To Stay Ahead Of The Curve -- Aurora Martinez, IJNet
Aurora Martinez -- International Journalist Network
March 15, 2024
Generative artificial intelligence (GenAI) may fundamentally change how people work.
In some ways, it already has – and the technology is rapidly evolving. Take for instance, the emergence of multimodal AI models, which can simultaneously combine images, text and speech to generate new content.
Concerns around GenAI notwithstanding – its potential to spread mis- and disinformation, and fears it could replace journalism jobs, among them – media leaders should consider how to use the technology to uplift their work. If utilized responsibly and ethically, AI can improve how news is produced and consumed.
“The future of AI — really we're just getting started on what is possible,” said Nikita Roy, an ICFJ Knight fellow and host of Newsroom Robots, a podcast featuring conversations with leaders in AI and journalism. “Things are getting exponentially better almost every week at this point in time.”
By approaching AI with an open mind and dedication to understanding its uses, journalists can better share valuable information with audiences and remain competitive in the ever-evolving media landscape, Roy said.
The following are tips for journalists to stay ahead of the AI curve:
Harnessing the benefits of AI
AI is not here to replace the role journalists play in their communities, Roy said. Instead, reporters can leverage the technology to optimize productivity, and make their news coverage more profound, engaging and dynamic.
“How can [AI] support the journalists in your newsroom to produce more interesting work? How can you create new news experiences with AI? How can you reimagine news with AI?” These are all questions journalists should be asking themselves now, Roy advised.
AI can help journalists carry out projects that wouldn’t have been possible otherwise. This can be especially helpful in smaller newsrooms with less capacity. For example, journalists can use natural language prompts in ChatGPT-4 to analyze large data sets, visualize information and identify trends, all of which could lead to story ideas.
Future innovations could spur even more benefits. Though still not available to the public, in February, OpenAI revealed Sora, a new tool that promises to create video from text in a matter of seconds. Sora could assist visual storytelling, but like other GenAI tools, it has sparked concerns about its possible negative consequences, such as whether it will perpetuate biases.
Journalists shouldn’t immediately jump to the negative aspects of new AI tools, however, urged Emilse Garzón, a digital journalist and professor specialized in AI and cybersecurity in Argentina: “We have to talk about the consequences without being catastrophic when it doesn’t imply a real catastrophe, and not [catastrophizing] because of clickbait.”
Newsroom leaders should encourage discussions on GenAI across teams, Roy said. Organization-wide, team-specific training can help build institutional knowledge and improve workflows.
Creating a Slack channel to share “AI wins” can help further encourage stronger AI practices, Roy suggested: “You’re creating a culture where people can be transparent and open about how they have been using generative AI if it falls within the guidelines.”
Utilizing AI ethically
Transparent, responsible utilization of GenAI tools will promote its ethical use by others. “Technology doesn’t harm us,” Garzón said. “What harms us is how it’s used, who creates it and who endorses its use.” On an individual level, if Garzón uses GenAI tools like Midjourney or Canva’s AI-powered apps to generate images from text, she discloses this information with her audience.
Newsrooms, Roy urged, should implement AI editorial guidelines to standardize best practices. Having rules in place can push back on taboos around using AI tools, make clear how people can use them, and address ethical concerns.
Guidelines should stress the importance of human involvement in GenAI-assisted content, including fact-checking before publication, and a zero tolerance policy for plagiarism. AI should never be used to create stories that go straight to publishing, Roy warned. “Everything has a right and a wrong use,” she said. “We have to understand how to use these [AI] tools in the right way.”
It’s up to individual newsrooms to dictate the scope of AI use among their staff: while the AP does not allow ChatGPT to be used in any published content, other newsrooms may only restrict using GenAI to author original reporting.
“Those guidelines don't have to be set in stone,” Roy said. They can be revised periodically based on how new technologies evolve and what each organization needs.
Fostering dialogue around AI
Journalists should be at the forefront of the many discussions being had around AI today, Garzón said.
Offering tips on how AI can be used to tackle disinformation, and diving into the complexities of why people should care about copyrights, data protection, and who is behind AI-driven technologies are all important discussions that reporters can drive through their coverage.
Journalists should offer a variety of perspectives and nuance in their reporting on the topic, and not just focus on the positives or negatives, Garzón added. Their duty is to inform the public – not tell them what to think about it. “We ought to have among us those debates that sometimes escape and go beyond our idea of the editorial line,” she said. “What we have to ensure is that there is diversity of opinions, even if we don’t like those opinions.”
Through their exploration of AI technologies, journalists can also identify where GenAI is producing misleading information or biased outputs. This can promote a deeper understanding of how AI models work and could be improved on, Roy said.
“We need to be informed citizens,” Roy said. “It is expected and important to do extensive research and contextualize stories about the development and evolution of today’s AI to show respect for the people who trust us.”
Linus Zoll & Google DeepMind / Better Images of AI / Generative Image models / CC-BY 4.0.
Read more articles by
Aurora Martínez is a former communications intern at ICFJ. She is originally from Nicaragua and is a student of journalism at the University of Florida
- CreatedWednesday, March 20, 2024
- Last modifiedWednesday, March 20, 2024
Electronic Frontier Foundation
- Free Speech Around the World | EFFector 36.6 May 08, 2024
- What Can Go Wrong When Police Use AI to Write Reports? May 08, 2024
- Speaking Freely : Nompilo Simanje May 07, 2024
- Podcast Episode: Building a Tactile Internet May 07, 2024
The Intercept
- As Biden Warns Against Rafah Invasion, AIPAC Pushes Congress to Support Israel’s Operation May 08, 2024
- Some Universities Chose Violence. Others Responded to Protests by Considering Student Demands. May 08, 2024
- 600,000 Palestinian Kids in Rafah Can’t “Evacuate” Safely, UNICEF Official Says May 08, 2024
- They Used to Say Arabs Can’t Have Democracy Because It’d Be Bad for Israel. Now the U.S. Can’t Have It Either. May 08, 2024
VTDigger
- Brenna Galdenzi: Our Voices Matter May 09, 2024
- Vermont Conversation: Dartmouth Professor Annelise Orleck was arrested but not silenced May 08, 2024
- Final Reading: Another side of Dick McCormack May 08, 2024
- Vermont lawmakers advance bill to increase supervision of people accused of crimes May 08, 2024