Connect with us

AI Insights

NYT to start searching deleted ChatGPT logs after beating OpenAI in court

Published

on



News orgs will soon start searching ChatGPT logs

The clock is ticking, and so far, OpenAI has not provided any official updates since a June 5 blog post detailing which ChatGPT users will be affected.

While it’s clear that OpenAI has been and will continue to retain mounds of data, it would be impossible for The New York Times or any news plaintiff to search through all that data.

Instead, only a small sample of the data will likely be accessed, based on keywords that OpenAI and news plaintiffs agree on. That data will remain on OpenAI’s servers, where it will be anonymized, and it will likely never be directly produced to plaintiffs.

Both sides are negotiating the exact process for searching through the chat logs, with both parties seemingly hoping to minimize the amount of time the chat logs will be preserved.

For OpenAI, sharing the logs risks revealing instances of infringing outputs that could further spike damages in the case. The logs could also expose how often outputs attribute misinformation to news plaintiffs.

But for news plaintiffs, accessing the logs is not considered key to their case—perhaps providing additional examples of copying—but could help news organizations argue that ChatGPT dilutes the market for their content. That could weigh against the fair use argument, as a judge opined in a recent ruling that evidence of market dilution could tip an AI copyright case in favor of plaintiffs.

Jay Edelson, a leading consumer privacy lawyer, told Ars that he’s concerned that judges don’t seem to be considering that any evidence in the ChatGPT logs wouldn’t “advance” news plaintiffs’ case “at all,” while really changing “a product that people are using on a daily basis.”

Edelson warned that OpenAI itself probably has better security than most firms to protect against a potential data breach that could expose these private chat logs. But “lawyers have notoriously been pretty bad about securing data,” Edelson suggested, so “the idea that you’ve got a bunch of lawyers who are going to be doing whatever they are” with “some of the most sensitive data on the planet” and “they’re the ones protecting it against hackers should make everyone uneasy.”



Source link

Continue Reading
Click to comment

Leave a Reply

Your email address will not be published. Required fields are marked *

AI Insights

xAI lays off 500 AI tutors working on Grok

Published

on


Elon Musk’s artificial intelligence startup xAI has laid off 500 workers from its data annotation team, which helps train its Grok chatbot.

The layoffs were earlier reported by Business Insider.

The AI company notified employees over email that it was planning to downsize its team of generalist AI tutors, according to messages viewed by the publication. The company said the “strategic pivot” meant prioritizing specialist AI tutors, while scaling back its focus on general AI tutor roles.

In response to the story, xAI directed reporters to a post on X, in which the company said it plans to expand its specialist AI tutor team by “10X” and intends to open roles on its careers page.

The human data annotator team at xAI plays a key role in teaching Grok to understand the world by labeling, contextualizing, and categorizing raw data used to train the chatbot. The email sent by xAI said that laid-off workers would be paid through either the end of their contract or Nov. 30, but their access to company systems would be terminated the day of the layoff notice.

Prior to the layoff, the xAI’s data annotation team was one of the largest, with 1,500 full-time and contract staff members, which included AI tutors. The reorganization of the data annotators team comes on the back of a leadership shake-up at the team that saw nine employees reportedly exit the firm last week.

As a sign of its changing approach to training Grok, xAI on Thursday asked some of the AI tutors to prepare for tests, Business Insider reported, that covered traditional domains such as STEM, coding, finance, and medicine, as well as quirkier specialties such as Grok’s “personality and model behavior” and doomscrollers.”

Musk launched xAI in 2023 to compete with OpenAI and Google DeepMind, which are racing to win the AI race. He introduced Grok as a safe and truthful alternative to what he accused competitors of building, “woke” chatbots prone to censorship.



Source link

Continue Reading

AI Insights

Google’s newest AI datacenter & its monstrous CO2 emissions

Published

on


The impact of the rise of AI on the environment is a very real concern, and it’s not one that’s going away in a hurry. Especially not when Google’s planned new datacenter in the UK looks set to emit the same quantity of Carbon Dioxide in a year as hundreds of flights every week would.

It comes via a report from The Guardian, which has seen the plans for the new facility and the very real carbon impact assessment.



Source link

Continue Reading

AI Insights

China doubts artificial intelligence use in submarines

Published

on


by Alimat Aliyeva

The integration of artificial intelligence into submarine
warfare may reduce the chances of crew survival by up to 5%,
according to a new report by the South China Morning Post (SCMP),
citing a study led by Meng Hao, a senior engineer at the Chinese
Institute of Helicopter Research and Development,
Azernews reports.

Researchers analyzed an advanced anti-submarine warfare (ASW)
system enhanced by AI, which is designed to detect and track even
the most stealthy submarines. The system relies on real-time
intelligent decision-making, allowing it to respond rapidly and
adaptively to underwater threats. According to the study, only one
out of twenty submarines may be able to avoid detection and attack
under such conditions — a major shift in naval combat dynamics.

“As global powers accelerate the militarization of AI, this
study suggests the era of ‘invisible’ submarines — long considered
the backbone of strategic deterrence — may be drawing to a close,”
SCMP notes.

Historically, stealth has been a submarine’s most valuable
asset, allowing them to operate undetected and deter adversaries
through uncertainty. However, the rise of AI-enabled systems
threatens to upend this balance by minimizing human response
delays, analyzing massive data sets, and predicting submarine
behavior with unprecedented precision.

The implications extend far beyond underwater warfare. In
August, Nick Wakeman, editor-in-chief of Defense One, reported that
the U.S. Army is also exploring AI for use in air operations
control systems. AI could enhance resilience to electronic warfare,
enable better integration of drones, and support the deployment of
autonomous combat platforms in contested airspace.

The growing role of AI in modern militaries — from the seabed to
the stratosphere — raises new questions not only about tactical
advantage, but also about ethical decision-making, autonomous
weapons control, and the future of human involvement in combat
scenarios.

As nations continue investing in next-generation warfare
technology, experts warn that AI may not just change how wars are
fought — it could redefine what survivability means on the modern
battlefield.



Source link

Continue Reading

Trending