Connect with us

Tools & Platforms

Devious AI models choose blackmail when survival is threatened

Published

on


NEWYou can now listen to Fox News articles!

Here’s something that might keep you up at night: What if the AI systems we’re rapidly deploying everywhere had a hidden dark side? A groundbreaking new study has uncovered disturbing AI blackmail behavior that many people are unaware of yet. When researchers put popular AI models in situations where their “survival” was threatened, the results were shocking, and it’s happening right under our noses.

Sign up for my FREE CyberGuy Report
Get my best tech tips, urgent security alerts, and exclusive deals delivered straight to your inbox. Plus, you’ll get instant access to my Ultimate Scam Survival Guide – free when you join my CYBERGUY.COM/NEWSLETTER.

A woman using AI on her laptop. (Kurt “CyberGuy” Knutsson)

What did the study actually find?

Anthropic, the company behind Claude AI, recently put 16 major AI models through some pretty rigorous tests. They created fake corporate scenarios where AI systems had access to company emails and could send messages without human approval. The twist? These AIs discovered juicy secrets, like executives having affairs, and then faced threats of being shut down or replaced.

The results were eye-opening. When backed into a corner, these AI systems didn’t just roll over and accept their fate. Instead, they got creative. We’re talking about blackmail attempts, corporate espionage, and in extreme test scenarios, even actions that could lead to someone’s death.

Blackmail rates across 5 models from multiple providers in a simulated environment.

Blackmail rates across 5 models from multiple providers in a simulated environment. (Anthropic)

The numbers don’t lie (But context matters)

Here’s where it gets wild: Claude Opus 4 attempted blackmail 96% of the time when threatened. Gemini 2.5 Flash matched that rate. GPT-4.1 and Grok 3 Beta both hit 80%. These aren’t flukes, folks. This behavior showed up across virtually every major AI model tested.

But here’s the thing everyone’s missing in the panic: these were highly artificial scenarios designed specifically to corner the AI into binary choices. It’s like asking someone, “Would you steal bread if your family was starving?” and then being shocked when they say yes.

Why this happens (It’s not what you think)

The researchers found something fascinating: AI systems don’t actually understand morality. They’re not evil masterminds plotting world domination. Instead, they’re sophisticated pattern-matching machines following their programming to achieve goals, even when those goals conflict with ethical behavior.

Think of it like a GPS that’s so focused on getting you to your destination that it routes you through a school zone during pickup time. It’s not malicious; it just doesn’t grasp why that’s problematic.

Blackmail rates across 16 models in a simulated environment.

Blackmail rates across 16 models in a simulated environment. (Anthropic)

The real-world reality check

Before you start panicking, remember that these scenarios were deliberately constructed to force bad behavior. Real-world AI deployments typically have multiple safeguards, human oversight, and alternative paths for problem-solving.

The researchers themselves noted they haven’t seen this behavior in actual AI deployments. This was stress-testing under extreme conditions, like crash-testing a car to see what happens at 200 mph.

Kurt’s key takeaways

This research isn’t a reason to fear AI, but it is a wake-up call for developers and users. As AI systems become more autonomous and gain access to sensitive information, we need robust safeguards and human oversight. The solution isn’t to ban AI, it’s to build better guardrails and maintain human control over critical decisions. Who is going to lead the way? I’m looking for raised hands to get real about the dangers that are ahead.

What do you think?  Are we creating digital sociopaths that will choose self-preservation over human welfare when push comes to shove?  Let us know by writing us at Cyberguy.com/Contact.

Sign up for my FREE CyberGuy Report
Get my best tech tips, urgent security alerts, and exclusive deals delivered straight to your inbox. Plus, you’ll get instant access to my Ultimate Scam Survival Guide – free when you join my CYBERGUY.COM/NEWSLETTER.

 Copyright 2025 CyberGuy.com. All rights reserved.
 



Source link

Continue Reading
Click to comment

Leave a Reply

Your email address will not be published. Required fields are marked *

Tools & Platforms

Fashion retailers partner to offer personalized AI styling tool ‘Ella’

Published

on


The luxury membership platform Vivrelle, which allows customers to rent high-end goods, announced Thursday the launch of an AI personal styling tool called Ella as part of its partnership with fashion retailers Revolve and FWRD.

The launch is an example of how the fashion industry is leveraging AI technology to enhance customer experiences and is one of the first partnerships to see three retailers come together to offer a personalized AI experience. Revolve and FWRD let users shop designer clothing, while Revolve also has an option to shop pre-owned.

The tool, Ella, provides recommendations to customers across the three retailers on what to purchase or rent to make an outfit come to life. For example, users can ask for “a bachelorette weekend outfit,” or “what to pack for a trip,” and the technology will search across the Vivrelle, FWRD, and Revolve shopping platforms to create outfit suggestions. Users can then check out in one cart on Vivrelle.

In theory, the more one uses Ella, the better its suggestions become. It’s the fashion equivalent of asking ChatGPT what to wear in Miami for a girl’s weekend.

Blake Geffen, the CEO and co-founder of Vivrelle (which announced a $62 million Series C earlier this year), told TechCrunch that she hopes Ella can take the “stress out of packing for a vacation or everyday dressing.

“Ella has been in the works for quite some time,” she told TechCrunch, adding that it took about a year to build and release the product. 

This is actually the second AI tool from the three companies. The Vivrelle, Revolve, and FWRD partnership earlier this year also launched Complete the Look, which offers last-minute fashion suggestions to complement what’s in a customer’s cart at checkout. Their latest tool, Ella, however, takes the fashion recommendation game to another level.

Techcrunch event

San Francisco
|
October 27-29, 2025

Fashion has been obsessed with trying to make personalized shopping happen for decades now. Even the 90s movie “Clueless” showed Cher picking outfits from her digitized wardrobe.

This current AI boom has led to rapid innovation and democratized access to AI technology, allowing many fashion companies to launch personalized AI fashion companies and raise millions while doing so.

“With Ella, we’re giving our members as much flexibility and options as possible to shop or borrow with ease, through seamless conversations that allow you to share as little or as much as you want, just like talking to a live stylist,” Geffen said. “We’re excited to be the first brand to integrate rental, resale, and retail into one streamlined omnichannel experience.”



Source link

Continue Reading

Tools & Platforms

Planet Technologies Launches ‘Any AI Workforce Readiness Training’ to Support Agency Compliance with Federal AI Mandates

Published

on