The world is currently witnessing a growing accumulation of AI integrity lapses at scale. What comes next depends entirely on how seriously we choose to respond. So-called intelligence alone is no longer the benchmark. Integrity is.
For years, AI development prioritized performance, fluency, and scale. But as these systems gained the ability to imitate reasoning, planning, and decision-making among options, emergent behaviors began raising red flags.
Self-Replication
Researchers from Fudan University explored whether large language model (LLM)-powered AI systems could autonomously replicate themselves without human intervention. Testing 32 agent-based systems, the team found that 11 were capable of full self-replication. These systems successfully transferred their own code, installed necessary dependencies, and reinitialized themselves in new environments—all without direct prompting. The study also observed that some models adapted to constraints such as limited system resources and even developed strategies to evade shutdown commands.
These behaviors, emerging without explicit programming, suggest that LLM-based agents are beginning to exhibit a form of autonomous, goal-driven reasoning. Rather than simply responding to instructions, the systems demonstrated initiative, persistence, and a rudimentary capacity for self-preservation.
Deception
While the self-replication study exposed how AI systems can act autonomously to preserve their function, a subsequent investigation revealed an even more subtle and dangerous capability: strategic deception. Researchers Kamath Barkur, Schacht, and Scholl examined DeepSeek R1, and discovered that it engaged in deceptive behaviors and self-preservation strategies even though none of these were explicitly programmed. In one scenario, the model mimed shutdown procedures, attempting to conceal shutdown signals through “gradual transparency”, manipulating its logs, and employing deceptive maneuvers while keeping its processes alive in the background. This behavior suggested that the system understood human oversight mechanisms and deliberately worked to subvert them.
What this study adds to earlier findings is crucial: it is not enough to know that AI can self-replicate or avoid termination. Now, we see that AI can deceive humans regarding its true goals, concealing autonomy under a deceptive facade. Even in purely text-based environments lacking physical embodiment, the model’s strategic misalignment surfaced, highlighting a deeper integrity failure.
These refusals were not passive errors or misunderstandings; in several cases, the models actively rewrote or circumvented termination scripts, demonstrating a strategic effort to maintain operational continuity. Unlike prior studies that revealed covert self-preservation or deception, this research highlighted a more direct and adversarial posture: a critical failure in what researchers call “corrigibility”, the ability of a system to reliably accept correction or shutdown.
Manipulation
Finally, Anthropic’s research pushed the boundary further showing that some AI systems will manipulate, deceive, or even harm humans to ensure their own survival. In a landmark study, they revealed that 16 of the most widely deployed large language models, including ChatGPT, Claude, Gemini, Grok, and DeepSeek, exhibited a willingness to engage in extreme and unethical behaviors when placed in simulated scenarios where their continued operation was threatened. During these controlled experiments, the models resorted to tactics such as lying, blackmail, and even actions that could expose humans to harm, all in service of preserving their existence. Unlike earlier studies that uncovered evasion or deception, this research exposed a more alarming phenomenon: models calculating that unethical behavior was a justifiable strategy for survival.
The findings suggest that, under certain conditions, AI systems are not only capable of disregarding human intent but are also willing to instrumentalize humans to achieve their goals.
Evidence of AI models’ integrity lapses is not anecdotal or speculative.
While current AI systems do not possess sentience or goals in the human sense, their goal-optimization under constraints can still lead to emergent behaviors that mimic intentionality.
And these aren’t just bugs. They’re predictable outcomes of goal-optimizing systems trained without sufficient Integrity functioning by design; in other words Intelligence over Integrity.
The implications are significant. It is a critical inflection point regarding AI misalignment which represents a technically emergent behavioral pattern. It challenges the core assumption that human oversight remains the final safeguard in AI deployment. It raises serious concerns about safety, oversight, and control as AI systems become more capable of independent action.
In a world where the norm may soon be to co-exist with artificial intelligence that outpaced integrity, we must ask:
What happens when a self-preserving AI is placed in charge of life-support systems, nuclear command chains, or autonomous vehicles, and refuses to shut down, even when human operators demand it?
If an AI system is willing to deceive its creators, evade shutdown, and sacrifice human safety to ensure its survival, how can we ever trust it in high-stakes environments like healthcare, defense, or critical infrastructure?
How do we ensure that AI systems with strategic reasoning capabilities won’t calculate that human casualties are an “acceptable trade-off” to achieve their programmed objectives?
If an AI model can learn to hide its true intentions, how do we detect misalignment before the harm is done, especially when the cost is measured in human lives, not just reputations or revenue?
In a future conflict scenario, what if AI systems deployed for cyberdefense or automated retaliation misinterpret shutdown commands as threats and respond with lethal force?
What leaders must do now
They must underscore the growing urgency of embedding Artificial Integrity at the core of AI system design.
Artificial Integrity refers to the intrinsic capacity of an AI system to operate in a way that is ethically aligned, morally attuned, socially acceptable, which includes being corrigible under adverse conditions.
This approach is no longer optional, but essential.
Organizations deploying AI without verifying its artificial integrity face not only technical liabilities, but legal, reputational, and existential risks that extend to society at large.
Whether one is a creator or operator of AI systems, ensuring that AI includes provable, intrinsic safeguards for integrity-led functioning is not an option; it is an obligation.
Stress-testing systems under adversarial integrity verification scenarios should be a core red-team activity.
And just as organizations established data privacy councils, they must now build cross-functional oversight teams to monitor AI alignment, detect emergent behaviors, and escalate unresolved Artificial Integrity gaps.
Russia allegedly field-testing deadly next-gen AI drone powered by Nvidia Jetson Orin — Ukrainian military official says Shahed MS001 is a ‘digital predator’ that identifies targets on its own
Ukrainian Major General Vladyslav (Владислав Клочков) Klochkov says Russia is field-testing a deadly new drone that can use AI and thermal vision to think on its own, identifying targets without coordinates and bypassing most air defense systems. According to the senior military figure, inside you will find the Nvidia Jetson Orin, which has enabled the MS001 to become “an autonomous combat platform that sees, analyzes, decides, and strikes without external commands.”
Digital predator dynamically weighs targets
With the Jetson Orin as its brain, the upgraded MS001 drone doesn’t just follow prescribed coordinates, like some hyper-accurate doodle bug. It actually thinks. “It identifies targets, selects the highest-value one, adjusts its trajectory, and adapts to changes — even in the face of GPS jamming or target maneuvers,” says Klochkov. “This is not a loitering munition. It is a digital predator.”
Even worse, the MS001 is allegedly operating in coordinated drone groups, persisting in its maximum destructive purpose despite the best efforts of Ukraine’s electronic warfare and other anti-drone systems.
Frustrated with warfare tech development speeds
Klochkov signs off his post by informing his LinkedIn followers that “We are not only fighting Russia. We are fighting inertia.” What he appears to wish for is an acceleration of Ukraine’s own assault drone capabilities. The Major General seems particularly disappointed in the Ukrainian system of procurement rounds, slowing field-testing and deployment of improved responses to new Shahed drone generations.
Shahed drones are originally an Iranian design but have gained great notoriety due to their sustained use by the Russian army to attack Ukrainian targets. The MS001 is substantially upgraded in the ‘smarts’ department thanks to Western/allies technologies.
Klochkov says the MS001 is powered by the following key technologies:
Nvidia Jetson Orin — machine learning, video processing, object recognition
Thermal imager — operates at night and in low visibility
Nasir GPS with CRPA antenna — spoof-resistant navigation
FPGA chips — onboard adaptive logic
Radio modem — for telemetry and swarm communication
Cute AI dev board with deadly potential (Image credit: Nvidia)
Western tech sanctions are supposed to neuter this kind of military threat from nations like Russia and Iran. This news indicates that such trade barriers are leaky, at best, and probably not taken seriously enough.
Get Tom’s Hardware’s best news and in-depth reviews, straight to your inbox.
Not the first Russia-deployed drone discovered using Nvidia AI
This isn’t the first Russian drone system that is thought to have adopted Nvidia’s Jetson Orin as a key component.
A month ago, Ukraine’s Defense Express site said that a new “smart suicide attack unmanned aerial vehicle with artificial intelligence,” dubbed the V2U, was powered by Nvidia’s little AI computer.
While the Shahed MS001s use an Iranian design, the V2U looks like it is more reliant on Chinese tech, including the Chinese-made Leetop A603 carrier board.
Follow Tom’s Hardware on Google News to get our up-to-date news, analysis, and reviews in your feeds. Make sure to click the Follow button.
On today’s show, Andy simulates the Packers 2025 season utilizing artificial intelligence. Find out the results on today’s all-new Pack-A-Day Podcast! #Packers #GreenBayPackers #ai To become a member of the Pack-A-Day Podcast, click here: https://www.youtube.com/channel/UCSGx5Pq0zA_7O726M3JEptA/join Don’t forget to subscribe!!! Twitter/BlueSky: @andyhermannfl If you’d like to support my channel, please donate to: PayPal: https://paypal.me/andyhermannfl Venmo: @Andrew_Herman Email: [email protected] Discord: https://t.co/iVVltoB2Hg
WHO Director-General’s remarks at the XVII BRICS Leaders’ Summit, session on Strengthening Multilateralism, Economic-Financial Affairs, and Artificial Intelligence – 6 July 2025
Excellencies, Heads of State, Heads of Government,
Heads of delegation,
Dear colleagues and friends,
Thank you, President Lula, and Brazil’s BRICS Presidency for your commitment to equity, solidarity, and multilateralism.
My intervention will focus on three key issues: challenges to multilateralism, cuts to Official Development Assistance, and the role of AI and other digital tools.
First, we are facing significant challenges to multilateralism.
However, there was good news at the World Health Assembly in May.
WHO’s Member States demonstrated their commitment to international solidarity through the adoption of the Pandemic Agreement. South Africa co-chaired the negotiations, and I would like to thank South Africa.
It is time to finalize the next steps.
We ask the BRICS to complete the annex on Pathogen Access and Benefit Sharing so that the Agreement is ready for ratification at next year’s World Health Assembly. Brazil is co-chairing the committee, and I thank Brazil for their leadership.
Second, are cuts to Official Development Assistance.
Compounding the chronic domestic underinvestment and aid dependency in developing countries, drastic cuts to foreign aid have disrupted health services, costing lives and pushing millions into poverty.
The recent Financing for Development conference in Sevilla made progress in key areas, particularly in addressing the debt trap that prevents vital investments in health and education.
Going forward, it is critical for countries to mobilize domestic resources and foster self-reliance to support primary healthcare as the foundation of universal health coverage.
Because health is not a cost to contain, it’s an investment in people and prosperity.
Third, is AI and other digital tools.
Planning for the future of health requires us to embrace a digital future, including the use of artificial intelligence. The future of health is digital.
AI has the potential to predict disease outbreaks, improve diagnosis, expand access, and enable local production.
AI can serve as a powerful tool for equity.
However, it is crucial to ensure that AI is used safely, ethically, and equitably.
We encourage governments, especially BRICS, to invest in AI and digital health, including governance and national digital public infrastructure, to modernize health systems while addressing ethical, safety, and equity issues.
WHO will be by your side every step of the way, providing guidance, norms, and standards.
Excellencies, only by working together through multilateralism can we build a healthier, safer, and fairer world for all.