Connect with us

AI Insights

AI Researchers Explore Whether Soft Robotics and Embodied Cognition Unlock Artificial General Intelligence

Published

on


IN A NUTSHELL
  • 🤖 Researchers explore whether AI needs a physical body to achieve true intelligence.
  • 🧠 The concept of embodied cognition suggests that sensing, acting, and thinking are interconnected.
  • 🐙 Soft robotics, inspired by creatures like the octopus, offer a new path for developing adaptive AI.
  • 🔄 Autonomous physical intelligence (API) allows materials to self-regulate and make decisions independently.

In the realm of artificial intelligence (AI), the concept of whether machines require physical bodies to achieve true intelligence has long been a topic of debate. Popular culture, from Rosie the robot maid in “The Jetsons” to the empathetic C-3PO in “The Empire Strikes Back,” has offered diverse interpretations of robots and AI. However, these fictional portrayals often overlook the complexities and limitations faced by real-world AI systems. With recent advancements in robotics and AI, researchers are revisiting the question of embodiment in AI, exploring whether a physical form could be essential for achieving artificial general intelligence (AGI). This exploration could redefine our understanding of cognition, intelligence, and the future of AI technology.

The Limits of Disembodied AI

Recent studies have highlighted the shortcomings of disembodied AI systems, particularly in their ability to perform complex tasks. A study from Apple on Large Reasoning Models (LRMs) found that while these systems can outperform standard language models in some scenarios, they struggle significantly with more complex problems. Despite having ample computing power, these models often collapse under complexity, revealing a fundamental flaw in their reasoning capabilities.

Unlike humans, who can reason consistently and algorithmically, these AI models lack internal logic in their “reasoning traces.” Nick Frosst, a former Google researcher, emphasized this discrepancy, noting that current AI systems merely predict the next most likely word rather than truly think like humans. This raises concerns about the viability of disembodied AI in replicating human-like intelligence.

“What we are building now are things that take in words and predict the next most likely word … That’s very different from what you and I do,” Frosst told The New York Times.

The limitations of disembodied AI underscore the need for exploring alternative approaches to achieve true cognitive abilities in machines.

“They Brought Something Back from Space” Dragon Capsule’s 6,700-Pound Cargo Contains Technology That Changes Everything About Earth’s Future

Cognition Is More Than Just Computation

Historically, artificial intelligence was developed under the paradigm of Good Old-Fashioned Artificial Intelligence (GOFAI), which treated cognition as symbolic logic. This approach assumed that intelligence could be built by processing symbols, akin to a computer executing code. However, real-world challenges exposed the limitations of this model, leading researchers to question whether intelligence could be achieved without a physical body.

Research from various disciplines, including psychology and neuroscience, suggests that intelligence is inherently linked to physical interactions with the environment. In humans, the enteric nervous system, often referred to as the “second brain,” operates independently, illustrating that intelligence can be distributed throughout an organism rather than centralized in a brain.

This has led to the concept of embodied cognition, where sensing, acting, and thinking are interconnected processes. As Rolf Pfeifer, Director of the University of Zurich’s Artificial Intelligence Laboratory, pointed out, “Brains have always developed in the context of a body that interacts with the world to survive.” This perspective challenges the traditional view of cognition and suggests that a physical body might be crucial for developing adaptable and intelligent systems.

Embodied Intelligence: A Different Kind of Thinking

The exploration of embodied intelligence has prompted researchers to consider new approaches to AI development. Cecilia Laschi, a pioneer in soft robotics, advocates for the use of soft-bodied machines inspired by organisms like the octopus. These creatures demonstrate a form of intelligence that is distributed throughout their bodies, allowing them to adapt and respond to their environments without centralized control.

“Robots Threaten Jobs,” as Parcel Delivery Costs Plummet 53% with Automation in the US

Laschi argues that smarter AI requires softer, more flexible bodies that can offload perception, control, and decision-making to the physical structure of the robot itself. This approach reduces the computational demands on the main AI system, enabling it to function more effectively in unpredictable environments.

In a May special issue of Science Robotics, Laschi explained that “motor control is not entirely managed by the computing system … motor behavior is partially shaped mechanically by external forces acting on the body.” This suggests that behavior and intelligence are shaped by experience and interaction with the environment, rather than pre-programmed algorithms.

The field of soft robotics, which employs materials like silicone and special fabrics, offers promising possibilities for creating adaptive, real-time learning systems. By integrating flexibility and adaptability into the physical form of AI, researchers are paving the way for machines that can think and learn in ways similar to living organisms.

Flesh and Feedback: How to Make Materials Think for Themselves

The development of soft robotics is also advancing the concept of autonomous physical intelligence (API), where materials themselves exhibit decision-making capabilities. Ximin He, an Associate Professor of Materials Science and Engineering at UCLA, has been at the forefront of this research, designing soft materials that not only react to stimuli but also regulate their movements using built-in feedback.

“We Built A Walking Robot From Just 18 Metal Parts”: Tokyo Engineers Create Open-Source Bipedal Robot That Anyone Can Assemble At Home

He’s approach involves embedding logic directly into the materials, allowing them to sense, act, and decide autonomously. This method contrasts with traditional robotics, which relies on external control systems to analyze sensory data and dictate actions. By incorporating nonlinear feedback mechanisms, soft robots can achieve rhythmic, controlled behaviors without external intervention.

He’s work has demonstrated the potential for soft materials to self-regulate their movements, a significant advancement toward creating lifelike autonomy in machines. This approach opens up new possibilities for AI systems that can adapt and respond to their environments in more natural and intuitive ways.

By integrating sensing, control, and actuation at the material level, researchers are moving closer to developing machines that can independently decide, adapt, and act, paving the way for a new era of intelligent robotics.

As researchers continue to explore the potential of embodied intelligence and soft robotics, the future of AI appears increasingly promising. These innovations could lead to breakthroughs in fields ranging from medicine to environmental exploration, offering machines that are not only intelligent but also capable of understanding and interacting with the world in new ways. However, questions remain about how these technologies will be integrated into society and the ethical implications of creating machines with lifelike autonomy. As we move forward, how will the intersection of AI and physical embodiment redefine our relationship with technology and the world around us?

This article is based on verified sources and supported by editorial technologies.

Did you like it? 4.7/5 (26)



Source link

Continue Reading
Click to comment

Leave a Reply

Your email address will not be published. Required fields are marked *

AI Insights

Opening the black box of machine learning-controlled plasma treatments

Published

on


Understanding machine learning modifies cold atmospheric plasma medicine delivery in cancer treatments without being trained on detailed plasma parameters.

Although artificial intelligence (AI) can adapt to changing conditions and achieve desired outcomes, how algorithms “understand” and adjust to inputs can be a mystery.

Lin et al. sought to uncover this “black box” in AI-controlled cold atmospheric plasma (CAP) treatments, an approach that induces apoptosis in diseased cells while preserving healthy ones. In previous work, they developed a machine learning (ML) system that predicts the post-treatment state of cancer cell targets and adjusts treatment accordingly. However, they didn’t know how the ML system achieved this outcome without an understanding of specific plasma parameters.

Using an AI-based optical emission spectroscopy (OES) spectra translation algorithm, the authors reverse engineered real-time chemical accumulations above cell medium surfaces. They found that, despite changing conditions, the ML algorithm alters experimental parameters to achieve the same therapeutic outcomes. The application of a Fourier transformation on OES spectra and chemical kinetics analysis revealed how the ML algorithm independently captured additional layers of physics information relying solely on cell viability status, without human input of this information, to achieve the precision and reliability of their AI-controlled CAP model.

“Beyond plasma medicine, similar approaches could advance machine learning-based control in fields like electric propulsion for satellites, plasma-based microfabrication, fusion reactor management, and many other plasma applications” said author Michael Keidar.

Next, the team looks to extend the scope of control that was demonstrated in this paper.

“Instead of limiting the AI to adjusting treatment duration, we plan to authorize and train the AI to control multiple plasma parameters simultaneously, including voltage, gas flow rate, and even additional external electric fields,” said author Li Lin. “In doing so, we aim to tailor therapy to the specific needs of each patient.”

Source: “Low-temperature plasma adaptation in the course of machine learning controls of plasma medicine,” by Li Lin, Qihui Wang, Zichao Hou, Michael Keidar, Physics of Plasmas (2025). The article can be accessed at https://doi.org/10.1063/5.0274614 .





Source link

Continue Reading

AI Insights

Ohio brings on artificial intelligence chatbot app to help fight crime, terrorism

Published

on


The adage when it comes to public safety has been “if you see something, say something.” Ohio is now employing a new tool where you can say something to an interactive artificial intelligence chatbot; an app that allows people to submit information about potential criminal activity.
 
Ohio Department of Public Safety Director Andy Wilson said the multi-lingual app Safeguard Ohio can allow anyone to upload video, audio, and photos of suspicious activity. Then it lets artificial intelligence to take it from there.
 
“Because AI is involved, it asks the follow-up questions,” Wilson said. “It asks basically everything that needs to be gathered from an informational point of view to get what we need to, number one, understand what’s going on and get it to the right folks.”
 
Users can select from eight categories to report a tip. Those include drug-related activity, human trafficking, terrorism, school threats, and crimes against children.

“People can submit suspicious activity reports using this bot, using this app, sending this information into homeland security and we will get it where it needs to go,” Wilson said.

Ohio Homeland Security (OHS) Director Mark Porter said up to this point, people who want to report suspicious activity would have to call or go to a static form online where they could enter information. He said authorities had seen a decrease in the number of reports over time, getting an average of 30 tips per month until Aug. 6. That’s when the new app went online.

“In the last 30 days, our numbers have tripled in what we are getting,” Porter said. He attributed the increase to the app’s capability to process multiple languages and younger people being more likely to file information using an app and chatbot.

Wilson said reports made via the app can still be made anonymously. But emergencies need to be handled as they always have been.

“This isn’t a substitute for 911. What this is is to catch more of the suspicious activity, not the imminent ‘Hey something is going down,’ but ‘my roommate has a manifesto’ or ‘I saw this person online basically threaten to kill so and so.’ That kind of stuff,” Wilson said. “The AI chatbot will direct the user in case of an emergency, something that’s an emergency or imminent, to call 911.”

Ohio Homeland Security paid approximately $200,000 to the software company Vigiliti for the initial development of the Safeguard Ohio chatbot, backend dashboard for OHS staff, and compatibility with OHS’s current case management system. OHS also signed a two-year contract for $250,000 per year with the company for maintenance of the system and 24/7 access to help resolve any technical issues.





Source link

Continue Reading

AI Insights

How Artificial Intelligence Is Revolutionizing Emergency Medicine

Published

on


Introduction
Applications of AI in emergency medicine
Benefits of AI in emergency care
Challenges and limitations
Conclusions
References
Further reading


Artificial intelligence is transforming emergency medicine by enhancing triage, diagnosis, and resource management, while also facing challenges related to ethics, bias, and regulation. This article explores its applications, benefits, and limitations in real-world clinical care.

Image Credit: JHEVPhoto / Shutterstock.com

Introduction

Artificial intelligence (AI) is an interdisciplinary field that integrates computer science, mathematics, and related disciplines to create algorithms that can perform tasks conventionally restricted to human intelligence. AI algorithms utilize data-driven analysis, probabilistic modelling, and iterative optimization to learn, solve problems, and make decisions.1

Unprecedented computational power, widely available and open-access electronic health data, as well as algorithmic breakthroughs, are rapidly transitioning AI from a conceptual technology to an integrated component of modern healthcare.1 Despite projected growth of the global AI healthcare market, its incorporation into clinical practice remains limited due to the relative nascency of this technology and lack of standardization.2

In emergency medicine, AI has gained traction not only in clinical decision support (CDS) but also in digital twin modeling of patients, predictive analytics for emergency department (ED) flow, and integration with prehospital emergency medical services (EMS).3,8,9

Additionally, recent primers emphasize the importance of familiarizing nonexpert clinicians with AI principles, terminology, and limitations to support safe and informed adoption.11

Applications of AI in emergency medicine

AI-driven triage algorithms can analyze large datasets without bias and with significantly greater depth than conventional models, enabling clinicians to prioritize patients more effectively compared to traditional methods.5 In fact, machine learning models consistently demonstrate superior discrimination and performance capabilities for predicting emergency outcomes like hospital admission or intensive care unit (ICU) transfer and conditions like stroke, sepsis, and myocardial infarction.4,5

Medical imaging and the interpretation of these images are among the most mature applications of AI, as numerous deep learning algorithms have been trained to analyze X-rays, computed tomography (CT) scans, and ultrasound images.1 For these applications, AI technologies have successfully detected abnormalities like intracranial hemorrhage, fractures, and pneumothorax with high accuracy to support clinicians and reduce conventional diagnostic delays.1 Explainable AI (XAI) methods are increasingly being incorporated into these models to enhance clinician trust by making diagnostic outputs more interpretable.7,11

AI-powered CDS systems have also been developed to integrate real-time data from electronic health records (EHRs) and provide timely recommendations.1 For example, AI models have been used to analyze electrocardiograms (ECGs) to predict impending cardiac arrest. Machine learning-assisted alerts have also been shown to improve the time to antibiotic administration.1 More recently, scoping reviews highlight that CDS tools in emergency departments have been used to improve sepsis management, diagnostic accuracy, and disposition planning.3 Published case examples include Duke’s “Sepsis Watch” system and Viz.ai for subdural hematoma detection, which illustrate real-world clinical adoption.11

AI-based predictive analytics can mitigate ED crowding by forecasting patient arrivals and anticipating surges. This application of AI allows hospitals to transition away from a reactive to a proactive staffing model that ensures the optimal allocation of limited resources like beds.1,6

AI-powered symptom checkers and chatbots can simultaneously guide patients in self-assessing the urgency of their condition. Emergency dispatchers can also utilize natural language processing to recognize conditions, such as out-of-hospital cardiac arrest, faster and more accurately, despite limitations in first-responder knowledge.1 EMS applications include AI-driven decision support for ambulance routing, prehospital risk stratification, and remote monitoring to improve patient outcomes before hospital arrival.6,11

Another emerging domain is the use of digital twins, virtual patient models that simulate disease progression and treatment response, which could help personalize emergency care interventions and optimize resource use.9

Benefits of AI in emergency care

AI algorithms can rapidly process and synthesize vast quantities of data, thereby leading to faster and more precise assessments.4 This significantly reduces conventional image interpretation delays, with some AI models demonstrating performance superior to that of human specialists in specific tasks.1

AI can provide several benefits to the existing public health infrastructure. By accurately predicting patient volume, AI can enable hospitals to better manage patient throughput, reduce system inefficiencies, alleviate overcrowding, and shorten patient wait times.6 These predictive tools also support disaster preparedness and surge capacity planning, strengthening system resilience.4,5

For administrative purposes, AI can automate routine and time-consuming tasks using ambient listening technologies and generative AI-based clinical summaries. The adoption of AI into these aspects of healthcare has the potential to reduce clinician burnout, as well as improve both patient satisfaction and provider well-being.1,4 Furthermore, AI can facilitate continuous quality improvement by identifying patterns in adverse events and enabling evidence-based policy development.7,11

High-tech hospital uses artificial intelligence in patient care

Challenges and limitations

Despite its future promise and validated benefits, the integration of AI into emergency medicine is associated with numerous technical, ethical, and legal challenges that must be addressed to ensure its safe and equitable deployment.1,4,6

A foundational principle of machine learning is that models are only as good as the data on which they are trained. Thus, models trained on historical health data containing latent biases, such as societal inequities or non-generalizable sampling designs, could learn and amplify these biases at scale.6 Unfortunately, these underrepresented are often the exact patient subpopulations like women, racial minorities, and other marginalized groups that would benefit the most from AI integration.2

A significant practical barrier, especially in developing and underdeveloped regions, is the difficulty of integrating novel AI systems into existing, often fragmented, hospital intelligence technologies (IT) infrastructure. The lack of data interoperability between different EHR systems makes it difficult to seamlessly integrate AI solutions, which could increase the complexity and associated costs of implementation.1 Even in advanced settings, CDS systems face challenges in workflow integration and clinician adoption, which can limit their real-world impact.3,11

AI models require access to massive datasets of sensitive patient information, which carries significant risks to patient privacy and data security.6,7 This is compounded by the “black box” problem, in which the internal decision-making processes of complex deep learning models are opaque and not readily interpretable. Explainability and transparency are therefore critical to support clinical accountability and medico-legal decision-making.7,11

Regulatory concerns are increasingly important: AI tools classified as software as a medical device (SaMD) fall under U.S. FDA oversight, requiring evidence of safety, effectiveness, and lifecycle monitoring.11

Both automation complacency, which reflects an over-reliance on AI, as well as selective adherence to only accept advice that confirms pre-existing beliefs, represent practical and ongoing challenges in clinical-AI interactions.1

Image Credit: Sutipond Somnam / Shutterstock.com

Conclusions

AI represents a transformative force in emergency medicine with the potential to accelerate and improve the accuracy of patient triage, diagnoses, and resource management, thereby leading to a more efficient and resilient global emergency care system. Nevertheless, the naivety and inherent limitations associated with AI emphasize the importance of using this technology as a tool to augment and empower human clinicians, rather than replace or undermine them. Future directions include broader evaluation of digital twins, real-world validation of CDS systems, EMS-focused AI interventions, and clinician education for nonexperts, which will be key to realizing AI’s full potential in emergency medicine.1,3,8,9,11

The role of digital twins in transforming emergency medicine.9

The role of digital twins in transforming emergency medicine.9

As these technologies continue to advance and become more readily accessible, policymakers, regulators, and healthcare leaders must collaborate to create robust ethical and legal frameworks that provide clear guidance on data privacy, algorithmic transparency, and legal liability. These efforts will ensure that the principles of safety, fairness, and accountability guide the gradual deployment of AI into the global healthcare sector.

References

  1. Chenais, G., Lagarde, E., & Gil-Jardiné, C. (2023). Artificial Intelligence in Emergency Medicine: Viewpoint of Current Applications and Foreseeable Opportunities and Challenges. Journal of Medical Internet Research, 25, e40031. DOI:10.2196/40031, https://www.jmir.org/2023/1/e40031
  2. Bajwa, J., Munir, U., Nori, A., & Williams, B. (2021). Artificial intelligence in healthcare: transforming the practice of medicine. Future Healthcare Journal, 8(2), e188-e194. DOI:10.7861/fhj.2021-0095, https://www.sciencedirect.com/science/article/pii/S2514664524005277?via%3Dihub
  3. Kareemi, H., Yadav, K., Price, C., et al. (2025). Artificial intelligence–based clinical decision support in the emergency department: A scoping review. Academic Emergency Medicine, 32(4), 386-395. DOI:10.1111/acem.15099, https://onlinelibrary.wiley.com/doi/full/10.1111/acem.15099
  4. Da’Costa, A., Teke, J., Origbo, J. E., et al. (2025). AI-driven triage in emergency departments: A review of benefits, challenges, and future directions. International Journal of Medical Informatics, 197, 105838. DOI:10.1016/j.ijmedinf.2025.105838, https://www.sciencedirect.com/science/article/pii/S1386505625000164
  5. Piliuk, K., & Tomforde, S. (2023). Artificial intelligence in emergency medicine. A systematic literature review. International Journal of Medical Informatics, 180, 105274. DOI:10.1016/j.ijmedinf.2023.105274, https://www.sciencedirect.com/science/article/pii/S1386505623002927
  6. Rosemaro, E., Anasica, & Zellar, I. (2025). AI-Based Decision Support Systems for Emergency Medical Services. International Journal of Recent Advances in Engineering and Technology, 13(1), 6-10.  https://journals.mriindia.com/index.php/ijraet/article/view/55
  7. Al Kuwaiti, A., Nazer, K., Al-Reedy, A., et al. (2023). A Review of the Role of Artificial Intelligence in Healthcare. Journal of Personalized Medicine, 13(6), 951. DOI:10.3390/jpm13060951, https://www.mdpi.com/2075-4426/13/6/951
  8. Li, F., Ruijs, N., & Lu, Y. (2022). Ethics & AI: A Systematic Review on Ethical Concerns and Related Strategies for Designing with AI in Healthcare. AI, 4(1), 28-53. DOI:10.3390/ai4010003, https://www.mdpi.com/2673-2688/4/1/3
  9. Li, H., Zhang, J., Zhang, N., & Zhu, B. (2025). Advancing Emergency Care With Digital Twins. JMIR Aging, 8, e71777. DOI:10.2196/71777, https://aging.jmir.org/2025/1/e71777/
  10. Smith, M. E., Zalesky, C. C., Lee, S., Gottlieb, M., Adhikari, S., Goebel, M., Wegman, M., Garg, N., Lam, S. H. F. (2025). Artificial Intelligence in Emergency Medicine: A Primer for the Nonexpert. JACEP Open, 6, 100051. DOI: 10.1016/j.acepjo.2025.100051, https://www.sciencedirect.com/science/article/pii/S2688115225000098

Further Reading

Last Updated: Sep 15, 2025



Source link

Continue Reading

Trending