Connect with us

AI Insights

AI shapes autonomous underwater “gliders” | MIT News

Published

on


Marine scientists have long marveled at how animals like fish and seals swim so efficiently despite having different shapes. Their bodies are optimized for efficient, hydrodynamic aquatic navigation so they can exert minimal energy when traveling long distances.

Autonomous vehicles can drift through the ocean in a similar way, collecting data about vast underwater environments. However, the shapes of these gliding machines are less diverse than what we find in marine life — go-to designs often resemble tubes or torpedoes, since they’re fairly hydrodynamic as well. Plus, testing new builds requires lots of real-world trial-and-error.

Researchers from MIT’s Computer Science and Artificial Intelligence Laboratory (CSAIL) and the University of Wisconsin at Madison propose that AI could help us explore uncharted glider designs more conveniently. Their method uses machine learning to test different 3D designs in a physics simulator, then molds them into more hydrodynamic shapes. The resulting model can be fabricated via a 3D printer using significantly less energy than hand-made ones.

The MIT scientists say that this design pipeline could create new, more efficient machines that help oceanographers measure water temperature and salt levels, gather more detailed insights about currents, and monitor the impacts of climate change. The team demonstrated this potential by producing two gliders roughly the size of a boogie board: a two-winged machine resembling an airplane, and a unique, four-winged object resembling a flat fish with four fins.

Peter Yichen Chen, MIT CSAIL postdoc and co-lead researcher on the project, notes that these designs are just a few of the novel shapes his team’s approach can generate. “We’ve developed a semi-automated process that can help us test unconventional designs that would be very taxing for humans to design,” he says. “This level of shape diversity hasn’t been explored previously, so most of these designs haven’t been tested in the real world.”

But how did AI come up with these ideas in the first place? First, the researchers found 3D models of over 20 conventional sea exploration shapes, such as submarines, whales, manta rays, and sharks. Then, they enclosed these models in “deformation cages” that map out different articulation points that the researchers pulled around to create new shapes.

The CSAIL-led team built a dataset of conventional and deformed shapes before simulating how they would perform at different “angles-of-attack” — the direction a vessel will tilt as it glides through the water. For example, a swimmer may want to dive at a -30 degree angle to retrieve an item from a pool.

These diverse shapes and angles of attack were then used as inputs for a neural network that essentially anticipates how efficiently a glider shape will perform at particular angles and optimizes it as needed.

Giving gliding robots a lift

The team’s neural network simulates how a particular glider would react to underwater physics, aiming to capture how it moves forward and the force that drags against it. The goal: find the best lift-to-drag ratio, representing how much the glider is being held up compared to how much it’s being held back. The higher the ratio, the more efficiently the vehicle travels; the lower it is, the more the glider will slow down during its voyage.

Lift-to-drag ratios are key for flying planes: At takeoff, you want to maximize lift to ensure it can glide well against wind currents, and when landing, you need sufficient force to drag it to a full stop.

Niklas Hagemann, an MIT graduate student in architecture and CSAIL affiliate, notes that this ratio is just as useful if you want a similar gliding motion in the ocean.

“Our pipeline modifies glider shapes to find the best lift-to-drag ratio, optimizing its performance underwater,” says Hagemann, who is also a co-lead author on a paper that was presented at the International Conference on Robotics and Automation in June. “You can then export the top-performing designs so they can be 3D-printed.”

Going for a quick glide

While their AI pipeline seemed realistic, the researchers needed to ensure its predictions about glider performance were accurate by experimenting in more lifelike environments.

They first fabricated their two-wing design as a scaled-down vehicle resembling a paper airplane. This glider was taken to MIT’s Wright Brothers Wind Tunnel, an indoor space with fans that simulate wind flow. Placed at different angles, the glider’s predicted lift-to-drag ratio was only about 5 percent higher on average than the ones recorded in the wind experiments — a small difference between simulation and reality.

A digital evaluation involving a visual, more complex physics simulator also supported the notion that the AI pipeline made fairly accurate predictions about how the gliders would move. It visualized how these machines would descend in 3D.

To truly evaluate these gliders in the real world, though, the team needed to see how their devices would fare underwater. They printed two designs that performed the best at specific points-of-attack for this test: a jet-like device at 9 degrees and the four-wing vehicle at 30 degrees.

Both shapes were fabricated in a 3D printer as hollow shells with small holes that flood when fully submerged. This lightweight design makes the vehicle easier to handle outside of the water and requires less material to be fabricated. The researchers placed a tube-like device inside these shell coverings, which housed a range of hardware, including a pump to change the glider’s buoyancy, a mass shifter (a device that controls the machine’s angle-of-attack), and electronic components.

Each design outperformed a handmade torpedo-shaped glider by moving more efficiently across a pool. With higher lift-to-drag ratios than their counterpart, both AI-driven machines exerted less energy, similar to the effortless ways marine animals navigate the oceans.

As much as the project is an encouraging step forward for glider design, the researchers are looking to narrow the gap between simulation and real-world performance. They are also hoping to develop machines that can react to sudden changes in currents, making the gliders more adaptable to seas and oceans.

Chen adds that the team is looking to explore new types of shapes, particularly thinner glider designs. They intend to make their framework faster, perhaps bolstering it with new features that enable more customization, maneuverability, or even the creation of miniature vehicles.

Chen and Hagemann co-led research on this project with OpenAI researcher Pingchuan Ma SM ’23, PhD ’25. They authored the paper with Wei Wang, a University of Wisconsin at Madison assistant professor and recent CSAIL postdoc; John Romanishin ’12, SM ’18, PhD ’23; and two MIT professors and CSAIL members: lab director Daniela Rus and senior author Wojciech Matusik. Their work was supported, in part, by a Defense Advanced Research Projects Agency (DARPA) grant and the MIT-GIST Program.



Source link

Continue Reading
Click to comment

Leave a Reply

Your email address will not be published. Required fields are marked *

AI Insights

The End of the Internet As We Know It

Published

on


The internet as we know it runs on clicks. Billions of them. They fuel ad revenue, shape search results, and dictate how knowledge is discovered, monetized, and, at times, manipulated. But a new wave of AI powered browsers is trying to kill the click. They’re coming for Google Chrome.

On Wednesday, the AI search startup Perplexity officially launched Comet, a web browser designed to feel more like a conversation than a scroll. Think of it as ChatGPT with a browser tab, but souped up to handle your tasks, answer complex questions, navigate context shifts, and satisfy your curiosity all at once.

Perplexity pitches Comet as your “second brain,” capable of actively researching, comparing options, making purchases, briefing you for your day, and analyzing information on your behalf. The promise is that it does all this without ever sending you off on a wild hyperlink chase across 30 tabs, aiming to collapse “complex workflows into fluid conversations.”

“Agentic AI”

The capabilities of browsers like Comet point to the rapid evolution of agentic AI. This is a cutting-edge field where AI systems are designed not just to answer questions or generate text, but to autonomously perform a series of actions and make decisions to achieve a user’s stated goal. Instead of you telling the browser every single step, an agentic browser aims to understand your intent and execute multi-step tasks, effectively acting as an intelligent assistant within the web environment. “Comet learns how you think, in order to think better with you,” Perplexity says.

Comet’s launch throws Perplexity into direct confrontation with the biggest gatekeeper of the internet: Google Chrome. For decades, Chrome has been the dominant gateway, shaping how billions navigate the web. Every query, every click, every ad. It’s all been filtered through a system built to maximize user interaction and, consequently, ad revenue. Comet is trying to blow that model up, fundamentally challenging the advertising-driven internet economy.

And it’s not alone in this ambitious assault. OpenAI, the maker of ChatGPT, is reportedly preparing to unveil its own AI powered web browser as early as next week, according to Reuters. This tool will likely integrate the power of ChatGPT with Operator, OpenAI’s proprietary web agent. Launched as a research preview in January 2025, OpenAI’s Operator is an AI agent capable of autonomously performing tasks through web browser interactions. It leverages OpenAI’s advanced models to navigate websites, fill out forms, place orders, and manage other repetitive browser-based tasks.

Operator is designed to “look” at web pages like a human, clicking, typing, and scrolling, aiming to eventually handle the “long tail” of digital use cases. If integrated fully into an OpenAI browser, it could create a full-stack alternative to Google Chrome and Google Search in one decisive move. In essence, OpenAI is coming for Google from both ends: the browser interface and the search functionality.

Goodbye clicks. Hello cognition

Perplexity’s pitch is simple and provocative: the web should respond to your thoughts, not interrupt them. “The internet has become humanity’s extended mind, while our tools for using it remain primitive,” the company stated in its announcement, advocating for an interface as fluid as human thought itself.

Instead of navigating through endless tabs and chasing hyperlinks, Comet promises to run on context. You can ask it to compare insurance plans. You can ask it to summarize a confusing sentence or instantly find that jacket you forgot to bookmark. Comet promises to “collapse entire workflows” into fluid conversations, turning what used to be a dozen clicks into a single, intuitive prompt.

If that sounds like the end of traditional Search Engine Optimization (SEO) and the death of the familiar “blue links” of search results, that’s because it very well could be. AI browsers like Comet don’t just threaten individual publishers and their traffic; they directly threaten the very foundation of Google Chrome’s ecosystem and Google Search’s dominance, which relies heavily on directing users to external websites.

Google’s Grip is Slipping

Google Search has already been under considerable pressure from AI native upstarts like Perplexity and You.com. Its own attempts at deeper AI integration, such as the Search Generative Experience (SGE), have drawn criticism for sometimes producing “hallucinations” (incorrect information) and awkward summaries. Simultaneously, Chrome, Google’s dominant browser, is facing its own identity crisis. It’s caught between trying to preserve its massive ad revenue pipeline and responding to a wave of AI powered alternatives that don’t rely on traditional links or clicks to deliver useful information.

Comet doesn’t just sidestep the old ad driven model, it fundamentally breaks it. There’s no need to sort through 10 blue links. No need to open 12 tabs to compare specifications, prices, or user reviews. With Comet, you just ask, and let the browser do the work.

OpenAI’s upcoming browser could deepen that transformative shift even further. If it is indeed designed to keep user interactions largely inside a ChatGPT-like interface instead of linking out, it could effectively create an entirely new, self-contained information ecosystem. In such a future, Google Chrome would no longer be the indispensable gateway for knowledge or commerce.

What’s at Stake: Redefining the Internet

If Comet or OpenAI’s browser succeed, the impact won’t be limited to just disrupting search. They will fundamentally redefine how the entire internet works. Publishers, advertisers, online retailers, and even traditional software companies may find themselves disintermediated—meaning their direct connection to users is bypassed—by AI agents. These intelligent agents could summarize their content, compare their prices, execute their tasks, and entirely bypass their existing websites and interfaces.

It’s a new, high-stakes front in the war for how humans interact with information and conduct their digital lives. The AI browser is no longer a hypothetical concept. It’s here.



Source link

Continue Reading

AI Insights

Google releases Gemma 3n models for on-device AI

Published

on


Google has released its Gemma 3n AI model, positioned as an advancement for on-device AI and bringing multimodal capabilities and higher performance to edge devices.

Previewed in May, Gemma 3n is multimodal by design, with native support for image, audio, video, and text inputs and outputs, Google said. Optimized for edge devices such as phones, tablets, laptops, desktops, or single cloud accelerators, Gemma 3n models are available in two sizes based on “effective” parameters, E2B and E4B. Whereas the raw parameter counts for E2B and E4B are 5B and 8B, respectively, these models run with a memory footprint comparable to traditional 2B and 4B models, running with as little as 2GB and 3GB of memory, Google said. 

Announced as a production release June 26, Gemma 3n models can be downloaded from Hugging Face and Kaggle. Developers also can try out Gemma 3n in Google AI Studio.



Source link

Continue Reading

AI Insights

Artificial intelligence drives the demand for the electric grid – Fox News

Published

on



Artificial intelligence drives the demand for the electric grid  Fox News



Source link

Continue Reading

Trending