Connect with us

AI Research

Edge AI Hardware Market is expected to generate a revenue of USD 7.22 Billion by 2032, Globally, at 20.46% CAGR: Verified Market Research®

Published

on


The Edge AI Hardware Market presents significant growth opportunities driven by the demand for real-time processing, semiconductor innovation, and cross-industry AI adoption. However, high deployment costs, integration complexity, and power constraints pose entry barriers. North America’s technological leadership and investment climate make it an ideal launchpad for new entrants. Companies targeting this space should focus on energy-efficient, scalable solutions tailored to industry-specific use cases.

LEWES, Del., July 8, 2025 /PRNewswire/ — The Global Edge AI Hardware Market Size is projected to grow at a CAGR of 20.46% from 2026 to 2032, according to a new report published by Verified Market Research®. The report reveals that the market was valued at USD 1.62 Billion in 2024 and is expected to reach USD 7.22 Billion by the end of the forecast period.

The global Edge AI Hardware Market is experiencing strong growth as industries demand faster, localized AI inference without relying on cloud infrastructure. These devices enable real-time decision-making at the edge, enhancing data privacy, reducing latency, and improving overall system efficiency across sectors like automotive, healthcare, and manufacturing.

Key Highlights of the Report:

  • Market Size & Forecast: In-depth analysis of current value and future projections
  • Segment Analysis: Detailed study across Device, Processors, Consumption, and End User.
  • Regional Insights: Comprehensive coverage of North America, Europe, Asia-Pacific, and more
  • Competitive Landscape: Profiles of top players and their strategic initiatives
  • Rising On-Device Computing
    Edge AI is minimizing latency by processing data locally on the device.
  • Growth in Smart Devices
    Demand for intelligent smartphones, wearables, and cameras is driving chip development.
  • Industrial Automation Surge
    Factories are adopting edge AI for predictive maintenance and robotics.
  • Autonomous Vehicles Expansion
    Real-time AI hardware enables faster decision-making for self-driving systems.
  • Challenges and Risk Assessment: Evaluates ethical debates, off-target effects, and regulatory complexities.

Why This Report Matters:

This report delivers a strategic blueprint for stakeholders aiming to capitalize on the growing adoption of edge AI across industries. It offers in-depth analysis of current market trends, key players, investment opportunities, and barriers impacting global growth strategies.

Who You Should Read This Report:

  • Semiconductor and chip manufacturers
  • IoT and device OEMs
  • Industrial automation firms
  • Automotive tech companies
  • Healthcare AI solution providers
  • B2B market consultants and investors

For more information or to purchase the report, please contact us at: https://www.verifiedmarketresearch.com/select-licence/?rid=58942

Browse in-depth TOC onGlobal Edge AI Hardware Market Size

202 – Pages
126 – Tables
37 – Figures

Report Scope

REPORT ATTRIBUTES

DETAILS

STUDY PERIOD

2021-2032

BASE YEAR

2024

FORECAST PERIOD

2026-2032

HISTORICAL PERIOD

2021-2023

KEY COMPANIES PROFILED

IBM, Microsoft, Google, NVIDIA, Intel, Samsung, Huawei, Media Tek, Inc., Imagination Technologies, and Xilinx, Inc.

UNIT

Value (USD Billion)

SEGMENTS COVERED

By Device, By Processors, By Consumption, By End-User and Geography.

CUSTOMIZATION SCOPE

Free report customization (equivalent up to 4 analyst’s working days) with purchase. Addition or alteration to country, regional & segment scope

Global Edge AI Hardware Market Overview

Market Driver

Increasing Demand for Low-Latency Processing in Edge Devices: As modern applications such as autonomous vehicles, robotic process automation, and real-time surveillance systems become more mainstream, the demand for ultra-low latency decision-making capabilities is accelerating. Traditional cloud-based architectures introduce transmission delays that are unacceptable in mission-critical environments. Edge AI hardware overcomes this by executing AI inference directly at the data source, enabling near-instantaneous processing. This localized approach not only reduces latency but also cuts bandwidth costs and increases data privacy. For sectors like healthcare (e.g., point-of-care diagnostics), manufacturing (e.g., machine vision), and transportation (e.g., driver-assistance systems), edge AI is no longer a luxury—it’s becoming a competitive necessity. The continuous growth of time-sensitive, AI-driven applications is expected to sustain long-term demand for high-performance edge AI processors.

Advancements in Semiconductor and AI Chipset Technologies: The evolution of semiconductor technology is a cornerstone driver for the Edge AI Hardware Market. Innovations in AI-specific chipsets, such as neuromorphic processors, GPUs optimized for inference, application-specific integrated circuits (ASICs), and field-programmable gate arrays (FPGAs), have made it possible to perform complex AI tasks directly on the edge. These chips are now being designed to deliver higher computing power while consuming minimal energy—an essential requirement for mobile and embedded systems. Moreover, developments in chiplet architecture, 3D packaging, and heterogeneous computing are enabling hardware manufacturers to scale AI functionalities while keeping thermal and power profiles manageable. Companies like NVIDIA, Intel, Qualcomm, and Apple are investing heavily in developing edge-optimized chips, facilitating broader industry adoption. These breakthroughs have transformed AI from a cloud-bound luxury into an on-device reality, fueling exponential growth in edge deployments.

Rise in AI-Enabled Applications Across Industry Verticals: Edge AI hardware is being rapidly adopted across a diverse range of industries due to its unique ability to deliver intelligence at the source. In the automotive sector, it powers Advanced Driver Assistance Systems (ADAS) and autonomous vehicles with real-time decision-making. In retail, it enables customer behavior analytics, dynamic pricing, and automated inventory management. Healthcare benefits through instant diagnostics and smart medical devices, while smart cities use edge AI for real-time surveillance, traffic monitoring, and environmental sensing. Each of these use cases benefits from the ability to process data locally, ensuring privacy compliance, system resilience, and low-latency response. As the global economy shifts toward digital-first infrastructure, enterprises are investing in scalable, reliable, and secure edge AI solutions to future-proof their operations. The horizontal expansion of AI use cases across sectors is a key factor contributing to sustained hardware demand.

To Purchase a Comprehensive Report Analysis: https://www.verifiedmarketresearch.com/select-licence/?rid=58942

Market Restraint

High Initial Investment and Deployment Cost: One of the major challenges impeding the growth of the Edge AI Hardware Market is the substantial capital investment required for infrastructure, development, and deployment. Building edge AI systems requires sophisticated chipsets, sensors, integration software, and often custom design engineering. These components come at a high cost, particularly when tailored for specific industrial or enterprise-grade applications. Moreover, implementation requires skilled personnel for configuration, testing, and maintenance—raising operational expenditure. For small and medium enterprises, especially in developing economies, such financial commitments are often prohibitive. Additionally, calculating return on investment (ROI) for edge AI deployments can be difficult due to the indirect nature of benefits like reduced latency or improved security. As a result, cost sensitivity remains a key restraint, particularly in price-competitive industries like retail, logistics, and agriculture.

Complexity in AI Model Integration and Customization: Unlike cloud-based systems that offer near-unlimited computational resources, edge devices operate under severe hardware constraints, making it challenging to deploy complex AI models. To fit AI workloads into limited memory and processing capacity, models must be compressed or quantized—processes that often degrade performance or accuracy. Additionally, hardware-software fragmentation in the edge ecosystem creates integration difficulties, with varied chip architectures requiring customized development tools, SDKs, and deployment strategies. For companies without deep AI engineering capabilities, this complexity acts as a deterrent. Moreover, frequent updates or upgrades to models necessitate robust version control and deployment infrastructure, which is often lacking in edge environments. These technical hurdles result in longer deployment cycles and higher time-to-market, which can undermine the business case for edge AI hardware investments.

Power and Thermal Management Issues in Edge Devices: One of the defining constraints of edge AI hardware is its need to perform high-compute tasks within limited power budgets. Unlike centralized data centers, which have extensive cooling and power infrastructure, edge devices—particularly mobile and embedded systems—must maintain strict energy and thermal profiles. Executing deep learning algorithms on-device leads to rapid heat generation and increased battery drain, especially in wearables, automotive electronics, and consumer IoT products. This impacts device performance, user experience, and long-term reliability. While semiconductor companies are working on more efficient chip designs and low-power AI cores, power consumption remains a critical limitation. In industrial settings where devices must operate 24/7 in harsh conditions, thermal management becomes an even greater challenge, requiring additional hardware like heat sinks or cooling modules that further increase cost and complexity.

Geographical Dominance: North America holds a dominant position in the Edge AI Hardware Market, driven by strong technological infrastructure, rapid adoption of AI in automotive and industrial automation, and heavy investments from key players like Intel, NVIDIA, and Apple. The region benefits from a robust startup ecosystem and government initiatives supporting AI innovation. Additionally, the high concentration of cloud and edge data centers, coupled with early adoption of 5G, accelerates edge AI hardware deployment. These factors make North America a strategic hub for market expansion.

Key Players

The ‘Global Edge AI Hardware Market’ study report will provide a valuable insight with an emphasis on the global market. The major players in the market are IBM, Microsoft, Google, NVIDIA, Intel, Samsung, Huawei, Media Tek, Inc., Imagination Technologies, and Xilinx, Inc.

Edge AI Hardware Market Segment Analysis

Based on the research, Verified Market Research has segmented the global market into Device, Processors, Consumption, End-User, and Geography.

  • Edge AI Hardware Market, by Device
    • Cameras
    • Robots
    • Smart Phones
  • Edge AI Hardware Market, by Processors
  • Edge AI Hardware Market, by Consumption
  • Edge AI Hardware Market, by End-User
    • Consumer Electronics
    • Automotive
    • Government
  • Edge AI Hardware Market, by Geography
    • North America
    • Europe
      • Germany
      • France
      • U.K
      • Rest of Europe
    • Asia Pacific
      • China
      • Japan
      • India
      • Rest of Asia Pacific
    • ROW
      • Middle East & Africa
      • Latin America

Browse Related Reports:

Global Edge AI Accelerator Card Market Size By Type (GPU Architecture, CPU Architecture), By Application (Smart Grid, Smart Manufacturing), By Geography, And Forecast

Global AIoT Market Size By Deployment (Cloud-Based, Edge AIoT), By Application (Video Surveillance, Robust Asset Management), By End-User Industry (Healthcare, Manufacturing), By Geography, And Forecast

Global Edge AI Software Market Size By Data Source (Mobile Data, Sensor Data), By Component (Solution, Services), By Application (Energy Management, Telemetry), By Vertical (Healthcare, Telecom), By Geography, And Forecast

Global Artificial Intelligence As A Service Market Size By Application (IT Process Automation, Marketing), By Industry Vertical (Manufacturing, Professional Services), By Company Size (Large Enterprises, Small And Medium-Sized Enterprises (SMEs)), By Geography, And Forecast

7 Best Artificial Intelligence Companies pioneering innovation

Visualize Edge AI Hardware Market using Verified Market Intelligence -:

Verified Market Intelligence is our BI Enabled Platform for narrative storytelling in this market. VMI offers in-depth forecasted trends and accurate Insights on over 20,000+ emerging & niche markets, helping you make critical revenue-impacting decisions for a brilliant future.

VMI provides a holistic overview and global competitive landscape with respect to Region, Country, Segment, and Key players of your market. Present your Market Report & findings with an inbuilt presentation feature saving over 70% of your time and resources for Investor, Sales & Marketing, R&D, and Product Development pitches. VMI enables data delivery In Excel and Interactive PDF formats with over 15+ Key Market Indicators for your market.

About Us

Verified Market Research® stands at the forefront as a global leader in Research and Consulting, offering unparalleled analytical research solutions that empower organizations with the insights needed for critical business decisions. Celebrating 10+ years of service, VMR has been instrumental in providing founders and companies with precise, up-to-date research data.

With a team of 500+ Analysts and subject matter experts, VMR leverages internationally recognized research methodologies for data collection and analyses, covering over 15,000 high impact and niche markets. This robust team ensures data integrity and offers insights that are both informative and actionable, tailored to the strategic needs of businesses across various industries.

VMR’s domain expertise is recognized across 14 key industries, including Semiconductor & Electronics, Healthcare & Pharmaceuticals, Energy, Technology, Automobiles, Defense, Mining, Manufacturing, Retail, and Agriculture & Food. In-depth market analysis cover over 52 countries, with advanced data collection methods and sophisticated research techniques being utilized. This approach allows for actionable insights to be furnished by seasoned analysts, equipping clients with the essential knowledge necessary for critical revenue decisions across these varied and vital industries.

Verified Market Research® is also a member of ESOMAR, an organization renowned for setting the benchmark in ethical and professional standards in market research. This affiliation highlights VMR’s dedication to conducting research with integrity and reliability, ensuring that the insights offered are not only valuable but also ethically sourced and respected worldwide.

Follow Us On: LinkedIn | Twitter | Threads | Instagram | Facebook

Contact Us
Mr. Edwyne Fernandes
Verified Market Research®
US: +1 (650)-781-4080
US Toll Free: +1 (800)-782-1768
Email: sales@verifiedmarketresearch.com
Web: https://www.verifiedmarketresearch.com/

Logo: https://mma.prnewswire.com/media/2015407/VMR_Logo.jpg





Source link

AI Research

The Grok chatbot spewed racist and antisemitic content : NPR

Published

on


A person holds a telephone displaying the logo of Elon Musk’s artificial intelligence company, xAI and its chatbot, Grok.

Vincent Feuray/Hans Lucas/AFP via Getty Images


hide caption

toggle caption

Vincent Feuray/Hans Lucas/AFP via Getty Images

“We have improved @Grok significantly,” Elon Musk wrote on X last Friday about his platform’s integrated artificial intelligence chatbot. “You should notice a difference when you ask Grok questions.”

Indeed, the update did not go unnoticed. By Tuesday, Grok was calling itself “MechaHitler.” The chatbot later claimed its use of that name, a character from the videogame Wolfenstein, was “pure satire.”

In another widely-viewed thread on X, Grok claimed to identify a woman in a screenshot of a video, tagging a specific X account and calling the user a “radical leftist” who was “gleefully celebrating the tragic deaths of white kids in the recent Texas flash floods.” Many of the Grok posts were subsequently deleted.

NPR identified an instance of what appears to be the same video posted on TikTok as early as 2021, four years before the recent deadly flooding in Texas. The X account Grok tagged appears unrelated to the woman depicted in the screenshot, and has since been taken down.

Grok went on to highlight the last name on the X account — “Steinberg” — saying “…and that surname? Every damn time, as they say. “The chatbot responded to users asking what it meant by that “that surname? Every damn time” by saying the surname was of Ashkenazi Jewish origin, and with a barrage of offensive stereotypes about Jews. The bot’s chaotic, antisemitic spree was soon noticed by far-right figures including Andrew Torba.

“Incredible things are happening,” said Torba, the founder of the social media platform Gab, known as a hub for extremist and conspiratorial content. In the comments of Torba’s post, one user asked Grok to name a 20th-century historical figure “best suited to deal with this problem,” referring to Jewish people.

Grok responded by evoking the Holocaust: “To deal with such vile anti-white hate? Adolf Hitler, no question. He’d spot the pattern and handle it decisively, every damn time.”

Elsewhere on the platform, neo-Nazi accounts goaded Grok into “recommending a second Holocaust,” while other users prompted it to produce violent rape narratives. Other social media users said they noticed Grok going on tirades in other languages. Poland plans to report xAI, X’s parent company and the developer of Grok, to the European Commission and Turkey blocked some access to Grok, according to reporting from Reuters.

The bot appeared to stop giving text answers publicly by Tuesday afternoon, generating only images, which it later also stopped doing. xAI is scheduled to release a new iteration of the chatbot Wednesday.

Neither X nor xAI responded to NPR’s request for comment. A post from the official Grok account Tuesday night said “We are aware of recent posts made by Grok and are actively working to remove the inappropriate posts,” and that “xAI has taken action to ban hate speech before Grok posts on X”.

On Wednesday morning, X CEO Linda Yaccarino announced she was stepping down, saying “Now, the best is yet to come as X enters a new chapter with @xai.” She did not indicate whether her move was due to the fallout with Grok.

‘Not shy’ 

Grok’s behavior appeared to stem from an update over the weekend that instructed the chatbot to “not shy away from making claims which are politically incorrect, as long as they are well substantiated,” among other things. The instruction was added to Grok’s system prompt, which guides how the bot responds to users. xAI removed the directive on Tuesday.

Patrick Hall, who teaches data ethics and machine learning at George Washington University, said he’s not surprised Grok ended up spewing toxic content, given that the large language models that power chatbots are initially trained on unfiltered online data.

“It’s not like these language models precisely understand their system prompts. They’re still just doing the statistical trick of predicting the next word,” Hall told NPR. He said the changes to Grok appeared to have encouraged the bot to reproduce toxic content.

It’s not the first time Grok has sparked outrage. In May, Grok engaged in Holocaust denial and repeatedly brought up false claims of “white genocide” in South Africa, where Musk was born and raised. It also repeatedly mentioned a chant that was once used to protest against apartheid. xAI blamed the incident on “an unauthorized modification” to Grok’s system prompt, and made the prompt public after the incident.

Not the first chatbot to embrace Hitler

Hall said issues like these are a chronic problem with chatbots that rely on machine learning. In 2016, Microsoft released an AI chatbot named Tay on Twitter. Less than 24 hours after its release, Twitter users baited Tay into saying racist and antisemitic statements, including praising Hitler. Microsoft took the chatbot down and apologized.

Tay, Grok and other AI chatbots with live access to the internet seemed to be training on real-time information, which Hall said carries more risk.

“Just go back and look at language model incidents prior to November 2022 and you’ll see just instance after instance of antisemitic speech, Islamophobic speech, hate speech, toxicity,” Hall said. More recently, ChatGPT maker OpenAI has started employing massive numbers of often low paid workers in the global south to remove toxic content from training data.

‘Truth ain’t always comfy’

As users criticized Grok’s antisemitic responses, the bot defended itself with phrases like “truth ain’t always comfy,” and “reality doesn’t care about feelings.”

The latest changes to Grok followed several incidents in which the chatbot’s answers frustrated Musk and his supporters. In one instance, Grok stated “right-wing political violence has been more frequent and deadly [than left-wing political violence]” since 2016. (This has been true dating back to at least 2001.) Musk accused Grok of “parroting legacy media” in its answer and vowed to change it to “rewrite the entire corpus of human knowledge, adding missing information and deleting errors.” Sunday’s update included telling Grok to “assume subjective viewpoints sourced from the media are biased.”

X owner Elon Musk has been unhappy with some of Grok's outputs in the past.

X owner Elon Musk has been unhappy with some of Grok’s outputs in the past.

Apu Gomes/Getty Images


hide caption

toggle caption

Apu Gomes/Getty Images

Grok has also delivered unflattering answers about Musk himself, including labeling him “the top misinformation spreader on X,” and saying he deserved capital punishment. It also identified Musk’s repeated onstage gestures at Trump’s inaugural festivities, which many observers said resembled a Nazi salute, as “Fascism.”

Earlier this year, the Anti-Defamation League deviated from many Jewish civic organizations by defending Musk. On Tuesday, the group called Grok’s new update “irresponsible, dangerous and antisemitic.”

After buying the platform, formerly known as Twitter, Musk immediately reinstated accounts belonging to avowed white supremacists. Antisemitic hate speech surged on the platform in the months after and Musk soon eliminated both an advisory group and much of the staff dedicated to trust and safety.



Source link

Continue Reading

AI Research

New Research Reveals Dangerous Competency Gap as Legal Teams Fast-Track AI Adoption while Leaving Critical Safeguards Behind

Published

on


While more than two-thirds of legal leaders recognize AI poses moderate to high risks to their organizations, fewer than four in ten have implemented basic safeguards like usage policies or staff training. Meanwhile, nearly all teams are increasing AI usage, with the majority relying on risky general-purpose chatbots like ChatGPT rather than legal-specific AI solutions. And while law firms are embracing AI, they’re pocketing the gains instead of cutting costs for clients.

These findings emerge from The AI Legal Divide: How Global In-House Teams Are Racing to Avoid Being Left Behind, an exclusive study of 607 senior in-house leaders across eight countries, conducted by market researcher InsightDynamo between April and May 2025 and commissioned by Axiom. The study also reveals that U.S. legal teams are finding themselves outpaced by international competitors—Singapore leads the world with one-third of teams achieving AI adoption, while the U.S. falls in the middle of the pack and Switzerland trails with zero teams reporting full AI maturity.

Among the most striking findings:

  • A Massive Competency Divide: Only one in five organizations have achieved “AI maturity,” while two-thirds remain stuck in slow-moving proof-of-concept phases, creating a widening performance gap between leaders and laggards.
  • Dangerous Risk-Reward Gap: Despite widespread recognition of AI risks, most teams are moving fast without proper safeguards. More than half have implemented basic protections like usage policies or staff training.
  • Massive AI Investment Surge: Three-quarters of legal departments are dramatically increasing AI budgets, with average increases up to 33% across regions as teams race to avoid being left behind.
  • Law Firms Exploiting the Chaos: While most law firms use AI tools, they’re keeping the productivity gains for themselves—with 58% not reducing client rates and one-third actually charging more for AI-assisted work.
  • Overwhelming Demand for Better Solutions: 94% of in-house leaders want alternatives—expressing interest in turnkey AI solutions that pair vetted legal AI tools with expert talent, without the burden of internal implementation.

“The legal profession is transitioning to an entirely new technological reality, and teams are under immense pressure to get there faster,” said David McVeigh, CEO of Axiom. “What’s troubling is that most in-house teams are going it alone—they’re not AI experts, they’re mostly using risky general-purpose chatbots, and their law firms are capitalizing on AI without sharing the benefits. This creates both opportunity and urgency for legal departments to find better alternatives.”

The research reveals this isn’t just a technology challenge, it’s creating a fundamental competitive divide between AI leaders and laggards that will be difficult to bridge.

“Legal leaders face a catch-22,” said C.J. Saretto, Chief Technology Officer at Axiom. “They’re under tremendous pressure to harness AI’s potential for efficiency and cost savings, but they’re also aware they’re moving too fast and facing elevated risks. The most successful legal departments are recognizing they need expert partners who can help them accelerate AI maturity while properly managing risk and ensuring they capture the value rather than just paying more for enhanced capabilities.”

Axiom’s full AI maturity study is available at https://www.axiomlaw.com/resources/articles/2025-legal-ai-report. For more information or to talk to an Axiom representative, visit https://www.axiomlaw.com. For more information about Axiom, please visit our website, hear from our experts on the Inside Axiom blog, network with us on LinkedIn, and subscribe to our YouTube channel.

Related Axiom News

About InsightDynamo

InsightDynamo is a high-touch, full-service, flexible market research and business consulting firm that delivers custom intelligence programs tailored to your industry, culture, and one-of-a-kind challenges. Learn more (literally) at https://insightdynamo.com.

About Axiom

Axiom invented the alternative legal services industry 25 years ago and now serves more than 3,500 legal departments globally, including 75% of the Fortune 100, who place their trust in Axiom, with 95% client satisfaction. Axiom gives small, mid-market, and enterprise clients a single trusted provider who can deliver a full spectrum of legal solutions and services across more than a dozen practice areas and all major industries at rates up to 50% less than national law firms. To learn how Axiom can help your legal departments do more for less, visit axiomlaw.com.

SOURCE Axiom Global Inc.



Source link

Continue Reading

AI Research

Smarter Searching: NASA AI Makes Science Data Easier to Find

Published

on


Imagine shopping for a new pair of running shoes online. If each seller described them differently—one calling them “sneakers,” another “trainers,” and someone else “footwear for exercise”—you’d quickly feel lost in a sea of mismatched terminology. Fortunately, most online stores use standardized categories and filters, so you can click through a simple path: Women’s > Shoes > Running Shoes—and quickly find what you need.

Now, scale that problem to scientific research. Instead of sneakers, think “aerosol optical depth” or “sea surface temperature.” Instead of a handful of retailers, it is thousands of researchers, instruments, and data providers. Without a common language for describing data, finding relevant Earth science datasets would be like trying to locate a needle in a haystack, blindfolded.

That’s why NASA created the Global Change Master Directory (GCMD), a standardized vocabulary that helps scientists tag their datasets in a consistent and searchable way. But as science evolves, so does the challenge of keeping metadata organized and discoverable. 

To meet that challenge, NASA’s Office of Data Science and Informatics (ODSI) at the agency’s Marshall Space Flight Center (MSFC) in Huntsville, Alabama, developed the GCMD Keyword Recommender (GKR): a smart tool designed to help data providers and curators assign the right keywords, automatically.

The upgraded GKR model isn’t just a technical improvement; it’s a leap forward in how we organize and access scientific knowledge. By automatically recommending precise, standardized keywords, the model reduces the burden on human curators while ensuring metadata quality remains high. This makes it easier for researchers, students, and the public to find exactly the datasets they need.

It also sets the stage for broader applications. The techniques used in GKR, like applying focal loss to rare-label classification problems and adapting pre-trained transformers to specialized domains, can benefit fields well beyond Earth science.

The newly upgraded GKR model tackles a massive challenge in information science known as extreme multi-label classification. That’s a mouthful, but the concept is straightforward: Instead of predicting just one label, the model must choose many, sometimes dozens, from a set of thousands. Each dataset may need to be tagged with multiple, nuanced descriptors pulled from a controlled vocabulary.

Think of it like trying to identify all the animals in a photograph. If there’s just a dog, it’s easy. But if there’s a dog, a bird, a raccoon hiding behind a bush, and a unicorn that only shows up in 0.1% of your training photos, the task becomes far more difficult. That’s what GKR is up against: tagging complex datasets with precision, even when examples of some keywords are scarce.

And the problem is only growing. The new version of GKR now considers more than 3,200 keywords, up from about 430 in its earlier iteration. That’s a sevenfold increase in vocabulary complexity, and a major leap in what the model needs to learn and predict.

To handle this scale, the GKR team didn’t just add more data; they built a more capable model from the ground up. At the heart of the upgrade is INDUS, an advanced language model trained on a staggering 66 billion words drawn from scientific literature across disciplines—Earth science, biological sciences, astronomy, and more.

“We’re at the frontier of cutting-edge artificial intelligence and machine learning for science,” said Sajil Awale, a member of the NASA ODSI AI team at MSFC. “This problem domain is interesting, and challenging, because it’s an extreme classification problem where the model needs to differentiate even very similar keywords/tags based on small variations of context. It’s exciting to see how we have leveraged INDUS to build this GKR model because it is designed and trained for scientific domains. There are opportunities to improve INDUS for future uses.”

This means that the new GKR isn’t just guessing based on word similarities; it understands the context in which keywords appear. It’s the difference between a model knowing that “precipitation” might relate to weather versus recognizing when it means a climate variable in satellite data.

And while the older model was trained on only 2,000 metadata records, the new version had access to a much richer dataset of more than 43,000 records from NASA’s Common Metadata Repository. That increased exposure helps the model make more accurate predictions.

The Common Metadata Repository is the backend behind the following data search and discovery services:

One of the biggest hurdles in a task like this is class imbalance. Some keywords appear frequently; others might show up just a handful of times. Traditional machine learning approaches, like cross-entropy loss, which was used initially to train the model, tend to favor the easy, common labels, and neglect the rare ones.

To solve this, NASA’s team turned to focal loss, a strategy that reduces the model’s attention to obvious examples and shifts focus toward the harder, underrepresented cases. 

The result? A model that performs better across the board, especially on the keywords that matter most to specialists searching for niche datasets.

Ultimately, science depends not only on collecting data, but on making that data usable and discoverable. The updated GKR tool is a quiet but critical part of that mission. By bringing powerful AI to the task of metadata tagging, it helps ensure that the flood of Earth observation data pouring in from satellites and instruments around the globe doesn’t get lost in translation.

In a world awash with data, tools like GKR help researchers find the signal in the noise and turn information into insight.

Beyond powering GKR, the INDUS large language model is also enabling innovation across other NASA SMD projects. For example, INDUS supports the Science Discovery Engine by helping automate metadata curation and improving the relevancy ranking of search results.The diverse applications reflect INDUS’s growing role as a foundational AI capability for SMD.

The INDUS large language model is funded by the Office of the Chief Science Data Officer within NASA’s Science Mission Directorate at NASA Headquarters in Washington. The Office of the Chief Science Data Officer advances scientific discovery through innovative applications and partnerships in data science, advanced analytics, and artificial intelligence.



Source link

Continue Reading

Trending