Connect with us

AI Research

Watermarking AI-generated text and video with SynthID

Published

on


Science

Published

Announcing our novel watermarking method for AI-generated text and video, and how we’re bringing SynthID to key Google products

Generative AI tools — and the large language model technologies behind them — have captured the public imagination. From helping with work tasks to enhancing creativity, these tools are quickly becoming part of products that are used by millions of people in their daily lives.

These technologies can be hugely beneficial but as they become increasingly popular to use, the risk increases of people causing accidental or intentional harms, like spreading misinformation and phishing, if AI-generated content isn’t properly identified. That’s why last year, we launched SynthID, our novel digital toolkit for watermarking AI-generated content.

Today, we’re expanding SynthID’s capabilities to watermarking AI-generated text in the Gemini app and web experience, and video in Veo, our most capable generative video model.

SynthID for text is designed to complement most widely-available AI text generation models and for deploying at scale, while SynthID for video builds upon our image and audio watermarking method to include all frames in generated videos. This innovative method embeds an imperceptible watermark without impacting the quality, accuracy, creativity or speed of the text or video generation process.

SynthID isn’t a silver bullet for identifying AI generated content, but is an important building block for developing more reliable AI identification tools and can help millions of people make informed decisions about how they interact with AI-generated content. Later this summer, we’re planning to open-source SynthID for text watermarking, so developers can build with this technology and incorporate it into their models.

How text watermarking works

Large language models generate sequences of text when given a prompt like, “Explain quantum mechanics to me like I’m five” or “What’s your favorite fruit?”. LLMs predict which token most likely follows another, one token at a time.

Tokens are the building blocks a generative model uses for processing information. In this case, they can be a single character, word or part of a phrase. Each possible token is assigned a score, which is the percentage chance of it being the right one. Tokens with higher scores are more likely to be used. LLMs repeat these steps to build a coherent response.

SynthID is designed to embed imperceptible watermarks directly into the text generation process. It does this by introducing additional information in the token distribution at the point of generation by modulating the likelihood of tokens being generated — all without compromising the quality, accuracy, creativity or speed of the text generation.

SynthID adjusts the probability score of tokens generated by a large language model.

The final pattern of scores for both the model’s word choices combined with the adjusted probability scores are considered the watermark. This pattern of scores is compared with the expected pattern of scores for watermarked and unwatermarked text, helping SynthID detect if an AI tool generated the text or if it might come from other sources.

A piece of text generated by Gemini with the watermark highlighted in blue.

The benefits and limitations of this technique

SynthID for text watermarking works best when a language model generates longer responses, and in diverse ways — like when it’s prompted to generate an essay, a theater script or variations on an email.

It performs well even under some transformations, such as cropping pieces of text, modifying a few words and mild paraphrasing. However, its confidence scores can be greatly reduced when an AI-generated text is thoroughly rewritten or translated to another language.

SynthID text watermarking is less effective on responses to factual prompts because there are fewer opportunities to adjust the token distribution without affecting the factual accuracy. This includes prompts like “What is the capital of France?” or queries where little or no variation is expected like “recite a William Wordsworth poem”.

Many currently available AI detection tools use algorithms for labeling and sorting data, known as classifiers. These classifiers often only perform well on particular tasks, which makes them less flexible. When the same classifier is applied across different types of platforms and content, its performance isn’t always reliable or consistent. This can lead to a text being mislabeled, which can cause problems, for example, where text might be incorrectly identified as AI-generated.

SynthID works effectively on its own, but it can also be combined with other AI detection approaches to give better coverage across content types and platforms. While this technique isn’t built to directly stop motivated adversaries like cyberattackers or hackers from causing harm, it can make it harder to use AI-generated content for malicious purposes.

How video watermarking works

At this year’s I/O we announced Veo, our most capable generative video model. While video generation technologies aren’t as widely available as image generation technologies, they’re rapidly evolving and it’ll become increasingly important to help people know if a video is generated by an AI or not.

Videos are composed of individual frames or still images. So we developed a watermarking technique inspired by our SynthID for image tool. This technique embeds a watermark directly into the pixels of every video frame, making it imperceptible to the human eye, but detectable for identification.

Empowering people with knowledge of when they’re interacting with AI-generated media can play an important role in helping prevent the spread of misinformation. Starting today, all videos generated by Veo on VideoFX will be watermarked by SynthID.

SynthID for video watermarking marks every frame of a generated video

Bringing SynthID to the broader AI ecosystem

SynthID’s text watermarking technology is designed to be compatible with most AI text generation models and for scaling across different content types and platforms. To help prevent widespread misuse of AI-generated content, we’re working on bringing this technology to the broader AI ecosystem.

This summer, we’re planning to publish more about our text watermarking technology in a detailed research paper, and we’ll open-source SynthID text watermarking through our updated Responsible Generative AI Toolkit, which provides guidance and essential tools for creating safer AI applications, so developers can build with this technology and incorporate it into their models.

Acknowledgements

The SynthID text watermarking project was led by Sumanth Dathathri and Pushmeet Kohli, with key research and engineering contributions from (listed alphabetically): Vandana Bachani, Sumedh Ghaisas, Po-Sen Huang, Rob McAdam, Abi See and Johannes Welbl.

Thanks to Po-Sen Huang and Johannes Welbl for helping initiate the project. Thanks to Brad Hekman, Cip Baetu, Nir Shabat, Niccolò Dal Santo, Valentin Anklin and Majd Al Merey for collaborating on product integration; Borja Balle, Rudy Bunel, Taylan Cemgil, Sven Gowal, Jamie Hayes, Alex Kaskasoli, Ilia Shumailov, Tatiana Matejovicova and Robert Stanforth for technical input and feedback. Thanks also to many others who contributed across Google DeepMind and Google, including our partners at Gemini and CoreML.

The SynthID video watermarking project was led by Sven Gowal and Pushmeet Kohli, with key contributions from (listed alphabetically): Rudy Bunel, Christina Kouridi, Guillermo Ortiz-Jimenez, Sylvestre-Alvise Rebuffi, Florian Stimberg and David Stutz. Additional thanks to Jamie Hayes and others listed above.

Thanks to Nidhi Vyas and Zahra Ahmed for driving SynthID product delivery.



Source link

Continue Reading
Click to comment

Leave a Reply

Your email address will not be published. Required fields are marked *

AI Research

If I Could Only Buy 1 Artificial Intelligence (AI) Chip Stock Over The Next 10 Years, This Would Be It (Hint: It’s Not Nvidia)

Published

on


While Nvidia continues to capture headlines, a critical enabler of the artificial intelligence (AI) infrastructure boom may be better positioned for long-term gains.

When investors debate the future of the artificial intelligence (AI) trade, the conversation generally finds its way back to the usual suspects: Nvidia, Advanced Micro Devices, and cloud hyperscalers like Microsoft, Amazon, and Alphabet.

Each of these companies is racing to design GPUs or develop custom accelerators in-house. But behind this hardware, there’s a company that benefits no matter which chip brand comes out ahead: Taiwan Semiconductor Manufacturing (TSM -3.05%).

Let’s unpack why Taiwan Semi is my top AI chip stock over the next 10 years, and assess whether now is an opportune time to scoop up some shares.

Agnostic to the winner, leveraged to the trend

As the world’s leading semiconductor foundry, TSMC manufactures chips for nearly every major AI developer — from Nvidia and AMD to Amazon’s custom silicon initiatives, dubbed Trainium and Inferentia.

Unlike many of its peers in the chip space that rely on new product cycles to spur demand, Taiwan Semi’s business model is fundamentally agnostic. Whether demand is allocated toward GPUs, accelerators, or specialized cloud silicon, all roads lead back to TSMC’s fabrication capabilities.

With nearly 70% market share in the global foundry space, Taiwan Semi’s dominance is hard to ignore. Such a commanding lead over the competition provides the company with unmatched structural demand visibility — a trend that appears to be accelerating as AI infrastructure spend remains on the rise.

Image source: Getty Images.

Scaling with more sophisticated AI applications

At the moment, AI development is still concentrated on training and refining large language models (LLMs) and embedding them into downstream software applications.

The next wave of AI will expand into far more diverse and demanding use cases — autonomous systems, robotics, and quantum computing remain in their infancy. At scale, these workloads will place greater demands on silicon than today’s chips can support.

Meeting these demands doesn’t simply require additional investments in chips. Rather, it requires chips engineered for new levels of efficiency, performance, and power management. This is where TSMC’s competitive advantages begin to compound.

With each successive generation of process technology, the company has a unique opportunity to widen the performance gap between itself and rivals like Samsung or Intel.

Since Taiwan Semi already has such a large footprint in the foundry landscape, next-generation design complexities give the company a chance to further lock in deeper, stickier customer relationships.

TSMC’s valuation and the case for expansion

Taiwan Semi may trade at a forward price-to-earnings (P/E) ratio of 24, but dismissing the stock as “expensive” overlooks the company’s extraordinary positioning in the AI realm. To me, the company’s valuation reflects a robust growth outlook, improving earnings prospects, and a declining risk premium.

TSM PE Ratio (Forward) Chart

TSM PE Ratio (Forward) data by YCharts

Unlike many of its semiconductor peers, which are vulnerable to cyclicality headwinds, TSMC has become an indispensable utility for many of the world’s largest AI developers, evolving into one of the backbones of the ongoing infrastructure boom.

The scale of investment behind current AI infrastructure is jaw-dropping. Hyperscalers are investing staggering sums to expand and modernize data centers, and at the heart of each new buildout is an unrelenting demand for more chips. Moreover, each of these companies is exploring more advanced use cases that will, at some point, require next-generation processing capabilities.

These dynamics position Taiwan Semi at the crossroad of immediate growth and enduring long-term expansion, as AI infrastructure swiftly evolves from a constant driver of growth today into a multidecade secular theme.

TSMC’s manufacturing dominance ensures that its services will continue to witness robust demand for years to come. For this reason, I think Taiwan Semi is positioned to experience further valuation expansion over the next decade as the infrastructure chapter of the AI story continues to unfold.

While there are many great opportunities in the chip space, TSMC stands alone. I see it as perhaps the most unique, durable semiconductor stock to own amid a volatile technology landscape over the next several years.

Adam Spatacco has positions in Alphabet, Amazon, Microsoft, and Nvidia. The Motley Fool has positions in and recommends Advanced Micro Devices, Alphabet, Amazon, Intel, Microsoft, Nvidia, and Taiwan Semiconductor Manufacturing. The Motley Fool recommends the following options: long January 2026 $395 calls on Microsoft, short August 2025 $24 calls on Intel, short January 2026 $405 calls on Microsoft, and short November 2025 $21 puts on Intel. The Motley Fool has a disclosure policy.



Source link

Continue Reading

AI Research

Researchers train AI to diagnose heart failure in rural patients using low-tech electrocardiograms

Published

on


WVU computer scientists are training AI models to diagnose heart failure using data generated by low-tech equipment widely available in rural Appalachian medical practices. Credit: WVU/Micaela Morrissette

Concerned about the ability of artificial intelligence models trained on data from urban demographics to make the right medical diagnoses for rural populations, West Virginia University computer scientists have developed several AI models that can identify signs of heart failure in patients from Appalachia.

Prashnna Gyawali, assistant professor in the Lane Department of Computer Science and Electrical Engineering at the WVU Benjamin M. Statler College of Engineering and Mineral Resources, said —a chronic, persistent condition in which the heart cannot pump enough blood to meet the body’s need for oxygen—is one of the most pressing national and global health issues, and one that hits rural regions of the U.S. especially hard.

Despite the outsized impact of heart failure on rural populations, AI models are currently being trained to diagnose the disease using data representing patients from urban and suburban areas like Stanford, California, Gyawali said.

“Imagine Jane Doe, a 62-year-old woman living in a rural Appalachian community,” he suggested. “She has limited access to specialty care, relies on a small local clinic, and her lifestyle, diet and health history reflect the realities of her environment: high physical labor, minimal preventive care, and increased exposure to environmental risk factors like coal dust or poor air quality. Jane begins to experience fatigue and shortness of breath—symptoms that could point to heart failure.

“An AI system, trained primarily on data from urban hospitals in more affluent, coastal areas, evaluates Jane’s lab results. But because the system was not trained on patients who share Jane’s socioeconomic and environmental context, it fails to recognize her condition as urgent or abnormal,” Gyawali said. “This is why this work matters. By training AI models on data from West Virginia patients, we aim to ensure people like Jane receive accurate diagnoses, no matter where they live or how their lives differ from national averages.”

The researchers identified the AI models that were most accurate at diagnosing heart failure in an anonymized sample of more than 55,000 patients who received medical care in West Virginia. They also pinpointed the exact parameters for providing the AI models with data that most enhanced diagnostic accuracy. The findings appear in Scientific Reports, a Nature portfolio journal.

Doctoral student Alina Devkota emphasized they trained the AI models to work from patients’ electrocardiogram results, rather than the echocardiogram readings typical for patient data from urban areas.

Electrocardiograms rely on round electrodes stuck to the patient’s torso to record electrical signals from the heart. According to Devkota, they don’t require specialized equipment or specialized training to operate, but they still provide valuable insights into heart function.

“One of the criteria to diagnose heart failure is by measuring the ‘ejection fraction,’ or how much blood is pumped out of the heart with every beat, and the gold standard for doing that is with echocardiography, which uses to create images of the heart and the blood flowing through its valves,” she said.

“But echocardiography is expensive, time-consuming and often unavailable to patients in the very same rural Appalachian states that have the highest prevalence of heart failure across the nation. West Virginia, for example, ranks first in the U.S. for the prevalence of heart attack and , but many West Virginians don’t have local access to high-tech echocardiograms. They do have access to inexpensive electrocardiograms, so we tested whether AI models could use electrocardiogram readings to predict a patient’s ejection fraction.”

Devkota, Gyawali and their colleagues trained several AI models on patient records from 28 hospitals across West Virginia. The AI models used either “deep learning,” which relies on multilayered neural networks, or “non-deep learning,” which relies on simpler algorithms, to analyze the patient records and draw conclusions.

The researchers found the models, particularly one called ResNet, did best at correctly predicting a patient’s ejection fraction based on data from 12-lead electrocardiograms, with the results suggesting that a larger dataset for training would yield even better results. They also found that providing the AI models with specific “leads,” or combinations of data from different electrode pairs, affected how accurate the models’ ejection fraction predictions were.

Gyawali said while AI models are not yet being used in due to reliability concerns, training an AI to successfully estimate from electrocardiogram signals could soon give clinicians an edge in protecting patients’ cardiac health.

“Heart failure affects more than six million Americans today, and factors like our aging population mean the risk is growing rapidly—approximately 1 in 4 people alive today will experience heart failure during their lifetimes. The prevalence is even higher in rural Appalachia, so it’s critical the people here do not continue to be overlooked.”

Additional WVU contributors to the research included Rukesh Prajapati, graduate research assistant; Amr El-Wakeel, assistant professor; Donald Adjeroh, professor and chair for computer science; and Brijesh Patel, assistant professor in the WVU Health Sciences School of Medicine.

More information:
AI analysis for ejection fraction estimation from 12-lead ECG, Scientific Reports (2025). DOI: 10.1038/s41598-025-97113-0scientific

Citation:
Researchers train AI to diagnose heart failure in rural patients using low-tech electrocardiograms (2025, August 31)
retrieved 31 August 2025
from https://medicalxpress.com/news/2025-08-ai-heart-failure-rural-patients.html

This document is subject to copyright. Apart from any fair dealing for the purpose of private study or research, no
part may be reproduced without the written permission. The content is provided for information purposes only.





Source link

Continue Reading

AI Research

Should artificial intelligence be embraced in the classroom? – CBS News

Published

on



Should artificial intelligence be embraced in the classroom?  CBS News



Source link

Continue Reading

Trending