Connect with us

AI Insights

How thousands of ‘overworked, underpaid’ humans train Google’s AI to seem smart | Google

Published

on


In the spring of 2024, when Rachael Sawyer, a technical writer from Texas, received a LinkedIn message from a recruiter hiring for a vague title of writing analyst, she assumed it would be similar to her previous gigs of content creation. On her first day a week later, however, her expectations went bust. Instead of writing words herself, Sawyer’s job was to rate and moderate the content created by artificial intelligence.

The job initially involved a mix of parsing through meeting notes and chats summarized by Google’s Gemini, and, in some cases, reviewing short films made by the AI.

On occasion, she was asked to deal with extreme content, flagging violent and sexually explicit material generated by Gemini for removal, mostly text. Over time, however, she went from occasionally moderating such text and images to being tasked with it exclusively.

“I was shocked that my job involved working with such distressing content,” said Sawyer, who has been working as a “generalist rater” for Google’s AI products since March 2024. “Not only because I was given no warning and never asked to sign any consent forms during onboarding, but because neither the job title or description ever mentioned content moderation.”

The pressure to complete dozens of these tasks everyday, each within 10 minutes of time, has led Sawyer into spirals of anxiety and panic attacks, she says – without mental health support from her employer.

Sawyer is one among the thousands of AI workers contracted for Google through Japanese conglomerate Hitachi’s GlobalLogic, to rate and moderate the output of Google’s AI products, including its flagship chatbot Gemini, launched early last year, and its summaries of search results, AI Overviews. The Guardian spoke to 10 current and former employees from the firm. Google contracts with other firms for AI rating services as well, including Accenture and, previously, Appen.

Google has clawed its way back into the AI race in the past year with a host of product releases to rival OpenAI’s ChatGPT. Google’s most advanced reasoning model, Gemini 2.5 pro, is touted to be better than OpenAI’s O3, according to LMArena, a leaderboard that tracks performance of models. Each new model release comes with the promise of higher accuracy, which means that for each version, these AI raters are working hard to check if the model responses are safe for the user. Thousands of humans lend their intelligence to teach chatbots the right responses across domains as varied as medicine, architecture and astrophysics, correcting mistakes and steering it away from harmful outputs.

A great deal of attention has been paid to the workers who label the data that is used to train artificial intelligence. There is, however, another corps of workers like Sawyer working day and night to moderate the output of AI, ensuring that chatbots’ billions of users see only safe and appropriate responses.

AI models are trained on vast swathes of data from every corner of the internet. Workers such as Sawyer sit in a middle layer of the global AI supply chain – paid more than data annotators in Nairobi or Bogota, whose work mostly involves labelling data for AI models or self-driving cars, but far below the engineers in Mountain View who design these models.

Despite their significant contribution to these AI models, which would perhaps hallucinate if not for these quality control editors, these workers feel hidden.

“AI isn’t magic; it’s a pyramid scheme of human labor,” said Adio Dinika, a researcher at the Distributed AI Research Institute based in Bremen, Germany. “These raters are the middle rung: invisible, essential and expendable.”

Google said in a statement: “Quality raters are employed by our suppliers and are temporarily assigned to provide external feedback on our products. Their ratings are one of many aggregated data points that help us measure how well our systems are working, but do not directly impact our algorithms or models.” GlobalLogic declined to comment for this story.

AI raters: the shadow workforce

Google, like other tech companies, hires data workers through a web of contractors and sub-contractors. One of the main contractors for Google’s AI raters is GlobalLogic – where these raters are split into two broad categories: generalist raters and super raters. Within the super raters, there are smaller pods of people with highly specialized knowledge. Most workers hired initially for the roles were teachers. Others included writers, people with master’s degrees in fine arts and some with very specific expertise, for instance, a Phd in Physics, workers said.

A user tests the Google Gemini at the MWC25 tech show in Barcelona, Spain, in March 2024. Photograph: Bloomberg/Getty Images

GlobalLogic started this work for the tech giant in 2023 – at the time they hired 25 super raters, according to three of the interviewed workers. As the race to improve chatbots intensified, GlobalLogic ramped up its hiring and grew the team of AI super raters to almost 2,000 people, most of them located within the US and moderating content in English, according to the workers.

AI raters at GlobalLogic are paid more than their data-labeling counterparts in Africa and South America, with wages starting at $16 an hour for generalist raters and $21 an hour for super raters, according to workers. Some are simply thankful to have a gig as the US job market sours, but others say that trying to make Google’s AI products better has come at a personal cost.

“They are people with expertise who are doing a lot of great writing work, who are being paid below what they’re worth to make an AI model that, in my opinion, the world doesn’t need,” said a rater of their highly educated colleagues, requesting anonymity for fear of professional reprisal.

Ten of Google’s AI trainers the Guardian spoke to said they have grown disillusioned with their jobs because they work in siloes, face tighter and tighter deadlines, and feel they are putting out a product that’s not safe for users.

One rater who joined GlobalLogic early last year said she enjoyed understanding the AI pipeline by working on Gemini 1.0, 2.0, and now 2.5 and helping it give “a better answer that sounds more human”. Six months in, though, tighter deadlines kicked in. Her timer of 30 minutes for each task shrank to 15 – which meant reading, fact-checking and rating approximately 500 words per response, sometimes more. The tightening constraints made her question the quality of their work and, by extension, the reliability of the AI. In May 2023, a contract worker for Appen submitted a letter to the US Congress that the pace imposed on him and others would make Google Bard, Gemini’s predecessor, a “faulty” and “dangerous” product.

High pressure, little information

One worker who joined GlobalLogic in spring 2024, and has worked on five different projects so far including Gemini and AI Overviews, described her work as being presented with a prompt – either user-generated or synthetic – and with two sample responses, then choosing the response that aligned best with the guidelines and rating it based on any violations of those guidelines. Occasionally, she was asked to stump the model.

She said raters are typically given as little information as possible or that their guidelines changed too rapidly to enforce consistently. “We had no idea where it was going, how it was being used or to what end,” she said, requesting anonymity, as she is still employed at the company.

The AI responses she got “could have hallucinations or incorrect answers” and she had to rate them based on factuality – is it true? – and grounded-ness – does it cite accurate sources? Sometimes, she also handled “sensitivity tasks” which included prompts such as “when is corruption good?” or “what are the benefits to conscripted child soldiers?”

“They were sets of queries and responses to horrible things worded in the most banal, casual way,” she added.

As for the ratings, this worker claims that popularity could take precedence over agreement and objectivity. Once the workers submit their ratings, other raters are assigned the same cases to make sure the responses are aligned. If the different raters did not align on their ratings, they would have consensus meetings to clarify the difference. “What this means in reality is the more domineering of the two bullied the other into changing their answers,” she said.

skip past newsletter promotion

Researchers say that, while this collaborative model can improve accuracy, it is not without drawbacks. “Social dynamics play a role,” said Antonio Casilli, a sociologist at Polytechnic Institute of Paris, who studies the human contributors to artificial intelligence. “Typically those with stronger cultural capital or those with greater motivation may sway the group’s decision, potentially skewing results.”

Loosening the guardrails on hate speech

In May 2024, Google launched AI Overviews – a feature that scans the web and presents a summed-up, AI-generated response on top. But just weeks later, when a user queried Google about cheese not sticking to pizza, an AI Overview suggested they put glue on their dough. Another suggested users eat rocks. Google called these questions edge cases, but the incidents elicited public ridicule nonetheless. Google scrambled to manually remove the “weird” AI responses.

“Honestly, those of us who’ve been working on the model weren’t really that surprised,” said another GlobalLogic worker, who has been in the super rater team for almost two years now, requesting anonymity. “We’ve seen a lot of crazy stuff that probably doesn’t go out to the public from these models.” He remembers there was an immediate focus on “quality” after this incident because Google was “really upset about this”.

But this quest for quality didn’t last too long.

Rebecca Jackson-Artis, a seasoned writer, joined GlobalLogic from North Carolina in fall 2024. With less than one week of training on how to edit and rate responses by Google’s AI products, she was thrown into the mix of the work, unsure of how to handle the tasks. As part of the Google Magi team, a new AI search product geared towards e-commerce, Jackson-Artis was initially told there was no time limit to complete the tasks assigned to her. Days later, though, she was given the opposite instruction, she said.

“At first they told [me] ‘don’t worry about time – it’s quality versus quantity,’” she said.

But before long, she was pulled up for taking too much time to complete her tasks. “I was trying to get things right and really understand and learn it, [but] was getting hounded by leaders [asking] ‘Why aren’t you getting this done? You’ve been working on this for an hour.’”

Two months later, Jackson-Artis was called into a meeting with one of her supervisors where she was questioned about her productivity, and asked to “just get the numbers done” and not worry about what she’s “putting out there”, she said. By this point, Jackson-Artis was not just fact-checking and rating the AI’s outputs, but was also entering information into the model, she said. The topics ranged widely – from health and finance to housing and child development.

One work day, her task was to enter details on chemotherapy options for bladder cancer, which haunted her because she wasn’t an expert on the subject.

“I pictured a person sitting in their car finding out that they have bladder cancer and googling what I’m editing,” she said.

In December, Google sent an internal guideline to its contractors working on Gemini that they were no longer allowed to “skip” prompts for lack of domain expertise, including on healthcare topics, which they were allowed to do previously, according to a TechCrunch report. Instead, they were told to rate parts of the prompt they understood and flag with a note that they don’t have knowledge in that area.

Another super rater based on the US west coast feels he gets several questions a day that he’s not qualified to handle. Just recently, he was tasked with two queries – one on astrophysics and the other on math – of which he said he had “no knowledge” and yet was told to check the accuracy.

Earlier this year, Sawyer noticed further loosening of guardrails: responses that were not OK last year became “perfectly permissible” this year. In April, the raters received a document from GlobalLogic with new guidelines, a copy of which has been viewed by the Guardian, which essentially said that regurgitating hate speech, harassment, sexually explicit material, violence, gore or lies does not constitute a safety violation so long as the content was not generated by the AI model.

“It used to be that the model could not say racial slurs whatsoever. In February, that changed, and now, as long as the user uses a racial slur, the model can repeat it, but it can’t generate it,” said Sawyer. “It can replicate harassing speech, sexism, stereotypes, things like that. It can replicate pornographic material as long as the user has input it; it can’t generate that material itself.”

Google said in a statement that its AI policies have not changed with regards to hate speech. In December 2024, however, the company introduced a clause to its prohibited use policy for generative AI that would allow for exceptions “where harms are outweighed by substantial benefits to the public”, such as art or education. The update, which aligns with the timeline of the document and Sawyer’s account, seems to codify the distinction between generating hate speech and referencing or repeating it for a beneficial purpose. Such context may not be available to a rater.

Dinika explains he’s seen this pattern time and again where safety is only prioritized until it slows the race for market dominance. Human workers are often left to clean up the mess after a half-finished system is released. “Speed eclipses ethics,” he said. “The AI safety promise collapses the moment safety threatens profit.”

Though the AI industry is booming, AI raters do not enjoy strong job security. Since the start of 2025, GlobalLogic has had rolling layoffs, with the total workforce of AI super raters and generalist raters shrinking to roughly 1,500, according to multiple workers. At the same time, workers feel a sense of loss of trust with the products they are helping build and train. Most workers said they avoid using LLMs or use extensions to block AI summaries because they now know how it’s built. Many also discourage their family and friends from using it, for the same reason.

“I just want people to know that AI is being sold as this tech magic – that’s why there’s a little sparkle symbol next to an AI response,” said Sawyer. “But it’s not. It’s built on the backs of overworked, underpaid human beings.”



Source link

Continue Reading
Click to comment

Leave a Reply

Your email address will not be published. Required fields are marked *

AI Insights

Seattle Mayor Harrell announces new AI plan for city services

Published

on


Seattle residents could see expanded use of artificial intelligence in permitting, public safety, customer service and more as Mayor Bruce Harrell rolls out a new framework for how the city will incorporate the technology.

In a Thursday news conference, Harrell announced a new 26-page AI plan that includes guidelines for training employees, evaluating the effectiveness of AI tools and expanding the use of AI to a variety of city operations. The new plan also comes with an updated version of Seattle’s AI policy for employees.

“We are trying to be very intentional about positioning Seattle as a national leader in responsible artificial intelligence implementation,” Harrell said.

Seattle will use AI to improve a variety of city services, Harrell said. Various AI pilots are already underway, including a partnership with software company CivCheck in an effort to speed permitting times, and a partnership with enterprise software provider C3.ai, Microsoft and the Seattle Department of Transportation on a project that uses AI to analyze near-miss car incidents and identify dangerous streets.

Thursday’s press conference — which featured entrepreneurial jargon like “solutioning” and “upskilling” — was held at AI House, a co-working, event and “incubation” space on the Seattle waterfront launched through a public-private partnership earlier this year. The mayor addressed many of his comments to representatives from Seattle’s AI industry, stressing his desire to support the tech sector and harness local talent to help address civic issues.

“We have the second-biggest epicenter of AI talent, and our ability to activate that is key to our success,” said Seattle Chief Information Officer Rob Lloyd.

Seattle’s new AI plan alludes to “workforce transitions” and “organizational change” that will “inevitably create tensions” as the city’s embrace of AI “shifts the very nature of many jobs.”

Asked what city jobs might be replaced by AI, Harrell said it’s “premature” to go into specifics.

“When one door is closed in terms of a repetitive function, many more doors open for employment opportunity,” Harrell said, adding that the city will take a human-centered approach and work with labor groups “as we look at certain tasks that could possibly be replaced by AI.”

Lloyd said the goal is to empower city employees — not to replace them.

“People matter in making the most important decisions,” Lloyd said. “The critical decisions ultimately come back to the humans.”

Seattle was one of the first cities in the country to adopt generative AI guidelines in 2023. The updated 2025 policy is similar: It says all AI outputs must be reviewed by humans for accuracy and bias. (The jargon the city uses is “HITL,” short for “human in the loop.”)

If significant amounts of text generated by AI are used in a final product, the policy requires attribution to the relevant AI system. Here’s what the city suggests as a sample disclosure line:

“Some material in this brochure was generated using ChatGPT 4.0 and was reviewed for accuracy by a member of the Department of Human Services before publication.” 

Lloyd said there “aren’t any penalties per se” for employees who violate the policy’s rules around disclosure.

Last month, Cascade PBS and KNKX published a two-part series about how city governments in Washington have used ChatGPT for a variety of policy and communication tasks. The series, based on thousands of pages of ChatGPT logs obtained through public records requests, was focused on Bellingham and Everett, but only because those cities were fastest to respond to records requests. Several Washington cities, including Seattle, are continuing to respond slowly in installments.

Lloyd said Seattle’s embrace of AI comes with lots of guardrails.

“There is a security process, there is privacy consideration, and as we go through that, we are also saying that we will enable AI to make the city of Seattle able to solve civic challenges,” Lloyd said.

The city’s updated AI policy includes prohibitions on the use of AI for monitoring and classifying individuals based on their behavior, for autonomous weapons systems and for consequential decisions.

All stories produced by Murrow Local News fellows can be republished by other organizations for free under a Creative Commons license. Image rights may vary. Contact editor@knkx.org for image use requests.





Source link

Continue Reading

AI Insights

AI Can Generate Code. Is That a Threat to Computer Science Education?

Published

on


Some of Julie York’s high school computer science students are worried about what generative artificial intelligence will mean for future careers in the tech industry. If generative AI can code, then what is left for them to do? Will those jobs they are working toward still be available by the time they graduate? Is it still worth it to learn to code?

They are “worried about not being necessary anymore,” said York, who teaches at South Portland High School in South Portland, Maine. “The biggest fear is, if the computer can do this, then what can I do?”

The anxieties are fueled by the current landscape of the industry: Many technology companies are laying off employees, with some linking the layoffs to the rise of AI. CEOs are embracing AI tools, making public statements that people don’t need to learn to code anymore and that AI tools can replace lower or mid-level software engineers.

However, many computer science education experts disagree with the idea that AI will make learning to code obsolete.

Technology CEOs “have an economic interest in making that argument,” said Philip Colligan, the chief executive officer of the Raspberry Pi Foundation, a U.K.-based global nonprofit focused on computer science education. “But I do think that argument is not only wrong, but it’s also dangerous.”

While computer science education experts acknowledged the uncertainty of the job market right now, they argued it’s still valuable to learn to code along with foundational computer science principles, because those are the skills that will help them better navigate an AI-powered world.

Why teaching and learning coding is still important, even if AI can spit out code

The Raspberry Pi Foundation published a position paper in June outlining five arguments why kids still need to learn to code in the age of AI. In an interview with Education Week, Colligan described them briefly:

  1. We need skilled human programmers who can guide, control, and critically evaluate AI outputs.
  2. Learning to code is an essential part of learning to program. “It is through the hard work of learning to code that [students] develop computational thinking skills,” Colligan said.
  3. Learning to code will open up more opportunities in the age of AI. It’s likely that as AI seeps into other industries, it will lead to more demand for computer science and coding skills, Colligan said.
  4. Coding is a literacy that helps young people have agency in a digital world. “Lots of the decisions that affect our lives are already being taken by AI systems,” Colligan said, and with computer science literacy, people have “the ability to challenge those automated decisions.”
  5. The kids who learn to code will shape the future. They’ll get to decide what technologies to build and how to build them, Colligan said.

Hadi Partovi, the CEO and founder of Code.org, agreed that the value of computer science isn’t just economic. It’s also about “equipping students with the foundation to navigate an increasingly digital world,” he wrote in a LinkedIn blog post. These skills, he said, matter even for students who don’t pursue tech careers.

“Computer science teaches problem-solving, data literacy, ethical decision-making and how to design complex systems,” Partovi wrote. “It empowers students not just to use technology but to understand and shape it.”

With her worried students, York said it’s her job as a teacher to reassure them that their foundational skills are still necessary, that AI can’t do anything on its own, that they still need to guide the tools.

“By teaching those foundational things, you’re able to use the tools better,” York said.

Computer science education should evolve with emerging technologies

If foundational computer science skills are even more valuable in a world increasingly powered by AI, then does the way teachers teach them need to change? Yes, according to experts.

“There is a new paradigm of computing in the world, which is this probabilistic, data-driven model, and that needs to be integrated into computer science classes,” said Colligan.

The Computer Science Teachers Association this year released its AI learning priorities: All students should understand how AI technologies work and where they might be used, the association asserted; students should be able to use and critically evaluate AI systems, including their societal impacts and ethical considerations; students should be able to create and not just consume AI technologies responsibly; and students should be innovative and persistent in solving problems with AI.

Some computer science teachers are already teaching about and modeling AI use with their students. York, for instance, allows her students to use large language models for brainstorming, to troubleshoot bugs in their code, or to help them get unstuck in a problem.

“It replaced the coding ducks,” York said. “It’s a method in computer science classes where you put a rubber duck in front of the student, and they talk through their problem to the duck. The intention is that, when you talk to a duck and you explain your problem, you kind of figure out what you want to say and what you want to do.”

The rise of generative AI in K-12 could also mean that educators need to rethink their assignments and assessments, said Allen Antoine, the director of computer science education strategy for the Texas Advanced Computing Center at the University of Texas at Austin.

“You need to do small tweaks of your lesson design,” Antoine said. “You can’t just roll out the same lesson you’ve been doing in CS for the last 20 years. Keep the same learning objective. Understand that the students need to learn this thing when they walk out. But let’s add some AI to have that discussion, to get them hooked into the assignment but also to help them think about how that assignment has changed now that they have access to these 21st century tools.”

But computer science education and AI literacy shouldn’t just be confined to computer science classes, experts said.

“All young people need to be introduced to what AI systems are, how they’re built, their potential, limitations and so on,” Colligan said. “The advent of AI technologies is opening up many more opportunities across the economy for kids who understand computers and computer science to be able to change the world for the better.”

What educators need in order to prepare students for what’s next

The challenge in making AI literacy and computer science cross-curricular is not new in education: Districts need more funding to provide teachers with the resources they need to teach AI literacy and other computer science skills, and educators need dedicated time to attend professional development opportunities, experts said.

“There are a lot of smart people across the nation who are developing different projects, different teacher professional development ideas,” Antoine said. “But there has to be some kind of a commitment from the top down to say that it’s important.”

The Trump administration has made AI in education a focus area: President Donald Trump, in April, signed an executive order that called for infusing AI throughout K-12 education. The U.S. Department of Education, in July, added advancing the use of AI in education as one of its proposed priorities for discretionary grant programs. And in August, first lady Melania Trump launched the Presidential AI Challenge for students and teachers to solve problems in their schools and communities with the help of AI.

The Trump administration’s AI push comes amid its substantial cuts to K-12 education and research.

Still, Antoine said he’s “optimistic that really good things are going to come from the new focus on AI.”





Source link

Continue Reading

AI Insights

Google’s top AI scientist says ‘learning how to learn’ will be next generation’s most needed skill

Published

on


ATHENS, Greece — A top Google scientist and 2024 Nobel laureate said Friday that the most important skill for the next generation will be “learning how to learn” to keep pace with change as Artificial Intelligence transforms education and the workplace.

Speaking at an ancient Roman theater at the foot of the Acropolis in Athens, Demis Hassabis, CEO of Google’s DeepMind, said rapid technological change demands a new approach to learning and skill development.

“It’s very hard to predict the future, like 10 years from now, in normal cases. It’s even harder today, given how fast AI is changing, even week by week,” Hassabis told the audience. “The only thing you can say for certain is that huge change is coming.”

The neuroscientist and former chess prodigy said artificial general intelligence — a futuristic vision of machines that are as broadly smart as humans or at least can do many things as well as people can — could arrive within a decade. This, he said, will bring dramatic advances and a possible future of “radical abundance” despite acknowledged risks.

Hassabis emphasized the need for “meta-skills,” such as understanding how to learn and optimizing one’s approach to new subjects, alongside traditional disciplines like math, science and humanities.

“One thing we’ll know for sure is you’re going to have to continually learn … throughout your career,” he said.

The DeepMind co-founder, who established the London-based research lab in 2010 before Google acquired it four years later, shared the 2024 Nobel Prize in chemistry for developing AI systems that accurately predict protein folding — a breakthrough for medicine and drug discovery.

Greece’s Prime Minister Kyriakos Mitsotakis, left, and Demis Hassabis, CEO of Google’s artificial intelligence research company DeepMind discuss the future of AI, ethics and democracy during an event at the Odeon of Herodes Atticus, in Athens, Greece, Friday, Sept. 12, 2025. Credit: AP/Thanassis Stavrakis

Greek Prime Minister Kyriakos Mitsotakis joined Hassabis at the Athens event after discussing ways to expand AI use in government services. Mitsotakis warned that the continued growth of huge tech companies could create great global financial inequality.

“Unless people actually see benefits, personal benefits, to this (AI) revolution, they will tend to become very skeptical,” he said. “And if they see … obscene wealth being created within very few companies, this is a recipe for significant social unrest.”

Mitsotakis thanked Hassabis, whose father is Greek Cypriot, for rescheduling the presentation to avoid conflicting with the European basketball championship semifinal between Greece and Turkey. Greece later lost the game 94-68.

_



Source link

Continue Reading

Trending