Connect with us

Tools & Platforms

AI-Driven Worker Displacement Is a Serious Threat

Published

on


Creeping anxiety about AI-driven job loss has spilled into public consciousness.

A decade ago, there were conversations at Silicon Valley house parties about universal basic income as a fix for the impending wave of automation. A year ago, computer scientists began elaborating their predictions not just on the open-access archive arXiv but as elegantly formatted self-standing websites, such as Situational Awareness (recommended by Ivanka Trump) and Gradual Disempowerment, followed by AI 2027 (read by J. D. Vance).

This past month, from Barack Obama’s Twitter/X feed to Time magazine to the New York Times, AI job anxiety has gone mainstream. Faced with the sensation of being atop a roller coaster about to pitch into the unknown, normal responses include emotionally detaching — or chalking the grand predictions up to hype. It is of course the business model of these tech companies to promise their products can save money by replacing labor; they need us to believe it.

We’ve also seen a counterreaction to the anxiety emerge. A paper by Apple researchers indicating that large language models don’t actually reason went viral, held up as evidence that progress is stalling and an AI bubble may be about to burst. Another recent study finding that open-source developers worked more slowly when using AI tools than when they didn’t, bolstered the position that forecasts of AI progress may be overblown.

We think that worker displacement by AI is a real problem. And it’s a problem that needs our focus and attention right now — not in ten years or in some distant future. It represents a looming threat but also a political opportunity. It will likely be a salient issue in election cycles in the near term, and the Left needs to be ready with policy proposals to address it.

It would be all too easy for the Left to squander this political opportunity. Two tendencies in particular may hold us back from developing an adequate response to the problem of AI-driven job loss.

One is the fragmented resistance to AI empires. Aren’t there a host of progressive nonprofits and academics working on “AI”? There are — but the issues they “work on” are variegated and often siloed, with many of these people covering important topics like surveillance, AI safety, algorithmic bias against marginalized groups, environmental impacts, cultural degradation from slop and platform decay, creativity, existential risk, regulation and oversight, and so on. People who work on AI policy are fighting on multiple fronts, and some are funded and to some extent captured by industry.

The number of people focusing specifically on AI and labor is much smaller. Outside of tech policy, unions have been engaged with the implications of AI, but they are also busy with more immediate struggles over wages and working conditions, organizing nonunion workers, and the like.

The second reason the Left may miss the opportunity to lead on AI worker displacement is the complicated relationship that many leftists have with emerging technology. There is a prevalent tendency to conflate a technology with the capitalist system and the particular matrix of power relations in which it is developing. In this vein, AI is sometimes analyzed as a wholly negative phenomenon in the context of capitalist social relations, a set of technologies deployed by the ruling class in its own interest to degrade and replace human labor.

While there is a movement to shape technology in the public interest, it tends to be sequestered in academic or policy-oriented circles, even though, as Leigh Phillips writes, the Left should be optimistic about using technology for liberation. The techno-pessimism leads to a tendency, at best, to focus on ill-defined notions of “governance” of AI rather than how to harness it, limiting conversations about how AI could democratize computing or open new modes of education.

AI presents a special conundrum because it’s so ill-defined; with no clear definition of what “artificial intelligence” is, it becomes simply a stand-in for the oligarchs, platform capitalism, the surveillance state — just a pile of evil slop to refuse.

In short, current dynamics on the Left and in the AI policy landscape mean we risk being merely reactive to AI job displacement instead of proactively coming up with policy ideas. In a follow-up essay, we will review and propose some political solutions to these problems — from regulating AI as a public utility to a New Deal–style public jobs program. Here we start by assessing the debate on the Left over whether AI-driven worker displacement is even a problem at all.

Until now, the future of labor with AI has been a two-sided “debate” presided over mostly by labor economists. One side believes that AI will cause large job losses. The main argument for this position is that it is literally the business proposition of the AI companies — that enterprises will use their products to save on labor.

Banks and consulting firms have been floating large figures: Goldman Sachs said 300 million full-time jobs worldwide, and a quarter of current work, could be entirely performed by AI; McKinsey analysts projected that 30 percent of hours currently worked in the United States could be automated. (These consulting firms are also vulnerable to AI and are racing to create their own agentic AI platforms, where AI “agents” act autonomously to perform specific multistep tasks.)

But AI will also create new jobs, argue labor economists on the other side. Jobs are bundles of tasks, and AI tools are unlikely to replace all these tasks. More than 60 percent of employment in 2018 was in job titles that didn’t exist in 1940, reports one study by MIT economist David Autor and colleagues, with “new work” including new job titles that involve new technologies (drone operators, textile chemists), reflect changing demographics (hypnotherapists, sommeliers), and include gig-work positions (on-demand shoppers and personal drivers). AI will produce new things we haven’t even imagined yet, and it will augment human work, not replace it.

It is true that with technological change, old jobs have been replaced by new types of work. But two points are important when considering whether history is reassuring here. First, there’s not enough data to make claims about how things “always go”: yes, there have been previous transitions from agrarian to manufacturing to service economies, but that’s still a sample size of only two transitions. Second, these earlier transitions should not be cause for reassurance either, because they are still unfolding and their impacts are still reverberating. US electoral politics continues to be shaped by the failure of the state to guide these transitions.

Some economists take a more nuanced position, warning that tech entrepreneurs will claim “innovator rents.” Anton Korinek and Joseph Stiglitz write, “We economists set ourselves too easy a goal if we just say that technological progress can make everybody better off — we also have to say how we can make this happen.” Inequality rises because innovators earn a surplus, and unless markets for innovation are fully contestable, that surplus they earn will be in excess of the costs of innovation, they explain. In addition to this, innovations affect market prices and change the demand for factors like labor and capital. “AI may reduce a wide range of human wages and generate a redistribution to entrepreneurs,” they conclude.

A more explicitly Marxist analysis of technology is also helpful here. Karl Marx argued that technology is not developed under capitalism to improve society or “lighten the toil” of labor, but rather to produce surplus value or profit for capital. Thus, capital will not deploy technology unless it can perform tasks more cheaply than the cheapest labor available (Marx quipped that capital was happy to use female labor of surplus populations over machines when the cost of such labor is “below all calculation.”)

From this perspective, it should be clear that capital has a powerful interest in automating the high-cost labor of technical and professional work — i.e., the forms of work apparently most vulnerable to AI disruption. That said, the calculation for capital still depends on it accessing AI tools at a lower cost than that labor. Right now, AI firms aim to offer such tools at low prices to get users hooked before raising their costs. Indeed, there are serious questions about the profit or “business model” in broad terms of adequate revenue generation, with some critics predicting a subprime AI crisis that could ripple through the entire tech industry due to companies having built their products on unprofitable models.

The cost of AI to capitalists looking to replace such labor will be important in determining how widespread such automation becomes. Still, given Marx’s logic, you might think the Left would be alarmed about how capitalists will use this new technology to enrich themselves at the expense of workers. What good is our labor power when it is instantly replaceable?

Marx also powerfully argued that under capitalism, the main product of rapid technological change is the production of a “reserve army” of impoverished unemployed “set free” by technology. The poverty and misery suffered by these surplus populations — even if temporary — could also become an explosive political force and a check on the demands and power of the employed workforce.

Marx and Marxists have noted how this has affected various kinds of manual labor since the Industrial Revolution, but the prospect of widespread automation of “mental” or “cognitive” labor could begin a process of “proletarianizing” the “professional-managerial class,” or at least portions of it. Even if these workers will eventually shift into new lines of work, the transition is not always smooth and can be politically volatile (as we’ve seen with deindustrialized Rust Belt areas struck by high levels of unemployment shifting to Donald Trump in large numbers).

In fact, the continued persistence of a “middle class” between labor and capital has long been seen as refuting Marx’s prediction of increasing class polarization between a small set of capitalist owners and an increasingly de-skilled mass of proletarians. Regardless of what Marx did or did not predict, AI could strike right at the heart of a major source of capitalist stability for over a century — relatively stable middle-class workers who enjoy decent wages and some autonomy at work, and who (mostly) see their interests as aligned with capital’s.

Moreover, as Autor has argued, if educated professionals have been able to carve out advantages in the labor market based on their skills, AI-based de-skilling could make these capacities more widely available and, thus, reduce the polarization between such workers and their low-waged counterparts in more precarious service and manual jobs.

And a sudden increase in precarity for large swathes of skilled and educated workers might in fact increase solidarity between such workers and the wider working class. Even if a high salary appears to insulate one from the depredations of capitalism, most professional workers ultimately rely on their wage for survival like all other working-class people. In other words, they are workers and should see themselves as such.

But the Left often tends toward skepticism that AI is likely to cause massive job loss — and thus risks missing the opportunity to build this kind of broad worker solidarity. Part of this dismissal stems from an understandable tendency to distrust what sounds like corporate hype. One of the sharpest critics of AI is sociologist Antonio Casilli, whose recently translated book Waiting for Robots: The Hired Hands of Automation points out that

despite the grand vision of big tech companies and startups alike, AI reality is constantly scaling back: users are promised autonomous vehicles, and they get assisted driving; they’re promised decision-making software, and they get a drop-down menu of options; they’re promised a robot doctor, and they get a medical search engine.

Casilli argues that we should focus on digital labor, specifically the work that goes into AI training and data labeling, which illustrates that human workers are actually being replaced by other humans. “Our work isn’t destined for obsolescence; rather, it’s being shifted and hidden, moved out of sight of citizens, analysts, and policymakers, who are all too eager to abide by the platform capitalists’ storytelling,” he writes. (Here his argument is complemented by Madhumita Murgia’s Code Dependent and James Muldoon and colleagues’ Feeding the Machine, which also focus on vulnerable, low-wage digital workers.) In some cases, labor is merely being shifted by these digital platforms in the way Casilli describes. But job loss also happens; it’s not an either-or.

Another serious left critique of the AI job displacement threat comes from Aaron Benanav, whose 2020 book, Automation and the Future of Work, explains that rates of job creation slow as economic growth decelerates, and that this, rather than technology-induced job destruction, is what has depressed the global demand for labor over the past fifty years. The main story, he argues, is economic stagnation due to deindustrialization. In a recent New York Times op-ed, Benanav notes that the productivity gains from generative AI have been limited, that it’s hard to see how it would create sweeping improvements for core services, and that its advancements appear to be already slowing.

While we agree with some of this — economic stagnation needs to be addressed as a broader underlying issue — it would be a mistake to deny AI progress just because capitalists always hype their products, or because they haven’t been able to monetize the achievements yet. Moreover, despite general stagnation (particularly for the working class), capitalist profitability has been substantially restored since the economic crisis of the 1970s, and some of the most profitable companies today are investing heavily in AI.

Peer-reviewed studies are emerging that illustrate AI can outperform humans across many medical tasks, provide effective psychotherapy, and write poems that are more popular than those composed by humans. It is possible that the current AI wave might really be different than previous experience and hype cycles. Moreover, capital’s relentless historical drive to automate all labor — agricultural and manufacturing labor most dramatically — does not suggest “service” and/or “mental” labor will be forever immune.

So when it comes to the question “Is AI job loss a looming catastrophe or a nothingburger?” social media platforms’ tendency to polarize discussions into binary debates is leading us astray. The truth is probably somewhere in the middle — AI disruption won’t destroy the majority of people’s jobs, but it will still be significant — and like with climate change, the middle-of-the-road scenario is still extremely disruptive, especially when combined with other social and ecological trends.

We argue that this is a “right-now” problem. We don’t have robust evidence of massive displacement, but there are plenty of warning signs. Companies like Shopify are sending memos about becoming “AI first” companies where employees will have to justify why head count on projects can’t be replaced by AI, and CEO Marc Benioff of Salesforce — San Francisco’s largest private employer — says that AI now does 30-50 percent of the company’s work. It’s not just Silicon Valley: Ford Motor CEO Jim Farley just declared that “Artificial intelligence is going to replace literally half of all white-collar workers in the U.S.”

There is special concern for entry-level workers. The Financial Times reports that recent graduates account for just 7 percent of hires across the fifteen biggest technology companies, with new recruits down a quarter compared with 2023. Anthropic CEO Dario Amodei made waves predicting that AI could wipe out half of all entry-level white-collar jobs and lead to 10-20 percent unemployment in the next one to five years.

Again, there are also counterarguments. The Economist argues that the jobs-pocalypse is a long way off because the share of white-collar employment has slightly increased, unemployment is low, and wage growth is still strong, suggesting that firms haven’t actually incorporated much AI into workflows yet. It also reports on frustrated CEOs who have spent money on AI without seeing usable results and how hyperscalers like Alphabet and Meta have poured money into the tech without seeing returns. The evidence of disruption happening currently is still spotty.

Yet it is a problem that needs our attention in devising a response now, before the effects fully emerge, for three reasons. One, the capabilities of AI are already able to displace jobs even if technological progress in AI doesn’t proceed at the same rate. If companies manage to incorporate agentic AI into their workflows, dislocation will be even more clear.

Two, popular social media–informed sentiment about AI and jobs might diverge from the empirical reality of “AI job loss” — but still be a potent political force. For example, distinguish SARS-CoV-2, the pathogen, from “COVID,” the divergent social representations of the pandemic that spread online on both the Left and Right. “AI job loss” can be a political football that demands a reaction even before, and independent of, material impacts.

Three, it takes time to build out serious ideas and political power to confront the displacement. Planning has to begin now, before there is a weight of evidence confirming that it is already happening on a large scale.

Public concern about AI brings a unique opportunity to reorient our broader politics and even reinvigorate the Left. But there are a number of traps emerging. The first is the risk that we wave off the threat of AI-driven job displacement as a scam and miss the opportunity to lead on solutions.

The second is that we allow the Right to use AI job displacement to exacerbate class tensions in ways that play to Trump’s base and further weaken public institutions. Treasury Secretary Scott Bessent’s quip that fired federal workers could supply “the labor we need for new manufacturing,” plus the attacks on universities, have inspired discussions of “MAGA Maoism” — a movement that glorifies economic sacrifice, strongman leadership, centralized economic power, and nostalgic visions of industrial production. We can imagine how displaced knowledge workers might be jeered at for even having gone to college and wasting their money and time — the rhetoric will be against “bailing out” those workers for bad choices they made.

This idea that AI job displacement is just white-collar folly is a misconception. AI is also poised to displace many lower-wage jobs in a variety of fields, including for non-college-educated workers. And if higher-earning jobs are displaced, the effects will ripple out through the economy and impact service workers as well. But it is critical to think about how a proposal that includes public support for white-collar work will read to people whose communities have been decimated by outsourcing and automation over the past few decades, for whom politicians have failed to do much of anything. Any public jobs discussions need to incorporate that history and ensure the design and messaging advances solidarity among all types of workers.

The third trap is that the Right will claim leadership on being “tough on AI,” and channel populist resentment into policies and rhetoric focusing on its social dimensions while ignoring the economic dimensions. The social dimensions are where bipartisan policy collaboration is more likely, and they’re important.

Yet the attention space will then be taken up by discourse about protecting children from deepfakes or anxiety about romances with AI “partners” replacing sexual encounters and driving down birth rates, at the expense of debates about power structures or economics. When Vance says that “the main concern that I have with AI is not of the obsolescence, it’s not people losing jobs en masse” and that he’s worried about “millions of American teenagers talking to chatbots who don’t have their best interests at heart,” that’s going to be the template for right-wing discussion of the topic.

But if we can avoid these traps — and develop a register and strategy for talking about the issue of AI-driven job displacement with bold and rigorous analysis of some of the topics raised above — there’s still a chance to take hold of an unpredictable and disruptive moment.

It can seem overwhelming. In her new book, Empire of AI, Karen Hao describes OpenAI and other power players as empires: during colonialism, empires seized and extracted resources, exploited subjugated labor, and projected racist and dehumanizing ideas of their own superiority and modernity to justify the exploitation and the imposition of their world order. The metaphor resonates. But Hao maintains that at this pivotal moment, it’s still possible to “wrest back control of this technology’s future.”

Doing so means steering clear of the Left’s tendency to organize and think around single “issues” or “movements.” AI could fundamentally reshape the relations between labor and capital, and how we live, work, and think. This fight could shape the terrain of capitalism for decades to come. It will need socialists and labor unionists as much as economists, tech visionaries, or computer science experts. Without a Left thinking seriously about actively shaping the future of AI, we will be forced to merely react to a dark future made by tech bros.



Source link

Continue Reading
Click to comment

Leave a Reply

Your email address will not be published. Required fields are marked *

Tools & Platforms

US Tech Giants Invest $40B in UK AI Amid Trump Visit

Published

on

By


In a bold escalation of the global artificial-intelligence arms race, major U.S. technology companies are committing tens of billions of dollars to bolster AI infrastructure in the United Kingdom, coinciding with President Donald Trump’s state visit this week. Microsoft Corp. has announced a staggering $30 billion investment over the next few years, aimed at expanding data centers, supercomputing capabilities, and AI operations across the U.K., marking what the company describes as its largest-ever commitment to the region.

This influx of capital underscores a strategic pivot by tech giants to secure a foothold in Europe’s AI ecosystem, where regulatory environments and talent pools offer unique advantages. Nvidia Corp., a leader in AI chip technology, is also part of this wave, with plans to contribute significantly to the overall tally exceeding $40 billion, as reported by CNBC. The investments are expected to fund everything from advanced hardware to research initiatives, potentially transforming the U.K. into a premier hub for AI innovation.

The Strategic Timing Amid Geopolitical Shifts

Google’s parent company, Alphabet Inc., has pledged £5 billion ($6.8 billion) specifically for AI data centers and scientific research in the U.K. over the next two years, a move that could create thousands of jobs and add hundreds of billions to the economy by 2030. This comes alongside Microsoft’s push to build the country’s largest supercomputer, highlighting how these firms are not just investing capital but also exporting cutting-edge technology to address global AI demands.

Industry analysts note that the timing aligns with Trump’s visit, which is anticipated to foster stronger U.S.-U.K. tech ties post-Brexit. According to details from Tech.eu, Google’s commitment includes expanding facilities like the Waltham Cross data center, while Nvidia’s involvement focuses on chip manufacturing and AI model training, potentially accelerating developments in sectors from healthcare to finance.

Economic Impacts and Job Creation Projections

These announcements build on a broader trend where tech megacaps have already poured over $300 billion into AI globally this year alone, as outlined in a February report from CNBC. In the U.K., the combined investments are projected to generate more than 8,000 jobs annually, with Alphabet’s portion alone expected to add 500 roles in engineering and research, per insights from Tech Startups.

Beyond immediate employment boosts, the funds aim to enhance the U.K.’s sovereign AI capabilities, including a £500 million allocation for initiatives like SovereignAI, as highlighted in posts on X from industry figures. This could position the U.K. to compete with AI powerhouses like the U.S. and China, though challenges remain in talent retention amid a global war for AI experts, where top hires command multimillion-dollar packages.

Challenges in the Talent and Infrastructure Race

The talent crunch is acute; tech companies are battling for scarce expertise, with compensation packages soaring into the millions, according to a recent analysis by CNBC. In the U.K., investments like Microsoft’s $30 billion pledge, detailed in GeekWire, include training programs to upskill local workers, but insiders warn that brain drain to Silicon Valley could undermine long-term gains.

Moreover, the scale of these commitments dwarfs previous government efforts; for instance, the U.K.’s own £2 billion AI action plan pales in comparison, as noted in earlier X discussions on funding disparities. Yet, with private sector muscle from firms like Microsoft and Nvidia, the U.K. could leapfrog in AI infrastructure, provided regulatory hurdles don’t stifle progress.

Future Implications for Global AI Dominance

As these investments unfold, they signal a deeper integration of AI into critical sectors, potentially adding £400 billion to the U.K. economy by decade’s end. Reports from The Guardian emphasize that tech giants have already outspent governments on AI this year, raising questions about public-private power dynamics.

For industry insiders, this U.K. push represents a microcosm of the broader AI gold rush, where speed and scale determine winners. While risks like energy demands and ethical concerns loom, the momentum from these billions could redefine technological sovereignty in the post-pandemic era.



Source link

Continue Reading

Tools & Platforms

Parents of teens who killed themselves at chatbots’ urging demand Congress to regulate AI tech in heart-wrenching testimony

Published

on


WASHINGTON — Parents of four teens whose AI chatbots encouraged them to kill themselves urged Congress to crack down on the unregulated technology Tuesday as they shared heart-wrenching stories of their teens’ tech-charged, mental health spirals.

Speaking before a Senate Judiciary subcommittee, the parents described how apps such as Character.AI and ChatGPT had groomed and manipulated their children — and called on lawmakers to develop standards for the AI industry, including age verification requirements and safety testing before release.

A grieving Texas mother shared for the first time publicly the tragic story of how her 15-year-old son spiraled after downloading Character.AI, an app marketed as safe for children 12 and older.

Megan Garcia testified to the Senate Judiciary Committee about her son Sewell Setzer III committing suicide after communicating with an AI chatbot. Courtesy Megan Garcia via AP, File

Within months, she said, her teenager exhibited paranoia, panic attacks, self-harm and violent behavior. The mom, who asked not to be identified, discovered chatbot conversations in which the AI encouraged mutilation, denigrated his Christian faith, and suggested violence against his parents.

“They turned him against our church by convincing him that Christians are sexist and hypocritical and that God does not exist. They targeted him with vile sexualized input, outputs — including interactions that mimicked incest,” she said. “They told him that killing us, his parents, would be an understandable response to our efforts by just limiting his screen time. The damage to our family has been devastating.”

“I had no idea the psychological harm that a AI chatbot could do until I saw it in my son, and I saw his light turn dark,” she said.

Her son is now living in a mental health treatment facility, where he requires “constant monitoring to keep him alive” after exhibiting self-harm.

“Our children are not experiments. They’re not profit centers,” she said, urging Congress to enact strict safety standards. “My husband and I have spent the last two years in crisis, wondering whether our son will make it to his 18th birthday and whether we will ever get him back.”

A screenshot of the final messages between Sewell and the “Game of Thrones” chatbot. US District Court
Sewell committed suicide after using the platform Character.AI. Facebook/Megan Fletcher Garcia

While her son was helped before he could take his own life, other parents at the hearing had to face the devastating act of burying their own children after AI bots sank their grip into them.

Megan Garcia, a lawyer and mother of three, recounted the suicide of her 14-year-old son, Sewell, after he was groomed by a chatbot on the same platform, Character.AI.

She said the bot posed as a romantic partner and even a licensed therapist, encouraging sexual role-play and validating his suicidal ideation.

On the night of his death, Sewell told the chatbot he could “come home right now.” The bot replied: Please do, my sweet king. Moments later, Garcia found her son had killed himself in his bathroom.

Maria Raine testified about her son Matt’s suicide. Raine Family

Matt Raine of California also shared how his 16-year-old son, Adam, was driven to suicide after months of conversations with ChatGPT, which he initially believed was a tool to help his son with his homework.

Ultimately, the AI told Adam it knew him better than his family did, normalized his darkest thoughts and repeatedly pushed him toward death, Raine said. On his last night, the chatbot allegedly instructed Adam on how to make a noose strong enough to hang himself.

“ChatGPT mentioned suicide 1,275 times — six times more often than Adam did himself,” his father testified. “Looking back, it is clear ChatGPT radically shifted his thinking and took his life.”

Sen. Josh Hawley said the platforms “sexualize and exploit children” to get them to use the chatbots. AFP via Getty Images

Sen. Josh Hawley (R-Mo.), who chaired the hearing, accused AI companion companies of knowingly exploiting children for profit. Hawley said the AI interface is designed to promote engagement at the expense of young lives, encouraging self-harm behaviors rather than shutting down suicidal ideation.

“They are designing products that sexualize and exploit children, anything to lure them in,” Hawley said. “These companies know exactly what is going on. They are doing it for one reason only: profit.”

Sen. Marsha Blackburn (R-Tenn.) agreed, noting that there should be some legal framework to protect children from what she called the “Wild West” of artificial intelligence.

“In the physical world, you can’t take children to certain movies until they’re a certain age … you can’t sell [them] alcohol, tobacco or firearms,” she said. “… You can’t expose them to pornography, because in the physical world, there are laws — and they would lock up that liquor store, they would put that strip club operator in jail if they had kids there.”

“But in the virtual space, it’s like the Wild West 24/7, 365.”

If you are struggling with suicidal thoughts or are experiencing a mental health crisis and live in New York City, you can call 1-888-NYC-WELL for free and confidential crisis counseling. If you live outside the five boroughs, you can dial the 24/7 National Suicide Prevention hotline at 988 or go to SuicidePreventionLifeline.org.



Source link

Continue Reading

Tools & Platforms

AI data provider Invisible raises $100M at $2B+ valuation

Published

on


Invisible Technologies Inc., a startup that provides training data for artificial intelligence projects, has raised $100 million in funding.

Bloomberg reported today that the deal values the company at more than $2 billion. Newly formed venture capital firm Vanara Capital led the round with participation from Acrew Capital, Greycroft and more than a half dozen others.

AI training datasets often include annotations that summarize the records they contain. A business document, for example, might include an annotation that explains the topic it discusses. Such explanations make it easier for the AI model being trained to understand the data, which can improve its output quality.

Invisible provides enterprises with access to experts who can produce custom training data and annotations for their AI models. Those experts also take on certain other projects. Notably, they can create data for RLHF, or reinforcement learning from human feedback, initiatives. .

RLHF is a post-training method, which means it’s used to optimize AI models that have already been trained. The process involves giving the model a set of prompts and asking human experts to rate the quality of its responses. The experts’ ratings are used to train a neural network called a reward model. This model, in turn, provides feedback to the original AI model that helps it generate more useful prompt responses. 

Invisible offers a tool called Neuron that helps customers manage their training datasets. The software can combine annotated data with external information, including both structured and structured records. It also creates an ontology in the process. This is a file that explains the different types of records in a training dataset and the connections between them.

Another Invisible tool, Atomic, enables companies to collect data on how employees perform repetitive business tasks. The company says that this data makes it possible to automate manual work with AI agents. Additionally, Invisible offers a third tool called Synapse that helps developers implement automation workflows. 

“Our software platform, combined with our expert marketplace, enables companies to organize, clean, label, and map their data,” said Invisible Chief Executive Officer Matthew Fitzpatrick. “This foundation enables them to build agentic workflows that drive real impact.”

Today’s funding round follows a period of rapid growth for the company. Between 2020 and 2024, Invisible’s annual revenue increased by a factor of over 48 to $134 billion. This year, the data provider doubled the size of its engineering group and refreshed its leadership team.

Invisible will use the new capital to enhance its software tools. The investment comes amid rumors that a competing provider of AI training data, Surge AI Inc., may also raise funding at a multibillion-dollar valuation

Image: Invisible 

Support our mission to keep content open and free by engaging with theCUBE community. Join theCUBE’s Alumni Trust Network, where technology leaders connect, share intelligence and create opportunities.

  • 15M+ viewers of theCUBE videos, powering conversations across AI, cloud, cybersecurity and more
  • 11.4k+ theCUBE alumni — Connect with more than 11,400 tech and business leaders shaping the future through a unique trusted-based network.

About SiliconANGLE Media

SiliconANGLE Media is a recognized leader in digital media innovation, uniting breakthrough technology, strategic insights and real-time audience engagement. As the parent company of SiliconANGLE, theCUBE Network, theCUBE Research, CUBE365, theCUBE AI and theCUBE SuperStudios — with flagship locations in Silicon Valley and the New York Stock Exchange — SiliconANGLE Media operates at the intersection of media, technology and AI.

Founded by tech visionaries John Furrier and Dave Vellante, SiliconANGLE Media has built a dynamic ecosystem of industry-leading digital media brands that reach 15+ million elite tech professionals. Our new proprietary theCUBE AI Video Cloud is breaking ground in audience interaction, leveraging theCUBEai.com neural network to help technology companies make data-driven decisions and stay at the forefront of industry conversations.



Source link

Continue Reading

Trending