Connect with us

Tools & Platforms

Why Coca-Cola CIO Neeraj Tolmare prioritizes big-impact AI pilot projects

Published

on


Neeraj Tolmare, the chief information officer at Coca-Cola, says the artificial intelligence pilots that he invests in only get authorized if the beverage giant can envision revenue generation or big efficiency gains once fully in production.

This strategy reflects a mandate from Coke’s CEO James Quincey and the broader leadership team that demands that AI experimentation “has to be tied to a real outcome that moves the needle for us,” says Tolmare. “We are all about scale.” 

That scale is massive for Coca-Cola, which is ranked 97th on the Fortune 500. Each AI bet by Tolmare has to weigh cost of implementation, as well as how the technology may be used internally across various functions ranging from software development to sales. On top of that, he must weigh how new technology advancements like agentic AI will change workflows and how data is shared within Coke and close key external partners, including the 200 bottlers and 950 production facilities that make up the entire system that Coke relies on to serve 2.2 billion beverages each day.

One pilot that Tolmare is particularly excited about is the development of an algorithm that Coke developed to help retail outlets better predict demand, using AI to ensure that their shelves are always appropriately stocked.

In the past, sales agents would visit stores every few weeks and at times, find themselves facing half-empty coolers. Coke had to rely on historical retail scan data to try to anticipate future demand. But with AI, the company was able to create an algorithm that triangulates historical data with weather patterns, which is an important consideration for beverage purchases, and geolocation data from Google to more precisely predict future sales. 

Those AI insights inform the messages Coca-Cola sends via WhatsApp to managers, advising them on when to stock more Sprite or Diet Coke before they run out. The pilot was tested in three countries and saw sales increase by 7% to 8% versus outlets that weren’t using the AI algorithm, according to Tolmare, who intends to broaden the application of this AI tool to more markets globally.

Another investment with big reach is Coke’s usage of AI for content creation. The beverage giant sells in more than 180 countries and creates marketing assets in over 130 languages. It’s a process that’s both time consuming and costly, because Coke has to adjust the materials to reflect not only language, but also cultural sensibilities. Coke recently created 20 AI-generated assets, all based on the company’s own proprietary intellectual property, and then created 10,000 variations to be used across many different languages and geographies.

Tolmare asserts that consumers are 20% more likely to engage with this content than prior iterations that weren’t crafted by AI, and that the AI-produced content was three times faster to generate. 

In both cases, he stresses the importance of keeping humans in the loop. “AI has proven that it can unlock value that a human being would not be able to unlock because the computation needed to mine through all this data and bring something meaningful is so complex,” says Tolmare, in regards to the algorithm to support retail sales.

With marketing, AI’s application is a bit trickier and Coke is still figuring out the right approach. An AI-generated Christmas spot generated some backlash last year and in April, a different campaign made a mistake featuring a quote from a book that novelist J.G. Ballard never wrote. Tolmare says Coke has strict guidelines in place to ensure that AI-generated content avoids social biases that would be harmful to consumers and is keenly aware of hallucinations or deepfakes that can occur when generating creative content with AI tools.

“That’s not to say that we won’t ever fall into it,” admits Tolmare. “But we have a mechanism in place that allows us to react in a responsible way if and when that happens.”

Most of Tolmare’s career has been in the technology industry, including working for gadget manufacturer Palm, which ultimately lost out on an effort to compete with Apple in the smartphone market. He also previously served as a director at networking-equipment company Cisco Systems and as a VP at HP during the years in which it split up the personal-computer and printer businesses from the corporate hardware and services arm. 

“I’ve spent more than two decades working for companies that develop and manufacture technology that the rest of the world uses,” says Tolmare. He joined Coca-Cola in 2018 to work at a company that consumes technology, rather than developing it, to grow their business.

Tolmare spearheaded the company’s all-in bet on cloud computing, retiring or selling all of Coke’s physical data centers. Roughly 80% of the company’s footprint is in Microsoft Azure with the rest split between Amazon Web Services and Google Cloud. “We continuously monitor which one is going to give us better efficiency by moving a workload from one cloud to another,” says Tolmare.

His hybrid cloud playbook is also applicable to how Coke thinks about generative AI experimentation. Coke works closely with Microsoft and its strategic partner OpenAI, but also taps other AI hyperscalers including Google, Meta, and Anthropic. “We don’t want to paint ourselves into a corner too soon,” he says. “The market itself is not consolidating.”

What’s next on deck is further exploration with agentic AI, systems that are designed to perform tasks autonomously or with little human intervention. The company is considering offerings from vendors including Microsoft, SAP, and Adobe, while developing its own AI agents trained on Coke’s data. All of these AI agents remain in the pilot phase, because Tolmare still wants to determine if these systems can be built at a cost effective price, while achieving the optimal business outcomes.

“We haven’t launched agentic in production yet, but we are very close,” says Tolmare. “I’m fascinated by what this can do for our business.”

John Kell

Send thoughts or suggestions to CIO Intelligence here.

NEWS PACKETS

Tech giants offer more discounts to the federal government. The Wall Street Journal reports that Oracle is cutting the cost of the company’s database software and cloud-computing service for the federal government. The discount on cloud infrastructure is the first of its kind for the entire government and could be replicated with other cloud providers in negotiations that are currently underway. Other recent price discounts that the General Services Administration were able to extract from vendors include a deal cut with Salesforce to trim the price of messaging app Slack by 90% through the end of November and lower costs from Google and Adobe. The GSA has opted to negotiate these savings through direct engagement with the tech giants, rather than discussing pricing with third parties. “Through procurement consolidation, we’re aiming to bring the leverage of the whole, commanding purchasing power of the federal wallet to these [technology providers] to get the best discounts for the taxpayers,” Federal Acquisition Service Commissioner Josh Gruenbaum told WSJ.

Ford CEO predicts AI will replace half of U.S. white-collar workers. CEOs have increasingly shared their dour prognostication of AI’s impact on the future of work and the latest diagnosis came from Jim Farley, the CEO of automaker Ford. Speaking at the Aspen Ideas Festival, Farley was making the case that the education system is too focused on pushing students toward a four-year college degree, which may be counterintuitive, given his view that “AI is gonna replace literally half of all white-collar workers in the U.S.” Farley makes a pitch for more investment in trade workers, those who may take a job on an auto factory floor or in construction, where there is a shortage of workers. He may have a point, as tech giants continue to slice jobs, with Microsoft earlier this month cutting another 9,000 jobs, bringing the grand total of jobs eliminated by the tech giant to 15,000 this year.

But, IT unemployment dropped to its lowest level this year in June. Much of the consternation about AI’s impact on work has been focused on tech jobs, especially as big employers like Microsoft and TikTok continue to generate national headlines when announcing layoffs. That said, across all sectors, businesses are on the hunt for talent and have been hiring for software developers and engineers, systems engineers and architects, and cybersecurity professionals, according to CIO Dive, which reported on insights from CompTIA. The IT trade organization reported that IT unemployment fell to 2.8% in June, as companies across all sectors added 90,000 net new tech jobs. Active employer job listings for tech positions reached 455,341 in June, with 47% of that total newly added in June. But the tech industry itself continues to be a weak point, as that sector reduced staffing by a net 7,256 positions in June.

Google parent company’s AI-powered drug firm gearing up for clinical trials. Alphabet’s drug discovery arm, Isomorphic Labs, is getting close to testing AI-designed drugs in humans, Fortune reports, as company president Colin Murdoch says “the next big milestone is actually going to clinical trials, starting to put these things into human beings. We’re staffing up now. We’re getting very close.” Isomorphic Labs, which was spun out of Google’s AI research lab DeepMind in 2021, raised $600 million in April to help propel its goal of combining the expertise of machine learning researchers and pharma veterans to design new treatments at a faster pace, more cheaply, and with a higher rate of success. Isomorphic supports existing pharma drug discovery programs, having already inked research collaborations with Novartis and Eli Lilly, and also develops its own internal drug candidates in areas such as oncology and immunology.

ADOPTION CURVE

Why CIOs need to focus on both data quality and permission. A new survey found that while 52% of companies are “highly reliant” on the consented data that’s used to craft personalized marketing materials and product development, there are still roadblocks as it pertains to how enterprises use this data. That’s important to get right if enterprises want to retain trust with the consumers that authorize the use of their data, while also adhering to state laws in places like California, which has given individuals greater control over their personal information.

Some big issues that CIOs face related to consent requirements include a lack of visibility into data flows (61%), followed closely by difficulty tracking user preferences across channels (59%). They’ll need to solve these problems quickly as 100% of the 265 survey respondents say that consented consumer data is foundational to their ability to succeed when it comes to pursuing AI initiatives. 

Kate Parker, CEO of Transcend, the data privacy management software provider that funded the study, tells Fortune that the findings highlight a shift in thinking that’s less focused on data quality and more about permission to use the consumer data in the first place. 

“If CIOs and digital leaders are in agreement that the consent of the data is the fundamental component of AI and personalization, but they also admit very openly that they do not have the infrastructure to solve that in scale, that’s going to be one of the biggest growth blockers,” says Parker.

Courtesy of Transcend

JOBS RADAR

Hiring:

H&H is seeking a CIO, based in New York City. Posted salary range: $200K-$250K/year.

Lease & LaBau is seeking a CIO, based in New York City. Posted salary range: $450K-$500K/year.

Vibrant Emotional Health is seeking a CTO in a remote-based role. Posted salary range: $261K-$350K/year.

Major League Soccer is seeking a chief information security officer, based in New York City. Posted salary range: $200K-$275K/year.

Hired:

Old National Bancorp named Matt Keen as CIO, taking on the C-suite leadership role after the Minneapolis-based bank acquired Bremer Bank, where Keen had served in the same role. Previously, Keen also served as a consultant for American Express as part of his tenure at PriceWaterhouseCoopers. He also previously served as CTO at real estate investment trust Two Harbors Investment Corp.

EProductivity Software Packaging announced the appointment of Scott Brown as CIO, joining the software provider for the packaging and print industries to oversee an expansion of the company’s cloud portfolio, as well as enhance security and compliance across all platforms. Brown joins from software provider Sciforma, where he served as CIO.

Rancher Government Solutions promoted Adam Toy to the role of CTO, where he will steer the strategic direction of the company’s tech portfolio, overseeing platform engineering, architecture, and product development. Toy had served in that role on an interim basis since March and initially joined Rancher, which helps government agencies modernize their IT infrastructure, in 2020 as a senior solutions architect.



Source link

Tools & Platforms

AI-generated child sexual abuse videos surging online, watchdog says | Internet

Published

on


The number of videos online of child sexual abuse generated by artificial intelligence has surged as paedophiles have pounced on developments in the technology.

The Internet Watch Foundation said AI videos of abuse had “crossed the threshold” of being near-indistinguishable from “real imagery” and had sharply increased in prevalence online this year.

In the first six months of 2025, the UK-based internet safety watchdog verified 1,286 AI-made videos with child sexual abuse material (CSAM) that broke the law, compared with two in the same period last year.

The IWF said just over 1,000 of the videos featured category A abuse, the classification for the most severe type of material.

The organisation said the multibillion-dollar investment spree in AI was producing widely available video-generation models that were being manipulated by paedophiles.

“It is a very competitive industry. Lots of money is going into it, so unfortunately there is a lot of choice for perpetrators,” said one IWF analyst.

The videos were found as part of a 400% increase in URLs featuring AI-made child sexual abuse in the first six months of 2025. The IWF received reports of 210 such URLs, compared with 42 last year, with each webpage featuring hundreds of images, including the surge in video content.

The IWF saw one post on a dark web forum where a paedophile referred to the speed of improvements in AI, saying how they had mastered one AI tool only for “something new and better to come along”.

IWF analysts said the images appeared to have been created by taking a freely available basic AI model and “fine-tuning” it with CSAM in order to produce realistic videos. In some cases these models had been fine-tuned with a handful of CSAM videos, the IWF said.

The most realistic AI abuse videos seen this year were based on real-life victims, the watchdog said.

Derek Ray-Hill, the IWF’s interim chief executive, said the growth in capability of AI models, their wide availability and the ability to adapt them for criminal purposes could lead to an explosion of AI-made CSAM online.

“There is an incredible risk of AI-generated CSAM leading to an absolute explosion that overwhelms the clear web,” he said, adding that a growth in such content could fuel criminal activity linked to child trafficking, child sexual abuse and modern slavery.

The use of existing victims of sexual abuse in AI-generated images meant that paedophiles were significantly expanding the volume of CSAM online without having to rely on new victims, he added.

The UK government is cracking down on AI-generated CSAM by making it illegal to possess, create or distribute AI tools designed to create abuse content. People found to have breached the new law will face up to five years in jail.

Ministers are also outlawing possession of manuals that teach potential offenders how to use AI tools to either make abusive imagery or to help them abuse children. Offenders could face a prison sentence of up to three years.

Announcing the changes in February, the home secretary, Yvette Cooper, said it was vital that “we tackle child sexual abuse online as well as offline”.

AI-generated CSAM is illegal under the Protection of Children Act 1978, which criminalises the taking, distribution and possession of an “indecent photograph or pseudo photograph” of a child.



Source link

Continue Reading

Tools & Platforms

Top AI performers are burned out and eyeing a better workplace

Published

on


Here’s data that’s sure to get the attention of HR leaders: The employees delivering your biggest AI-driven productivity gains are twice as likely to quit as everyone else.

A new study by the Upwork Research Institute, the research arm of the remote job platform Upwork, reveals that nearly 9 in 10 top AI performers are burned out and eyeing the exits. Meanwhile, more than two-thirds say they trust the technology more than they do their coworkers — with 64% finding machines to be more polite and empathetic.

AI is “unlocking speed and scale but also reshaping how we collaborate and connect as humans,” said Kelly Monahan, managing director of the Upwork Research Institute. “The productivity paradox we’re seeing may be a natural growing pain of traditional work systems, ones that reward output with AI, but overlook the human relationships behind that work.”

According to the study, based on the perspectives of 2,500 workers globally, the emotional dimension around AI runs deeper than many employers may realize. Nearly half of those surveyed say “please” and “thank you” with every request they submit to AI, while 87% phrase their requests as if speaking to a human—an anthropomorphizing of AI tools indicating that employees are forming more genuine emotional connections with their digital assistants than with their colleagues.

“[AI is] unlocking speed and scale but also reshaping how we collaborate and connect as humans.”

Kelly Monahan,

managing director, Upwork Research Institute

Colin Rocker, a content creator specializing in career development, makes the point that “AI will always be the most agreeable coworker, but we have to also be mindful that it’s a system that, by nature, will agree with and amplify whatever is said to it.”

The study also revealed a disconnect between individual AI adoption and organizational strategy. While employees are racing ahead with AI integration, 62% of high-performing AI users say they don’t understand how their daily AI use aligns with company goals. That misalignment creates a dangerous scenario where the most productive employees feel isolated from the broader organizational mission, even as they’re delivering exceptional results.

The contrast with freelancers is illuminating, meanwhile. Unlike full-time employees, independent contractors appear to thrive alongside AI, with nearly nine in 10 reporting a positive impact on their work. These workers use AI primarily as a learning partner, with 90% saying it helps them acquire new skills faster and 42% crediting it with helping them specialize in a particular niche — suggesting that the problem is not technology itself but, rather, how it’s being integrated into traditional organizational structures.

Ultimately, the survey suggests, the path to sustainable, AI-empowered businesses requires reimagining work as a collaboration between the technology and the people who use it; cultivating flexible and resilient talent ecosystems; and redefining AI strategies around relationships, emerging AI roles and responsible governance.

To lead effectively in the age of AI, Monahan suggests that employers “need to redesign work in ways that support not just efficiency but also well-being, trust and long-term resilience.”



Source link

Continue Reading

Tools & Platforms

In test-obsessed Korea, AI boom arrives in exams, ahead of the technology itself

Published

on


Over 500 new AI certifications have sprung up in Korea in two years, but few are trusted or even taken

Students at Sangincheon Middle School in Incheon take part in an AI class in February to prepare for KT’s nationally accredited AICE (AI Certificate for Everyone) Junior certification exam. (KT)

A wave of artificial intelligence certifications has flooded the market in South Korea over the past two years.

But according to government data, most of these tests exist only on paper, and have never been used by a single person.

As of Wednesday, there were 505 privately issued AI-related certifications registered with the Korea Research Institute for Professional Education and Training, a state-funded body under the Prime Minister’s Office.

This is nearly five times the number recorded in 2022, before tools like ChatGPT captured global attention. But more than 90 percent of those certifications had zero test-takers as of late last year, the institute’s own data shows.

Many of the credentials are loosely tied to artificial intelligence in name only. Among recent additions are titles like “AI Brain Fitness Coach,” “AI Art Storybook Author,” and “AI Trainer,” which often have no connection to real AI technology.

KT’s AICE (AI Certificate for Everyone) is South Korea’s only nationally accredited AI certification, offering five levels of exams that assess real-world AI understanding and skills, from block coding for elementary students to Python-based modeling for professionals. (KT)
KT’s AICE (AI Certificate for Everyone) is South Korea’s only nationally accredited AI certification, offering five levels of exams that assess real-world AI understanding and skills, from block coding for elementary students to Python-based modeling for professionals. (KT)

Only one of the 505 AI-related certifications — KT’s AICE exam — has received official recognition from the South Korean government. The rest have been registered by individuals, companies, or private organizations, with no independent oversight or quality control.

In 2024, just 36 of these certifications held any kind of exam. Only two had more than 1,000 people apply. Fourteen had a perfect 100 percent pass rate. And 20 were removed from the registry that same year.

For test organizers, the appeal is often financial. One popular certification that attracted around 500 candidates last year charged up to 150,000 won ($110) per person, including test fees and course materials. The content reportedly consisted of basic instructions on how to use existing tools like ChatGPT or Stable Diffusion. Some issuers even promote these credentials as qualifications to teach AI to students or the general public.

The people signing up tend to be those anxious about keeping up in an AI-driven world. A survey released this week by education firm Eduwill found that among 391 South Koreans in their 20s to 50s, 39.1 percent said they planned to earn an AI certificate to prepare for the digital future. Others (27.6 percent) said they were taking online AI courses or learning how to use automation tools like Notion AI.

Industry officials warn that most of these certificates hold little value in the job market. Jeong Sung-hoon, communications manager at Seoul-based AI startup Wrtn, told The Korea Herald that these credentials are often “window dressing” for resumes.

Wrtn ranked second in generative AI app usage among Koreans under 30 this March, according to local mobile analytics firm Wiseapp.

“Most private AI certifications aren’t taken seriously by hiring managers,” Jeong said. “Even for non-technical jobs like communications or marketing, what matters more is whether someone actually understands the AI space. That can’t be faked with a certificate.”

mjh@heraldcorp.com



Source link

Continue Reading

Trending