Connect with us

AI Research

GenAI Gains Importance to CFOs as Workforce Needs Change

Published

on


The ranks of CFOs who consider generative artificial intelligence (GenAI) as “very” or “extremely” important to their enterprises have about doubled across use cases, according to the upcoming CAIO report from PYMNTS Intelligence.

These high-impact companies are also pulling away from the pack when it comes to deploying GenAI effectively in their businesses. While most companies use GenAI to generate content, high-impact firms tend to use it for more strategic and advanced functions while lower-impact firms mainly are using the technology for entry-level tasks.

These are among the findings of the report, based on interviews with 60 CFOs at U.S. firms with over $1 billion in annual revenue.

Among high-impact companies, the top three use cases for GenAI are innovating of products and services; generating real-time, automated responses to customer queries; and creating contextualized emails or recommendations. Each were cited by 87.5% of CFOs.

The importance of these three use cases have nearly or more than doubled from a year ago, according to the CFOs surveyed.

Among medium- and low-impact companies — those in which GenAI still remains in early to mid-stage deployment or relegated to entry-level tasks — CFOs generally see smaller gains or even declines in the impact of GenAI on their businesses, year-over-year.

Top use cases in these two groups are:

  • Medium-impact companies: answering customer queries in real time, writing emails and recommendations, and a tie between identifying fraud, errors and consistencies and helping customers and employees access data.
  • Low-impact firms: answering customer queries in real time, helping workers and employees access data, and run an automated workflow management system.

When looking at all companies regardless of GenAI’s impact, the top use case for the technology is financial planning and analysis, with 90% of CFOs saying it is highly important compared with 46.7% saying the same thing a year ago. In cost management, the percentages are 85% vs. 40%; in treasury management, it is 58.3% compared with 25%.

The report found that GenAI is being adopted for a wide array of tasks across enterprises. Its use in spotting fraud and errors is up 31% since March 2024, and use for product and service innovation has risen by 21% in the same time frame.

GenAI Budget Shrinks

But despite a sharp increase in the use and perceived importance of GenAI across large U.S. enterprises, investment growth in GenAI is slowing across industries. Only 27% of CFOs say their companies are raising their GenAI budgets this year — down from 53% in 2024.

The two trends underscore a corporate balancing act: While GenAI is beneficial to business, an uncertain macroeconomic environment amid two global wars is making companies more fiscally cautious.

The willingness to expand AI investment also varied depending on perceived ROI. Among firms that reported very positive ROI from GenAI, 50% planned to boost their budgets, down from 68.8% a year ago.

In contrast, only 16.7% of those with negligible ROI from GenAI expected to do the same, down from 50% in 2024. Among those that have seen “somewhat positive” ROI, 22.2% plan to spend more on GenAI, down from 47% last year.

Automation level also played a key role. One-third of highly automated enterprises — those with low reliance on human operators — planned to raise GenAI spending. Only 21% of firms in low-automation environments intended to do so.

While investment levels are cooling, the demand for employees skilled in AI is surging. Every CFO surveyed said the implementation of GenAI has increased their company’s need for more analytically skilled workers. The 100% showing compares to 58.3% saying the same last year.

In addition, 96.7% said the mix of skills needed across their organizations has changed in the GenAI era. They need more skilled workers and fewer entry-level employees: 83.3% of CFOs said they need a larger tech or IT support team, up from 58.1% in 2024.

Meanwhile, 61.7% of CFOs said they need fewer lower-skilled workers, compared to 64.3% in 2024. Also, 45% said they don’t need to hire as many people in certain parts of the business, up from 40% a year ago.

Overall, the findings point to a recognition among CFOs that GenAI is a strategic imperative instead of a fad. However, they still face constraints in budgeting as macroeconomic and geopolitical issues emerge.

At the same time, CFOs have to carefully calibrate their labor needs as they pivot from lower-skilled, entry-level workers to AI talent for their tech or IT support teams — with ROI squarely in the crosshairs.

Read more: AI at the Crossroads: Agentic Ambitions Meet Operational Realities

Read more: From Spark to Strategy: How Product Leaders Are Using GenAI to Gain a Competitive Edge

Read more: The Investment Impact of GenAI Operating Standards on Enterprise Adoption

 



Source link

AI Research

I asked ChatGPT to help me pack for my vacation – try this awesome AI prompt that makes planning your travel checklist stress-free

Published

on


It’s that time of year again, when those of us in the northern hemisphere pack our sunscreen and get ready to venture to hotter climates in search of some much-needed Vitamin D.

Every year, I book a vacation, and every year I get stressed as the big day gets closer, usually forgetting to pack something essential, like a charger for my Nintendo Switch 2, or dare I say it, my passport.



Source link

Continue Reading

AI Research

Sakana AI: Think LLM dream teams, not single models

Published

on


Enterprises may want to start thinking of large language models (LLMs) as ensemble casts that can combine knowledge and reasoning to complete tasks, according to Japanese AI lab Sakana AI.

Sakana AI in a research paper outlined a method called Multi-LLM AB-MCTS (Adaptive Branching Monte Carlo Tree Search) that uses a collection of LLMs to cooperate, perform trial-and-error and leverage strengths to solve complex problems.

In a post, Sakana AI said:

“Frontier AI models like ChatGPT, Gemini, Grok, and DeepSeek are evolving at a breathtaking pace amidst fierce competition. However, no matter how advanced they become, each model retains its own individuality stemming from its unique training data and methods. We see these biases and varied aptitudes not as limitations, but as precious resources for creating collective intelligence. Just as a dream team of diverse human experts tackles complex problems, AIs should also collaborate by bringing their unique strengths to the table.”

Sakana AI said AB-MCTS is a method for inference-time scaling to enable frontier AIs to cooperate and revisit problems and solutions. Sakana AI released the algorithm as an open source framework called TreeQuest, which has a flexible API that allows users to use AB-MCTS for tasks with multiple LLMs and custom scoring.

What’s interesting is that Sakana AI gets out of that zero-sum LLM argument. The companies behind LLM training would like you to think there’s one model to rule them all. And you’d do the same if you were spending so much on training models and wanted to lock in customers for scale and returns.

Sakana AI’s deceptively simple solution can only come from a company that’s not trying to play LLM leapfrog every few minutes. The power of AI is in the ability to maximize the potential of each LLM. Sakana AI said:

“We saw examples where problems that were unsolvable by any single LLM were solved by combining multiple LLMs. This went beyond simply assigning the best LLM to each problem. In (an) example, even though the solution initially generated by o4-mini was incorrect, DeepSeek-R1-0528 and Gemini-2.5-Pro were able to use it as a hint to arrive at the correct solution in the next step. This demonstrates that Multi-LLM AB-MCTS can flexibly combine frontier models to solve previously unsolvable problems, pushing the limits of what is achievable by using LLMs as a collective intelligence.”

A few thoughts:

  • Sakana AI’s research and move to emphasize collective intelligence over on LLM and stack is critical to enterprises that need to create architectures that don’t lock them into one provider.
  • AB-MCTS could play into what agentic AI needs to become to be effective and complement emerging standards such as Model Context Protocol (MCP) and Agent2Agent.
  • If combining multiple models to solve problems becomes frictionless, the costs will plunge. Will you need to pay up for OpenAI when you can leverage LLMs like DeepSeek combined with Gemini and a few others? 
  • Enterprises may want to start thinking about how to build decision engines instead of an overall AI stack. 
  • We could see a scenario where a collective of LLMs achieves superintelligence before any one model or provider. If that scenario plays out, can LLM giants maintain valuations?
  • The value in AI may not be in the infrastructure or foundational models in the long run, but the architecture and approaches.

More:



Source link

Continue Reading

AI Research

Positive attitudes toward AI linked to more prone to problematic social media use

Published

on


People who have a more favorable view of artificial intelligence tend to spend more time on social media and may be more likely to show signs of problematic use, according to new research published in Addictive Behaviors Reports.

The new study was designed to explore a question that, until now, had been largely overlooked in the field of behavioral research. While many factors have been identified as risk factors for problematic social media use—including personality traits, emotional regulation difficulties, and prior mental health issues—no research had yet explored whether a person’s attitude toward artificial intelligence might also be linked to unhealthy social media habits.

The researchers suspected there might be a connection, since social media platforms are deeply intertwined with AI systems that drive personalized recommendations, targeted advertising, and content curation.

“For several years, I have been interested in understanding how AI shapes societies and individuals. We also recently came up with a framework called IMPACT to provide a theoretical framework to understand this. IMPACT stand for the Interplay of Modality, Person, Area, Country/Culture and Transparency variables, all of relevance to understand what kind of view people form regarding AI technologies,” said study author Christian Montag, a distinguished professor of cognitive and brain sciences at the Institute of Collaborative Innovation at University of Macau.

Artificial intelligence plays a behind-the-scenes role in nearly every major social media platform. Algorithms learn from users’ behavior and preferences in order to maximize engagement, often by showing content that is likely to capture attention or stir emotion. These AI-powered systems are designed to increase time spent on the platform, which can benefit advertisers and the companies themselves. But they may also contribute to addictive behaviors by making it harder for users to disengage.

Drawing from established models in psychology, the researchers proposed that attitudes toward AI might influence how people interact with social media platforms. In this case, people who trust AI and believe in its benefits might be more inclined to embrace AI-powered platforms like social media—and potentially use them to excess.

To investigate these ideas, the researchers analyzed survey data from over 1,000 adults living in Germany. The participants were recruited through an online panel and represented a wide range of ages and education levels. After excluding incomplete or inconsistent responses and removing extreme outliers (such as those who reported using social media for more than 16 hours per day), the final sample included 1,048 people, with roughly equal numbers of men and women.

Participants completed a variety of self-report questionnaires. Attitudes toward artificial intelligence were measured using both multi-item scales and single-item ratings. These included questions such as “I trust artificial intelligence” and “Artificial intelligence will benefit humankind” to assess positive views, and “I fear artificial intelligence” or “Artificial intelligence will destroy humankind” to capture negative perceptions.

To assess social media behavior, participants were asked whether they used platforms like Facebook, Instagram, TikTok, YouTube, or WhatsApp, and how much time they spent on them each day, both for personal and work purposes. Those who reported using social media also completed a measure called the Social Networking Sites–Addiction Test, which includes questions about preoccupation with social media, difficulty cutting back, and using social media to escape from problems.

Overall, 956 participants said they used social media. Within this group, the researchers found that people who had more positive attitudes toward AI also tended to spend more time on social media and reported more problematic usage patterns. This relationship held for both men and women, but it was stronger among men. In contrast, negative attitudes toward AI showed only weak or inconsistent links to social media use, suggesting that it is the enthusiastic embrace of AI—not fear or skepticism—that is more closely associated with excessive use.

“It is interesting to see that the effect is driven by the male sample,” Montag told PsyPost. “On second thought, this is not such a surprise, because in several samples we saw that males reported higher positive AI attitudes than females (on average). So, we must take into account gender for research questions, such as the present one.”

“Further I would have expected that negative AI attitudes would have played a larger role in our work. At least for males we observed that fearing AI went also along with more problematic social media use, but this effect was mild at best (such a link might be explained via negative affect and escapism tendencies). I would not be surprised if such a link becomes more visible in future studies. Let’s keep in mind that AI attitudes might be volatile and change (the same of course is also true for problematic social media use).”

To better understand how these variables were related, the researchers conducted a mediation analysis. This type of analysis can help clarify whether one factor (in this case, time spent on social media) helps explain the connection between two others (positive AI attitudes and problematic use).

The results suggested that people with positive attitudes toward AI tended to spend more time on social media, and that this increased usage was associated with higher scores on the addiction measure. In other words, time spent on social media partly accounted for the link between AI attitudes and problematic behavior.

“I personally believe that it is important to have a certain degree of positive attitude towards benevolent AI technologies,” Montag said. “AI will profoundly change our personal and business lives, so we should better prepare ourselves for active use of this technology. This said, our work shows that positive attitudes towards AI, which are known to be of relevance to predict AI technology use, might come with costs. This might be in form of over-reliance on such technology, or in our case overusing social media (where AI plays an important role in personalizing content). At least we saw this to be true for male study participants.”

Importantly, the researchers emphasized that their data cannot establish cause and effect. Because the study was cross-sectional—that is, based on a single snapshot in time—it is not possible to say whether positive attitudes toward AI lead to excessive social media use, or whether people who already use social media heavily are more likely to hold favorable views of AI. It’s also possible that a third factor, such as general interest in technology, could underlie both tendencies.

The study’s sample, while diverse in age and gender, skewed older on average, with a mean age of 45. This may limit the generalizability of the findings, especially to younger users, who are often more active on social media and may have different relationships with technology. Future research could benefit from focusing on younger populations or tracking individuals over time to see how their attitudes and behaviors change.

“In sum, our work is exploratory and should be seen as stimulating discussions. For sure, it does not deliver final insights,” Montag said.

Despite these limitations, the findings raise important questions about how people relate to artificial intelligence and how that relationship might influence their behavior. The authors suggest that positive attitudes toward AI are often seen as a good thing—encouraging people to adopt helpful tools and new innovations. But this same openness to AI might also make some individuals more vulnerable to overuse, especially when the technology is embedded in products designed to maximize engagement.

The researchers also point out that people may not always be aware of the role AI plays in their online lives. Unlike using an obvious AI system, such as a chatbot or virtual assistant, browsing a social media feed may not feel like interacting with AI. Yet behind the scenes, algorithms are constantly shaping what users see and how they engage. This invisible influence could contribute to compulsive use without users realizing how much the technology is guiding their behavior.

The authors see their findings as a starting point for further exploration. They suggest that researchers should look into whether positive attitudes toward AI are also linked to other types of problematic online behavior, such as excessive gaming, online shopping, or gambling—especially on platforms that make heavy use of AI. They also advocate for studies that examine whether people’s awareness of AI systems influences how those systems affect them.

“In a broader sense, we want to map out the positive and negative sides of AI technology use,” Montag explained. “I think it is important that we use AI in the future to lead more productive and happier lives (we investigated also AI-well-being in this context recently), but we need to be aware of potential dark sides of AI use.”

“We are happy if people are interested in our work and if they would like to support us by filling out a survey. Here we do a study on primary emotional traits and AI attitudes. Participants also get as a ‘thank you’ insights into their personality traits: https://affective-neuroscience-personality-scales.jimdosite.com/take-the-test/).”

The study, “The darker side of positive AI attitudes: Investigating associations with (problematic) social media use,” was authored by Christian Montag and Jon D. Elhai.



Source link

Continue Reading

Trending