AI Insights
Artificial Intelligence News for the Week of September 5; Updates from DeepL, Denodo, Snowflake & More

Solutions Review Executive Editor Tim King curated this list of notable artificial intelligence news for the week of September 5, 2025.
Keeping tabs on all the most relevant artificial intelligence news can be a time-consuming task. As a result, our editorial team aims to provide a summary of the top headlines from the last week in this space. Solutions Review editors will curate vendor product news, mergers and acquisitions, venture capital funding, talent acquisition, and other noteworthy artificial intelligence news items.
For early access to all the expert insights published on Solutions Review, join Insight Jam, a community dedicated to enabling the human conversation on AI.
Artificial Intelligence News for the Week of September 5, 2025
Apiiro Study: AI Coding Assistants Boost Velocity but Multiply Vulnerabilities
A new research blog from Apiiro shows that while AI coding assistants increase engineering velocity by up to 4x, they also introduce tenfold more vulnerabilities in code. Organizations are urged to invest in code review automation, secure coding training, and risk mitigation strategies to close the expanding security gap.
Read more → https://apiiro.com/blog/4x-velocity-10x-vulnerabilities-ai-coding-assistants-are-shipping-more-risks/
Cisco and NVIDIA Expand Secure AI Factory to Accelerate Agentic AI for Enterprises
Cisco, in collaboration with NVIDIA and VAST Data, has announced an expanded Secure AI Factory designed to make enterprise data more accessible, secure, and actionable for agentic AI. The solution leverages the VAST Data InsightEngine, NVIDIA AI Data Platform reference design, and Cisco AI PODs with NVIDIA RTX PRO 6000 Blackwell Server GPUs to drastically reduce latency in retrieval-augmented generation (RAG) pipelines—from minutes to seconds.
Read more → https://www.channele2e.com/news/cisco-secure-ai-factory-with-nvidia-unlocks-enterprise-data-for-agentic-ai
Concentric AI Integrates with OpenAI ChatGPT for Smart Data Security
Concentric AI is now integrating with OpenAI’s ChatGPT to deliver intelligent context-aware data protection capabilities. This partnership enables enterprises to leverage advanced NLP-driven suggestions for sensitive data classification, automated policy enforcement, and proactive risk management in cloud and SaaS environments.
Read more → https://concentric.ai/press-release/concentric-ai-announces-integration-with-openais-chatgpt/
DeepL Unveils Autonomous AI Agent for Business
DeepL has announced its new autonomous AI agent for business applications, offering advanced natural language understanding, translation, and information retrieval. The agent is designed to support customer service, content generation, and multilingual communications for global enterprises.
Read more → https://www.prnewswire.com/news-releases/deepl-unveils-autonomous-ai-agent-for-businesses-302543914.html
DE-CIX Launches World’s First Fully Integrated AI Internet Exchange
DE-CIX has introduced the world’s first AI-integrated Internet Exchange (AI-IX), enabling efficient cross-network connectivity for model training and AI inference. The platform unifies edge, cloud, and analytics traffic to boost throughput, reduce latency, and accelerate enterprise AI adoption.
FICO & Corinium Study: Responsible AI Standards Surpass GenAI Hype for Finance
FICO and Corinium’s latest study finds that global financial institutions are moving beyond GenAI hype, prioritizing responsible AI standards to ensure trust, compliance, and value. Leaders now favor transparency, ethical model development, and rigorous governance to sustain adoption and minimize risk.
Geniez AI Raises $6M Seed Round for AI-Driven Process Automation
Geniez AI, a startup developing automation solutions powered by generative AI, has raised $6 million in seed funding. The capital will be used to advance Geniez’s AI agent platform for automating complex business processes and workflows, focusing on enterprise adoption and rapid go-to-market acceleration.
Read more → https://www.finsmes.com/2025/09/geniez-ai-raises-6m-in-seed-funding.html
Google Cloud Report: 52% of Executives Have Deployed AI Agents for Business Value
A Google Cloud study finds that 52 percent of global executives have already deployed AI agents in their organizations—unlocking significant new business value in customer engagement, productivity, and automation. The survey highlights a growing wave of adoption as leaders recognize agentic AI’s potential to drive competitive transformation.
Gravity Launches Orion Multi-Agent AI for Enterprise Decision Intelligence
Gravity unveiled Orion, a multi-agent AI platform to transform enterprise decision intelligence. Orion uses coordinated AI agents to autonomously collect, enrich, and analyze data across business units, empowering executives with actionable insights for complex decision-making. The platform marks a new milestone in scalable, agent-driven enterprise analytics.
Laserfiche AI Workplace Survey Finds Automation Boosts Retention, Productivity
Laserfiche’s “AI Workplace Survey 2025” highlights that organizations implementing AI-driven automation report significant gains in employee retention and productivity. However, survey respondents emphasize the need for upskilling, change management, and clearer strategies to avoid AI change fatigue. Leaders are encouraged to balance automation with proactive workforce development practices.
Read more → https://www.laserfiche.com/resources/press-center/press/ai-workplace-survey-2025/
Pearson and Cognizant Announce Strategic Partnership for AI-Era Learning
Pearson and Cognizant have formed a global partnership to create new learning and workforce development programs supporting early career growth in the age of AI. The collaboration combines educational technology, professional services, and enterprise AI expertise to deliver career-ready content and improve access to high-demand digital skills.
Qlik Achieves AWS Generative AI Competency
Qlik has attained AWS Generative AI Competency, affirming its excellence in deploying generative AI solutions within the AWS ecosystem. This milestone validates Qlik’s ability to empower customers with scalable generative analytics, AI-driven automation, and secure cloud data integrations.
Read more → https://www.qlik.com/us/news/company/press-room/press-releases/qlik-achieves-aws-generative-ai-competency
Redis Acquires Decodable to Expand Real-Time Context and Memory for AI Agents
Redis has announced the acquisition of real-time data platform Decodable. The move expands Redis’ capabilities for AI, bringing streaming context and memory solutions for agentic AI systems—including advanced retrieval, embedding, and real-time data enrichment. The integration positions Redis as a foundational component for next-generation, memory-enabled enterprise AI.
WisdomAI Unleashes Autonomous AI Data Analyst
WisdomAI has launched its AI-powered data analyst, enabling businesses to automate complex analytics tasks, including report generation, trend analysis, and on-demand insights. The agent simplifies data exploration for non-technical users and speeds up the business intelligence lifecycle.
Read more → https://www.enterprisetimes.co.uk/2025/09/04/wisdomai-unleashes-its-ai-data-analyst/
Expert Insights
Watch this space each week as our editors will share upcoming events, new thought leadership, and the best resources from Insight Jam, Solutions Review’s enterprise tech community where the human conversation around AI is happening. The goal? To help you gain a forward-thinking analysis and remain on-trend through expert advice, best practices, predictions, and vendor-neutral software evaluation tools.
What to Expect at Solutions Review‘s Spotlight with SoftwareOne on September 10: How to Use Copilot Agents
Generative AI has evolved from a tool we harness for productivity to a true collaborator. With Copilot Agents, AI works alongside us to boost creativity, streamline workflows, and provide real-time coaching—becoming a valuable teammate and ushering in the next era of AI-driven productivity.
What to Expect at Solutions Review‘s Industry Trends Session with Denodo on September 16: AI Deep Research
We’ll explore the kinds of questions Deep Research can tackle: Why is customer churn rising in a specific region? How will new regulations impact our supply chain? What is the optimal strategy for entering a new market? These are the complex, open-ended questions that go far beyond dashboards and canned reports, questions that require reasoning, synthesis, and real-time, trusted data.
New Episode of The Cyber Circuit with Michael Morgenstern: The AI Revolution Needs Better Asset Management
From shadow AI implementations draining sensitive data to why “we went too fast with cloud and we’re doing it again,” discover why the organizations that can’t answer “what assets do we have?” are the ones AI will hurt most. Parker reveals why AI represents the fourth industrial revolution and what security leaders must do now before regulatory hammers drop.
SR Thought Leaders: Agentic AI Governance: 4 Criteria to Evaluate Tools by Kevin Petrie
As companies mature, they are taking a more sober view of the opportunity that agentic AI represents. In fact, only 27 percent of executives say they would trust fully autonomous agents for enterprise use, down from 43 percent one year ago, according to a recent Capgemini survey. They recognize the need to manage risks before reaping the innovation benefits of agentic AI. This requires a robust governance program, supported by effective technology.
Editorial: Soft Skills Crisis? Why Gen Z Struggles – and How Leaders Can Help by Dr. Laurie Cure
Soft skills have always played a quiet but powerful role in workplace success. After all, these are the often invisible qualities that allow people to collaborate, resolve issues, and lead with confidence — often the most essential elements for leadership success.
Editorial: The Evolution of Enterprise Data Archives in the AI Era by Archive360’s George Tziahanas
For decades, enterprise data archives have occupied an understated position within organizational IT infrastructure. These vast repositories of information were treated as necessary for meeting legal and regulatory obligations, but only modestly accessed once data was safely stored away. The prevailing approach was simple: keep costs low, ensure compliance, and preserve the data until a court order or regulatory investigation demanded its retrieval.
Editorial: Why Data Quality is the Make-or-Break Factor for AI Success by Semarchy’s Craig Gravina
According to a recent survey of 1,050 senior business leaders across the US, UK, and France, only 46 percent express confidence in the quality of their data. This lack of trust in data quality represents the Achilles’ heel of many promising AI strategies, underscoring a critical truth: without trustworthy data, even the most sophisticated AI initiatives risk falling short of their potential.
For consideration in future artificial intelligence news roundups, send your announcements to the editor: tking@solutionsreview.com.
AI Insights
Generative vs. agentic AI: Which one really moves the customer experience needle?
Artificial intelligence, first coined by John McCarthy in 1956, lay dormant for decades before exploding into a cultural and business phenomenon post-2012. From predictive algorithms to chatbots and creative tools, AI has evolved rapidly. Now, two powerful paradigms are shaping its future: generative AI, which crafts content from text to art, and agentic AI, which acts autonomously to solve complex tasks. But should businesses pit generative AI against agentic AI or combine them to innovate? The answer isn’t binary, because these technologies aren’t competing forces. In fact, they often complement each other in powerful ways, especially when it comes to transforming customer engagement.
The rise of generative AI: Creativity meets scale
Generative AI is all about creation; it represents the imaginative side of artificial intelligence. From producing marketing copy and designing campaign visuals to generating product descriptions and chat responses, generative AI has unlocked new possibilities for enterprises looking to scale content and personalisation like never before.
Fuelled by powerful models like ChatGPT, DALL·E, and MidJourney, these systems have entered the enterprise stack at speed. Marketing teams are using them to brainstorm ideas and accelerate go-to-market efforts. Customer support teams are deploying them to enhance chatbot interactions with more human-like language. Product teams are using generative AI to auto-draft FAQs or documentation. And sales teams are experimenting with tailored email pitches generated from past deal data.
At the heart of this capability is the model’s ability to learn from massive datasets, analysing and replicating patterns in text, visuals, and code to produce new, relevant content on demand. This has made generative AI a valuable tool in customer engagement workflows where speed, relevance, and personalisation are paramount. But while generative AI can start the conversation, it rarely finishes it. That’s where its limitations show up.
For instance, it can draft a beautifully written response to a billing query, but it can’t resolve the issue by accessing the customer’s account, applying credits, or triggering workflows across enterprise systems. In other words, it creates the message but not the outcome. This creative strength makes generative AI a powerful enabler of customer engagement but not a complete solution. To drive real business value, measured in resolution rates, retention, and revenue, enterprises need to go beyond content generation and toward intelligent action. This is where agentic AI comes into play.
How agentic AI is redefining enterprise and consumer engagement
As the need for deeper automation grows, agentic AI is taking centre stage. Agentic AI is built to act; it makes decisions, takes autonomous actions, and adapts in real time to achieve goals. For businesses, this marks a transformative shift. Generative AI has empowered enterprises to accelerate communication, generate insights, and personalise engagement. Agentic AI, on the other hand, goes beyond assistance to autonomy. Imagine a virtual enterprise assistant that doesn’t just draft emails but manages entire customer service workflows — triggering follow-ups, updating CRM systems, and escalating issues when needed.
In industries like supply chain, finance, and telecom, agentic AI can dynamically reconfigure networks, detect anomalies, or reroute deliveries—all with minimal human input. It’s a new era of AI-driven execution. On the consumer front, agentic AI takes engagement from passive response to proactive assistance. Think of a digital concierge that not only understands your intent but acts on your behalf — tracking shipments or negotiating a better mobile plan based on usage patterns.
A new layer of intelligence — with responsibility
The increased autonomy of agentic AI raises important questions around trust, governance, and accountability. Who’s liable when an agentic system makes an error or an ethically questionable decision? Enterprises adopting such systems will need to ensure alignment with human values, transparency in decision-making, and robust fail-safes.
Generative and agentic AI are not rivals — they’re complementary forces that, together, enable a new era of intelligent enterprise and consumer engagement.
When generative meets agentic AI
Generative AI and agentic AI may serve different functions. However, rather than operating in isolation, these technologies frequently collaborate, enhancing both communication and execution.
Take, for example, a virtual customer service agent. The agentic AI manages the flow of interaction, makes decisions, and determines next steps, while generative AI crafts clear, personalised responses tailored to the conversation in real time.
This collaborative dynamic also plays out in robotics. Imagine a robot chef: generative AI could invent creative recipes based on user tastes and available ingredients, while agentic AI would take over the cooking, executing the recipe with precision and adapting to real-time conditions in the kitchen.
Summing Up
As AI continues to evolve, the boundaries between generative and agentic systems will become increasingly fluid. We’re heading toward a future where AI doesn’t just imagine possibilities but also brings them to life, merging creativity with execution in a seamless loop. This fusion holds immense promise across industries, from streamlining healthcare operations to revolutionising manufacturing workflows.
However, with such transformative power comes great responsibility. Ethical development, transparency, and accountability must remain non-negotiable, especially when it comes to safeguarding consumer data. As these systems take on more autonomous roles, ensuring privacy, security, and user consent will be critical to building trust.
By understanding the distinct roles and combined potential of generative and agentic AI, we can shape a future where technology enhances human capability responsibly, meaningfully, and with integrity at its core.
This article is authored by Harsha Solanki, VP GM Asia, Infobip.
Disclaimer: The views expressed in this article are those of the author/authors and do not necessarily reflect the views of ET Edge Insights, its management, or its members
AI Insights
How AI is eroding human memory and critical thinking
AI Insights
The human thinking behind artificial intelligence

Artificial intelligence is built on the thinking of intelligent humans, including data labellers who are paid as little as US$1.32 per hour. Zena Assaad, an expert in human-machine relationships, examines the price we’re willing to pay for this technology. This article was originally published in the Cosmos Print Magazine in December 2024.
From Blade Runner to The Matrix, science fiction depicts artificial intelligence as a mirror of human intelligence. It’s portrayed as holding a capacity to evolve and advance with a mind of its own. The reality is very different.
The original conceptions of AI, which hailed from the earliest days of computer science, defined it as the replication of human intelligence in machines. This definition invites debate on the semantics of the notion of intelligence.
Can human intelligence be replicated?
The idea of intelligence is not contained within one neat definition. Some view intelligence as an ability to remember information, others see it as good decision making, and some see it in the nuances of emotions and our treatment of others.
As such, human intelligence is an open and subjective concept. Replicating this amorphous notion in a machine is very difficult.
Software is the foundation of AI, and software is binary in its construct; something made of two things or parts. In software, numbers and values are expressed as 1 or 0, true or false. This dichotomous design does not reflect the many shades of grey of human thinking and decision making.
Not everything is simply yes or no. Part of that nuance comes from intent and reasoning, which are distinctly human qualities.
To have intent is to pursue something with an end or purpose in mind. AI systems can be thought to have goals, in the form of functions within the software, but this is not the same as intent.
The main difference is goals are specific and measurable objectives whereas intentions are the underlying purpose and motivation behind those actions.
You might define the goals as ‘what’, and intent as ‘why’.
To have reasoning is to consider something with logic and sensibility, drawing conclusions from old and new information and experiences. It is based on understanding rather than pattern recognition. AI does not have the capacity for intent and reasoning and this challenges the feasibility of replicating human intelligence in a machine.
There is a cornucopia of principles and frameworks that attempts to address how we design and develop ethical machines. But if AI is not truly a replication of human intelligence, how can we hold these machines to human ethical standards?
Can machines be ethical?
Ethics is a study of morality: right and wrong, good and bad. Imparting ethics on a machine, which is distinctly not human, seems redundant. How can we expect a binary construct, which cannot reason, to behave ethically?
Similar to the semantic debate around intelligence, defining ethics is its own Pandora’s box. Ethics is amorphous, changing across time and place. What is ethical to one person may not be to another. What was ethical 5 years ago may not be considered appropriate today.
These changes are based on many things; culture, religion, economic climates, social demographics, and more. The idea of machines embodying these very human notions is improbable, and so it follows that machines cannot be held to ethical standards. However, what can and should be held to ethical standards are the people who make decisions for AI.
Contrary to popular belief, technology of any form does not develop of its own accord. The reality is their evolution has been puppeteered by humans. Human beings are the ones designing, developing, manufacturing, deploying and using these systems.
If an AI system produces an incorrect or inappropriate output, it is because of a flaw in the design, not because the machine is unethical.
The concept of ethics is fundamentally human. To apply this term to AI, or any other form of technology, anthropomorphises these systems. Attributing human characteristics and behaviours to a piece of technology creates misleading interpretations of what that technology is and is not capable of.
Decades long messaging about synthetic humans and killer robots have shaped how we conceptualise the advancement of technology, in particular, technology which claims to replicate human intelligence.
AI applications have scaled exponentially in recent years, with many AI tools being made freely available to the general public. But freely accessible AI tools come at a cost. In this case, the cost is ironically in the value of human intelligence.
The hidden labour behind AI
At a basic level, artificial intelligence works by finding patterns in data, which involves more human labour than you might think.
ChatGPT is one example of AI, referred to as a large language model (LLM). ChatGPT is trained on carefully labelled data which adds context, in the form of annotations and categories, to what is otherwise a lot of noise.
Using labelled data to train an AI model is referred to as supervised learning. Labelling an apple as “apple”, a spoon as “spoon”, a dog as “dog”, helps to contextualise these pieces of data into useful information.
When you enter a prompt into ChatGPT, it scours the data it has been trained on to find patterns matching those within your prompt. The more detailed the data labels, the more accurate the matches. Labels such as “pet” and “animal” alongside the label “dog” provide more detail, creating more opportunities for patterns to be exposed.
Data is made up of an amalgam of content (images, words, numbers, etc.) and it requires this context to become useful information that can be interpreted and used.
As the AI industry continues to grow, there is a greater demand for developing more accurate products. One of the main ways for achieving this is through more detailed and granular labels on training data.
Data labelling is a time consuming and labour intensive process. In absence of this work, data is not usable or understandable by an AI model that operates through supervised learning.
Despite the task being essential to the development of AI models and tools, the work of data labellers often goes entirely unnoticed and unrecognised.
Data labelling is done by human experts and these people are most commonly from the Global South – Kenya, India and the Philippines. This is because data labelling is labour intensive work and labour is cheaper in the Global South.
Data labellers are forced to work under stressful conditions, reviewing content depicting violence, self-harm, murder, rape, necrophilia, child abuse, bestiality and incest.
Data labellers are pressured to meet high demands within short timeframes. For this, they earn as little as US$1.32 per hour, according to TIME magazine’s 2023 reporting, based on an OpenAI contract with data labelling company Sama.
Countries such as Kenya, India and the Philippines incur less legal and regulatory oversight of worker rights and working conditions.
Similar to the fast fashion industry, cheap labour enables cheaply accessible products, or in the case of AI, it’s often a free product.
AI tools are commonly free or cheap to access and use because costs are being cut around the hidden labour that most people are unaware of.
When thinking about the ethics of AI, cracks in the supply chain of development rarely come to the surface of these discussions. People are more focused on the machine itself, rather than how it was created. How a product is developed, be it an item of clothing, a TV, furniture or an AI-enabled capability, has societal and ethical impacts that are far reaching.
A numbers game
In today’s digital world, organisational incentives have shifted beyond revenue and now include metrics around the number of users.
Releasing free tools for the public to use exponentially scales the number of users and opens pathways for alternate revenue streams.
That means we now have a greater level of access to technology tools at a fraction of the cost, or even at no monetary cost at all. This is a recent and rapid change in the way technology reaches consumers.
In 2011, 35% of Americans owned a mobile phone. By 2024 this statistic increased to a whopping 97%. In 1973, a new TV retailed for $379.95 USD, equivalent to $2,694.32 USD today. Today, a new TV can be purchased for much less than that.
Increased manufacturing has historically been accompanied by cost cutting in both labour and quality. We accept poorer quality products because our expectations around consumption have changed. Instead of buying things to last, we now buy things with the expectation of replacing them.
The fast fashion industry is an example of hidden labour and its ease of acceptance in consumers. Between 1970 and 2020, the average British household decreased their annual spending on clothing despite the average consumer buying 60% more pieces of clothing.
The allure of cheap or free products seems to dispel ethical concerns around labour conditions. Similarly, the allure of intelligent machines has created a facade around how these tools are actually developed.
Achieving ethical AI
Artificial intelligence technology cannot embody ethics; however, the manner in which AI is designed, developed and deployed can.
In 2021, UNESCO released a set of recommendations on the ethics of AI, which focus on the impacts of the implementation and use of AI. The recommendations do not address the hidden labour behind the development of AI.
Misinterpretations of AI, particularly those which encourage the idea of AI developing with a mind of its own, isolate the technology from the people designing, building and deploying that technology. These are the people making decisions around what labour conditions are and are not acceptable within their supply chain, what remuneration is and isn’t appropriate for the skills and expertise required for data labelling.
If we want to achieve ethical AI, we need to embed ethical decision making across the AI supply chain; from the data labellers who carefully and laboriously annotate and categorise an abundance of data through to the consumers who don’t want to pay for a service they have been accustomed to thinking should be free.
Everything comes at a cost, and ethics is about what costs we are and are not willing to pay.
-
Business1 week ago
The Guardian view on Trump and the Fed: independence is no substitute for accountability | Editorial
-
Tools & Platforms4 weeks ago
Building Trust in Military AI Starts with Opening the Black Box – War on the Rocks
-
Ethics & Policy1 month ago
SDAIA Supports Saudi Arabia’s Leadership in Shaping Global AI Ethics, Policy, and Research – وكالة الأنباء السعودية
-
Events & Conferences4 months ago
Journey to 1000 models: Scaling Instagram’s recommendation system
-
Jobs & Careers2 months ago
Mumbai-based Perplexity Alternative Has 60k+ Users Without Funding
-
Education2 months ago
VEX Robotics launches AI-powered classroom robotics system
-
Podcasts & Talks2 months ago
Happy 4th of July! 🎆 Made with Veo 3 in Gemini
-
Funding & Business2 months ago
Kayak and Expedia race to build AI travel agents that turn social posts into itineraries
-
Education2 months ago
Macron says UK and France have duty to tackle illegal migration ‘with humanity, solidarity and firmness’ – UK politics live | Politics
-
Podcasts & Talks2 months ago
OpenAI 🤝 @teamganassi