Connect with us

Business

Colorado Delays AI Act; Businesses Get Breathing Room

Published

on


The five-month extension gives lawmakers more time to consider changes during the 2026 regular session, though similar efforts since the law’s passage in May 2024 have repeatedly stalled. Even with the new deadline, the timeline remains tight: Colorado’s session is scheduled to adjourn in May 2026, which would leave only weeks for the attorney general to issue rules and for employers to adapt compliance programs unless another delay is approved.

The Colorado AI Act is the first broad U.S. law regulating “high-risk” AI systems used in consequential decisions, from hiring and housing to lending, healthcare and education. It requires developers to disclose training data summaries, known risks and usage instructions, while employers and other deployers must adopt risk-management programs, conduct annual impact assessments and notify consumers when AI plays a role in significant decisions. The statute also creates appeal rights for individuals affected by AI-based outcomes, and violations may be pursued as unfair or deceptive trade practices, according to KPMG.

For businesses, the delay provides a narrow but valuable window to prepare. Analysts caution that documentation, auditing and monitoring obligations could impose significant operational costs. A study by the Common Sense Institute estimated that even a 1% increase in production costs tied to compliance could ripple across industries such as financial services, healthcare, housing and insurance. For FinTechs that build AI-enabled HR or credit-decisioning tools, the law is both a hurdle and an opportunity to differentiate products by meeting Colorado’s higher bar.

Colorado’s framework is being closely monitored nationally as other states, including California, New York and Illinois, weigh similar legislation. At the federal level, the Federal Trade Commission has signaled that it will examine how companies’ market and deploy AI, focusing on misleading claims and practices that could harm consumers or distort competition and the Equal Employment Opportunity Commission has signaled growing oversight of AI in employment.

The delay reflects broader uncertainty around AI regulation, as companies weigh compliance strategies against shifting legislative timelines. Still, as PYMNTS has reported, the direction is clear: Businesses will be held accountable for how their AI systems affect workers and consumers. The question for companies is not whether compliance obligations will arrive, but how quickly they will need to adapt, and which firms can turn early readiness into a competitive advantage.



Source link

Business

Robinhood CEO Says AI Won’t Fully Take Over Trading

Published

on


AI isn’t ready to take over all the human aspects of trading, said Robinhood’s CEO.

In an interview with Bloomberg Wealth released on Wednesday, Vlad Tenev reiterated his view that people don’t always trade to make money and that there will always be a human aspect to business.

“Most of the time you’re not doing it just because you want to make money,” he said. “You also love trading and you’re you’re extremely passionate about it.”

He added: “I don’t think there’s going to be a future where AI just does all of your thinking, all of your financial planning, all the strategizing for you.”

Tenev, who cofounded the brokerage platform in 2013, said that AI could be a bigger platform shift than mobile and cloud technologies. He said that while every company will quickly become an AI company, AI won’t completely take over trading.

“It’ll be a helpful assistant to a trader and also to your broader financial life,” he said. “But I think the humans will ultimately be calling the shots.”

Tenev made similar remarks about how investors “legitimately enjoy trading” in an August interview with Axios.

On Tuesday, Robinhood announced it was building a social media platform where users can post their trades and track what other investors, including politicians, are buying or selling.

Other CEOs are also wary of pronouncing AI the future of trading.

Ken Griffin, the founder and CEO of Citadel, said he doesn’t think AI will revolutionize the investment business.

“Do we use it in our investment business? A little bit, a little bit. I can’t say it’s been game-changing,” Griffin said in a May interview with Stanford Graduate School of Business.

“It saves some time. It’s a productivity enhancement tool. It’s nice, I don’t think it’s going to revolutionize most of what we do in finance,” Griffin added.

On the topic of AI and trading, Goldman Sachs CEO David Solomon has said that AI has been a big boost to productivity in the investing business.

In an interview with CNBC last year, he said that 40 years ago, when he started the banking business, it took six hours to compare two stocks. Now, it takes an instant, he said.

The firm has already launched ventures that could change how Wall Street makes deals. Louisa AI, a startup founded within the firm six years ago, helps bankers and investors analyze millions of articles and employees’ knowledge to identify deals.





Source link

Continue Reading

Business

Aha moments, the ‘first ten hours’, and other pro tips from business leaders building AI-ready workforces

Published

on


As businesses face pressure to bring new AI tools on board, they have the dual challenge of effectively incorporating the technology into their operations and of helping their workforce make the best use of the technology. 

Longstanding methods for assessing the skills and performance of an employee, as well as hiring practices, are being upended and re-imagined, according to business leaders who spoke at the Fortune Brainstorm Tech conference on Tuesday in Park City, Utah. 

Technical skills, contrary to what you might think, are not paramount in the age of AI. In fact, for many employers, technical skills are becoming less important.

“For the first time this summer on our platform we saw a shift,” said Hayden Brown, CEO of Upwork, an online jobs marketplace for freelancers. In the past, when Upwork asked employers on its platform about the most important skills they were hiring for, the answer invariably involved deep expertise in certain technical areas, Brown said. “For the first time this summer, it’s now soft skills. It’s human skills; it’s things like problem solving, judgement, creativity, taste.” 

Jim Rowan, the head of AI at consulting firm Deloitte, which sponsored the Brainstorm discussion, said an employee’s “fluency” should not be an end goal in itself. More important is intellectual curiosity around new tools and technology.

And that’s something that needs to start at the top.

“We’ve done a lot of work with executive teams to make sure the top levels of the organization and the boards are actually familiar with AI,” said Rowan. “That helps because then they can communicate better with their teams and see what they’re doing.” 

For Toni Vanwinkle, VP of Digital Employee Experience at Adobe, it’s critical for employees at all levels of an organization to have an “aha moment” with AI technology. And the best way to bring that about is for each employee to get their “first ten hours” in. 

“Go play with it,” Vanwinkle says. “Sort your email box, take the notes in your meeting, create a marketing campaign, whatever it is that you do.” Through that initial process of personal exploration, you start to understand the potential of the technology, she says.

The next step, Vanwinkle says, is collaboration, discussions, and experimentation among colleagues within the same departments or functionalities.

“This whole spirit of experiment, learn fast. That twitch muscle can turn into something of value when people talk openly,” Vanwinkle says.

The importance of embracing experimentation, and fostering it as a value within the organization, was echoed by Indeed chief information officer Anthony Moisant.

“I think about the pilots we run, most of them fail. And I’m not embarrassed at all to say that,” Moisant says. It all comes down to what a particular organization is optimizing for, and in the case of Indeed, Moisant says, “what we go for is fast twitch muscle. Can we move faster?”

By encouraging more low stakes experiments with AI, companies can gain valuable insights and experience that employees can leverage quickly when it counts. “The only way to move faster is to take a few bets early on, without real long term strategic ROI,” says Moisant.

Workday Vice President of AI Kathy Pham emphasizes that with new tools like AI, getting a full picture of an employee’s value and performance may take a bit longer than some people are used to. “Part of the measurement is better understanding what the return is and over what period of time,” she said.

Fortune Global Forum returns Oct. 26–27, 2025 in Riyadh. CEOs and global leaders will gather for a dynamic, invitation-only event shaping the future of business. Apply for an invitation.



Source link

Continue Reading

Business

AI agents in business create new risks & urgent security needs

Published

on


Radware has released new research analysing the cybersecurity risks linked to the growing adoption of agentic artificial intelligence (AI) systems in enterprise environments.

The report, titled ‘The Internet of Agents: The Next Threat Surface,’ examines how AI agents driven by large language models (LLMs) are being integrated into business operations. These systems differ from standard chatbots by acting autonomously, executing tasks, and collaborating with other digital agents using protocols such as the Model Context Protocol (MCP) and Agent-to-Agent (A2A).

Expanded attack surface

Organisations are increasingly deploying LLM-powered AI agents into customer service, development, and operational workflows. Unlike traditional software, these agents are capable of reasoning, executing commands, and initiating actions autonomously across enterprise networks.

The report notes that as these agents interact with business systems, they establish complex transitive chains of access to sensitive resources. This can complicate tracking and securing business processes with existing cybersecurity measures. According to Radware, these pathways present “complex pathways into sensitive enterprise resources that are difficult to track or secure with existing defences.”

New protocols and exposures

The adoption of protocols such as MCP and A2A enables enhanced interoperability and scalability for AI agents across different business processes, but this also introduces new risks. The report highlights threats such as prompt injection, tool poisoning, lateral compromise, and malicious handshakes, which take advantage of these emerging protocols.

Prompt injection attacks, in particular, are identified as a growing risk. By embedding covert instructions in content like emails or web pages, attackers can manipulate AI agents to exfiltrate data or initiate unauthorised actions – often without any indication to the end user. The research states: “Adversaries can embed hidden instructions in emails, web pages, or documents. When an AI agent processes that content, it may unwittingly exfiltrate data or trigger unauthorized actions – without the user ever clicking a link or approving a request.”

Lower barriers to cybercrime

The report observes that new “dark AI ecosystems” are emerging, which lower the technical barriers to cybercrime. Black-hat platforms such as XanthoroxAI provide access to offensive AI tools that automate previously manual attacks, including malware creation and phishing campaigns. These tools, offered on a subscription basis, enable less-experienced attackers to develop and deploy exploits more easily.

Radware’s analysis also shows that AI accelerates the weaponisation of new vulnerabilities. The report references research demonstrating that GPT-4 can develop working exploits for recently disclosed vulnerabilities more rapidly than experienced human researchers, reducing the window for IT teams to patch vulnerable systems before attackers strike.

Changing digital landscape

The emergence of the so-called ‘Internet of Agents’ is likened to previous digital shifts, such as the rise of the Internet of Things. In this new ecosystem, autonomous digital actors with memory, reasoning, and action capabilities are increasingly interconnected, resulting in greater operational efficiency but also expanded risk exposure.

Radware’s report argues that organisations must adjust their security models to account for the new roles played by AI agents in the enterprise. With these systems acting as decision-makers, intermediaries, and operational partners, the need for effective governance and security oversight is heightened.

“We are not entering an AI future – we are already living in it,” said [Insert Radware Spokesperson]. “The agentic ecosystem is expanding rapidly across industries, but without strong security and oversight, these systems risk becoming conduits for cybercrime. The businesses that succeed will be those capable of delivering trustworthy, secure AI experiences.”

Security recommendations

The report sets out a series of recommendations for enterprises to protect against the unique risks posed by autonomous AI agents. These include:

  • Treating LLMs and AI agents as privileged actors, subject to strict governance and access controls.
  • Integrating red-teaming and prompt evaluation exercises into software development lifecycles.
  • Evaluating protocols such as MCP and A2A as security-critical interfaces, rather than mere productivity tools.
  • Monitoring dark AI ecosystems to stay aware of how adversaries are adapting and exploiting new tools.
  • Investing in detection, sandboxing, and behavioural monitoring technologies tailored for autonomous AI systems.
  • Recognising that AI-powered defensive capabilities will play an increasingly important role in combating AI-driven threats.

The report concludes by noting that AI agents represent a significant technology shift for businesses. Although these systems hold potential for efficiency and economic growth, they also introduce risks that enterprises must urgently address as the boundaries between helpful tools and security threats continue to blur.



Source link

Continue Reading

Trending