Connect with us

AI Insights

Metaplanet Holders Approve Fresh Funding Tools to Buy Bitcoin

Published

on




Japanese Bitcoin treasury Metaplanet Inc. secured shareholder approval for a proposal enabling it to raise as much as ¥555 billion ($3.8 billion) via preferred shares, in a bid to expand its financing options after its stock slumped.



Source link

Continue Reading
Click to comment

Leave a Reply

Your email address will not be published. Required fields are marked *

AI Insights

Cerillion hails recognition in Gartner artificial intelligence reports

Published

on


(Alliance News) – Cerillion PLC on Tuesday said it has been recognised in two recently published artificial intelligence reports from Gartner Inc, a Connecticut-based research and advisory firm.

The London-based billing, charging and customer relationship management software said it was named in the Gartner’s Magic Quadrant for AI in communication service provider customer and business operations report, and in a report on critical capabilities for AI in the same sector.

Cerillion said Gartner evaluated vendors across criteria including market understanding, product strategy, sales strategy, innovation and customer experience.

The firm said it believes its inclusion in the reports follows its ongoing investment in AI-powered capabilities.

“We’re delighted to be recognised in these new Magic Quadrant and Critical Capabilities reports for AI in CSP customer and business operations,” said Chief Executive Officer Louis Hall.

“We believe it validates our ongoing strategy of embedding advanced AI into our [software-as-a-service]-based, composable [business support systems]/[operations support systems] suite to help CSPs streamline operations, enhance customer experience and accelerate innovation.”

Shares in Cerillion were up 2.0% at 1,397.50 pence in London on Tuesday morning.

By Michael Hennessey, Alliance News reporter

Comments and questions to newsroom@alliancenews.com

Copyright 2025 Alliance News Ltd. All Rights Reserved.



Source link

Continue Reading

AI Insights

Salesforce cuts 4,000 jobs with AI — CEO calls AGI overhyped

Published

on


At the beginning of this year, Salesforce CEO Marc Benioff indicated that the company was seriously debating hiring software engineers in 2025. Consequently, the executive revealed that the tech firm was using AI to do up to 50% of its work, citing incredible productivity gains due to agentic AIs.

During a recent episode of The Logan Bartlett Show, Benioff revealed that AI is on course to replace humans at the workplace, specifically indicating that the technology is helping bolster the company’s sales by augmenting the customer support division, prompting it to cut support staff from 9,000 to 5,000 (via Business Insider).

It’s been eight of the most exciting months of my career. I was able to rebalance my head count on my support. I’ve reduced it from 9,000 heads to about 5,000 because I need less heads.

Salesforce CEO, Marc Benioff



Source link

Continue Reading

AI Insights

AI companies must honour a foundational rule of the internet: Respecting site owners’ wishes on content

Published

on


Open this photo in gallery:

AI tools are diverting traffic and attention away from Wikipedia and other websites, starving them of the revenue that they need to survive, researchers have concluded.Gregory Bull/The Associated Press

Viet Vu is manager of economic research at the Dais at Toronto Metropolitan University.

One foundational rule of the internet has been the social contract between website owners and search giants such as Google.

Owners would agree to let search-engine bots crawl and index their websites for free, in return for their sites showing up in search results for relevant queries, driving traffic. If they didn’t want their website to be indexed, they could politely note it in what’s called a “robots.txt” file on their server, and upon seeing it, the bot would leave the site unindexed.

Without intervention from any political or legal bodies, most companies have complied (with some hiccups) with the voluntary standard since its introduction in 1994. That is, until the latest generation of large language models, or LLMs, emerged.

LLMs are data-hungry. A previous version of ChatGPT, OpenAI’s chatbot model, was reportedly trained with data equivalent to 10 trillion words, or enough to fill more than 13 billion Globe and Mail op-eds. It would take a columnist writing daily for 36.5 million years to generate sufficient data to train that model.

Opinion: AI will ruin art, and it will ruin the sparkle in our lives

To satisfy this need, artificial-intelligence companies have to look to the broader internet for high-quality text input. As it turns out, there aren’t nearly enough websites that allow bots to collect data to create better chatbots.

Some AI companies state that they respect robots.txt. In some cases, their public statements are allegedly at odds with their actual practices. But even when these pledges are genuine, companies benefit from another loophole in robots.txt: To block a bot, the website owner must be able to specify the bot’s name. And many AI companies’ bots’ names are only disclosed or discovered after they have crawled freely through the internet.

The impact of these bots is profound. Take Wikipedia, for example: Its content is freely licensed, allowing bots to crawl through its trove of high-quality information. Between January, 2024, and April, 2025, the non-profit found that its multimedia downloads increased by 50 per cent, driven in large part by AI companies’ bots downloading licence-free images to train their products.

Wikipedia says the bot traffic is adding to its operating costs. This ultimately could have been profitable for the site if these chatbots directed users to a page inviting them to contribute to the donation-funded organization.

Instead, researchers have concluded that as AI tools divert traffic and attention away from Wikipedia and other websites, those sites are increasingly starved of the revenue that they need to survive, even as the tools themselves rely on those sites for input. Figures published by Similarweb indicate a 22.7-per-cent drop in Wikipedia’s overall traffic between 2022 and 2025.

Instead of a quid pro quo of letting search engines crawl websites and serving traffic to them through search results, the current arrangement means websites face increased costs while seeing fewer actual visitors, and while also providing the resources to train ever-improving AI chatbots. Without intervention, this threatens to create a vicious cycle where AI products drive websites to shut down, destroying the very data AI needs to improve further.

In response, companies such as Cloudflare, a web-services provider, have started treating AI bots as hackers trying to compromise their customers’ cybersecurity. In August, Cloudflare alleged that Perplexity, a leading AI company, is actively developing new ways to hide its crawling activities to circumvent existing cybersecurity barriers.

A technical showdown against the largest and most well-resourced technology companies in the world isn’t sustainable. The obvious solution would be for these AI companies to honour the trust-based mechanism of robots.txt. However, given competitive pressures, companies aren’t incentivized to do the right thing.

Individual countries are also struggling to find ways to protect individual content creators. For example, Canada’s Online News Act was an attempt to compel social-media companies to compensate news organizations for lost revenue. Instead of achieving that goal, Ottawa learned that companies such as Meta would rather remove Canadians’ access to news on their platforms than compensate publishers.

Our next-best bet is an international agreement, akin to the Montreal Protocol, which bound countries to co-ordinate laws phasing out substances that eroded the Earth’s ozone layer. Absent American leadership, Canada should lead in establishing a similar protocol for AI bots. It could encourage countries to co-ordinate legislative efforts compelling companies to honour robots.txt instructions. If all tech companies around the world had to operate under common rules, it would level the playing field by removing the competitive pressure to race to the bottom.

AI technology can, and should, benefit the world – but it cannot do so by breaking the internet.



Source link

Continue Reading

Trending