[To share your insights with us as part of editorial or sponsored content, please write to psen@itechseries.com]
AI Research
Mithril arrives, expanding omnicloud AI compute platform trusted by Cursor, LG AI Research, Arc Institute and others

Foundry rebrands as Mithril, announcing 550% platform growth and new inference service that cuts large-scale AI data processing costs by more than 50%
Mithril announced its rebrand from Foundry, reflecting its mission to simplify access to distributed AI compute. The Mithril brand embodies the platform’s mission: transforming powerful but traditionally rigid AI infrastructure into a fluid, universally accessible resource for AI practitioners.
Since launching in limited preview in August 2024, Mithril’s platform usage has grown by over 550%. Today, the platform provides critical AI compute to customers including Cursor, LG AI Research, and Arc Institute. In parallel to the rebrand, Mithril is introducing its highly demanded batch inference service, which improves price-performance by 2–5x for large-scale AI workloads such as document analysis, content generation, and multimodal annotations.
“Traditional AI cloud procurement models mean that AI-native companies need to constantly shop around for the best compute prices and often need to buy more than what they actually need,” said Jared Quincy Davis, founder and CEO of Mithril. “By building on our omnicloud platform, Mithril customers have been able to massively simplify getting access to the compute they need. Rather than spending effort negotiating with every cloud, they can provision compute for a few hours or multiple months through a single platform with transparent, market-based pricing.”
Also Read: AiThority Interview with Suzanne Livingston, Vice President, IBM Watsonx Orchestrate Agent Domains
While the GPU market, driven by the rapid adoption of AI, is set to reach nearly $230 billion by 2030, the cost per unit of performance for GPU compute is declining rapidly. Mithril’s omnicloud platform is designed to harness this new paradigm, virtualizing capacity across multiple providers and distributing it through a fluid market with transparent pricing. By aggregating capacity in this way, Mithril is able to provide customers with the best economics for AI compute without bespoke negotiations or long-term contracts.
Building on this foundation, Mithril today unveiled its new batch inference API, purpose-built to orchestrate large-scale inference workloads across its omnicloud platform. The service enables customers to run compute-intensive tasks—such as document summarization, video transcription, multimodal annotation, and synthetic data generation—at a fraction of the traditional cost. By leveraging Mithril’s distributed compute marketplace, the batch inference API delivers 2–5x better price-performance than conventional inference services, while eliminating the need for customers to manage complex infrastructure or negotiate access to scarce GPU capacity. With simple, token-based pricing and support for both off-the-shelf and custom models, teams can submit jobs of virtually any scale and rely on Mithril to handle orchestration, scaling, and infrastructure reliability.
Also Read: C-Gen.AI Emerges from Stealth to End Infrastructure Limitations Affecting AI Workloads
With the addition of batch inference, Mithril extends its core mission: to make compute as universally available and accessible as electricity. Whether provisioning GPUs for foundational research or scaling inference workloads across enterprise datasets, Mithril’s omnicloud platform provides AI teams with the flexibility, efficiency, and simplicity they need to push the boundaries of what’s possible.
“With Mithril, we trained the first textless conversational audio base model for under $200,000, instead of the $10 million or more we would have needed to raise to reserve multi-year capacity,” says Devansh Pandey, co-founder of Standard Intelligence, a startup building foundational models towards AGI. “Burst compute, which Mithril excels at, gives even small startups a leg up against incumbents.”
“The tremendous demand for AI compute capacity and the large fraction of idle time makes sharing a perfect solution,” said Paul Milgrom, a Stanford economics professor, Nobel laureate, and Mithril advisor who also has helped Google with its market-leading ad platform. “Mithril’s innovative market is the right approach to making that happen.”
AI Research
Canada invests $28.7M to train clean energy workers and expand AI research

The federal government is investing $28.7 million to equip Canadian workers with skills for a rapidly evolving clean energy sector and to expand artificial intelligence (AI) research capacity.
The funding, announced Sept. 9, includes more than $9 million over three years for the AI Pathways: Energizing Canada’s Low-Carbon Workforce project. Led by the Alberta Machine Intelligence Institute (Amii), the initiative will train nearly 5,000 energy sector workers in AI and machine learning skills for careers in wind, solar, geothermal and hydrogen energy. Training will be offered both online and in-person to accommodate mid-career workers, industry associations, and unions across Canada.
In addition, the government is providing $19.7 million to Amii through the Canadian Sovereign AI Compute Strategy, expanding access to advanced computing resources for AI research and development. The funding will support researchers and businesses in training and deploying AI models, fostering innovation, and helping Canadian companies bring AI-enabled products to market.
“Canada’s future depends on skilled workers. Investing and upskilling Canadian workers ensures they can adapt and succeed in an energy sector that’s changing faster than ever,” said Patty Hajdu, Minister of Jobs and Families and Minister responsible for the Federal Economic Development Agency for Northern Ontario.
Evan Solomon, Minister of Artificial Intelligence and Digital Innovation, added that the investment “builds an AI-literate workforce that will drive innovation, create sustainable jobs, and strengthen our economy.”
Amii CEO Cam Linke said the funding empowers Canada to become “the world’s most AI-literate workforce” while providing researchers and businesses with a competitive edge.
The AI Pathways initiative is one of eight projects funded under the Sustainable Jobs Training Fund, which supports more than 10,000 Canadian workers in emerging sectors such as electric vehicle maintenance, green building retrofits, low-carbon energy, and carbon management.
The announcement comes as Canada faces workforce shifts, with an estimated 1.2 million workers retiring across all sectors over the next three years and the net-zero transition projected to create up to 400,000 new jobs by 2030.
The federal investments aim to prepare Canadians for the jobs of the future while advancing research, innovation, and commercialization in AI and clean energy.
AI Research
OpenAI and NVIDIA will join President Trump’s UK state visit

U.S. President Donald Trump is about to do something none of his predecessors have — make a second full state visit to the UK. Ordinarily, a President in a second term of office visits, meets with the monarch, but doesn’t get a second full state visit.
On this one it seems he’ll be accompanied by two of the biggest faces in the ever-growing AI race; OpenAI CEO, Sam Altman, and NVIDIA CEO, Jensen Huang.
This is according to a report by the Financial Times, which claims that the two are accompanying President Trump to announce a “large artificial intelligence infrastructure deal.”
The deal is said to support a number of data center projects in the UK, another deal towards developing “sovereign” AI for another of the United States’ allies.
The report claims that the two CEOs will announce the deal during the Trump state visit, and will see OpenAI supply the technology, and NVIDIA the hardware. The UK will supply all the energy required, which is handy for the two companies involved.
UK energy is some of the most expensive in the world (one reason I’m trying to use my gaming PC with an RTX 5090 a lot less!)
The exact makeup of the deal is still unknown, and, naturally, neither the U.S. nor UK governments have said anything at this point.
The UK government, like many others, has openly announced its plans to invest in AI. As the next frontier for tech, you either get on board or you get left behind. And President Trump has made no secret of his desires to ensure the U.S. is a world leader.
OpenAI isn’t the only company that could provide the software side, but it is the most established. While Microsoft may be looking towards a future where it is less reliant on the tech behind ChatGPT for its own AI ambitions, it makes total sense that organizations around the world would be looking to OpenAI.
NVIDIA, meanwhile, continues to be the runaway leader on the hardware front. We’ve seen recently that AMD is planning to keep pushing forward, and a recent Chinese model has reportedly been built to run specifically without NVIDIA GPUs.
But for now, everything runs best on NVIDIA, and as long as it can keep churning out enough GPUs to fill these data centers, it will continue to print money.
The state visit is scheduled to begin on Wednesday, September 17, so I’ll be keeping a close eye out for when this AI deal gets announced.
AI Research
100x Faster Brain-Inspired AI Model

In the rapidly evolving field of artificial intelligence, a new contender has emerged from China’s research labs, promising to reshape how we think about energy-efficient computing. The SpikingBrain-7B model, developed by the Brain-Inspired Computing Lab (BICLab) at the Chinese Academy of Sciences, represents a bold departure from traditional large language models. Drawing inspiration from the human brain’s neural firing patterns, this system employs spiking neural networks to achieve remarkable efficiency gains. Unlike conventional transformers that guzzle power, SpikingBrain-7B mimics biological neurons, firing only when necessary, which slashes energy consumption dramatically.
At its core, the model integrates hybrid-linear attention mechanisms and conversion-based training techniques, allowing it to run on domestic MetaX chips without relying on NVIDIA hardware. This innovation addresses a critical bottleneck in AI deployment: the high energy demands of training and inference. According to a technical report published on arXiv, the SpikingBrain series, including the 7B and 76B variants, demonstrates over 100 times faster first-token generation at long sequence lengths, making it ideal for edge devices in industrial control and mobile applications.
Breaking Away from Transformer Dominance
The genesis of SpikingBrain-7B can be traced to BICLab’s GitHub repository, where the open-source code reveals a sophisticated architecture blending spiking neurons with large-scale model training. Researchers at the lab, led by figures like Guoqi Li and Bo Xu, have optimized for non-NVIDIA clusters, overcoming challenges in parallel training and communication overhead. This approach not only enhances stability but also paves the way for neuromorphic hardware that prioritizes energy optimization over raw compute power.
Recent coverage in Xinhua News highlights how SpikingBrain-1.0, the foundational system, breaks from mainstream models like ChatGPT by using spiking networks instead of dense computations. This brain-inspired paradigm allows the model to train on just a fraction of the data typically required—reports suggest as little as 2%—while matching or exceeding transformer performance in benchmarks.
Efficiency Gains and Real-World Applications
Delving deeper, the model’s spiking mechanism enables asynchronous processing, akin to how the brain handles information dynamically. This is detailed in the arXiv report, which outlines a roadmap for next-generation hardware that could integrate seamlessly into sectors like healthcare and transportation. For instance, in robotics, SpikingBrain’s low-power profile supports real-time decision-making without the need for massive data centers.
Posts on X (formerly Twitter) from AI enthusiasts, such as those praising its 100x speedups, reflect growing excitement. Users have noted how the model’s hierarchical processing mirrors neuroscience findings, with emergent brain-like patterns in its structure. This sentiment aligns with broader neuromorphic computing trends, as seen in a Nature Communications Engineering article on advances in robotic vision, where spiking networks enable efficient AI in constrained environments.
Challenges and Future Prospects
Despite its promise, deploying SpikingBrain-7B isn’t without hurdles. The arXiv paper candidly discusses adaptations needed for CUDA and Triton operators in hybrid attention setups, underscoring the technical feats involved. Moreover, training on MetaX clusters required custom optimizations to handle long-sequence topologies, a feat that positions China at the forefront of independent AI innovation amid global chip restrictions.
In industry circles, this development is seen as a catalyst for shifting AI paradigms. A NotebookCheck report emphasizes its potential for up to 100x performance boosts over conventional systems, fueling discussions on sustainable AI. As neuromorphic computing gains traction, SpikingBrain-7B could inspire a wave of brain-mimicking models, reducing the environmental footprint of AI while expanding its reach to everyday devices.
Implications for Global AI Research
Beyond technical specs, the open-sourcing of SpikingBrain-7B via GitHub invites global collaboration, with the repository already garnering attention for its spike-driven transformer implementations. This mirrors earlier BICLab projects like Spike-Driven-Transformer-V2, building a continuum of research toward energy-efficient intelligence.
Looking ahead, experts anticipate integrations with emerging hardware, as outlined in PMC’s coverage of spike-based dynamic computing. With SpikingBrain’s bilingual capabilities and industry validations, it stands as a testament to how bio-inspired designs can democratize AI, challenging Western dominance and fostering a more inclusive technological future.
-
Business2 weeks ago
The Guardian view on Trump and the Fed: independence is no substitute for accountability | Editorial
-
Tools & Platforms1 month ago
Building Trust in Military AI Starts with Opening the Black Box – War on the Rocks
-
Ethics & Policy2 months ago
SDAIA Supports Saudi Arabia’s Leadership in Shaping Global AI Ethics, Policy, and Research – وكالة الأنباء السعودية
-
Events & Conferences4 months ago
Journey to 1000 models: Scaling Instagram’s recommendation system
-
Jobs & Careers3 months ago
Mumbai-based Perplexity Alternative Has 60k+ Users Without Funding
-
Podcasts & Talks2 months ago
Happy 4th of July! 🎆 Made with Veo 3 in Gemini
-
Education2 months ago
VEX Robotics launches AI-powered classroom robotics system
-
Education2 months ago
Macron says UK and France have duty to tackle illegal migration ‘with humanity, solidarity and firmness’ – UK politics live | Politics
-
Podcasts & Talks2 months ago
OpenAI 🤝 @teamganassi
-
Funding & Business3 months ago
Kayak and Expedia race to build AI travel agents that turn social posts into itineraries