Connect with us

AI Research

A language model built for the public good

Published

on


Earlier this week in Geneva, around 50 leading global initiatives and organisations dedicated to open-source LLMs and trustworthy AI convened at the International Open-Source LLM Builders Summit. Hosted by the AI centres of EPFL and ETH Zurich, the event marked a significant step in building a vibrant and collaborative international ecosystem for open foundation models. Open LLMs are increasingly viewed as credible alternatives to commercial systems, most of which are developed behind closed doors in the United States or China.

Participants of the summit previewed the forthcoming release of a fully open, publicly developed LLM — co-created by researchers at EPFL, ETH Zurich and other Swiss universities in close collaboration with engineers at CSCS. Currently in final testing, the model will be downloadable under an open license. The model focuses on transparency, multilingual performance, and broad accessibility.

The model will be fully open: source code and weights will be publicly available, and the training data will be transparent and reproducible, supporting adoption across science, government, education, and the private sector. This approach is designed to foster both innovation and accountability.

“Fully open models enable high-trust applications and are necessary for advancing research about the risks and opportunities of AI. Transparent processes also enable regulatory compliance,” says Imanol Schlag, research scientist at the ETH AI Center, who is leading the effort alongside EPFL AI Center faculty members and professors Antoine Bosselut and Martin Jaggi.

Multilingual by design

A defining characteristic of the LLM is its fluency in over 1000 languages. “We have emphasised making the models massively multilingual from the start,” says Antoine Bosselut.

Training of the base model was done on a large text dataset in over 1500 languages — approximately 60% English and 40% non-English languages — as well as code and mathematics data. Given the representation of content from all languages and cultures, the resulting model maintains the highest global applicability.

Designed for scale and inclusion

The model will be released in two sizes — 8 billion and 70 billion parameters, meeting a broad range of users’ needs. The 70B version will rank among the most powerful fully open models worldwide. The number of parameters reflects a model’s capacity to learn and generate complex responses.

High reliability is achieved through training on over 15 trillion high-quality training tokens (units representing a word or part of the word), enabling robust language understanding and versatile use cases.

Responsible data practices

The LLM is being developed with due consideration to Swiss data protection laws, Swiss copyright laws, and the transparency obligations under the EU AI Act. In a external page recent study, the project leaders demonstrated that for most everyday tasks and general knowledge acquisition, respecting web crawling opt-outs during data acquisition produces virtually no performance degradation.

Supercomputer as an enabler of sovereign AI

The model is trained on the “Alps” supercomputer at CSCS in Lugano, one of the world’s most advanced AI platforms, equipped with over 10,000 NVIDIA Grace Hopper Superchips. The system’s scale and architecture made it possible to train the model efficiently using 100% carbon-neutral electricity.

The successful realisation of “Alps” was significantly facilitated by a long-standing collaboration spanning over 15 years with NVDIA and HPE/Cray. This partnership has been pivotal in shaping the capabilities of “Alps”, ensuring it meets the demanding requirements of large-scale AI workloads, including the pre-training of complex LLMs.

“Training this model is only possible because of our strategic investment in ‘Alps’, a supercomputer purpose-built for AI,” says Thomas Schulthess, Director of CSCS and professor at ETH Zurich. “Our enduring collaboration with NVIDIA and HPE exemplifies how joint efforts between public research institutions and industry leaders can drive sovereign infrastructure, fostering open innovation — not just for Switzerland, but for science and society worldwide.”

Public access and global reuse

In late summer, the LLM will be released under the Apache 2.0 License. Accompanying documentation will detail the model architecture, training methods, and usage guidelines to enable transparent reuse and further development.

“As scientists from public institutions, we aim to advance open models and enable organiations to build on them for their own applications”, says Antoine Bosselut.

“By embracing full openness — unlike commercial models that are developed behind closed doors — we hope that our approach will drive innovation in Switzerland, across Europe, and through multinational collaborations. Furthermore, it is a key factor in attracting and nurturing top talent,” says EPFL professor Martin Jaggi.



Source link

AI Research

Nuclear energy plan unveiled by UK and US, promising thousands of jobs

Published

on


Charlotte EdwardsBusiness reporter, BBC News

Getty Images Aerial landscape view of Drax power station with thick white steam rising from cooling towers on a sunny dayGetty Images

The UK and US are set to sign a landmark agreement aimed at accelerating the development of nuclear power.

The move is expected to generate thousands of jobs and strengthen Britain’s energy security.

It is expected to be signed off during US President Donald Trump’s state visit this week, with both sides hoping it will unlock billions in private investment.

Prime Minister Sir Keir Starmer said the two nations were “building a golden age of nuclear” that would put them at the “forefront of global innovation”.

The government has said that generating more power from nuclear can cut household energy bills, create jobs, boost energy security, and tackle climate change.

The new agreement, known as the Atlantic Partnership for Advanced Nuclear Energy, aims to make it quicker for companies to build new nuclear power stations in both the UK and the US.

It will streamline regulatory approvals, cutting the average licensing period for nuclear projects from up to four years to just two.

‘Nuclear renaissance’

The deal is also aimed at increasing commercial partnerships between British and American companies, with a number of deals set to be announced.

Key among the plans is a proposal from US nuclear group X-Energy and UK energy company Centrica to build up to 12 advanced modular nuclear reactors in Hartlepool, with the potential to power 1.5 million homes and create up to 2,500 jobs.

The broader programme could be worth up to £40bn, with £12bn focused in the north east of England.

Other plans include multinational firms such as Last Energy and DP World working together on a micro modular reactor at London Gateway port. This is backed by £80m in private investment.

Elsewhere, Holtec, EDF and Tritax are also planning to repurpose the former Cottam coal-fired plant in Nottinghamshire into a nuclear-powered data centre hub.

This project is estimated to be worth £11bn and could create thousands of high-skilled construction jobs, as well as permanent jobs in long-term operations.

Beyond power generation, the new partnership includes collaboration on fusion energy research, and an end to UK and US reliance on Russian nuclear material by 2028.

Commenting on the agreement, Energy Secretary Ed Miliband said: “Nuclear will power our homes with clean, homegrown energy and the private sector is building it in Britain, delivering growth and well-paid, skilled jobs for working people.”

And US Energy Secretary Chris Wright described the move as a “nuclear renaissance”, saying it would enhance energy security and meet growing global power demands, particularly from AI and data infrastructure.

Sir Keir has previously said he wants the UK to return to being “one of the world leaders on nuclear”.

In the 1990s, nuclear power generated about 25% of the UK’s electricity but that figure has fallen to around 15%, with no new power stations built since then and many of the country’s ageing reactors due to be decommissioned over the next decade.

In November 2024, the UK and 30 other countries signed a global pledge to triple their nuclear capacity by 2050.

And earlier this year, the government announced a deal with private investors to build the Sizewell C nuclear power station in Suffolk.

Its nuclear programme also includes the UK’s first small modular reactors (SMRs), which will be built by UK firm Rolls Royce.



Source link

Continue Reading

AI Research

Researchers ‘polarised’ over use of AI in peer review

Published

on


Researchers appear to be becoming more divided over whether generative artificial intelligence should be used in peer review, with a survey showing entrenched views on either side.

A poll by IOP Publishing found that there has been a big increase in the number of scholars who are positive about the potential impact of new technologies on the process, which is often criticised for being slow and overly burdensome for those involved.

A total of 41 per cent of respondents now see the benefits of AI, up from 12 per cent from a similar survey carried out last year. But this is almost equal to the proportion with negative opinions which stands at 37 per cent after a 2 per cent year-on-year increase.

This leaves only 22 per cent of researchers neutral or unsure about the issue, down from 36 per cent, which IOP said indicates a “growing polarisation in views” as AI use becomes more commonplace.

Women tended to have more negative views about the impact of AI compared with men while junior researchers tended to have a more positive view than their more senior colleagues.

Nearly a third (32 per cent) of those surveyed say they already used AI tools to support them with peer reviews in some form.

Half of these say they apply it in more than one way with the most common use being to assist with editing grammar and improving the flow of text.

A minority used it in more questionable ways such as the 13 per cent who asked the AI to summarise an article they were reviewing – despite confidentiality and data privacy concerns – and the 2 per cent who admitted to uploading an entire manuscript into a chatbot so it could generate a review on their behalf.

IOP – which currently does not allow AI use in peer reviews – said the survey showed a growing recognition that the technology has the potential to “support, rather than replace, the peer review process”.

But publishers must fund ways to “reconcile” the two opposing viewpoints, the publisher added.

A solution could be developing tools that can operate within peer review software, it said, which could support reviewers without positing security or integrity risks.

Publishers should also be more explicit and transparent about why chatbots “are not suitable tools for fully authoring peer review reports”, IOP said.

“These findings highlight the need for clearer community standards and transparency around the use of generative AI in scholarly publishing. As the technology continues to evolve, so too must the frameworks that support ethical and trustworthy peer review,” Laura Feetham-Walker, reviewer engagement manager at IOP and lead author of the study, said.

tom.williams@timeshighereducation.com



Source link

Continue Reading

AI Research

Amazon Employing AI to Help Shoppers Comb Reviews

Published

on

By


Amazon earlier this year began rolling out artificial intelligence-voiced product descriptions for select customers and products.

Now, the company’s “Hear the Highlights” feature has extended to all users, CNBC reported Sunday (Sept. 14), arguing this could replace user-created reviews as the main source of product information.

Among the advantages here, the report added, is that artificial intelligence (AI) won’t suffer from cognitive overload from combing through thousands of reviews. 

“It’s important to recognize where AI is currently strong, such as in automation and pattern recognition, and where it still falls short, like in judgment-heavy tasks,” said Ankur Edkie, co-founder and CEO of Murf AI, which develops AI voiceovers. “A key question is whether there’s a way to factor in customer context as an input while generating these summaries.”

The value of AI, according to Edkie, is determining the right “problem-capability fit.” Without that, he added, a sense of “gimmickry” is likely to filter through thanks to AI fatigue, which he says consumers are likely feeling by now.

PYMNTS has contacted Amazon for comment but has not yet gotten a reply.

The CNBC report also noted that the tendency of AI to focus on common themes can water down responses even as it summarizes them, taking out the detailed personal experiences found in human reviews.

“AI might overlook unique insights or niche needs that don’t align with the majority of responses,” said Brian Numainville, principal at consumer research firm Feedback Group. “Additionally, the ability to critically interpret reviews — like spotting biases or trusting certain reviewers — is diminished with AI summaries.”

Nauman Dawalatabad, a research scientist at Zoom Communications, offered his opinion that the technology is on its way to improving customer experience.

“I take it as technology helping us to make informed decisions,” he said, pointing to the mental fatigue and wasted time that can result from working through customer reviews.

Meanwhile, recent research by PYMNTS Intelligence shows that AI shopping adoption has begun to gain traction with younger and middle-aged consumers. The research found that 32% of all consumers said they have used or would use generative AI for shopping.

“Bridge millennials — older millennials straddling Gen X — lead the way, with 38% reporting AI use for shopping,” PYMNTS wrote last month. “Zillennials are close behind at 36%, followed by millennials at 35%. Gen X is next, at 33%, while Gen Z comes in at 31%. Baby boomers show some traction as well, with 28% using gen AI for shopping.”



Source link

Continue Reading

Trending