Connect with us

AI Research

Smarter nucleic acid design with NucleoBench and AdaBeam

Published

on


We introduced ordered and unordered beam search algorithms, staples from computer science, to test how fixing the order of sequence edits compares to a more flexible, random-order approach. We also created Gradient Evo, a novel hybrid that enhances the directed evolution algorithm by using model gradients to guide its mutations to independently evaluate how important gradients were for edit location selection versus selecting a specific edit.

We also developed AdaBeam, a hybrid adaptive beam search algorithm that combines the most effective elements of unordered beam search with AdaLead, a top-performing, non-gradient design algorithm. Adaptive search algorithms don’t typically explore randomly; instead, their behavior changes as a result of the search to focus their efforts on the most promising areas of the sequence space. AdaBeam’s hybrid approach maintains a “beam”, or a collection of the best candidate sequences found so far, and greedily expands on particularly promising candidates until they’ve been sufficiently explored.

In practice, AdaBeam begins with a population of candidate sequences and their scores. In each round, it first selects a small group of the highest-scoring sequences to act as “parents”. For each parent, AdaBeam generates a new set of “child” sequences by making a random number of random-but-guided mutations. It then follows a short, greedy exploration path, allowing the algorithm to quickly “walk uphill” in the fitness landscape. After sufficient exploration, all the newly generated children are pooled together, and the algorithm selects the absolute best ones to form the starting population for the next round, repeating the cycle. This process of adaptive selection and targeted mutation allows AdaBeam to efficiently focus on high-performing sequences.

Computer-assisted design tasks pose difficult engineering problems, owing to the incredibly large search space. These difficulties become more acute as we attempt to design longer sequences, such as mRNA sequences, and use modern, large neural networks to guide the design. AdaBeam is particularly efficient on long sequences by using fixed-compute probabilistic sampling instead of computations that scale with sequence length. To enable AdaBeam to work with large models, we reduce peak memory consumption during design by introducing a trick we call “gradient concatenation.” However, existing design algorithms that don’t have these features have difficulty scaling to long sequences and large models. Gradient-based algorithms are particularly affected. To facilitate a fair comparison, we limit the length of the designed sequences, even though AdaBeam can scale longer and larger. For example, even though the DNA expression prediction model Enformer runs on ~200K nucleotide sequences, we limit design to just 256 nucleotides.



Source link

Continue Reading
Click to comment

Leave a Reply

Your email address will not be published. Required fields are marked *

AI Research

Measuring Machine Intelligence Using Turing Test 2.0

Published

on

By


In 1950, British mathematician Alan Turing (1912–1954) proposed a simple way to test artificial intelligence. His idea, known as the Turing Test, was to see if a computer could carry on a text-based conversation so well that a human judge could not reliably tell it apart from another human. If the computer could “fool” the judge, Turing argued, it should be considered intelligent.

For decades, Turing’s test shaped public understanding of AI. Yet as technology has advanced, many researchers have asked whether imitating human conversation really proves intelligence — or whether it only shows that machines can mimic certain human behaviors. Large language models like ChatGPT can already hold convincing conversations. But does that mean they understand what they are saying?

In a Mind Matters podcast interview, Dr. Georgios Mappouras tells host Robert J. Marks that the answer is no. In a recent paper, The General Intelligence Threshold, Mappouras introduces what he calls Turing Test 2.0. This updated approach sets a higher bar for intelligence than simply chatting like a human. It asks whether machines can go beyond imitation to produce new knowledge.

From information to knowledge

At the heart of Mappouras’s proposal is a distinction between two kinds of information, non-functional vs. functional:

  • Non-functional information is raw data or observations that don’t lead to new insights by themselves. One example would be noticing that an apple falls from a tree.
  • Functional information is knowledge that can be applied to achieve something new. When Isaac Newton connected the falling apple to the force of gravity, he transformed ordinary observation into scientific law.

True intelligence, Mappouras argues, is the ability to transform non-functional information into functional knowledge. This creative leap is what allows humans to build skyscrapers, develop medicine, and travel to the moon. A machine that merely rearranges words or retrieves facts cannot be said to have reached the same level.

The General Intelligence Threshold

Mappouras calls this standard the General Intelligence Threshold. His threshold sets a simple challenge: given existing knowledge and raw information, can the system generate new insights that were not directly programmed into it?

This threshold does not require constant displays of brilliance. Even one undeniable breakthrough — a “flash of genius” — would be enough to demonstrate that a machine possesses general intelligence. Just as a person may excel in math but not physics, a machine would only need to show creativity once to prove its potential.

Creativity and open problems

One way to apply the new test is through unsolved problems in mathematics. Throughout history, breakthroughs such as Andrew Wiles’s proof of Fermat’s Last Theorem or Grigori Perelman’s solution to the Poincaré Conjecture marked milestones of human creativity. If AI could solve open problems like the Riemann Hypothesis or the Collatz Conjecture — problems that no one has ever solved before — it would be strong evidence that the system had crossed the threshold into true intelligence.

Large language models already solve equations and perform advanced calculations, but solving a centuries-old unsolved problem would show something far deeper: the ability to create knowledge that has never existed before.

Beyond symbol manipulation

Mappouras also draws on philosopher John Searle’s famous “Chinese Room” thought experiment. In the scenario, a person who does not understand Chinese sits in a room with a rulebook for manipulating Chinese characters. By following instructions, the person produces outputs that convince outsiders he understands the language, even though he does not.

This scenario, Searle argued, shows that a computer might appear intelligent without real understanding. Mappouras agrees but goes further. For him, real intelligence is proven not just by producing outputs, but by acting on new knowledge. If the instructions in the Chinese Room included a way to escape, the person could only succeed if he truly understood what the words meant. In the same way, AI must demonstrate it can act meaningfully on information, not just shuffle symbols.

Image Credit: top images – Adobe Stock

Can AI pass the new test?

So far, Mappouras does not think modern AI has passed the General Intelligence Threshold. Systems like ChatGPT may look impressive, but their apparent creativity usually comes from patterns in the massive data sets on which they were trained. They have not shown the ability to produce new, independent knowledge disconnected from prior inputs.

That said, Mappouras emphasizes that success would not require constant novelty. One true act of creativity — an undeniable demonstration of new knowledge — would be enough. Until that happens, he remains cautious about claims that today’s AI is truly intelligent.

A shift in the debate

The debate over artificial intelligence is shifting. The original Turing Test asked whether machines could fool us into thinking they were human. Turing Test 2.0 asks a harder question: can they discover something new?

Mappouras believes this is the real measure of intelligence. Intelligence is not imitation — it is innovation. Whether machines will ever cross that line remains uncertain. But if they do, the world will not just be talking with computers. We will be learning from them.

Final thoughts: Today’s systems, tomorrow’s threshold

Models like ChatGPT and Grok are remarkable at conversation, summarization, and problem-solving within known domains, but their strengths still reflect pattern learning from vast training data. By Mappouras’s standard, they will cross the General Intelligence Threshold only when they produce a verifiable breakthrough — an insight not traceable to prior text or human scaffolding, such as an original solution to a major open problem. Until then, they remain powerful imitators and accelerators of human work — impressive, useful, and transformative, but not yet creators of genuinely new knowledge.

Additional Resources

Podcast Transcript Download



Source link

Continue Reading

AI Research

UTM Celebrates Malaysia’s Youngest AI Researcher Recognised at IEEE AI-SI 2025 – UTM NewsHub

Published

on


KUALA LUMPUR, 28 August 2025 – Universiti Teknologi Malaysia (UTM) proudly hosted the Institute of Electrical and Electronics Engineers (IEEE) International Conference on Artificial Intelligence for Sustainable Innovation (AI-SI) 2025, themed “Empowering Innovation for a Sustainable Future.” The conference gathered global experts, academics, and industry leaders to explore how Artificial Intelligence (AI) can address sustainability challenges. Among its highlights was the remarkable achievement of 17-year-old Malaysian researcher, Charanarravindaa Suriess, who was celebrated as the youngest presenter and awarded Best Presenter for his groundbreaking paper on adversarial robustness in neural networks. His recognition reflected not only individual brilliance but also Malaysia’s growing strength in the global AI research landscape.

Charanarravindaa’s presentation, titled “Two-Phase Evolutionary Framework for Adversarial Robustness in Neural Networks,” introduced an innovative framework designed to improve AI systems’ ability to defend against adversarial attacks. His contribution addressed one of the most pressing challenges in AI, ensuring resilience and trustworthiness of machine learning models in real-world applications. Born in Johor Bahru, his journey into science and computing began early; by primary school, he was already troubleshooting computers and experimenting with small websites. At just 15 years old, he graduated early, motivated by a passion for deeper challenges. Participation in international hackathons, including DeepLearning Week at Nanyang Technological University (NTU) Singapore, strengthened his resolve and provided the encouragement that led to his first academic paper, now internationally recognised at IEEE AI-SI 2025.

Charanarravindaa Suriess, 17, youngest and Best Presenter at IEEE AI-SI 2025

Beyond academia, Charanarravindaa has also demonstrated entrepreneurial spirit by founding Cortexa, a startup dedicated to advancing AI robustness, architectures, and applied AI for scientific discovery. His long-term vision is to integrate artificial intelligence with quantum computing and theoretical physics to expand the boundaries of knowledge. This ambition is a testament to the potential of Malaysia’s youth in contributing to frontier technologies. His recognition at IEEE AI-SI 2025 reflects IEEE’s mission of advancing technology for humanity, where innovation is seen as a universal endeavour not limited by age. By honouring a young researcher, IEEE underscored its commitment to empowering future generations of scientists and innovators to shape technology for global good.

Charanarravindaa Suriess, 17, recognised as the youngest participant and Best Presenter at IEEE AI-SI 2025
Charanarravindaa Suriess, 17, recognised as the youngest participant and Best Presenter at IEEE AI-SI 2025

During the conference, the Faculty of Artificial Intelligence (FAI), UTM, represented by Associate Professor Dr. Noor Azurati Ahmad, extended an invitation to Charanarravindaa to explore possible research collaborations. This initiative aligns with FAI’s vision to be a leader in AI education, research, and innovation, with a particular focus on trustworthy, robust, and sustainable AI. Early discussions centred on aligning his research interests with UTM’s expertise in advanced architectures and digital sustainability. Such collaboration exemplifies how institutions and young talent can come together to accelerate innovation, while also strengthening Malaysia’s position as an emerging hub for AI research and talent cultivation.

At the national level, this achievement resonates strongly with the Malaysia National Artificial Intelligence Roadmap (2021–2025), which identifies talent development as a central pillar in building an AI-ready nation. Prime Minister Datuk Seri Anwar Ibrahim has repeatedly highlighted the urgency of nurturing local talent to enhance competitiveness and leadership in the global digital economy. Charanarravindaa’s success demonstrates tangible progress in this direction, showcasing how Malaysia can produce young innovators capable of contributing to both national aspirations and international scientific advancement. Through platforms such as IEEE AI-SI 2025, UTM reaffirms its role as a catalyst for excellence in AI research and talent development, embodying its mission to prepare the next generation of scholars and innovators who will drive sustainable futures.



Source link

Continue Reading

AI Research

Databricks at a crossroads: Can its AI strategy prevail without Naveen Rao?

Published

on


“Databricks is in a tricky spot with Naveen Rao stepping back. He was not just a figurehead, but deeply involved in shaping their AI vision, particularly after MosaicML,” said Robert Kramer, principal analyst at Moor Insights & Strategy.

“Rao’s absence may slow the pace of new innovation slightly, at least until leadership stabilizes. Internal teams can keep projects on track, but vision-driven leaps, like identifying the ‘next MosaicML’, may be harder without someone like Rao at the helm,” Kramer added.

Rao became a part of Databricks in 2023 after the data lakehouse provider acquired MosaicML, a company Rao co-founded, for $1.3 billion. During his tenure, Rao was instrumental in leading research for many Databricks products, including Dolly, DBRX, and Agent Bricks.



Source link

Continue Reading

Trending