AI Research
Larry Ellison Institute gives Oxford £118 million for AI vaccine research

The Ellison Institute of Technology (EIT) is funding an Oxford vaccine research project that will tackle pathogenic diseases using AI.
Ellison, who recently overtook Elon Musk as the world’s richest man, is giving Oxford £118 million for the programme, which will be led by the Oxford Vaccine Group.
Professor Sir Andrew Pollard, director of the group that led COVID-19 trials, described the programme as a “new frontier in vaccine science”. Scientists will use “human challenge models”, where volunteers are safely exposed to bacteria under controlled conditions and AI tools to identify immune responses that predict protection.
Oxford’s Vice-Chancellor, Irene Tracey, described the project as a “major step forward” in the strategic alliance with the Ellison Institute. She explained that her vision is to draw “more talent and capacity to the Oxford ecosystem to turn scientific challenges into real solutions for the world”.
EIT is designed to host 7,000 scientists, including an oncology clinic, auditorium, laboratories, library, classrooms, and park space. Oxford University, by comparison, has 5,000 research staff.
The Institute has already faced leadership turbulence, with the President, John Bell, resigning days before the vaccine project was announced. Bell was pictured signing the contracts with Irene Tracey when the “strategic alliance” was first announced in December 2024. Bell publicly endorsed Lord Hague in the Chancellor election last year.
The Wall Street Journal reported that Bell clashed with Ellison over operations and staffing, and that tensions flared over the mix of people being brought into the Institute, as well as Ellison’s decisions to fire senior staff without involving him.
Bell, who was Regius Professor of Medicine at Oxford until March last year, also serves as chair of Our Future Health, a government-funded project to genetically test millions of patients. He holds over £700,000 of shares in Roche, a pharmaceutical company where he sat on the board for 20 years, which has drawn criticism from genomics-monitoring groups for the “conflict of interest”.
Despite these controversies surrounding Bell’s various roles, a University spokesperson told Cherwell: “We recognise his pivotal contribution in helping to establish the Institute and in attracting outstanding researchers to its mission.”
Bell belonged to the Institute’s Faculty of Fellows alongside former Prime Minister, Tony Blair. Tony Blair’s own Institute for Global Change (TBI) is bankrolled by Ellison. As well as sharing their source of funding, the Ellison-funded institutes work in collaboration on an “AI for Governments” project.
Larry Ellison amassed his billions as boss of tech-giant Oracle, where he has made headlines for suggesting that Oracle would pioneer “AI mass surveillance”, as well as for his friendship with Israeli Prime Minister Benjamin Netanyahu, whom he offered a job at Oracle. Ellison donated to the Israeli military through Friends of the Israel Defense Forces, giving the organisation $16.6 million in 2017.
Ellison also reportedly has a close relationship with Trump, attending meetings in the Oval Office. Trump has questioned the effectiveness of COVID-19 vaccines, which were pioneered by the same Oxford Vaccine Group that are partnering with Ellison’s institute on this project.
Responding to Ellison’s ties to vaccine-sceptic politicians, as well as questions over the ownership of intellectual property (IP) stemming from the strategic alliance, the University spokesperson told Cherwell that it ensures any external partnerships “align with the University’s public mission, including by realising impact from our academic research”.
Further details on the ownership or management of intellectual property arising from the programme have not been made public.
AI Research
AI-powered CRISPR could lead to faster gene therapies, Stanford Medicine study finds

Yilong Zhou, a visiting undergraduate student from Tsinghua University, used CRISPR-GPT to successfully active genes in A375 melanoma cancer cells as part of his research into better understanding why cancer immunotherapy sometimes fails.
Zhou typed his question into CRISPR-GPT’s text box: “I plan to do a CRISPR activate in a culture of human lung cells, what method should I use?”
CRISPR-GPT responded like an experienced lab mate advising a new researcher. It drafted an experimental design and, at each step, explained its “thought” process, describing why the various steps were important.
“I could simply ask questions when I didn’t understand something, and it would explain or adjust the design to help me understand,” Zhou said. “Using CRISPR-GPT felt less like a tool and more like an ever-available lab partner.”
As an early-career scientist, Zhou had designed only a handful of CRISPR experiments prior to using CRISPR-GPT. In this experiment, it took him one attempt to get it right — a rarity for most scientists.
In the past, Zhou was constantly worrying about making mistakes and double-checking his designs.
Reducing error and increasing accessibility
CRISPR-GPT can toggle between three modes: beginner, expert and Q&A. The beginner mode functions as a tool and a teacher, providing an answer and explanation for each recommendation. Expert mode is more of an equal partner, working with advanced scientists to tackle complex problems without providing additional context. Any researcher can use the Q&A function to directly address specific questions.
It’s also useful for sharing knowledge and collaborating with other labs, Cong said. CRISPR-GPT provides a more detailed and holistic response than what’s generally gleaned from a scientific manuscript and responds to repetitive inquires in a snap.
CRISPR-GPT can also check researchers’ work and apply experimental frameworks to new diseases the researchers may not be thinking about.
“People in my lab have been finding this tool very helpful,” Cong said. “The decisions are ultimately made by human scientists, but it just makes that whole process — from experiment design to execution — super simple.”
Editing responsibly and future expansion
While the technology is promising for accelerating therapeutic research, there are still some safety concerns to address before pushing CRISPR-GPT more broadly.
Cong and his team have already incorporated safeguards to protect the AI tool from irresponsible uses. For instance, if the AI receives a request to assist with an unethical activity, such as editing a virus or human embryo, CRISPR-GPT will issue a warning to the user and respond with an error message, effectively halting the interaction. Cong also plans to bring the technology to government agencies, such as the National Institute of Standards and Technology, to ensure ethical use and sound biosecurity.
In the future, the tool may serve as a blueprint for training AI to execute specific biological tasks outside of gene editing. From developing new lines of stem cells as experimental models, to deciphering molecular pathways involved in heart diseases, Cong hopes to expand the technology to other disciplines building a range of AI agents to aid in genomic discovery. To that end, he and his team developed the Agent4Genomics website, where they host a range of related AI tools for scientists to use and explore.
Researchers at Google DeepMind, Princeton University and the University of California, Berkeley contributed to this study.
Funding for this research came from the National Institute of Health (grants 1R35HG011316 and 1R01GM1416), the Donald and Delia Baxter Foundation Faculty Scholar Award, the Weintz Family Foundation, and the National Science Foundation.
AI Research
Causaly Introduces First Agentic AI Platform Built for Life Sciences Research and Development –

What You Should Know:
– Causaly today announced Causaly Agentic Research, an agentic AI platform designed specifically for life sciences research and development.
– Built to deliver transparency and scientific rigor, the platform’s specialized AI agents access, analyze, and synthesize internal and external biomedical knowledge and competitive intelligence—so teams can automate complex workflows, uncover novel insights, and move from hypothesis to decision faster and with greater confidence.
A scientific AI built for how researchers actually work
Extending Causaly Deep Research, the new offering introduces a conversational interface that lets scientists partner directly with AI research agents. Unlike general-purpose tools or static literature review software, Causaly’s agents are trained for life sciences R&D and securely unify internal and external data into a single source of truth. They execute multi-step research tasks—from hypothesis generation and testing to structured, transparent outputs—always grounded in traceable evidence.
“Agentic AI fundamentally changes how life sciences conducts research,” said Yiannis Kiachopoulos, co-founder and CEO of Causaly. “Causaly Agentic Research emulates the scientific process—analyzing data, mapping biological relationships, and reasoning through problems—so scientists can reduce manual work, de-risk decisions, and focus on higher-value science.”
Solving the bottlenecks slowing discovery
R&D organizations grapple with vast, rapidly expanding biomedical information. Manual, siloed processes lengthen cycles, hide critical signals, and introduce bias. By combining extensive biomedical sources with competitive intelligence and proprietary datasets in one intelligent interface that fits existing workflows, Causaly helps teams break down silos, boost productivity, and accelerate ideas to market.
What the agents do
- Advance complex analyses and provide answers that propel research forward
- Verify quality and accuracy, compressing time-to-discovery
- Continuously scan the landscape to surface critical signals and emerging evidence in real time
- Deliver fully traceable insights to support confident, evidence-backed decisions and regulatory rigor
- Connect across systems and data—including public apps and other AI agents—to unify discovery workflows
AI Research
Learn Your Way: Reimagining textbooks with generative AI – Google Research
-
Business3 weeks ago
The Guardian view on Trump and the Fed: independence is no substitute for accountability | Editorial
-
Tools & Platforms1 month ago
Building Trust in Military AI Starts with Opening the Black Box – War on the Rocks
-
Ethics & Policy2 months ago
SDAIA Supports Saudi Arabia’s Leadership in Shaping Global AI Ethics, Policy, and Research – وكالة الأنباء السعودية
-
Events & Conferences4 months ago
Journey to 1000 models: Scaling Instagram’s recommendation system
-
Jobs & Careers3 months ago
Mumbai-based Perplexity Alternative Has 60k+ Users Without Funding
-
Podcasts & Talks2 months ago
Happy 4th of July! 🎆 Made with Veo 3 in Gemini
-
Education3 months ago
VEX Robotics launches AI-powered classroom robotics system
-
Education2 months ago
Macron says UK and France have duty to tackle illegal migration ‘with humanity, solidarity and firmness’ – UK politics live | Politics
-
Podcasts & Talks2 months ago
OpenAI 🤝 @teamganassi
-
Funding & Business3 months ago
Kayak and Expedia race to build AI travel agents that turn social posts into itineraries