Connect with us

AI Research

Smart Biology on a Budget: Why Researchers Should Leverage AI to Rethink Experimental Design

Published

on


Most biology labs today—particularly smaller or early-career groups and those outside major funding hubs—face a tough reality: shrinking budgets, rising costs, limited access to cutting-edge technologies, and mounting pressures to publish. The challenge isn’t just keeping pace with innovation—it’s staying in the game at all. In this climate, simulation-first approaches are a powerful equalizer. By allowing scientists to test biological hypotheses computationally before committing to costly experiments, these tools enable smart science even on a tight budget.

Constraint often drives creativity. One strategy gaining momentum is AI-guided predictive modeling, which relies on data-driven tools that simulate biological behavior to unlock new insights. Built using advanced techniques like generative modeling (which simulates how a cell might respond to genetic or environmental changes) and probabilistic inference (which estimates the likelihood of different outcomes), these models are reshaping how we interpret complex data.

AI models are already used to forecast weather, simulate disease outbreaks, and model ecosystems. In single-cell biology, using programs such as CellOracle, scGen, and scVI, researchers can model how cells respond to perturbations, reprogramming, or developmental cues. Similarly, tools like DeepCell apply deep learning to image-based single-cell phenotyping, automating what used to require hours of manual microscopy. Meanwhile, AlphaFold has revolutionized structural biology by predicting protein structures in silico—compressing what once took months or years of lab work into minutes.

The trend is clear: AI is becoming a critical partner in formulating and refining biological hypotheses before stepping into the lab. These models often require large, high-quality training datasets to perform well, posing challenges for researchers working in under-sampled systems or with limited starting material. Even so, when good data are available, or we intentionally generate AI-compatible datasets, these tools offer a faster, more cost-effective way to ask complex biological questions, generate hypotheses, prioritize experiments, and deepen our understanding of biology.

Predictive modeling doesn’t replace wet lab experiments; it makes them sharper, faster, and smarter. This approach is already reshaping workflows. For example, Samantha Morris, a biologist at Harvard Medical School, and her team used CellOracle to computationally infer gene regulatory networks from single-cell multi-omics data and simulate transcription factor perturbations in silico, helping identify developmental regulators most worthy of experimental follow-up.1 This allowed them to predict key shifts in cell identity without performing initial perturbation experiments, which they later validated in the lab—demonstrating a powerful strategy for prioritizing follow-up studies while conserving time and resources. Similarly, Fabian Theis, director at HelmholtzAI, and his team created scGen to predict cellular responses to disease and drug treatments, providing a map of likely outcomes before costly screening.2 Like any model, AI tools are only as good as the data and assumptions behind them. Their predictions should guide, not replace, biological reasoning.

Beyond improving efficiency and experimental design, these AI-driven approaches are also compelling because they democratize access to powerful biological insights and to science in general. Once trained, researchers, even those with limited resources, can run these predictive models and generate predictions using public datasets. The Human Cell Atlas, for instance, offers comprehensive single-cell profiles across tissues and conditions, while Gene Expression Omnibus (GEO) hosts thousands of curated transcriptomic datasets. Tabula Sapiens provides a human cell-type atlas across multiple organs, a valuable reference for cross-tissue comparisons. These datasets, combined with pre-trained models, can power meaningful discovery even without needing immediate access to a physical lab or multi-million budgets.

While AI tools are powerful, they do require some tuning, context, and computational literacy, and are not yet entirely plug-and-play for all users. Barriers to widespread adoption remain, including steep learning curves, uncertainty in interpreting model predictions, and the disciplinary divide, marked by communication gaps, between experimental and computational scientists. But the single-cell biology landscape is evolving fast: Documentation is improving, community awareness and support are growing, and most tools include training tutorials.

Adoption of AI in biology also requires a cultural shift. Many biologists still view modeling as speculative or overly abstract. But with the growing integration of AI in biomedical science, learning to collaborate across disciplines—and to trust simulations as a complementary part of the research pipeline—can open new doors. Predictive modeling allows us not just to reduce redundancy or lower costs, but to ask smarter questions, explore rare or difficult systems, and build more reproducible, generalizable models of biology. It’s a way to work more strategically, inclusively, and sustainably.

Increasingly, biologists are recognizing that our ability to ask meaningful questions is no longer constrained by what we can measure—it’s constrained by what we can model. That “simulation-first science” sentiment captures a growing recognition: In the era of abundant data, insight comes from interpretation, not just generation.

Biology’s future is predictive, not just descriptive. AI-driven models connect raw data to meaningful insight, empowering researchers to ask smarter questions and make faster progress. In an era of shrinking budgets and expanding ambition, adopting these tools isn’t a luxury—it’s the new foundation of biological discovery.



Source link

Continue Reading
Click to comment

Leave a Reply

Your email address will not be published. Required fields are marked *

AI Research

The Machine Learning Lessons I’ve Learned This Month

Published

on


in machine learning are the same.

Coding, waiting for results, interpreting them, returning back to coding. Plus, some intermediate presentations of one’s progress. But, things mostly being the same does not mean that there’s nothing to learn. Quite on the contrary! Two to three years ago, I started a daily habit of writing down lessons that I learned from my ML work. In looking back through some of the lessons from this month, I found three practical lessons that stand out:

  1. Keep logging simple
  2. Use an experimental notebook
  3. Keep overnight runs in mind

Keep logging simple

For years, I used Weights & Biases (W&B)* as my go-to experiment logger. In fact, I have once been in the top 5% of all active users. The stats in below figure tell me that, at that time, I’ve trained close to 25000 models, used a cumulative 5000 hours of compute, and did more than 500 hyperparameter searches. I used it for papers, for big projects like weather prediction with large datasets, and for tracking countless small-scale experiments.

My once upon a time stats of using W&B for experiment logging. Image by the author.

And W&B really is a great tool: if you want beautiful dashboards and are collaborating** with a team, W&B shines. And, until recently, while reconstructing data from trained neural networks, I ran multiple hyperparameter sweeps and W&B’s visualization capabilities were invaluable. I could directly compare reconstructions across runs.

But I realized that for most of my research projects, W&B was overkill. I rarely revisited individual runs, and once a project was done, the logs just sat there, and I did nothing with them ever after. When I then refactored the mentioned data reconstruction project, I thus explicitly removed the W&B integration. Not because anything was wrong with it, but because it wasn’t necessary.

Now, my setup is much simpler. I just log selected metrics to CSV and text files, writing directly to disk. For hyperparameter searches, I rely on Optuna. Not even the distributed version with a central server — just local Optuna, saving study states to a pickle file. If something crashes, I reload and continue. Pragmatic and sufficient (for my use cases).

The key insight here is this: logging is not the work. It’s a support system. Spending 99% of your time deciding on what you want to log — gradients? weights? distributions? and at which frequency? — can easily distract you from the actual research. For me, simple, local logging covers all needs, with minimal setup effort.

Maintain experimental lab notebooks

In December 1939, William Shockley wrote down an idea into his lab notebook: replace vacuum tubes with semiconductors. Roughly 20 years later, Shockley and two colleagues at Bell Labs were awarded Nobel Prizes for the invention of the modern transistor.

While most of us aren’t writing Nobel-worthy entries into our notebooks, we can still learn from the principle. Granted, in machine learning, our laboraties don’t have chemicals or test tubes, as we all envision when we think about a laboratory. Instead, our labs often are our computers; the same device that I use to write these lines has trained countless models over the years. And these labs are inherently portably, especially when we are developing remotely on high-performance compute clusters. Even better, thanks to highly-skilled administrative stuff, these clusters are running 24/7 — so there’s always time to run an experiment!

But, the question is, which experiment? Here, a former colleague introduced me to the idea of mainting a lab notebook, and lately I’ve returned to it in the simplest form possible. Before starting long-running experiments, I write down:

what I’m testing, and why I’m testing it.

Then, when I come back later — usually the next morning — I can immediately see which results are ready and what I had hoped to learn. It’s simple, but it changes the workflow. Instead of just “rerun until it works,” these dedicated experiments become part of a documented feedback loop. Failures are easier to interpret. Successes are easier to replicate.

Run experiments overnight

That’s a small, but painful lessons that I (re-)learned this month.

On a Friday evening, I discovered a bug that might affect my experiment results. I patched it and reran the experiments to validate. By Saturday morning, the runs had finished — but when I inspected the results, I realized I had forgotten to include a key ablation. Which meant … another full day of waiting.

In ML, overnight time is precious. For us programmers, it’s rest. For our experiments, it’s work. If we don’t have an experiment running while we sleep, we’re effectively wasting free compute cycles.

That doesn’t mean you should run experiments just for the sake of it. But whenever there is a meaningful one to launch, starting them in the evening is the perfect time. Clusters are often under-utilized and resources are more quickly available, and — most importantly — you will have results to analyse the next morning.

A simple trick is to plan this deliberately. As Cal Newport mentions in his book “Deep Work”, good workdays start the night before. If you know tomorrow’s tasks today, you can set up the right experiments in time.


* That ain’t bashing W&B (it would have been the same with, e.g., MLFlow), but rather asking users to evaluate what their project goals are, and then spend the majority of time on pursuing that goals with utmost focus.

** Footnote: mere collaborating is in my eyes not enough to warrant using such shared dashboards. You need to gain more insights from such shared tools than the time spent setting them up.



Source link

Continue Reading

AI Research

How is artificial intelligence affecting job searches?

Published

on


Artificial intelligence programs like ChatGPT use AI to do thinking or writing or creating for you. Pretty amazing, but also a little terrifying. What happens to the people who used to do those jobs?

Olivia Fair graduated four years ago. “I’ve applied to probably over a hundred jobs in the past, I don’t know, six months,” she said. “And yeah, none of them are landing.”

She’s had a series of short-term jobs – one was in TV production, transcribing interviews. “But now they don’t have a bunch of people transcribing,” she said. “They have maybe one person overseeing all of that, and AI doing the rest. Which I think is true for a lot of entry-level positions. And it can be a very useful tool for those people doing that work. But then there’s less people needed.”

According to Laura Ullrich, director of economic research at Indeed, the job-listings website, job postings have declined year over year by 6.7 percent. “This is a tough year,” she said. “Younger job seekers, specifically those who are recent grads, are having a harder time finding work.”

Asked if there is a correlation between the rise in AI and the decline in jobs for recent graduates, Ullrich said, “I think there is a cause-and-effect, but it’s maybe not as significant as a lot of people would think. If you look specifically at tech jobs, job postings are down 36% compared to pre-pandemic numbers. But that decline started happening prior to AI becoming commonly used.”

Ullrich said in 2021-22, as the effects of the COVID pandemic began to ebb, there was a hiring boom in some sectors, including tech: “Quite frankly, I think some companies overhired,” she said.

The uncertain national situation (tariffs, taxes, foreign policy) doesn’t help, either. Ullrich said, “Some other people have used the analogy of, like, driving through fog. If it’s foggy, you slow down a bit. But if it’s really foggy, you pull over. And unfortunately, some companies have pulled over to sit and wait to see what is gonna happen.”

That sounds a little more nuanced than some recent headlines, which make it pretty clear that AI is taking jobs:

“I read today an interview with a guy who said, you know, ‘By 2027, we will be jobless, lonely, crime on the streets,'” said David Autor, a labor economist at MIT. “And I said, ‘How do I take the other side of that bet?’ ‘Cause that’s just not true. I’m sure of that. My view is, look, there is great potential and great risk. I think that it’s not nearly as imminent on either direction as most people think.”

I said, “But what it does seem to do is relieve the newcomers, the beginning, incoming novices we don’t need anymore.”

“This is really a concern,” Autor said. “Judgment, expertise, it’s acquired slowly. It’s a product of immersion, right? You know, how do I care for this patient, or land this plane, or remodel this building? And it’s possible that we could strip out so much of the supporting work, that people never get the expertise. I don’t think it’s an insurmountable concern. But we shouldn’t take for granted that it will solve itself.”

Let’s cut to the chase. What are the jobs we’re going to lose? Laura Ullrich said, “We analyzed 2,800 specific skills, and 30% of them could be, at least partially, done by AI.” (Which means, 70% of job skills are not currently at risk of AI.) 

So, which jobs will AI be likely to take first? Most of it is jobs in front of a screen:

  • Coding
  • Accounting
  • Copy writing
  • Translation
  • Customer service
  • Paralegal work
  • Illustration
  • Graphic design
  • Songwriting
  • Information management

As David Autor puts it: “What will market demand be for this thing? How much should we order? How much should we keep in stock?”

AI will have a much harder time taking jobs requiring empathy, creative thinking, or physicality:

  • Healthcare
  • Teaching
  • Social assistance
  • Mental health
  • Police and fire
  • Engineering
  • Construction
  • Wind and solar
  • Tourism
  • Trades (like plumbing and electrical)

And don’t forget about the new job categories that AI will create. According to Autor, “A lot of the work that we do is in things that we just didn’t do, you know, 50 or 100 years ago – all this work in solar and wind generation, all types of medical specialties that were unthinkable.”

I asked, “You can’t sit here and tell me what the new fields and jobs will be?”

“No. We’re bad at predicting where new work will appear, what skills it will need, how much there will be,” Autor said, adding, “There will be new things, absolutely.”

“So, it sounds like you don’t think we are headed to becoming a nation of people who cannot find any work, who spend the day on the couch watching Netflix?”

Autor said, “No, I don’t see that. Of course, people will be displaced, certain types of occupations will disappear. People will lose careers. That’s going to happen. But we might actually get much better at medicine. We might figure out a way to generate energy more cheaply and with less pollution. We might figure out a better way to do agriculture that isn’t land-intensive and so ecologically intensive.”

Whatever is going to happen, will likely take a while to happen. The latest headlines look like these:

Until then, Laura Ullrich has some advice for young job seekers: “The number one piece of advice I would give is, move forward. So, whether that is getting another job, getting a part-time job, finding a post-graduate internship – reach out to the professors that you had. They have a whole network of former students, right? Reach out to other alumni who graduated from the school you went to, or majored in the same thing you majored in. It might be what gets you a job this year.”

So far, Olivia Fair is doing all of the above. I asked her, “You’re interested in creativity and writing and production. So let me hear, as a human, your pitch, why you’d be better than AI doing those jobs?”

“Okay,” Fair replied. “Hmm. I’m a person, and not a robot?”

      
For more info:

     
Story produced by Gabriel Falcon. Editor: Chad Cardin.

     
See also:



Source link

Continue Reading

AI Research

Bay Area home sales are cooling — but AI-bolstered SF is heating up

Published

on


The Bay Area housing market is in something of a lull, with sales down slightly this year compared with 2024. But sales in San Francisco are on the rise, a trend real estate agents attribute to the artificial intelligence boom and renewed optimism about the city’s future.

The number of Bay Area homes, including condominiums and co-ops, sold from January to July is down more than 2% from the same period last year, according to data from online real estate brokerage Redfin. But in San Francisco, sales are up 5%, rising from about 2,870 in 2024 to 3,010 in 2025.

The rise of AI companies in San Francisco, and the city’s affordable housing shortage, has already contributed to a surge in rents. While the growth in sales hasn’t yet translated to rising home prices — the city’s typical home value of $1.27 million is about 1% lower than it was last year, according to real estate company Zillow — some real estate agents believe that could soon change.

“I don’t want to say (the market’s) hot, but it’s very, very, very warm,” said Ruth Krishnan, a San Francisco real estate agent with Compass.

When mortgage rates spiked in 2022, home sales in San Francisco — and just about everywhere else — plummeted. Markets heated up again after those rates dipped in 2023, with prices soaring in Silicon Valley, where return-to-office policies and strong tech stock growth drove competition among buyers. This year, the market cooled again, thanks to a combination of tech layoffs, volatility in the stock market and the unlikelihood that mortgage rates will fall much further.

But San Francisco is proving to be a partial exception. Sales have gradually recovered over the past two years, even nearing pre-pandemic levels. Krishnan said renewed enthusiasm about the city and its new mayor, as well as AI companies’ move to the city, have led to a jump in sales. Some buyers may also be capitalizing on the dip in San Francisco home prices suspecting that prices will soon increase, she added.

The single-family home market is the primary driver behind the increase in sales, Redfin data shows, with condo sales practically flat from last year. Condo listings in San Francisco and San Mateo County have actually dipped, said Redfin senior economist Asad Khan, possibly indicating that some condo sellers have simply given up on finding a buyer. A May report by the company showed that more than a third of for-sale condos in the San Francisco metropolitan area were at risk of selling at a loss.

The condo market could eventually shift as competition for homes near tech offices heats up, said Patrick Carlisle, chief market analyst at Compass. But the biggest impact — at least at first — will probably be on rents, he added.

Besides San Francisco, only a few mid- and large-size Bay Area cities have seen home sales rise from last year, with sales in Vacaville and Oakland rising by about 10% and 5%, respectively. But both cities are much further from their pre-pandemic sales numbers than San Francisco.

In Oakland, those numbers reflect a market that remains fairly soft, according to East Bay real estate broker Daniel Winkler. An increase in inventory over the past couple of years has forced many sellers to offer deals, such as paying buyers’ closing costs or subsiding their mortgages.



Source link

Continue Reading

Trending