A junior researcher and a high school student just did something that would normally take months of specialized programming work. Using generative AI, they built prediction models for preterm birth in weeks—and their results matched or beat those from experienced human teams who'd spent far longer on the same problem.
This wasn't a controlled lab experiment. It was a real test. Scientists at UC San Francisco and Wayne State University gave eight different AI systems the same medical datasets that had challenged over 100 teams in a global research competition called DREAM. The human teams had taken months to analyze vaginal microbiome data and blood samples, trying to predict which pregnancies would end prematurely and when babies would actually be born.
The AI systems, guided by carefully written prompts, generated working code in minutes—something experienced programmers normally spend hours or days building. Only four of the eight AI tools produced usable results, but those that worked didn't need large teams of specialists hovering over them. The entire process, from first prompt to published findings, took six months. The original human competition followed by analysis and publication had taken nearly two years.
We're a new kind of news feed.
Regular news is designed to drain you. We're a non-profit built to restore you. Every story we publish is scored for impact, progress, and hope.
Start Your News DetoxWhy Speed Matters Here
Preterm birth kills more newborns than anything else and leaves many survivors with lasting cognitive and movement challenges. In the US alone, roughly 1,000 babies are born too early each day. Researchers still don't fully understand what triggers it. To find answers, they need to analyze enormous datasets—in this case, microbiome data pooled from about 1,200 pregnant women across nine separate studies. That kind of work only happens when researchers share data openly and collaborate globally.
But analyzing that much complex information has always been the bottleneck. You need skilled data scientists just to build the analytical pipeline—the code that sorts and processes the raw data—before you can even start looking for patterns. That's where generative AI changes the equation.
Marina Sirota, who co-led the original DREAM challenges at UCSF, put it plainly: these tools could "relieve one of the biggest bottlenecks in data science." The speed matters because patients need answers now, not in two years.
The Catch
Scientists are careful to note that AI still needs human judgment. These systems can produce misleading results if not properly overseen. The fact that only half the AI tools tested actually generated usable code underscores that point. But the researchers see the real win differently: AI isn't replacing expertise, it's freeing experts from tedious debugging so they can spend time on what they're actually good at—asking the right questions and interpreting what the data means.
Adi Tarca from Wayne State, who led two of the original DREAM challenges, sees a future where researchers without deep coding skills don't have to build massive teams just to analyze data. They can focus on the science instead.
The next phase is already underway: scaling this approach to larger datasets and testing whether these AI-generated models hold up when applied to new patient populations. If they do, the timeline for medical discoveries could shift dramatically.











