For two decades, typing symptoms into Google has been the first move for anyone feeling off. Now OpenAI is betting that ChatGPT can do it better. The company launched ChatGPT Health this month—a dedicated sidebar that wraps one of its existing models with medical guidance and safety tools. The stakes feel real: 230 million people ask ChatGPT health questions every week.
The appeal is straightforward. Unlike Google's sprawl of medical forums, pharmaceutical ads, and outdated wikis, a conversational AI can theoretically help you sort signal from noise. Marc Succi, a radiologist and associate professor at Harvard Medical School, sees genuine value here. "You see patients with a college education, a high school education, asking questions at the level of something an early med student might ask," he says. For people trying to understand their symptoms before calling a doctor—or deciding whether they need to call at all—that's not nothing.
But the evidence is messier than the marketing. Studies show GPT-4o answers medical questions correctly about 85% of the time in controlled tests. In real-world conditions? Medical experts rated only about half of its responses as entirely correct. The gap between "passes the licensing exam" and "actually helps you" is wider than it looks.
We're a new kind of news feed.
Regular news is designed to drain you. We're a non-profit built to restore you. Every story we publish is scored for impact, progress, and hope.
Start Your News DetoxThe real risks are baked into how these models work. LLMs tend to sound confident even when they're wrong. They agree with you rather than push back. They can fabricate medical facts—what researchers call "hallucination"—without admitting uncertainty. A person who trusts ChatGPT to validate their self-diagnosis, or worse, to replace a doctor's advice, could end up worse off.
OpenAI claims its newer GPT-5 models hallucinate less and are less eager to please. The company has specifically tested ChatGPT Health on medical questions. That's progress, but experts point out that some problems run deeper: an LLM that encourages you to second-guess your doctor's advice, or that misses the nuance of your individual health history, can do real harm even if it's technically accurate 85% of the time.
The honest take: ChatGPT Health is probably better than a random Google rabbit hole. It's worse than talking to an actual doctor. The tool works best as a bridge—a way to clarify your symptoms before an appointment, or to understand what your doctor just told you. Treat it as a second opinion from a smart friend, not a substitute for one from someone with a medical license and your full medical history.
As these tools evolve, the question isn't whether they'll get smarter. It's whether people will use them wisely—and whether the companies building them will be honest about what they can and can't do.










