Researchers at Emory University have figured out something fundamental about how AI systems should be designed: knowing what information to discard is just as important as knowing what to keep.
They've developed a mathematical framework that acts like a control knob for AI. Turn it one way and the system prioritizes compression — stripping away unnecessary details to work with smaller datasets. Turn it another and it prioritizes reconstruction — keeping enough information to solve the specific problem at hand. The framework, called the Variational Multivariate Information Bottleneck, is their attempt to unify how different AI methods approach this tradeoff.
"Our framework is essentially like a control knob," says co-author Michael Martini. "You can 'dial the knob' to determine the information to retain to solve a particular problem."
We're a new kind of news feed.
Regular news is designed to drain you. We're a non-profit built to restore you. Every story we publish is scored for impact, progress, and hope.
Start Your News DetoxThe breakthrough came while first author Eslam Abdelaleem was leaving campus one day. His smartwatch, misreading his racing heartbeat as three hours of cycling, accidentally captured the moment: excitement about discovering a unifying principle that could reshape how AI systems are built.
Why this matters
When AI systems encode unnecessary features, they waste computational power. That sounds like a technical problem, but it has real consequences. Less efficient AI means more electricity, more cooling, more environmental cost. The framework helps researchers avoid that waste by being intentional about what their systems actually need to learn.
The researchers tested their approach on dozens of existing AI methods and found it could derive more efficient loss functions — the mathematical rules that guide how AI systems learn — particularly when training data is limited. In fields like biology and cognitive science, where gathering large datasets is expensive and time-consuming, this efficiency gain could be significant.
"By helping guide the best AI approach, the framework helps avoid encoding features that are not important," says senior author Ilya Nemenman. "The less data required for a system, the less computational power required to run it, making it less environmentally harmful."
The real value here is that the framework provides a shared language. Instead of researchers in different fields reinventing the wheel each time they build a new AI system, they can use this unified principle to tailor algorithms to their specific questions. A neuroscientist studying brain function and a biologist analyzing protein structures could both use the same underlying framework, just dialed differently.
The researchers are publishing their work openly, hoping other teams will use it to build more efficient, more targeted AI systems for their own research. The framework doesn't solve AI — but it gives the next generation of builders a clearer map of what they're actually trying to do.










