Carnegie Mellon University

Pradeep Ravikumar and IBM on Machine Learning Models

May 30, 2018

Pradeep Ravikumar and IBM on Machine Learning Models

Professor Pradeep Ravikumar along IBM Researchers Explain Machine Learning Models By Exploring What Isn't There

Byron Spice
  • Director of Media Relations
  • 412-268-9068

The fact that humans are able to make deductions and learn from something that’s missing isn’t something that’s yet been widely applied to machine learning, but that’s something that a team of researchers a IBM want to change. In a paper published earlier this year, the team outlined a means of using missing results to get a better understanding of how machine learning models work.

“One of the pitfalls of deep learning is that it’s more or less black box,” explained Amit Dhurandhar, one of the members of the research team. “So it’s hard to determine why a certain decision was arrived at. The answer might be accurate but in many human-critical applications–like medicine–that’s insufficient.”

To help get a better understanding of how machine learning algorithms arrive at their decisions, the IBM team created a system for "contrastive explanations"–looking for information that was missing in order to better understand how a machine learning model arrived at its conclusion. What this means in practice is that, for example, if a machine learning model is identifying photos of a dog, this method can be used to show not only what the machine model is using to identify a dog (like fur and eyes) but also what things have to be absent for the model to identify a dog (like it doesn’t have wings.)

“It’s a simple idea, but it’s an important one, that I think others have missed,” said Pradeep Ravikumar, Associate Professor in the Machine Learning Department at Carnegie Mellon University, who is not affiliated with the IBM team.

Professor Ravikumar notes that IBM’s approach is ideally suited for making determinations in areas where a machine learning model is making binary distinctions–something's either there or it isn’t–which means that, for example, if someone was denied a loan, that could be explained not only by what’s present in a credit report (like a default) but what isn’t (like a person doesn’t have a college degree.)

Would you like to learn more regarding IBM’s research on Machine Learning Models?

Read the full article:

Article, courtesy of Alex Knapp, Editor at Forbes