AI & RoboticsNews

MIT AI system knows when to make a medical diagnosis or defer to an expert

AI can now detect lung, breast, brain, skin and cervical cancer. But in the world of medical AI, figuring out when to rely on experts versus algorithms is still tricky. It’s not simply a matter of who is “better” at making a diagnosis or prediction. Factors like how much time medical professionals have and their level of expertise also come into play. To address this, researchers from MIT’s Computer Science and Artificial Intelligence Lab (CSAIL) developed a machine learning system that can decide to either make a prediction or defer to an expert.

Most importantly, the system can adapt when and how often it defers to a human expert, based on that teammate’s availability, experience and scope of practice. For instance, in a busy hospital setting, the system may ask for human assistance only when it’s absolutely necessary.

The researchers trained the system on multiple tasks, including looking at chest X-rays to diagnose conditions like a collapsed lung. When asked to diagnose cardiomegaly (an enlarged heart), the human-AI hybrid model performed eight percent better than either the AI or medical professionals could on their own.

“There are many obstacles that understandably prohibit full automation in clinical settings, including issues of trust and accountability,” says David Sontag, lead author of a paper that the CSAIL team presented at the International Conference on Machine Learning. “We hope that our method will inspire machine learning practitioners to get more creative in integrating real-time human expertise into their algorithms.”

Next, the researchers will test a system that works with and defers to several experts at once. For instance, the AI might collaborate with different radiologists who are more experienced with different patient populations. 

The team also believes their system could have implications for content moderation because it’s able to detect offensive text and images. As social media companies struggle to remove misinformation and hate, a tool like this could help alleviate some of the burden on content moderators without resorting to full automation.


Author: Christine Fisher, @cfisherwrites
August 3, 2020

.
Source: Engadget

Related posts
AI & RoboticsNews

Nvidia and DataStax just made generative AI smarter and leaner — here’s how

AI & RoboticsNews

OpenAI opens up its most powerful model, o1, to third-party developers

AI & RoboticsNews

UAE’s Falcon 3 challenges open-source leaders amid surging demand for small AI models

DefenseNews

Army, Navy conduct key hypersonic missile test

Sign up for our Newsletter and
stay informed!