AI & RoboticsNews

MIT AI system knows when to make a medical diagnosis or defer to an expert

AI can now detect lung, breast, brain, skin and cervical cancer. But in the world of medical AI, figuring out when to rely on experts versus algorithms is still tricky. It’s not simply a matter of who is “better” at making a diagnosis or prediction. Factors like how much time medical professionals have and their level of expertise also come into play. To address this, researchers from MIT’s Computer Science and Artificial Intelligence Lab (CSAIL) developed a machine learning system that can decide to either make a prediction or defer to an expert.

Most importantly, the system can adapt when and how often it defers to a human expert, based on that teammate’s availability, experience and scope of practice. For instance, in a busy hospital setting, the system may ask for human assistance only when it’s absolutely necessary.

The researchers trained the system on multiple tasks, including looking at chest X-rays to diagnose conditions like a collapsed lung. When asked to diagnose cardiomegaly (an enlarged heart), the human-AI hybrid model performed eight percent better than either the AI or medical professionals could on their own.

“There are many obstacles that understandably prohibit full automation in clinical settings, including issues of trust and accountability,” says David Sontag, lead author of a paper that the CSAIL team presented at the International Conference on Machine Learning. “We hope that our method will inspire machine learning practitioners to get more creative in integrating real-time human expertise into their algorithms.”

Next, the researchers will test a system that works with and defers to several experts at once. For instance, the AI might collaborate with different radiologists who are more experienced with different patient populations. 

The team also believes their system could have implications for content moderation because it’s able to detect offensive text and images. As social media companies struggle to remove misinformation and hate, a tool like this could help alleviate some of the burden on content moderators without resorting to full automation.


Author: Christine Fisher, @cfisherwrites
August 3, 2020

.
Source: Engadget

Related posts
AI & RoboticsNews

Remaining Windsurf team and tech acquired by Cognition, makers of Devin: ‘We’re friends with Anthropic again’

AI & RoboticsNews

Anthropic launches finance-specific Claude with built-in data connectors, higher limits and prompt libraries

AI & RoboticsNews

Finally, a dev kit for designing on-device, mobile AI apps is here: Liquid AI’s LEAP

AI & RoboticsNews

Salesforce used AI to cut support load by 5% — but the real win was teaching bots to say ‘I’m sorry’

Sign up for our Newsletter and
stay informed!