Harvard boffins build multimodal AI system to predict cancer

#artificialintelligence 

Multimodal AI models, trained on numerous types of data, could help doctors screen patients at risk of developing multiple different cancers more accurately.. Researchers from the Brigham and Women's Hospital part of Harvard University's medical school developed a deep learning model capable of identifying 14 types of cancer. Most AI algorithms are trained to spot signs of disease from a single source of data, like medical scans, but this one can take inputs from multiple sources. Predicting whether someone is at risk of developing cancer isn't always as straightforward, doctors often have to consult various types of information like a patient's healthcare history or perform other tests to detect genetic biomarkers. These results can help doctors figure out the best treatment for a patient as they monitor the progression of the disease, but their interpretation of the data can be subjective, Faisal Mahmood, an assistant professor working at the Division of Computational Pathology at the Brigham and Women's Hospital, explained. "Experts analyze many pieces of evidence to predict how well a patient may do. These early examinations become the basis of making decisions about enrolling in a clinical trial or specific treatment regimens. But that means that this multimodal prediction happens at the level of the expert. We're trying to address the problem computationally," he said in a statement.

Duplicate Docs Excel Report

Title
None found

Similar Docs  Excel Report  more

TitleSimilaritySource
None found