How can scientists deal with the huge volume of new research publish on a daily basis? How can computers go further than merely parsing scientific papers, and actually suggest hypotheses themselves? When will we see a computer as another member of the lab team, serving hundreds of scientists simultaneously from its huge data set of extant research? This is the work of John Bachman, a systems biology PhD from Harvard Medical School, and Ben Giori, a postdoctoral fellow at Harvard Medical School's systems pharmacology lab. They're part of Darpa's Big Mechanism project, which is developing technology to read research abstracts and papers to extract pieces of causal mechanisms, then to assemble these pieces into more complete causal models, and to produce explanations.
Mar-10-2017, 07:15:04 GMT