Artificially intelligent algorithms can learn to identify amazingly subtle information, enabling them to distinguish between people in photos or to screen medical images as well as a doctor. But in most cases their ability to perform such feats relies on training that involves thousands to trillions of data points. This means artificial intelligence doesn't work all that well in situations where there is very little data, such as drug development. Vijay Pande, professor of chemistry at Stanford University, and his students thought that a fairly new kind of deep learning, called one-shot learning, that requires only a small number of data points might be a solution to that low-data problem. Stanford chemistry Professor Vijay Pande and his students see a future for machine learning in the early stages of drug development.
As new and tailored variants of medicines are developed can traditional clinical trial management solutions still meet the evolving needs of life sciences companies? However, what if different types of content and workflow management solutions could dynamically alter flows based on its own perceptible AI? However, there is an opportunity with AI for process flows to dynamically target the production of different types of flu vaccine automatically. Today, Enterprise Content Management (ECM) processes still require someone with intelligence to understand the defined flow (for example managing clinical trials), as it is a highly regulatory process that has a lot of governance and control.
The researchers constructed a bacterium's "genetic software" and transplanted it into a host cell. But the researchers hope eventually to design bacterial cells that will produce medicines and fuels and even absorb greenhouse gases. Dr Venter told BBC News: "We've now been able to take our synthetic chromosome and transplant it into a recipient cell - a different organism. Dr Helen Wallace from Genewatch UK, an organisation that monitors developments in genetic technologies, told BBC News that synthetic bacteria could be dangerous.