(PhysOrg.com) -- Scientists are developing a computer that can read vast amounts of scientific literature, make connections between facts and develop hypotheses.
Ask any biomedical scientist whether they manage to keep on top of reading all of the publications in their field, let alone an adjacent field, and few will say yes. New publications are appearing at a double-exponential rate, as measured by MEDLINE the US National Library of Medicines biomedical bibliographic database which now lists over 19 million records and adds up to 4,000 new records daily.
For a prolific field such as cancer research, the number of publications could quickly become unmanageable and important hypothesis-generating evidence may be missed. But what if scientists could instruct a computer to help them?
To be useful, a computer would need to trawl through the literature in the same way that a scientist would: reading the literature to uncover new knowledge, evaluating the quality of the information, looking for patterns and connections between facts, and then generating hypotheses to test. Not only might such a program speed up the progress of scientific discovery but, with the capacity to consider vast numbers of factors, it might even discover information that could be missed by the human brain.
The aim of Dr. Anna Korhonen and researchers in the Natural Language and Information Processing Group in the University of Cambridges Computer Laboratory is to develop computers that can understand written language in the same way that humans do. One of the projects she is involved in has recently developed a method of text mining one of the most literature-dependent areas of biomedicine: cancer risk assessment of chemicals.
Every year, thousands of new chemicals are developed, any one of which might pose a potential risk to human health. Complex risk assessment procedures are in place to determine the relationship between exposure and the likelihood of developing cancer, but its a lengthy process, as Royal Society University Research Fellow Dr Korhonen explained: The first stage of any risk assessment is a literature review. Its a major bottleneck. There could be tens of thousands of articles for a single chemical. Performed manually, its expensive and, because of the rising number of publications, its becoming too challenging to manage.
CRAB, the tool her team has developed in collaboration with Professor Ulla Stenius group at the Institute of Environmental Medicine at Swedens Karolinska Institutet, is a novel approach to cancer risk assessment that could help risk assessors move beyond manual literature review.
The approach is based on text-mining technology, which has been pioneered by computer scientists, and involves developing programs that can analyse natural language texts, despite their complexity, inconsistency and ambiguity. The tool Dr. Korhonen has developed with her colleagues is the first text-mining tool aimed at aiding literature review in chemical risk assessment.
At the heart of CRAB, the development of which was funded by the Medical Research Council and the Swedish Research Council among others, is a taxonomy that specifies scientific evidence used in cancer risk assessment, including key events that may result in cancer formation. The system takes the textual content of each relevant MEDLINE abstract and classifies it according to the taxonomy. At the press of a button, a profile is rapidly built for any particular chemical using all of the available literature, describing highly specific patterns of connections between chemicals and toxicity.
Although still under development, the system can be used to make connections that would be difficult to find, even if it had been possible to read all the documents, added Dr. Korhonen. In a recent experiment, we studied a group of chemicals with unknown mode of action and used the CRAB tool to suggest a new hypothesis that might explain their male-specific carcinogenicity in the pancreas.
The tool will be available for end-users via an online web interface. However, research into improving text mining will continue. One of the biggest current challenges is to develop adaptive technology that can be ported easily between different text types, tasks and scientific fields.
One day, rather than being at the mercy of the flourishing rate of publication, scientists will have at their fingertips a system to work alongside them that will not only point them towards those references that are relevant to their search, but will also tell them why.
Explore further:
Surgical skills with the click of a button
Nerdyguy
Wait. Stop there. Who gets to decide which are the "facts" upon which this system will build its ultimate conclusions? Setting aside for a moment such emotionally debated topics as climate change, what about the volume of just poorly-designed studies which are later shown to be invalid? Presumably, this would just be a faster method of dispensing bad science to the world. On the other hand, that would make it no different from the current system, except for the speed.
Jeddy_Mctedder
Jotaf
The inconsistencies in the data are presumably treated in the same way as in the natural language (they're bundled together in the articles). So any outlier which doesn't agree with most other studies won't be taken into account, same way as a portion of writing that doesn't make sense to the system.
In the worst-case scenario that the majority of studies are wrong but in exactly the same way, the system can't do better than human scientists would. The general assumption is that repeatable experiments are correct, and you need a very, very good reason to dispute that.
I certainly wouldn't mind an automated companion to give me a general overview of a lot of papers at once!
Nerdyguy
Makes sense to me. May still have the same biases as humans, but will be faster.
Sean_W
rwinners
hush1
An ambitious goal. Let me understand how humans understand written language first. Then use this knowledge to help humans understand written language. Then develop computers that can understand written language.
Of course text mining is orders of magnitude below this ambitious goal.
Seeker2
Maybe we could dispense of bad science period.
Seeker2
I think I can draw my own conclusions, if appropriate. Just give me the relevant facts. As for making hypotheses man that would be scary. Should be good for laughs though.
Seeker2
Poorly designed studies should be identified before they can spread. Especially controversial issues like climate change.