Artificial intelligence outperforms the repetitive animal tests in identifying toxic chemicals

July 25, 2018 by Thomas Hartung, The Conversation
These rats are in special cages for urine collection. Every year, millions of animals are used for testing chemicals that are used in industrial products. Credit: By unoL/shutterstock.com

Most consumers would be dismayed with how little we know about the majority of chemicals. Only 3 percent of industrial chemicals – mostly drugs and pesticides – are comprehensively tested. Most of the 80,000 to 140,000 chemicals in consumer products have not been tested at all or just examined superficially to see what harm they may do locally, at the site of contact and at extremely high doses.

I am a physician and former head of the European Center for the Validation of Alternative Methods of the European Commission (2002-2008), and I am dedicated to finding faster, cheaper and more accurate methods of testing the safety of chemicals. To that end, I now lead a new program at Johns Hopkins University to revamp the safety sciences.

As part of this effort, we have now developed a computer method of testing chemicals that could save more than a US$1 billion annually and more than 2 million animals. Especially in times where the government is rolling back regulations on the industry, new methods to identify dangerous substances are critical for human and environmental health.

How the computer took over from the lab rat

Our computerized testing is possible because of Europe's REACH (Registration, Evaluation, Authorizations and Restriction of Chemicals) legislation: It was the first worldwide regulation to systematical log existing . Over a period of one decade from 2008 to 2018, at least those chemicals produced or marketed at more than 1 ton per year in Europe had to be registered with increasing safety test information depending on the quantity sold.

Our team published a critical analysis of European testing demands in 2009 that concluded the demands of the legislation could only be met by adopting new methods of chemical analysis. Europe does not track new chemicals below an annual market or production volume of 1 ton. But the similar size U.S. brings about 1,000 chemicals at this tonnage range to the market each year. However, Europe does a much better job in requesting safety data. This also highlights how many new substances should be assessed every year even when they are produced in small quantities below 1 ton, which are not regulated in Europe. Inexpensive and fast computer methods lend themselves to this purpose.

Our group took advantage of the fact that REACH made its safety data on registered chemicals publicly available. In 2016, we reformatted the REACH data, making it machine-readable and creating the largest toxicological database ever. It logged 10,000 chemicals and connected them to the 800,000 associated studies.

This laid the foundation for testing whether animals tests – considered the gold standard for safety testing – were reproducible. Some chemicals were tested surprisingly often in the same animal test. For example, two chemicals were tested more than 90 times in rabbit eyes; 69 chemicals were tested more than 45 times. This enormous waste of animals, however, enabled us to study whether these animal tests yielded consistent results.

Our analysis showed that these tests, which consume more than 2 million animals per year worldwide, are simply not very reliable – when tested in animals a chemical known to be toxic is only proven so in about 70 percent of repeated animal tests. These were animal tests done according to OECD test guidelines under Good Laboratory Practice – which is to say, the best you can get. This clearly shows that the quality of these tests is overrated and agencies must try alternative strategies to assess the toxicity of various compounds.

This graphic reveals a small part of the chemical universe. Each dot represents a different chemical. Chemicals that are close together have similar structures and often properties. Credit: Thomas Hartung, CC BY-SA
Big data more reliable than animal testing

Following the vision of Toxicology for the 21st Century, a movement led by U.S. agencies to revamp , important work was carried out by my Ph.D. student Tom Luechtefeld at the Johns Hopkins Center for Alternatives to Animal Testing. Teaming up with Underwriters Laboratories, we have now leveraged an expanded database and machine learning to predict toxic properties. As we report in the journal Toxicological Sciences, we developed a novel algorithm and database for analyzing chemicals and determining their toxicity – what we call read-across structure activity relationship, RASAR.

To do this, we first created an enormous database with 10 million chemical structures by adding more public databases filled with chemical data, which, if you crunch the numbers, represent 50 trillion pairs of chemicals. A supercomputer then created a map of the chemical universe, in which chemicals are positioned close together if they share many structures in common and far where they don't. Most of the time, any molecule close to a toxic molecule is also dangerous. Even more likely if many are close, harmless substances are far. Any substance can now be analyzed by placing it into this map.

If this sounds simple, it's not. It requires half a billion mathematical calculations per chemical to see where it fits. The chemical neighborhood focuses on 74 characteristics which are used to predict the properties of a substance. Using the properties of the neighboring chemicals, we can predict whether an untested chemical is hazardous. For example, for predicting whether a chemical will cause eye irritation, our computer program not only uses information from similar chemicals, which were tested on rabbit eyes, but also information for skin irritation. This is because what typically irritates the skin also harms the eye.

How well does the computer identify toxic chemicals?

This method will be used for new untested substances. However, if you do this for chemicals for which you actually have data, and compare prediction with reality, you can test how well this prediction works. We did this for 48,000 chemicals that were well characterized for at least one aspect of toxicity, and we found the toxic substances in 89 percent of cases.

This is clearly more accurate that the corresponding which only yield the correct answer 70 percent of the time. The RASAR shall now be formally validated by an interagency committee of 16 U.S. agencies, including the EPA and FDA, that will challenge our computer program with chemicals for which the outcome is unknown. This is a prerequisite for acceptance and use in many countries and industries.

The potential is enormous: The RASAR approach is in essence based on chemical data that was registered for the 2010 and 2013 REACH deadlines. If our estimates are correct and chemical producers would have not registered chemicals after 2013, and instead used our RASAR program, we would have saved 2.8 million and $490 million in testing costs – and received more reliable data. We have to admit that this is a very theoretical calculation, but it shows how valuable this approach could be for other regulatory programs and safety assessments.

In the future, a chemist could check RASAR before even synthesizing their next chemical to check whether the new structure will have problems. Or a product developer can pick alternatives to toxic substances to use in their products. This is a powerful technology, which is only starting to show all its potential.

Explore further: Database analysis more reliable than animal testing for toxic chemicals

Related Stories

Public health researchers map world's 'chemical landscape'

February 12, 2016

Researchers from the Johns Hopkins Bloomberg School of Public Health have created a map of the world's chemical landscape, a catalogue of 10,000 chemicals for which there is available safety data that they say can predict ...

Recommended for you

A decade on, smartphone-like software finally heads to space

March 20, 2019

Once a traditional satellite is launched into space, its physical hardware and computer software stay mostly immutable for the rest of its existence as it orbits the Earth, even as the technology it serves on the ground continues ...

Tiny 'water bears' can teach us about survival

March 20, 2019

Earth's ultimate survivors can weather extreme heat, cold, radiation and even the vacuum of space. Now the U.S. military hopes these tiny critters called tardigrades can teach us about true toughness.

Researchers find hidden proteins in bacteria

March 20, 2019

Scientists at the University of Illinois at Chicago have developed a way to identify the beginning of every gene—known as a translation start site or a start codon—in bacterial cell DNA with a single experiment and, through ...

Turn off a light, save a life, says new study

March 20, 2019

We all know that turning off lights and buying energy-efficient appliances affects our financial bottom line. Now, according to a new study by University of Wisconsin-Madison researchers, we know that saving energy also saves ...

0 comments

Please sign in to add a comment. Registration is free, and takes less than a minute. Read more

Click here to reset your password.
Sign in to get notified via email when new comments are made.