Success by deception

February 13, 2017, ETH Zurich
Credit: Public Domain

Theoretical physicists from ETH Zurich deliberately misled intelligent machines, and thus refined the process of machine learning. They created a new method that allows computers to categorize data—even when humans have no idea what this categorization might look like.

When computers independently identify bodies of water and their outlines in satellite images, or beat the world's best professional players at the board game Go, then adaptive algorithms are working in the background. Programmers supply these algorithms with known examples in a training phase: images of bodies of water and land, or sequences of Go moves that have led to success or failure in tournaments. Similarly to how our produce new networks during learning processes, the special algorithms adapt in the learning phase based on the examples presented to them. This continues until they are able to differentiate bodies of water from land in unknown photos, or successful sequences of moves from unsuccessful ones.

Until now, these have been used in machine learning with a known decision-making criterion: we know what a body of water is and which sequences of moves were successful in Go tournaments.

Separating wheat from chaff

Now, a group of scientists working under Sebastian Huber, Professor of Condensed Matter Theory and Quantum Optics at ETH Zurich, have expanded the applications for these neural networks by developing a method that not only allows categorisation of any data, but also recognises whether complex datasets contain categories at all.

Questions of this kind arise in science: for example, the method could be useful for analysis of measurements from particle accelerators or astronomical observations. Physicists could thus filter out the most promising measurements from their often unmanageable quantities of measurement data. Pharmacologists could extract molecules with a certain probability of having a specific pharmaceutical effect or side-effect from large molecular databases. And data scientists could sort huge masses of disordered data ripples and obtain usable information (data mining).

Search for a boundary

The ETH researchers applied their method to an intensively researched phenomenon of : a many-body system of interacting magnetic dipoles that never reaches a state of equilibrium - even in the long term. Such systems have been described recently, but it is not yet known in detail which quantum physical properties prevent a many-body system from entering a state of equilibrium. In particular, it is unclear where exactly the boundary lies between systems that reach equilibrium and those that do not.

In order to locate this boundary, the scientists developed the "act as if" principle: taking data from quantum systems, they established an arbitrary boundary based on one parameter and used it to divide the data into two groups. They then trained an artificial neural network by pretending to it that one group reached a state of equilibrium while the other did not. Thus, the researchers acted as if they knew where the boundary was.

Scientists confused the system

They trained the network countless times overall, with a different boundary each time, and tested the network's ability to sort data after each session. The result was that, in many cases, the network struggled to classify the data as the scientists had. But in some cases, the division into the two groups was very accurate.

The researchers were able to show that this sorting performance depends on the location of the boundary. Evert van Nieuwenburg, a doctoral student in Huber's group, explains this as follows: "By choosing to train with a boundary far away from the actual boundary (which I don't know), I am able to mislead the network. Ultimately we're training the network incorrectly - and incorrectly trained networks are very bad at classifying ." However, if by chance a boundary is chosen close to the actual boundary, a highly efficient algorithm is produced. By determining the algorithm's performance, the researchers were able to track down the boundary between quantum systems that reach equilibrium and those that do not: the boundary is located where the network's sorting performance is highest.

The researchers also demonstrated the capabilities of their new method using two further questions from theoretical physics: topological phase transitions in one-dimensional solids and the Ising model, which describes magnetism inside solids.

Categorisation without prior knowledge

The new method can also be illustrated in simplified form with a thought experiment, where we want to classify red, reddish, bluish and blue balls into two groups. We assume that we have no idea of how such a classification might reasonably look.

If a neural network is trained by telling it that the dividing line lies somewhere in the red region, then this will confuse the network. "You try to teach the network that blue and reddish balls are the same and ask it to differentiate between red and red balls, which it simply isn't able to do," says Huber.

On the other hand, if you place the boundary in the violet colour spectrum, the network learns an actual difference and sorts the balls into red and blue groups. However, one does not need to know in advance that the dividing line should be in the violet region. By comparing the sorting performance at a variety of chosen boundaries, this boundary can be found with no prior knowledge.

Explore further: Researchers use artificial neural network to simulate a quantum many-body system

More information: van Nieuwenburg EPL, Liu YH, Huber SD: Learning phase transitions by confusion. Nature Physics, 13 February 2017, DOI: 10.1038/nphys4037

Related Stories

Looking for entangled atoms in a Bose-Einstein condensate

February 3, 2017

Using a Bose-Einstein condensate composed of millions of sodium atoms, researchers at the Georgia Institute of Technology have observed a sharp magnetically-induced quantum phase transition where they expect to find entangled ...

The thermodynamics of learning

February 6, 2017

(—While investigating how efficiently the brain can learn new information, physicists have found that, at the neuronal level, learning efficiency is ultimately limited by the laws of thermodynamics—the same principles ...

Recommended for you

CMS gets first result using largest-ever LHC data sample

February 15, 2019

Just under three months after the final proton–proton collisions from the Large Hadron Collider (LHC)'s second run (Run 2), the CMS collaboration has submitted its first paper based on the full LHC dataset collected in ...

Gravitational waves will settle cosmic conundrum

February 14, 2019

Measurements of gravitational waves from approximately 50 binary neutron stars over the next decade will definitively resolve an intense debate about how quickly our universe is expanding, according to findings from an international ...

1 comment

Adjust slider to filter visible comments by rank

Display comments: newest first

not rated yet Feb 15, 2017
This seems more of a discussion of imprecision in learning rather than a new technique. Categorization of social or physical objects are precise. Either the computer builds the tree or the tree is defined by a set of rules. So I see only a requirement that defines the axiomatic structure of any object and all objects will be defined by only two axioms and all objects may be broken into it's component parts. We deal with a finite set of objects with almost an infinite set of possibilities. Try from the bottom up instead of top down!

Please sign in to add a comment. Registration is free, and takes less than a minute. Read more

Click here to reset your password.
Sign in to get notified via email when new comments are made.