The significant advance, by a team at the University of New South Wales (UNSW) in Sydney appears today in the international journal Nature.
"What we have is a game changer," said team leader Andrew Dzurak, Scientia Professor and Director of the Australian National Fabrication Facility at UNSW.
"We've demonstrated a two-qubit logic gate - the central building block of a quantum computer - and, significantly, done it in silicon. Because we use essentially the same device technology as existing computer chips, we believe it will be much easier to manufacture a full-scale processor chip than for any of the leading designs, which rely on more exotic technologies.
"This makes the building of a quantum computer much more feasible, since it is based on the same manufacturing technology as today's computer industry," he added.
The advance represents the final physical component needed to realise the promise of super-powerful silicon quantum computers, which harness the science of the very small - the strange behaviour of subatomic particles - to solve computing challenges that are beyond the reach of even today's fastest supercomputers.
In classical computers, data are rendered as binary bits, which are always in one of two states: 0 or 1. However, a quantum bit (or 'qubit') can exist in both of these states at once, a condition known as a superposition. A qubit operation exploits this quantum weirdness by allowing many computations to be performed in parallel (a two-qubit system performs the operation on 4 values, a three-qubit system on 8, and so on).
"If quantum computers are to become a reality, the ability to conduct one- and two-qubit calculations are essential," said Dzurak, who jointly led the team in 2012 who demonstrated the first ever silicon qubit, also reported in Nature.
Until now, it had not been possible to make two quantum bits 'talk' to each other - and thereby create a logic gate - using silicon. But the UNSW team - working with Professor Kohei M. Itoh of Japan's Keio University - has done just that for the first time.
The result means that all of the physical building blocks for a silicon-based quantum computer have now been successfully constructed, allowing engineers to finally begin the task of designing and building a functioning quantum computer.
A key advantage of the UNSW approach is that they have reconfigured the 'transistors' that are used to define the bits in existing silicon chips, and turned them into qubits. "The silicon chip in your smartphone or tablet already has around one billion transistors on it, with each transistor less than 100 billionths of a metre in size," said Dr Menno Veldhorst, a UNSW Research Fellow and the lead author of the Nature paper.
"We've morphed those silicon transistors into quantum bits by ensuring that each has only one electron associated with it. We then store the binary code of 0 or 1 on the 'spin' of the electron, which is associated with the electron's tiny magnetic field," he added.
Dzurak noted that that the team had recently "patented a design for a full-scale quantum computer chip that would allow for millions of our qubits, all doing the types of calculations that we've just experimentally demonstrated."
He said that a key next step for the project is to identify the right industry partners to work with to manufacture the full-scale quantum processor chip.
Such a full-scale quantum processor would have major applications in the finance, security and healthcare sectors, allowing the identification and development of new medicines by greatly accelerating the computer-aided design of pharmaceutical compounds (and minimizing lengthy trial and error testing); the development of new, lighter and stronger materials spanning consumer electronics to aircraft; and faster information searching through large databases.
Explore further:
Physicists set new records for silicon quantum computing
More information:
A two-qubit logic gate in silicon, DOI: 10.1038/nature15263

marko
not rated yet Oct 05, 2015A Nobel Prize for Physics must be around the corner.
If the Australian Government had any brains, it would insist the chips and computers be made in Australia and the technological output of this breakthrough be used there first.
There's probably some stupid clause in one of those free trade agreements which might stop this happening.
The timely solutions to intractable problems could be worth more than the sale of the hardware.
Just think about optimising scheduling for global transport operations.
EyeNStein
not rated yet Oct 05, 2015If a multi qbit array can factor this into two primes (13x17) near instantly on a single chip then the computer world will take notice very rapidly, as all domestic encryption depends on the difficulty of factoring larger primes.
The 8 bit examples of primes multiplied should be an excellent demo of what quantum computers should do readily. The algorithms are waiting for the hardware.
docile
Oct 05, 2015qitana
1 / 5 (1) Oct 05, 2015The maximum number of cubits used for computation has been 7.
And now, it could be might be possible to scale it up to 1000's of qubits?
Is that true? No problem with noise destroying the superposition states? No problem with loss of quantum information?
t689
1 / 5 (2) Oct 05, 2015Thatpaper describes a similar scenario of a 2-qubit network, or multiple 2-qubit logic gates linked to form a quantum computer. In that paper, it describes how they believe it is possible using what we have here (pure silicon as your base to put you logic gate on) to amass over 100 of the two-qubit gates to form a usable quantum computer. The video has the scientist stating millions of them, but I haven't looked for a paper that says that.
ProcrastinationAccountNumber3659
2.5 / 5 (8) Oct 05, 2015Quantum based computers and classical computers are used for different applications as mentioned in the article. There is no proposed switch from classic to quantum in this case, these computers are meant for applications where classical computers have difficulty solving problems in any reasonable amount of time.
ProcrastinationAccountNumber3659
1.6 / 5 (7) Oct 05, 2015t689
1 / 5 (2) Oct 05, 2015From what I am reading and the paper I posted in my previous post and others today, a 2-qubit logic gate is sufficient and I believe the minimum to form a universal quantum computer. And I think that the team at Sydney has performed a significant advance in placing the logic gate on a pure form of silicon at each gate, which from my understanding makes for stable computations. The other article I read developed a scenario with quantum dots on the same form of pure silicon and their research says they could link over 100 of those quantum gates to perform reliable quantum computations. There is no mention of quantum dots in this article, might be wrong, but I see no problem from the other article that they can't link 100 in a completely stable way, which in my opinion would lend credence to the scientist talking about significant advances in healthcare and materials science.
qitana
5 / 5 (1) Oct 06, 2015The D-Wave quantum computer isn't considered to be a true quantum computer
docile
Oct 06, 2015Returners
1 / 5 (5) Oct 06, 2015We can only hope that muslims are never allowed access to quantum computing technology. There needs to be some sort of universal U.N. ban on muslims using Quantum technology, as well as a constitutional amendment in the U.S. doing the same. We can't have these lunatics developing weapons or computer hacks that could ruin what's left of western civilization.
Returners
1 / 5 (4) Oct 06, 2015The reason I say they should be limited to the same actions per minute is because I define game balance as "equal effort provides equal results and equal opportunity to win".
Strangely enough, Blizzard and most of their fan base do not seem to define game balance that way.
ProcrastinationAccountNumber3659
2.2 / 5 (10) Oct 06, 2015ProcrastinationAccountNumber3659
2.3 / 5 (9) Oct 06, 2015The first article you linked to doe not demonstrate the point you are trying to make. It just shows the DWAVE-one is not as useful as the company claims.
The second article you linked to is not about using indeterminism. It is pointing out well known thing in engineering should be applied in computing more. Approximate where you can and only be precise when you have to. It basically breaks down to them saying hey guys you know you can approximate 3.1 to 3 for many applications without it mattering.
The method you posted for encryption is interesting. The article does not say they have the same limits. It just shows that classical physics can also be used to create an unbreakable encryption.
docile
Oct 06, 2015docile
Oct 06, 2015docile
Oct 06, 2015t689
1 / 5 (2) Oct 06, 2015Returners
1 / 5 (4) Oct 06, 2015I'll never forget the videos of all those "moderate" muslims dancing in the streets celebrating on 9/11/2001.
Race has nothing to do with it, as the evil religion of islam has spread it's festering cancer to every race in the world.
You can't afford to be so naive.
The extremists get their food, clothing, and shelter from somewhere...
Oh yeah, *cough, cough* they get it from the "moderates" donations to their organizations.
Antisorc
3.5 / 5 (8) Oct 06, 2015You are one, very deluded individual. You can't afford to be so ignorant in this day and age. Firstly, why even bring up religion here? Secondly, don't forget that it was the very Muslims you are vilifying who ensured this research could take place...hundreds of years ago when they saved Greek works from destruction and added their own scientific breakthroughs in math, optics, physics etc.
Open your eyes man and stop watching Fox News
ProcrastinationAccountNumber3659
2.1 / 5 (7) Oct 06, 2015They are talking about a communications channel not computing. They do state that quantum techniques offer improvement for computing.
The DWAVE implementation is limited by noise and is not a general purpose processing unit. The article here talks about the building block of a general purpose processing unit.
ProcrastinationAccountNumber3659
2.1 / 5 (7) Oct 06, 2015ProcrastinationAccountNumber3659
2.1 / 5 (7) Oct 06, 2015That is verbatim from another website.
docile
Oct 06, 2015ProcrastinationAccountNumber3659
2.1 / 5 (7) Oct 06, 2015You do understand the DWAVE machine is not a great example of quantum computing, right? The people in the field of quantum computing do not expect a generalized processor to be realized until the 20s.
docile
Oct 06, 2015ProcrastinationAccountNumber3659
2.1 / 5 (7) Oct 06, 2015Returners
1 / 5 (3) Oct 06, 2015What a joke.
Virtually all development in any of those areas was done by Babylonians (not Muslims), Greeks, Romans, and other Europeans all the way up to and past Newton.
The "Muslim" contributions include some trinkets and a few machines which were actually already known and used in construction projects and other purposes several thousand years earlier by the Jews, the Babylonians, and the Egyptians during the reign of the Pharaohs.
Islam doesn't contribute anything. It mostly just steals and destroys.
65% of Nobel winners self-identified as Christian.
docile
Oct 06, 2015Returners
3 / 5 (4) Oct 06, 2015in order to get close to the theoretical limit you would need a machine specifically designed for solving the problem you have in mind; for example, solving the game of chess for the best possible game assuming both players always make the best move, and play to win, up to a forced 3 fold repetition, and of course doing this solution in the fewest number of total calculations.
Well, classical computer algorithms attempt to do that, but quantum computer algorithms are supposedly (theoretically) able to solve this in far fewer calculations...in an ideal scenario.
The theoretical mathematical advantage to the quantum computer depends on the type of problem you are attempting to solve, and the type of quantum computer you are using.
however, the claim of a QC being able to do infinite calculations, etc, etc, is BS.
Returners
4 / 5 (4) Oct 06, 2015In addition, if you wanted to use quantum computing for sorting and searching databases you would need to design the database to be used in that manner, and it will not necessarily work any faster (to the end user) than the existing indexing methods used in classical computers, for example by Google and other search engines. Again, it will depend on the situation. In some situations it might save you a lot of time, and in others it might not make a noticeable difference.
The real issue is that there are certain classes of calculations which supposedly cannot be done on a classical computer at all, which I doubt for reasons I'll specify later, and the quantum computer can supposedly solve some of these problems relatively quickly.
Returners
3 / 5 (4) Oct 06, 2015Why?
Because a classical computer can theoretically model any mathematical construct, the model in this case would just run slower than a normal, physical quantum computer.
it would be like an emulator or virtual engine.
ProcrastinationAccountNumber3659
2.5 / 5 (8) Oct 06, 2015ProcrastinationAccountNumber3659
2.5 / 5 (8) Oct 06, 2015Exactly and your previous two comments are right on. Its not that the problems cannot be done on a classical computer, it would just take an unreasonably long amount of time.
I still cannot believe you insist on those racist comments. You need to get to know more Muslims, the majority I know are good people. I personally dislike all religions and think they are more harmful than helpful. This is not the place to discuss it though.
docile
Oct 07, 2015ProcrastinationAccountNumber3659
2.3 / 5 (6) Oct 07, 2015Traditional computing is based on bits that can be either a 0 or a 1 and operations are performed on these bits with logic gates (AND, OR, NOR, NOT)
Quantum computers use qubits that can be a 0, 1, or both. That both is an important aspect of quantum computing. I believe the logic gates will be the same.
Neural networks: These function similar to the human brain. They are much more similar to an analog computer, but more general purpose.
Analog computing: This class of computing is too application specific to be useful. You supposedly see it in some applications.
DNA Computing: I am not familiar with this one.
Computing types are based on the math behind it not the physical system. The computer system is designed to do specific mathematical operations and then we attempt to formulate our real world problems in ways that those mathematical operations can be used.
jerromyjon
not rated yet Oct 10, 2015Classical computers deal with a low ratio of errors in data, which is dealt with by checksum. When you look at the probabilistic nature of quantum objects you should readily see that a much more involved system of error checking is required. As a simple example you could use 3 quantum bits to represent 1 classical bit of data and use a "2 out of 3" verification.
That is what I am curious about related to this article, what is the error rate on the silicon q-bits?
ProcrastinationAccountNumber3659
1.8 / 5 (5) Oct 10, 2015Google researchers have identified one method to identify and correct errors
http://googlerese...ion.html
Spaced out Engineer
3 / 5 (2) Oct 12, 2015The trick with quantum annealing is representing the problem space so that there are local minima for the space fetched and also being capable of fetching a subspace to represent without disturbing the "game board". You are correct that without the mathematics it's complete folly.
Hilbert computing devices are on the way. As is more advanced mathematics beyond imaginable processes.