# Defining Quantum Computing

The next big thing in computing is very small. Professor Michelle Simmons explains quantum computing, a field so complex that one pundit says you never fully understand - you just learn to live with it.

**Quantum Computer**

A quantum computer is a device for computation that makes direct use of *quantum mechanical phenomena*, such as superposition and entanglement, to perform operations on data. Quantum computers are different from digital computers based on *transistors*. Whereas digital computers require data to be encoded into binary digits (*bits*), quantum computation utilizes quantum properties to represent data and perform operations on these data. A theoretical model is the quantum Turing machine, also known as the universal quantum computer. Quantum computers share theoretical similarities with *non-deterministic and probabilistic computers*, like the ability to be in more than one state simultaneously. The field of quantum computing was first introduced by Richard Feynman in 1982.

Although quantum computing is still in its infancy, experiments have been carried out in which quantum computational operations were executed on a very small number of *qubits (quantum bits)*. Both practical and theoretical research continues, and many national government and military funding agencies support quantum computing research to develop quantum computers for both civilian and national security purposes, such as cryptanalysis.

Large-scale quantum computers could be able to solve certain problems much faster than any classical computer by using the best currently known algorithms, like integer factorization using Shor’s algorithm or the simulation of *quantum many-body systems*. There exist *quantum algorithms*, such as Simon’s algorithm, which run faster than any possible probabilistic classical algorithm. Given unlimited resources, a classical computer can simulate an arbitrary quantum algorithm so quantum computation does not violate the *Church–Turing thesis*. However, in practice infinite resources are never available and the computational basis of 500 qubits, for example, would already be too large to be represented on a classical computer because it would require 2500 complex values to be stored. (For comparison, a terabyte of digital information stores only 243 discrete on/off values) Nielsen and Chuang point out that “*Trying to store all these complex numbers would not be possible on any conceivable classical computer*.”