Simon's problem explained

In computational complexity theory and quantum computing, Simon's problem is a computational problem that is proven to be solved exponentially faster on a quantum computer than on a classical (that is, traditional) computer. The quantum algorithm solving Simon's problem, usually called Simon's algorithm, served as the inspiration for Shor's algorithm.[1] Both problems are special cases of the abelian hidden subgroup problem, which is now known to have efficient quantum algorithms.

The problem is set in the model of decision tree complexity or query complexity and was conceived by Daniel R. Simon in 1994.[2] Simon exhibited a quantum algorithm that solves Simon's problem exponentially faster with exponentially fewer queries than the best probabilistic (or deterministic) classical algorithm. In particular, Simon's algorithm uses a linear number of queries and any classical probabilistic algorithm must use an exponential number of queries.

This problem yields an oracle separation between the complexity classes BPP (bounded-error classical query complexity) and BQP (bounded-error quantum query complexity).[3] This is the same separation that the Bernstein–Vazirani algorithm achieves, and different from the separation provided by the Deutsch–Jozsa algorithm, which separates P and EQP. Unlike the Bernstein–Vazirani algorithm, Simon's algorithm's separation is exponential.

Because this problem assumes the existence of a highly-structured "black box" oracle to achieve its speedup, this problem has little practical value.[4] However, without such an oracle, exponential speedups cannot easily be proven, since this would prove that P is different from PSPACE.

Problem description

Given a function (implemented by a black box or oracle)

f:\{0,1\}n\{0,1\}n

with the promise that, for some unknown

s\in\{0,1\}n

, for all

x,y\in\{0,1\}n

,

f(x)=f(y)

if and only if

xy\in\{0n,s\}

,

where

denotes bitwise XOR. The goal is to identify

s

by making as few queries to

f(x)

as possible. Note that

ab=0n

if and only if

a=b

Furthermore, for some

x

and

s

in

xy=s

,

y

is unique (not equal to

x

) if and only if

s0n

. This means that

f

is two-to-one when

s0n

, and one-to-one when

s=0n

. It is also the case that

xy=s

implies

y=sx

, meaning thatf(x) = f(y) = f(x \oplus)which shows how

f

is periodic.

The associated decision problem formulation of Simon's problem is to distinguish when

s=0n

(

f

is one-to-one), and when

s0n

(

f

is two-to-one).

Example

The following function is an example of a function that satisfies the required property for

n=3

:

x

f(x)

000101
001010
010000
011110
100000
101110
110101
111010

In this case,

s=110

(i.e. the solution). Every output of

f

occurs twice, and the two input strings corresponding to any one given output have bitwise XOR equal to

s=110

.

For example, the input strings

010

and

100

are both mapped (by

f

) to the same output string

000

. That is,

{\displaystylef(010)=000}

and

{\displaystylef(100)=000}

. Applying XOR to 010 and 100 obtains 110, that is

{\displaystyle010 ⊕ 100=110=s}.

s=110

can also be verified using input strings 001 and 111 that are both mapped (by f) to the same output string 010. Applying XOR to 001 and 111 obtains 110, that is

001 ⊕ 111=110=s

. This gives the same solution

s=110

as before.

In this example the function f is indeed a two-to-one function where

{\displaystyles0n

}.

Problem hardness

Intuitively, this is a hard problem to solve in a "classical" way, even if one uses randomness and accepts a small probability of error. The intuition behind the hardness is reasonably simple: if you want to solve the problem classically, you need to find two different inputs

x

and

y

for which

f(x)=f(y)

. There is not necessarily any structure in the function

f

that would help us to find two such inputs: more specifically, we can discover something about

f

(or what it does) only when, for two different inputs, we obtain the same output. In any case, we would need to guess

{\displaystyle\Omega({\sqrt{2n

}})} different inputs before being likely to find a pair on which

f

takes the same output, as per the birthday problem. Since, classically to find s with a 100% certainty it would require checking

{\displaystyle\Theta({\sqrt{2n

}})} inputs, Simon's problem seeks to find s using fewer queries than this classical method.

Simon's algorithm

The algorithm as a whole uses a subroutine to execute the following two steps:

  1. Run the quantum subroutine an expected

O(n)

times to get a list of linearly independent bitstrings

y1,...,yn

.
  1. Each

yk

satisfies

yks=0

, so we can solve the system of equations this produces to get

s

.

Quantum subroutine

The quantum circuit (see the picture) is the implementation of the quantum part of Simon's algorithm. The quantum subroutine of the algorithm makes use of the Hadamard transformH^|k\rangle = \frac \sum_^ (-1)^ |j\ranglewhere

kj=k1j1\ldotsknjn

, where

denotes XOR.

First, the algorithm starts with two registers, initialized to

|0\rangle|0\rangle

. Then, we apply the Hadamard transform to the first register, which gives the state
1
\sqrt{2n
} \sum_^ |k\rangle |0\rangle^.

Query the oracle

Uf

to get the state
1
\sqrt{2n
} \sum_^ |k\rangle |f(k)\rangle.

Apply another Hadamard transform to the first register. This will produce the state

1
\sqrt{2n
} \sum_^ \left[\frac{1}{\sqrt{2^n}} \sum_{j = 0}^{2^n - 1} (-1)^{j \cdot k} |j\rangle \right] |f(k)\rangle= \sum_^ |j\rangle \left[\frac{1}{2^n} \sum_{k = 0}^{2^n - 1} (-1)^{j \cdot k} |f(k)\rangle \right].

Finally, we measure the first register (the algorithm also works if the second register is measured before the first, but this is unnecessary). The probability of measuring a state

|j\rangle

is\left|\left| \frac \sum_^ (-1)^ |f(k)\rangle \right|\right|^2This is due to the fact that taking the magnitude of this vector and squaring it sums up all the probabilities of all the possible measurements of the second register that must have the first register as

|j\rangle

. There are two cases for our measurement:

s=0n

and

f

is one-to-one.

s0n

and

f

is two-to-one.

For the first case, \left|\left| \frac \sum_^ (-1)^ |f(k)\rangle \right|\right|^2 = \fracsince in this case,

f

is one-to-one, implying that the range of

f

is

\{0,1\}n

, meaning that the summation is over every basis vector. For the second case, note that there exist two strings,

x1

and

x2

, such that

f(x1)=f(x2)=z

, where

z\inrange(f)

. Thus, \left|\left| \frac \sum_^ (-1)^ |f(k)\rangle \right|\right|^2= \left|\left| \frac \sum_ ((-1)^ + (-1)^) |z\rangle \right|\right|^2Furthermore, since

x1x2=s

,

x2=x1s

, and so\begin\left|\left| \frac \sum_ ((-1)^ + (-1)^) |z\rangle \right|\right|^2&= \left|\left| \frac \sum_ ((-1)^ + (-1)^) |z\rangle \right|\right|^2 \\&= \left|\left| \frac \sum_ ((-1)^ + (-1)^) |z\rangle \right|\right|^2 \\&= \left|\left| \frac \sum_ (-1)^(1 + (-1)^) |z\rangle \right|\right|^2\endThis expression is now easy to evaluate. Recall that we are measuring

j

. When

js=1

, then this expression will evaluate to

0

, and when

js=0

, then this expression will be

2-n

.

Thus, both when

s=0n

and when

s0n

, our measured

j

satisfies

js=0

.

Classical post-processing

We run the quantum part of the algorithm until we have a linearly independent list of bitstrings

y1,\ldots,yn

, and each

yk

satisfies

yks=0

. Thus, we can efficiently solve this system of equations classically to find

s

.

The probability that

y1,y2,...,yn-1

are linearly independent is at least\prod_^\infty \left(1 - \frac \right) = 0.288788\dotsOnce we solve the system of equations, and produce a solution

s'

, we can test if

f(0n)=f(s')

. If this is true, then we know

s'=s

, since

f(0n)=f(0ns)=f(s)

. If it is the case that

f(0n)f(s')

, then that means that

s=0n

, and

f(0n)f(s')

since

f

is one-to-one.

We can repeat Simon's algorithm a constant number of times to increase the probability of success arbitrarily, while still having the same time complexity.

Explicit examples of Simon's algorithm for few qubits

One qubit

Consider the simplest instance of the algorithm, with

n=1

. In this case evolving the input state through an Hadamard gate and the oracle results in the state (up to renormalization):

|0\rangle|f(0)\rangle+|1\rangle|f(1)\rangle.

If

s=1

, that is,

f(0)=f(1)

, then measuring the second register always gives the outcome

|f(0)\rangle

, and always results in the first register collapsing to the state (up to renormalization):

|0\rangle+|1\rangle.

Thus applying an Hadamard and measuring the first register always gives the outcome

|0\rangle

. On the other hand, if

f

is one-to-one, that is,

s=0

, then measuring the first register after the second Hadamard can result in both

|0\rangle

and

|1\rangle

, with equal probability.

We recover

s

from the measurement outcomes by looking at whether we measured always

|0\rangle

, in which case

s=1

, or we measured both

|0\rangle

and

|1\rangle

with equal probability, in which case we infer that

s=0

. This scheme will fail if

s=0

but we nonetheless always found the outcome

|0\rangle

, but the probability of this event is

2-N

with

N

the number of performed measurements, and can thus be made exponentially small by increasing the statistics.

Two qubits

Consider now the case with

n=2

. The initial part of the algorithm results in the state (up to renormalization):|00\rangle|f(00)\rangle +|01\rangle|f(01)\rangle +|10\rangle|f(10)\rangle +|11\rangle|f(11)\rangle.If

s=(00)

, meaning

f

is injective, then finding

|f(x)\rangle

on the second register always collapses the first register to

|x\rangle

, for all

x\in\{0,1\}2

. In other words, applying Hadamard gates and measuring the first register the four outcomes

00,01,10,11

are thus found with equal probability.

Suppose on the other hand

s(00)

, for example,

s=(01)

. Then measuring

|f(00)\rangle

on the second register collapses the first register to the state

|00\rangle+|10\rangle

. And more generally, measuring

|f(xy)\rangle

gives

|x,y\rangle+|x,y ⊕ 1\rangle=|x\rangle(|0\rangle+|1\rangle)

on the first register. Applying Hadamard gates and measuring on the first register can thus result in the outcomes

00

and

10

with equal probabilities.

Similar reasoning applies to the other cases: if

s=(10)

then the possible outcomes are

00

and

01

, while if

s=(11)

the possible outcomes are

00

and

11

, compatibly with the

js=0

rule discussed in the general case.

To recover

s

we thus only need to distinguish between these four cases, collecting enough statistics to ensure that the probability of mistaking one outcome probability distribution for another is sufficiently small.

Complexity

Simon's algorithm requires

O(n)

queries to the black box, whereas a classical algorithm would need at least

\Omega(2n/2)

queries. It is also known that Simon's algorithm is optimal in the sense that any quantum algorithm to solve this problem requires

\Omega(n)

queries.

See also

Notes and References

  1. Shor. Peter W.. 1999-01-01. Polynomial-Time Algorithms for Prime Factorization and Discrete Logarithms on a Quantum Computer. SIAM Review. 41. 2. 303–332. 10.1137/S0036144598347011. 0036-1445. quant-ph/9508027.
  2. Simon. Daniel R.. 1997-10-01. On the Power of Quantum Computation. SIAM Journal on Computing. 26. 5. 1474–1483. 10.1137/S0097539796298637. 0097-5397.
  3. Book: Preskill, John. Lecture Notes for Physics 229: Quantum Information and Computation. 1998. 273-275.
  4. Book: Aaronson, Scott. Introduction to Quantum Information Science Lecture Notes. 2018. 144-151.