PrettyGood Tomography - PowerPoint PPT Presentation

1 / 19
About This Presentation
Title:

PrettyGood Tomography

Description:

Pretty-Good Tomography. Scott Aaronson. MIT. There's a problem... [Watrous 2000]; meant to showcase pretty-good tomography with non-commuting measurements. ... – PowerPoint PPT presentation

Number of Views:84
Avg rating:3.0/5.0
Slides: 20
Provided by: scottaa
Category:

less

Transcript and Presenter's Notes

Title: PrettyGood Tomography


1
Pretty-Good Tomography
?
  • Scott Aaronson
  • MIT

2
Theres a problem
To do tomography on an entangled state of n
qubits, we need exp(n) measurements Does this
mean that a generic state of (say) 10,000
particles can never be learned within the
lifetime of the universe? If so, this is
certainly a practical problembut to me, its a
conceptual problem as well
3
What is a quantum state?
A state of the world? A state of knowledge?
Whatever else it is, should at least be a useful
hypothesis that encapsulates previous
observations and lets us predict future ones
How useful is a hypothesis that takes 105000
bits even to write down?
Seems to bolster the arguments of quantum
computing skeptics who think quantum mechanics
will break down in the large N limit
4
Really were talking about Humes Problem of
Induction
You see 500 ravens. Every one is black. Why
does that give you any grounds whatsoever for
expecting the next raven to be black?
?
The answer, according to computational learning
theory In practice, we always restrict attention
to some class of hypotheses vastly smaller than
the class of all logically conceivable hypotheses
5
Probably Approximately Correct (PAC) Learning
Set S called the sample space Probability
distribution D over S Class C of hypotheses
functions from S to 0,1 Unknown function
f?C Goal Given x1,,xm drawn independently from
D, together with f(x1),,f(xm), output a
hypothesis h?C such that with probability at
least 1-? over x1,,xm
6
Occams Razor Theorem
But the number of quantum states is infinite!
Valiant 1984 If the hypothesis class C is
finite, then any hypothesis consistent with
And even if we discretize, its still doubly
exponential in the number of qubits!
random samples will also be consistent with a 1-?
fraction of future data, with probability at
least 1-? over the choice of samples
Compression implies prediction
7
A Hint of Whats Possible
Theorem A. 2004 Any n-qubit quantum state can
be simulated using O(n log n log m) classical
bits, where m is the number of (binary)
measurements whose outcomes we care about.
Let E(E1,,Em) be two-outcome POVMs on an
n-qubit state ?. Then given (classical
descriptions of) E and ?, we can produce a
classical string of
bits, from which Tr(Ei?) can be estimated to
within additive error ? given any Ei (without
knowing ?).
8
Quantum Occams Razor TheoremA. 2006
Let ? be an n-qubit state, and let D be a
distribution over two-outcome measurements. Suppos
e we draw measurements E1,,Em independently from
D, and then find a hypothesis state ? that
minimizes
(bi outcome of Ei)
Then
with probability at least 1-? over E1,,Em,
provided
(C a constant)
9
Beyond the Bayesian and Max-Lik creeds a third
way?
Were not assuming any prior over states Removes
a lot of problems! Instead we assume a
distribution over measurements Why might that be
preferable for some applications? We can control
which measurements to apply, but not what the
state is
10
Extension to process tomography?
No! Suppose Ux?(-1)f(x)x?, for some random
Boolean function f0,1n?0,1 Then the values
of f(x) constitute 2n independently accessible
bits to be learned about Yet each measurement
provides at most n of the bits Hence, no analogue
of my learning theorem is possible
11
Extension to k-outcome measurements?
Sure, if we increase the number of sample
measurements m by a poly(k) factor Note that
theres no hope of learning to simulate
2n-outcome measurements (i.e. measurements on all
n qubits) after poly(n) sample measurements
12
How do we actually find ??
This is a convex programming problem, which can
be solved in time polynomial in the Hilbert space
dimension N2n In general, we cant hope for
better than thisfor basic computational
complexity reasons
13
Custom Convex Programming MethodE. Hazan, 2008
Let
Set S0 I/N For t0 to ? Compute smallest
eigenvector vt of ?f(St) Compute step size ?t
that minimizes f(St?t(vtvt-St)) Set St1 St
?t(vtvt-St)
Theorem (Hazan) This algorithm returns an
?-optimal solution after only log(m)/?2
iterations.
14
ImplementationA. Dechter 2008
We implemented Hazans algorithm in MATLAB Code
available on request Using MITs computing
cluster, we then did numerical simulations to
check experimentally that the learning theorem is
true
15
Experiments We Ran
1. Classical States (sanity check). States have
form ?x??x, measurements check if ith bit is 1
or 0, distribution over measurements is
uniform. 2. Linear Cluster States. States are n
qubits, prepared by starting with ??n and then
applying conditional phase (Pxy?(-1)xyxy?) to
each neighboring pair. Measurements check three
randomly-chosen neighboring qubits, in a basis
like 0??0?,1??1?,0?-?1?. Acceptance
probability is always ¾.
16
3. Z2n Subgroup States. Let H be a subgroup of
GZ2n of order 2n-1. States ?H??H are equal
superpositions over H. Theres a measurement Eg
for each element g?G, which checks whether g?H
where Ugh?gh? for all h?G. Eg accepts with
probability 1 if g?H, or ½ if g?H.
Inspired by Watrous 2000 meant to showcase
pretty-good tomography with non-commuting
measurements.
17
(No Transcript)
18
(No Transcript)
19
Open Problems
Find more convincing applications of our learning
theorem Find special classes of states for which
learning can be done using computation time
polynomial in the number of qubits Improve the
parameters of the learning theorem Experimental
demonstration!
Write a Comment
User Comments (0)
About PowerShow.com