Machine Intelligence and Quantum Information Processing - quinfog

8 downloads 269 Views 5MB Size Report
Jan 13, 2017 - Aristotle wrote in his Posterior Analytics. (4th century B.C.) that “We may ... competing hypotheses, t
Machine Intelligence and Quantum Information Processing Peter Wittek ICFO-The Institute of Photonic Sciences and University of Bor˚ as

January 2017

AI and Machine Learning

Quantum-Enhanced Learning

Learning in QIP

Towards Quantum AI

Summary

Directions •Quantum algorithms in machine learning •Improved sample and computational complexity

Machine learning and artificial intelligence

Quantum information processing and quantum computing

•Reinforcement learning in control problems •Deep learning and neural networks as representation

Peter Wittek

Machine Intelligence and Quantum Information Processing

AI and Machine Learning

Quantum-Enhanced Learning

Learning in QIP

Towards Quantum AI

Summary

How do AI and machine learning relate?

Source: https://blogs.nvidia.com/blog/2016/07/29/ whats-difference-artificial-intelligence-machine-learning-deep-learning-ai/ Peter Wittek

Machine Intelligence and Quantum Information Processing

AI and Machine Learning

Quantum-Enhanced Learning

Learning in QIP

Towards Quantum AI

Summary

What makes a good theory in the first place? Aristotle wrote in his Posterior Analytics (4th century B.C.) that “We may assume the superiority ceteris paribus [other things being equal] of the demonstration which derives from fewer postulates or hypotheses.” Occam’s razor (1300s): “Among competing hypotheses, the one with the fewest assumptions should be selected.”

Peter Wittek

Machine Intelligence and Quantum Information Processing

AI and Machine Learning

Quantum-Enhanced Learning

Learning in QIP

Towards Quantum AI

Summary

Major schools of thought in machine learning and AI Connectionism: neural networks and empirical risk minimization. Structural risk minimization: support vector machines. Probabilistic inference: Bayesian and Markov networks. Symbolic AI: first-order logic or other form of formal reasoning.

Pneumonia

Class 1 Class 2 Decision surface Margin

Peter Wittek

Tuberculosis

Lung I nf iltrates XRay

Sputum Smear

Machine Intelligence and Quantum Information Processing

AI and Machine Learning

Quantum-Enhanced Learning

Learning in QIP

Towards Quantum AI

Summary

Statistics and sample complexity Descriptive and inferential statistics. Assumptions derive from probability theory. Parameters enter through assumed probability distributions. It is often assumed that the data is generated by certain probability distributions described by a finite number of unknown parameters.

Law of large numbers. Concentration theorems.

Independent and identically distributed (IID) random variables. We can establish guarantees on accuracy based on the sample size. Example: metrology. Standard quantum limit: scales as 1/N. Heisenberg limit: scales as 1/N 2 . Peter Wittek

Machine Intelligence and Quantum Information Processing

AI and Machine Learning

Quantum-Enhanced Learning

Learning in QIP

Towards Quantum AI

Summary

Symbolic AI and Entailment Knowledge base (KB) in first-order logic: Every referee is competent: ∀x,y (Referees(x,y)⇒Competent(x)) Referees of physicists are physicists: ∀x,y (Referees(x,y)∧Physicist(y)⇒Physicist(x)) Let us have a formula from outside the KB: F: Bob referees Alice. Bob is not competent. Referees(Bob, Alice)∧¬Competent(Bob) Problem of entailment: KBF. What about contradictions? Combinatorial explosion and uncertainty. What is computable? What is the computational complexity of entailment and inference? Optimization: interesting problems tend to be NP-hard. Peter Wittek

Machine Intelligence and Quantum Information Processing

AI and Machine Learning

Quantum-Enhanced Learning

Learning in QIP

Towards Quantum AI

Summary

A Measure of Model Complexity: Rademacher Complexity Empirical Rademacher complexity of a set H of functions mapping X 7→ [a, b] ⊂ R, with a sample S = {X1 , . . . , XN }, where Xi ∼ X : # " m X 2 b S (H) = E sup σi h(Xi ) R N h∈H i=1

where σ1 , σ2 , . . . , σN are independent random variables such that P(σi = +1) = P(σi = −1) = 1/2 for i = 1, 2, . . . , N. Intuition: capacity of the learner, including ability to pick up random noise. Rademacher complexity: h i b S (H) Average over possible samples: RN (H) = ES R

Peter Wittek

Machine Intelligence and Quantum Information Processing

AI and Machine Learning

Quantum-Enhanced Learning

Learning in QIP

Towards Quantum AI

Summary

The essence of statistical learning theory Elements of balance: Sample complexity. Computational complexity. Model complexity. Generalization bounds E (h) ≤ ES (h) + F (model complexity of function family H, N) H is the function family we optimize over. F goes to zero as N → ∞. IID assumption. E (h) is the error of the learned function h over the whole distribution given the sample. ES is the error on the sample, the training error, the empirical risk. Peter Wittek

Machine Intelligence and Quantum Information Processing

AI and Machine Learning

Quantum-Enhanced Learning

Learning in QIP

Towards Quantum AI

Summary

What’s an easy problem? What’s a hard one? From easier to harder: Supervised learning: this is where deep learning shines.

Structure data.

Semi-supervised learning.

Unstructured data: text, images, video, LHC collisions. . .

Unsupervised learning. Generative models.

Semistructure data.

Reinforcement learning. the foundations of quantum physics are fermionic data processes that complete convergence of a result of the first environment. we finally articousl define a number of the contriting contraction of the constrained for any classical algorithm for quantum. . . Peter Wittek

Machine Intelligence and Quantum Information Processing

AI and Machine Learning

Quantum-Enhanced Learning

Learning in QIP

Towards Quantum AI

Summary

How to get started with machine learning? By doing it: The lingua franca is Python. Sebastian Raschka: Python Machine Learning, along with its git repository. Learn a deep learning framework. Perhaps Keras is the easiest to pick up today (13/01/2017). Sign up to Kaggle and start running kernels. E.g., the quant-ph dataset.

Peter Wittek

Machine Intelligence and Quantum Information Processing

AI and Machine Learning

Quantum-Enhanced Learning

Learning in QIP

Towards Quantum AI

Summary

Some synergies machine learning

quantum machine learning

quantum information processing

annealing quantum annealing

simulated annealing

markov chain monte-carlo

quantum BM

feed forward neural net

quantum perceptron

neural nets

quantum gibbs sampling

quantum topological algorithms

quantum PCA quantum SVM quantum NN classification quantum clustering quantum data fitting

Quantum ODE solvers

quantum rejection sampling / HHL control and metrology reinforcement learning

quantum control phase estimation hamiltonian learning

Peter Wittek

tomography

Machine Intelligence and Quantum Information Processing

AI and Machine Learning

Quantum-Enhanced Learning

Learning in QIP

Towards Quantum AI

Summary

Quantum Data

Exponential speedup: Quantum principal component analysis and quantum singular value decomposition. Quantum support vector machines: sparsity lost. Topological data analysis. Core ideas: Quantum matrix inversion is fast: HHL algorithm. Simulation of sparse matrixes is efficient. Non-sparse density matrices reveal the eigenstructure exponentially faster than in classical algorithms.

Peter Wittek

Machine Intelligence and Quantum Information Processing

AI and Machine Learning

Quantum-Enhanced Learning

Learning in QIP

Towards Quantum AI

Summary

Other gate-based models

Grover’s search as a building block: Discrete search space. Quantum associative memories. Quantum deep learning: Prepare-and-measure model for deep belief networks.

Peter Wittek

Machine Intelligence and Quantum Information Processing

AI and Machine Learning

Quantum-Enhanced Learning

Learning in QIP

Towards Quantum AI

Summary

How about regression? Learning unitaries: Given N disposals of a black-box unitary, reproduce its effect K times. Inductive or transductive strategy? Counter-intuitively, inductive is optimal.

Unlabeled instances Class 1 Class 2

Peter Wittek

Machine Intelligence and Quantum Information Processing

AI and Machine Learning

Quantum-Enhanced Learning

Learning in QIP

Towards Quantum AI

Summary

Quantum annealing and Gibbs sampling thermal annealing

quantum annealing

A

B

thermal state

Solve a nonconvex optimization directly: optimal sparsity of model. Generate samples of a Gibbs distribution: Boltzmann machines. Peter Wittek

Machine Intelligence and Quantum Information Processing

AI and Machine Learning

Quantum-Enhanced Learning

Learning in QIP

Towards Quantum AI

Summary

Tensor networks and deep learning

Learn phases transitions. arXiv:1605.01735 ’Solve’ the many-body problem. arXiv:1606.02318 Peter Wittek

Machine Intelligence and Quantum Information Processing

AI and Machine Learning

Quantum-Enhanced Learning

Learning in QIP

Towards Quantum AI

Summary

Foundational questions What does ’induction’ mean? Training and application phases. Translates to a non-signalling condition. Quantum de Finetti theorem for channels: asymptotic equivalence between classical and quantum cases. arXiv:1605.07541 Is there anything we can learn with quantum resources that we cannot with classical ones? A class of problems can be constructed. . . Complexity measures? Polynomial relationship between number of queries for classical learning problems. Sample complexity: constant factor difference in PAC setting. Peter Wittek

Machine Intelligence and Quantum Information Processing

AI and Machine Learning

Quantum-Enhanced Learning

Learning in QIP

Towards Quantum AI

Summary

Neural Networks Implementations Towards a Feasible Implementation of Quantum Neural Networks Using Quantum Dots (2015). DOI:10.1063/1.4943622 Photonic Delay Systems as Machine Learning Implementations (2015). A Coherent Perceptron for All-Optical Learning (2015). DOI:10.1140/epjqt/s40507-015-0023-3 Adiabatic Quantum Optimization for Associative Memory Recall (2014). DOI:10.3389/fphy.2014.00079 Quantum pattern recognition with liquid-state nuclear magnetic resonance (2009). DOI:10.1103/PhysRevA.79.042321 Associative memory based image and object recognition by quantum holography (2004). DOI:10.1023/B:OPSY.0000047571.17774.8d Peter Wittek

Machine Intelligence and Quantum Information Processing

AI and Machine Learning

Quantum-Enhanced Learning

Learning in QIP

Towards Quantum AI

Summary

Implementations of Not Neural Networks

Entanglement-Based Quantum Machine Learning (2014). K-means clustering, eight dimensions, photonic implementation. DOI:10.1103/physrevlett.114.110504 Experimental Realization of a Quantum Support Vector Machine (2014). Four-qubit NMR. DOI:10.1103/physrevlett.114.140504 On the Robustness of Bucket Brigade Quantum RAM (2015). DOI:10.1088/1367-2630/17/12/123010 And a suggestion with continuous-variable quantum optics: claimed exponential speedup. arXiv:1603.06222

Peter Wittek

Machine Intelligence and Quantum Information Processing

AI and Machine Learning

Quantum-Enhanced Learning

Learning in QIP

Towards Quantum AI

Summary

Reinforcement and Supervised Learning in Quantum Control Reinforcement learning Why not use it for controlling the classical part of some quantum process? Adaptive quantum phase estimation (2010). Quantum computations with an agent (2014). Producing Bose-Einstein condensates (2015). How about quantum agents? (2015) Supervised learning Learning mapping gate-based models to spin systems (2015-2016) Recurrent neural networks in dynamic decoupling (2016). Peter Wittek

Machine Intelligence and Quantum Information Processing

AI and Machine Learning

Quantum-Enhanced Learning

Learning in QIP

Towards Quantum AI

Summary

Reinforcement Learning in Adaptive Quantum Phase Estimation

Policy search in the processing unit (PU). Nonconvex objective function: Particle-swarm optimization. Differential evolution.

O(n7 ) computational complexity. Peter Wittek

Machine Intelligence and Quantum Information Processing

AI and Machine Learning

Quantum-Enhanced Learning

Learning in QIP

Towards Quantum AI

Summary

Results

0 SQL DE σ=0,η=0 PSO σ=0.2,η=0 DE σ=0.2,η=0 HL DE σ=0.2, η=0.2 Hill Climbing σ=0.0, η=0.0

−0.5

logVH

−1

−1.5

−2

−2.5 4

10

25

50

100

N

Peter Wittek

Machine Intelligence and Quantum Information Processing

AI and Machine Learning

Quantum-Enhanced Learning

Learning in QIP

Towards Quantum AI

Summary

Quantum Gate Learning

Alternative to gate-based quantum computing: pairwise always-on interactions between qubits. Computation by combining natural couplings with moderate external control. Can we do it without external modulation? “Computation by waiting.” Task: implemented a target unitary gate U acting on a set of qubits Q with a set A auxiliary qubits. exp(ıtHQA ) = U ⊗ VA after some time t.

Peter Wittek

Machine Intelligence and Quantum Information Processing

AI and Machine Learning

Quantum-Enhanced Learning

Learning in QIP

Towards Quantum AI

Summary

Quantum Gate Learning

Idea: coupling weights are just like a feedforward network. Algorithm: 1

Choose an initial parameter set λ at random; choose an initial learning rate ;

2

Repeat

3

Generate a random ψi;

4

Update the coupling strengths as λ 7→ λ + ∇λ hψ|U † ελ [|ψihψ|]U|ψi;

5

decrease ;

6

until convergence.

Peter Wittek

Machine Intelligence and Quantum Information Processing

AI and Machine Learning

Quantum-Enhanced Learning

Learning in QIP

Towards Quantum AI

Summary

Probabilistic Graphical Models Uncertainty (probabilities) and logical structure (independence constraints). Goal: compact representation of a joint probability distribution. For {X1 , . . . , XN } binary random variables, there are 2N assignments.

Complexity is dealt through graph theory. Factorization: compactness. Inference: reassembling factors. Conditional independence (X ⊥ Y |W ): P(X = x, Y = y |W = w ) = P(X = x|W = w )P(Y = y |W = w ) ∀x ∈ X , y ∈ Y , w ∈ W Peter Wittek

Machine Intelligence and Quantum Information Processing

AI and Machine Learning

Quantum-Enhanced Learning

Learning in QIP

Towards Quantum AI

Summary

Markov Networks Aka Markov random fields Direction may not be important. E.g., computer vision. See also pattern recognition.

D: set of random variables. Ideally a clique in the graph. Factor: π : Val(D) 7→ R+ A distribution factorizes over G if: P(X1 , . . . , XN ) = Z1 P 0 (X1 , . . . , XN ), where P 0 (X1 , . . . , XN ) = π1 [C1 ] × . . . × πk [Ck ] and Ci is a clique in G .

If all probabilities are Ppositive: P(X = x) = Z1 exp( k wj gj (x)). Connection to Boltzmann machines Peter Wittek

Machine Intelligence and Quantum Information Processing

AI and Machine Learning

Quantum-Enhanced Learning

Learning in QIP

Towards Quantum AI

Summary

Probabilistic Inference

How to apply the learned model? NP-hard, in fact, it is in #P. Contrast this to the cost of applying a model in other machine learning models.

Two types of queries: ,e) Conditional probability: P(Y |E = e) = P(Y P(e) . Maximum a posteriori: P argmaxy P(y |e) = argmaxy W P(y , W |e).

Inference by Markov chain Monte Carlo Gibbs sampling.

Peter Wittek

Machine Intelligence and Quantum Information Processing

AI and Machine Learning

Quantum-Enhanced Learning

Learning in QIP

Towards Quantum AI

Summary

Markov Logic Networks Real world can never match a KB. Weight each formula in a KB: high weight indicates high probability. Markov Logic Network Apply a KB {Fi } with matching weights {wi } to a finite set of constants C to define a Markov network: Add a binary node for each possible grounding for each atom. Add a binary feature for each possible grounding of each formula Fi . It is like a template to generate Markov networks.

Peter Wittek

Machine Intelligence and Quantum Information Processing

AI and Machine Learning

Quantum-Enhanced Learning

Learning in QIP

Towards Quantum AI

Summary

Prepare a Thermal State

Best known state preparation protocol (arXiv:1603.02940): p p O( d n β/Z polylog( d n β/Z /))

Where: n is number of subsystems and d local Hilbert space dimension. The good: Huge improvement over ∼ d n / runtime of classical simulated annealing. The bad: Still exponential in n. The ugly: Requires essentially a universal quantum computer.

Peter Wittek

Machine Intelligence and Quantum Information Processing

AI and Machine Learning

Quantum-Enhanced Learning

Learning in QIP

Towards Quantum AI

Summary

Can we hope for something better?

Not really. . . Finding and sampling from thermal states of (quantum) many-body problems is usually hard even though they have much more structure: Very small cliques Sites on a simple lattice Cliques geometrically local

For ground states we even have rigorous hardness results: Classical 3-SAT is NP-complete. Estimating quantum ground state energy is QMA-hard.

Peter Wittek

Machine Intelligence and Quantum Information Processing

AI and Machine Learning

Quantum-Enhanced Learning

Learning in QIP

Towards Quantum AI

Summary

Open Questions: Long Term

Strong AI Deep learning has its limits. What enabled its success? Generalization performance and computational complexity will play a role. ML/AI as an enabling technology in universal quantum computing See measurement-based and spin-based models. . . Reinforcement learning crops up in quantum metrology, error correction. . .

Peter Wittek

Machine Intelligence and Quantum Information Processing

AI and Machine Learning

Quantum-Enhanced Learning

Learning in QIP

Towards Quantum AI

Summary

Open Questions: Mid-Term

Quantum annealing and interesting learning/AI models. Other experimentally relevant learning protocols. Structural risk minimization and quantum resources. More contemporary approach to reinforcement learning in quantum control. Quantum learning in other quantum protocols. Establish mutual foundations of deep learning and tensor networks.

Peter Wittek

Machine Intelligence and Quantum Information Processing