Skip to content

Commit

Permalink
Merge branch 'master' into dev
Browse files Browse the repository at this point in the history
  • Loading branch information
actions-user committed Oct 1, 2024
2 parents 2adb533 + 6deb7de commit 1059369
Show file tree
Hide file tree
Showing 11 changed files with 384 additions and 0 deletions.
Binary file added _static/authors/jorge_martinez_de_lejarza.jpg
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
4 changes: 4 additions & 0 deletions _static/authors/jorge_martinez_de_lejarza.txt
Original file line number Diff line number Diff line change
@@ -0,0 +1,4 @@
.. bio:: Jorge Martinez de Lejarza
:photo: ../_static/authors/jorge_martinez_de_lejarza.jpg

Jorge is a PhD student at IFIC-Instituto de Fisica Corpuscular, where he works on quantum algorithms for High Energy Physics. He is also a summer resident at Xanadu.
Binary file added _static/authors/serene_shum.jpg
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
4 changes: 4 additions & 0 deletions _static/authors/serene_shum.txt
Original file line number Diff line number Diff line change
@@ -0,0 +1,4 @@
.. bio:: Serene Shum
:photo: ../_static/authors/serene_shum.jpg

Serene Shum is a PhD student in the Department of Physics at the University of Toronto working on quantum simulation algorithms. She is currently a resident at Xanadu.
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
84 changes: 84 additions & 0 deletions demonstrations/tutorial_qnn_multivariate_regression.metadata.json
Original file line number Diff line number Diff line change
@@ -0,0 +1,84 @@
{
"title": "Multidimensional regression with a variational quantum circuit",
"authors": [
{
"username": "gmlejarza"
},
{
"username": "quantumpenguin"
}
],
"dateOfPublication": "2024-10-01T00:00:00+00:00",
"dateOfLastModification": "2024-10-01T00:00:00+00:00",
"categories": [
"Quantum Computing",
"Quantum Machine Learning",
"Getting Started"
],
"tags": [],
"previewImages": [
{
"type": "thumbnail",
"uri": "/_static/demo_thumbnails/regular_demo_thumbnails/thumbnail_qnn_multivariate_regression.png"
},
{
"type": "large_thumbnail",
"uri": "/_static/demo_thumbnails/large_demo_thumbnails/thumbnail_large_qnn_multivariate_regression.png"
}
],
"seoDescription": "Learn to use a quantum neural network to fit a multivariate function.",
"doi": "",
"canonicalURL": "/qml/demos/tutorial_qnn_multivariate_regression",
"references": [
{
"id": "schuld",
"type": "article",
"title": "The effect of data encoding on the expressive power of variational quantum machine learning models",
"authors": "Maria Schuld, Ryan Sweke, Johannes Jakob Meyer",
"year": "2021",
"publisher": "",
"url": "https://arxiv.org/pdf/2008.08605"
},
{
"id": "qibo",
"type": "article",
"title": "Tutorial: Quantum Fourier Iterative Amplitude Estimation",
"authors": "Jorge J. Martinez de Lejarza",
"year": "2023",
"publisher": "",
"url": "hhttps://qibo.science/qibo/stable/code-examples/tutorials/qfiae/qfiae_demo.html"
},
{
"id": "demoschuld",
"type": "other",
"title": "Tutorial: Quantum models as Fourier series",
"authors": "Johannes Jakob Meyer, Maria Schuld",
"year": "2021",
"publisher": "",
"url": "https://pennylane.ai/qml/demos/tutorial_expressivity_fourier_series/"
},
{
"id": "demojax",
"type": "other",
"title": "How to optimize a QML model using JAX and Optax",
"authors": "Josh Izaac, Maria Schuld",
"year": "2024",
"publisher": "",
"url": "ttps://pennylane.ai/qml/demos/tutorial_How_to_optimize_QML_model_using_JAX_and_Optax/"
}
],
"basedOnPapers": [],
"referencedByPapers": [],
"relatedContent": [
{
"type": "demonstration",
"id": "tutorial_How_to_optimize_QML_model_using_JAX_and_Optax",
"weight": 1.0
},
{
"type": "demonstration",
"id": "tutorial_expressivity_fourier_series",
"weight": 1.0
}
]
}
292 changes: 292 additions & 0 deletions demonstrations/tutorial_qnn_multivariate_regression.py
Original file line number Diff line number Diff line change
@@ -0,0 +1,292 @@
r"""
Multidimensional regression with a variational quantum circuit
===========================================================
In this tutorial, we show how to use a variational quantum circuit to fit the simple multivariate function
.. math:: f(x_1, x_2) = \frac{1}{2} \left( x_1^2 + x_2^2 \right).
In [#schuld]_ it has been shown that, under some conditions, there exist variational quantum circuits that are expressive enough to realize any possible set
of Fourier coefficients. We will use a simple two-qubit parameterized quantum circuit to construct a partial Fourier series for fitting
the target function.
The main outline of the process is as follows:
1. Build a circuit consisting of layers of alternating data-encoding and parameterized training blocks.
2. Optimize the expectation value of the circuit output against a target function to be fitted.
3. Obtain a partial Fourier series for the target function. Since the function is not periodic, this partial Fourier series will only approximate the function in the region we will use for training.
4. Plot the optimized circuit expectation value against the exact function to compare the two.
What is a quantum model?
------------------------
A quantum model :math:`g_{\vec{\theta}}(\vec{x})` is the expectation value of some observable :math:`M` estimated
on the state prepared by a parameterized circuit :math:`U(\vec{x}, \vec{\theta})`:
.. math:: g_{\vec{\theta}}(\vec{x}) = \langle 0 | U^\dagger (\vec{x}, \vec{\theta}) M U(\vec{x}, \vec{\theta}) | 0 \rangle.
By repeatedly running the circuit with a set of parameters :math:`\vec{\theta}` and set of data points :math:`\vec{x}`, we can
approximate the expectation value of the observable :math:`M` in the state :math:`U(\vec{x}, \vec{\theta}) | 0 \rangle.` Then, the parameters can be
optimized to minimize some loss function.
Building the variational circuit
--------------------------------
In this example, we will use a variational quantum circuit to find the Fourier series that
approximates the function :math:`f(x_1, x_2) = \frac{1}{2} \left( x_1^2 + x_2^2 \right)`. The variational circuit that we are using is made up of :math:`L` layers. Each layer consists of a *data-encoding block*
:math:`S(\vec{x})` and a *training block* :math:`W(\vec{\theta})`. The overall circuit is:
.. math:: U(\vec{x}, \vec{\theta}) = W^{(L+1)}(\vec{\theta}) S(\vec{x}) W^{(L)} (\vec{\theta}) \ldots W^{(2)}(\vec{\theta}) S(\vec{x}) W^{(1)}(\vec{\theta}).
The training blocks :math:`W(\vec{\theta})` depend on the parameters :math:`\vec{\theta}` that can be optimized classically.
.. figure:: ../_static/demonstration_assets/qnn_multivariate_regression/qnn_circuit.png
:align: center
:width: 90%
We will build a circuit such that the expectation value of the :math:`Z\otimes Z` observable is a partial Fourier series
that approximates :math:`f(\vec{x})`, i.e.,
.. math:: g_{\vec{\theta}}(\vec{x})= \sum_{\vec{\omega} \in \Omega} c_\vec{\omega} e^{i \vec{\omega} \vec{x}} \approx f(\vec{x}).
Then, we can directly plot the partial Fourier series. We can also apply a Fourier transform to
:math:`g_{\vec{\theta}}`, so we can obtain the Fourier coefficients, :math:`c_\vec{\omega}`. To know more about how to obtain the
Fourier series, check out these two related tutorials [#demoschuld]_, [#demoqibo]_.
Constructing the quantum circuit
--------------------------------
First, let's import the necessary libraries and seed the random number generator. We will use Matplotlib for plotting and JAX [#demojax]_ for optimization.
We will also define the device, which has two qubits, using :func:`~.pennylane.device`.
"""

import matplotlib.pyplot as plt
import pennylane as qml
from pennylane import numpy as pnp
import jax
from jax import numpy as jnp
import optax

pnp.random.seed(42)

dev = qml.device('default.qubit', wires=2)

######################################################################
# Now we will construct the data-encoding circuit block, :math:`S(\vec{x})`, as a product of :math:`R_z` rotations:
#
# .. math::
# S(\vec{x}) = R_z(x_1) \otimes R_z(x_2).
#
# Specifically, we define the :math:`S(\vec{x})` operator using the :class:`~.pennylane.AngleEmbedding` function.

def S(x):
qml.AngleEmbedding( x, wires=[0,1],rotation='Z')

######################################################################
# For the :math:`W(\vec{\theta})` operator, we will use an ansatz that is available in PennyLane, called :class:`~.pennylane.StronglyEntanglingLayers`.

def W(params):
qml.StronglyEntanglingLayers(params, wires=[0,1])

######################################################################
# Now we will build the circuit in PennyLane by alternating layers of :math:`W(\vec{\theta})` and :math:`S(\vec{x})` layers. On this prepared state, we estimate the expectation value of the :math:`Z\otimes Z` operator, using PennyLane's :func:`~.pennylane.expval` function.

@qml.qnode(dev,interface="jax")
def quantum_neural_network(params, x):
layers=len(params[:,0,0])-1
n_wires=len(params[0,:,0])
n_params_rot=len(params[0,0,:])
for i in range(layers):
W(params[i,:,:].reshape(1,n_wires,n_params_rot))
S(x)
W(params[-1,:,:].reshape(1,n_wires,n_params_rot))

return qml.expval(qml.PauliZ(wires=0)@qml.PauliZ(wires=1))

######################################################################
# The function we will be fitting is :math:`f(x_1, x_2) = \frac{1}{2} \left( x_1^2 + x_2^2 \right)`, which we will define as ``target_function``:

def target_function(x):
f=1/2*(x[0]**2+x[1]**2)
return f

######################################################################
# Now we will specify the range of :math:`x_1` and :math:`x_2` values and store those values in an input data vector. We are fitting the function for :math:`x_1, x_2 \in [-1, 1]` using 30 evenly spaced samples for each variable.

x1_min=-1
x1_max=1
x2_min=-1
x2_max=1
num_samples=30

######################################################################
# Now we build the training data with the exact target function :math:`f(x_1, x_2)`. To do so, it is convenient to
# create a two-dimensional grid to make sure that, for each value of
# :math:`x_1,` we perform a sweep over all the values of :math:`x_2` and viceversa.

x1_train=pnp.linspace(x1_min,x1_max, num_samples)
x2_train=pnp.linspace(x2_min,x2_max, num_samples)
x1_mesh,x2_mesh=pnp.meshgrid(x1_train, x2_train)

######################################################################
# We define ``x_train`` and ``y_train`` using the above vectors, reshaping them for our convenience
x_train=pnp.stack((x1_mesh.flatten(), x2_mesh.flatten()), axis=1)
y_train = target_function([x1_mesh,x2_mesh]).reshape(-1,1)
# Let's take a look at how they look like
print("x_train:\n", x_train[:5])
print("y_train:\n", y_train[:5])

######################################################################
# Optimizing the circuit
# ----------------------
#
# We want to optimize the circuit above so that the expectation value of :math:`Z \otimes Z`
# approximates the exact target function. This is done by minimizing the mean squared error between
# the circuit output and the exact target function. In particular, the optimization
# process to train the variational circuit will be performed using JAX, an auto differentiable machine learning framework
# to accelerate the classical optimization of the parameters. Check out [#demojax]_
# to learn more about
# how to use JAX to optimize your QML models.
#
# .. figure:: ../_static/demonstration_assets/qnn_multivariate_regression/qnn_diagram.jpg
# :align: center
# :width: 90%
#

@jax.jit
def mse(params,x,targets):
# We compute the mean square error between the target function and the quantum circuit to quantify the quality of our estimator
return (quantum_neural_network(params,x)-jnp.array(targets))**2
@jax.jit
def loss_fn(params, x,targets):
# We define the loss function to feed our optimizer
mse_pred = jax.vmap(mse,in_axes=(None, 0,0))(params,x,targets)
loss = jnp.mean(mse_pred)
return loss

#######################################################################
#Here, we are choosing an Adam optimizer with a learning rate of 0.05 and 300 steps.

opt = optax.adam(learning_rate=0.05)
max_steps=300

@jax.jit
def update_step_jit(i, args):
# We loop over this function to optimize the trainable parameters
params, opt_state, data, targets, print_training = args
loss_val, grads = jax.value_and_grad(loss_fn)(params, data, targets)
updates, opt_state = opt.update(grads, opt_state)
params = optax.apply_updates(params, updates)

def print_fn():
jax.debug.print("Step: {i} Loss: {loss_val}", i=i, loss_val=loss_val)
# if print_training=True, print the loss every 50 steps
jax.lax.cond((jnp.mod(i, 50) == 0 ) & print_training, print_fn, lambda: None)
return (params, opt_state, data, targets, print_training)

@jax.jit
def optimization_jit(params, data, targets, print_training=False):
opt_state = opt.init(params)
args = (params, opt_state, jnp.asarray(data), targets, print_training)
# We loop over update_step_jit max_steps iterations to optimize the parameters
(params, opt_state, _, _, _) = jax.lax.fori_loop(0, max_steps+1, update_step_jit, args)
return params

######################################################################
# Now we will train the variational circuit with 4 layers and obtain a vector :math:`\vec{\theta}` with the optimized parameters.

wires=2
layers=4
params_shape = qml.StronglyEntanglingLayers.shape(n_layers=layers+1,n_wires=wires)
params=pnp.random.default_rng().random(size=params_shape)
best_params=optimization_jit(params, x_train, jnp.array(y_train), print_training=True)

######################################################################
# If you run this yourself, you'll see that the training step with JAX is extremely fast!
# Once the optimized :math:`\vec{\theta}` has been obtained, we can use those parameters to build our fitted version of the function.

def evaluate(params, data):
y_pred = jax.vmap(quantum_neural_network, in_axes=(None, 0))(params, data)
return y_pred
y_predictions=evaluate(best_params,x_train)

######################################################################
# To compare the fitted function to the exact target function, let's take a look at the :math:`R^2` score:

from sklearn.metrics import r2_score
r2 = round(float(r2_score(y_train, y_predictions)),3)
print("R^2 Score:", r2)

######################################################################
# Let's now plot the results to visually check how good our fit is!

fig = plt.figure()
# Target function
ax1 = fig.add_subplot(1, 2, 1, projection='3d')
ax1.plot_surface(x1_mesh,x2_mesh, y_train.reshape(x1_mesh.shape), cmap='viridis')
ax1.set_zlim(0,1)
ax1.set_xlabel('$x$',fontsize=10)
ax1.set_ylabel('$y$',fontsize=10)
ax1.set_zlabel('$f(x,y)$',fontsize=10)
ax1.set_title('Target ')

# Predictions
ax2 = fig.add_subplot(1, 2, 2, projection='3d')
ax2.plot_surface(x1_mesh,x2_mesh, y_predictions.reshape(x1_mesh.shape), cmap='viridis')
ax2.set_zlim(0,1)
ax2.set_xlabel('$x$',fontsize=10)
ax2.set_ylabel('$y$',fontsize=10)
ax2.set_zlabel('$f(x,y)$',fontsize=10)
ax2.set_title(f' Predicted \nAccuracy: {round(r2*100,3)}%')

# Show the plot
plt.tight_layout(pad=3.7)

######################################################################
# Cool! We have managed to successfully fit a multidimensional function using a parametrized quantum circuit!

######################################################################
# Conclusions
# ------------------------------------------
# In this demo, we've shown how to utilize a variational quantum circuit to solve a regression problem for a two-dimensional function.
# The results show a good agreement with the target function and the model
# can be trained further, increasing the number of iterations in the training to maximize the accuracy. It also
# paves the way for addressing a regression problem for an :math:`N`-dimensional function, as everything presented
# here can be easily generalized. A final check that could be done is to obtain the Fourier coefficients of the
# trained circuit and compare it with the Fourier series we obtained directly from the target function.
#
# References
# ----------
#
# .. [#schuld]
#
# Maria Schuld, Ryan Sweke, and Johannes Jakob Meyer
# "The effect of data encoding on the expressive power of variational quantum machine learning models.",
# `arXiv:2008.0865 <https://arxiv.org/pdf/2008.08605>`__, 2021.
#
# .. [#demoschuld]
#
# Johannes Jakob Meyer, Maria Schuld
# “Tutorial: Quantum models as Fourier series”,
# `Pennylane: Quantum models as Fourier series <https://pennylane.ai/qml/demos/tutorial_expressivity_fourier_series/>`__, 2021.
#
# .. [#demoqibo]
#
# Jorge J. Martinez de Lejarza
# "Tutorial: Quantum Fourier Iterative Amplitude Estimation",
# `Qibo: Quantum Fourier Iterative Amplitude Estimation <https://qibo.science/qibo/stable/code-examples/tutorials/qfiae/qfiae_demo.html>`__, 2023.
# .. [#demojax]
#
# Josh Izaac, Maria Schuld
# "How to optimize a QML model using JAX and Optax",
# `Pennylane: How to optimize a QML model using JAX and Optax <https://pennylane.ai/qml/demos/tutorial_How_to_optimize_QML_model_using_JAX_and_Optax/>`__, 2024
#
# About the authors
# -----------------

0 comments on commit 1059369

Please sign in to comment.