Cookies Policy
The website need some cookies and similar means to function. If you permit us, we will use those means to collect data on your visits for aggregated statistics to improve our service. Find out More
Accept Reject
  • Menu
About

About

Luís Paulo Santos is and Assistant Professor of the Department of Informatics, Universidade do Minho and researcher of CSIG, INESC-TEC. His research area is rendering and global illumination, focusing on algorithms' performance and heterogeneous parallel computing (CPU + GPU + Knights Landing) to reduce convergence tiome towards perceptually correct solutions. He published several papers in the most relevant international fora of Computer Graphics (conferences and journals), and authored a book on Bayesian Monte Carlo Rendering. He nelongs to the Program Committee of several international conferences, having chaired a few of these and organized 6 such events in Portugal.

He has been Vicer Director of the Department, and the Informatics Engineering degree. He was the Director of the Doctoral Programme on Informatics. He integrated the Committe designted by the Rector to install an United Nations University Operational Unit on Electronic Governance in Guimarães, Portugal, and is currently a member of the direction of the unit responsible for the interface between the 2 institutions.

He is Associate Editor of the Computers & Graphics Elsevier journal and President of the Portuguese Group of Computer Graphics, formally the portuguese chapter of Eurographics, for the period of 2017-2018.

Interest
Topics
Details

Details

  • Name

    Luís Paulo Santos
  • Role

    Senior Researcher
  • Since

    01st January 2017
Publications

2024

On Quantum Natural Policy Gradients

Authors
Sequeira, A; Santos, LP; Barbosa, LS;

Publication
IEEE TRANSACTIONS ON QUANTUM ENGINEERING

Abstract
This article delves into the role of the quantum Fisher information matrix (FIM) in enhancing the performance of parameterized quantum circuit (PQC)-based reinforcement learning agents. While previous studies have highlighted the effectiveness of PQC-based policies preconditioned with the quantum FIM in contextual bandits, its impact in broader reinforcement learning contexts, such as Markov decision processes, is less clear. Through a detailed analysis of L & ouml;wner inequalities between quantum and classical FIMs, this study uncovers the nuanced distinctions and implications of using each type of FIM. Our results indicate that a PQC-based agent using the quantum FIM without additional insights typically incurs a larger approximation error and does not guarantee improved performance compared to the classical FIM. Empirical evaluations in classic control benchmarks suggest even though quantum FIM preconditioning outperforms standard gradient ascent, in general, it is not superior to classical FIM preconditioning.

2024

VQC-based reinforcement learning with data re-uploading: performance and trainability

Authors
Coelho, R; Sequeira, A; Santos, LP;

Publication
QUANTUM MACHINE INTELLIGENCE

Abstract
Reinforcement learning (RL) consists of designing agents that make intelligent decisions without human supervision. When used alongside function approximators such as Neural Networks (NNs), RL is capable of solving extremely complex problems. Deep Q-Learning, a RL algorithm that uses Deep NNs, has been shown to achieve super-human performance in game-related tasks. Nonetheless, it is also possible to use Variational Quantum Circuits (VQCs) as function approximators in RL algorithms. This work empirically studies the performance and trainability of such VQC-based Deep Q-Learning models in classic control benchmark environments. More specifically, we research how data re-uploading affects both these metrics. We show that the magnitude and the variance of the model's gradients remain substantial throughout training even as the number of qubits increases. In fact, both increase considerably in the training's early stages, when the agent needs to learn the most. They decrease later in the training, when the agent should have done most of the learning and started converging to a policy. Thus, even if the probability of being initialized in a Barren Plateau increases exponentially with system size for Hardware-Efficient ansatzes, these results indicate that the VQC-based Deep Q-Learning models may still be able to find large gradients throughout training, allowing for learning.

2023

Policy gradients using variational quantum circuits

Authors
Sequeira, A; Santos, LP; Barbosa, LS;

Publication
QUANTUM MACHINE INTELLIGENCE

Abstract
Variational quantum circuits are being used as versatile quantum machine learning models. Some empirical results exhibit an advantage in supervised and generative learning tasks. However, when applied to reinforcement learning, less is known. In this work, we considered a variational quantum circuit composed of a low-depth hardware-efficient ansatz as the parameterized policy of a reinforcement learning agent. We show that an epsilon-approximation of the policy gradient can be obtained using a logarithmic number of samples concerning the total number of parameters. We empirically verify that such quantum models behave similarly to typical classical neural networks used in standard benchmarking environments and quantum control, using only a fraction of the parameters. Moreover, we study the barren plateau phenomenon in quantum policy gradients using the Fisher information matrix spectrum.

2022

Ensemble Metropolis Light Transport

Authors
Bashford Rogers, T; Santos, LP; Marnerides, D; Debattista, K;

Publication
ACM TRANSACTIONS ON GRAPHICS

Abstract
This article proposes a Markov Chain Monte Carlo (MCMC) rendering algorithm based on a family of guided transition kernels. The kernels exploit properties of ensembles of light transport paths, which are distributed according to the lighting in the scene, and utilize this information to make informed decisions for guiding local path sampling. Critically, our approach does not require caching distributions in world space, saving time and memory, yet it is able to make guided sampling decisions based on whole paths. We show how this can be implemented efficiently by organizing the paths in each ensemble and designing transition kernels for MCMC rendering based on a carefully chosen subset of paths from the ensemble. This algorithm is easy to parallelize and leads to improvements in variance when rendering a variety of scenes.

2022

Foreword to the special section on Recent Advances in Graphics and Interaction

Authors
Rodrigues, N; Mendes, D; Santos, LP; Bouatouch, K;

Publication
COMPUTERS & GRAPHICS-UK

Abstract

Supervised
thesis

2023

Quantum Reinforcement Learning: Foundations, algorithms, applications

Author
André Manuel Resende Sequeira

Institution
UM

2023

Algoritmos e aplicações de estimativa de amplitude quântica

Author
Alexandra Francisco Ramôa da Costa Alves

Institution
UM

2023

Classification and Clustering using Swap Test as distance metric

Author
Tomás Rodrigues Alves de Sousa

Institution
UM

2023

Algoritmos de otimização quântica

Author
Mafalda Francisco Ramôa da Costa Alves

Institution
UM

2022

Quantum Reinforcement Learning: a heuristic approach to solve deterministic MDPs

Author
Renato Alberto Soares de Brito

Institution
UM