Cookies Policy
The website need some cookies and similar means to function. If you permit us, we will use those means to collect data on your visits for aggregated statistics to improve our service. Find out More
Accept Reject
  • Menu
About

About

Miguel Guimarães completed his Bachelor's and Master's Degrees in Computer Science in 2022 at the School of Management and Technology, of the Polytechnic of Porto. His master's work was developed in the field of Machine Learning (ML). Specifically, he addresses the issue of explainability in ML and explores meta-learning as a way to address some of the current challenges of ML, namely with streaming data. Currently, he is an Invited Assistant Professor at the same institution. He has a strong passion for research, having participated as a Research Fellow in 2 projects. Miguel is the author of several publications in the fields of machine learning and hybrid systems. Published 5 articles in journals, 6 book chapters, 3 conference papers and received 2 best paper awards.

Also, he won a research grant at INESC TEC to work on the development and application of artificial intelligence techniques in the industrial environment, thus gaining knowledge in a work context, reconciling academic theory with practical experience. Recently, he has been investigating the application and drawbacks of Generative AI (GAI) models in the industry.

Specifically, he focuses in having a human-centric vision and organization context, incorporating socio-technical factors into the GAI lifecycle, in order to adapt the model's output to the specific context. He hopes this will encourage a wider adoption of GAI like Large Language Models (LLMs) in industrial environments.

Interest
Topics
Details

Details

  • Name

    Miguel Ângelo Guimarães
  • Role

    Research Assistant
  • Since

    01st June 2023
Publications

2024

Supervised and unsupervised techniques in textile quality inspections

Authors
Ferreira, HM; Carneiro, DR; Guimaraes, MA; Oliveira, FV;

Publication
5TH INTERNATIONAL CONFERENCE ON INDUSTRY 4.0 AND SMART MANUFACTURING, ISM 2023

Abstract
Quality inspection is a critical step in ensuring the quality and efficiency of textile production processes. With the increasing complexity and scale of modern textile manufacturing systems, the need for accurate and efficient quality inspection and defect detection techniques has become paramount. This paper compares supervised and unsupervised Machine Learning techniques for defect detection in the context of industrial textile production, in terms of their respective advantages and disadvantages, and their implementation and computational costs. We explore the use of an autoencoder for the detection of defects in textiles. The goal of this preliminary work is to find out if unsupervised methods can successfully train models with good performance without the need for defect labelled data. (c) 2023 The Authors. Published by Elsevier B.V.

2023

Real-Time Algorithm Recommendation Using Meta-Learning

Authors
Palumbo, G; Guimaraes, M; Carneiro, D; Novais, P; Alves, V;

Publication
AMBIENT INTELLIGENCE-SOFTWARE AND APPLICATIONS-13TH INTERNATIONAL SYMPOSIUM ON AMBIENT INTELLIGENCE

Abstract
As the field of Machine Learning evolves, the number of available learning algorithms and their parameters continues to grow. On the one hand, this is positive as it allows for the finding of potentially more accurate models. On the other hand, however, it also makes the process of finding the right model more complex, given the number of possible configurations. Traditionally, data scientists rely on trial-and-error or brute force procedures, which are costly, or on their own intuition or expertise, which is hard to acquire. In this paper we propose an approach for algorithm recommendation based on meta-learning. The approach can be used in real-time to predict the best n algorithms (based on a selected performance metric) and their configuration, for a given ML problem. We evaluate it through cross-validation, and by comparing it against an Auto ML approach, in terms of accuracy and time. Results show that the proposed approach recommends algorithms that are similar to those of traditional approaches, in terms of performance, in just a fraction of the time.

2023

Algorithm Recommendation and Performance Prediction Using Meta-Learning

Authors
Palumbo, G; Carneiro, D; Guimares, M; Alves, V; Novais, P;

Publication
INTERNATIONAL JOURNAL OF NEURAL SYSTEMS

Abstract
In the last years, the number of machine learning algorithms and their parameters has increased significantly. On the one hand, this increases the chances of finding better models. On the other hand, it increases the complexity of the task of training a model, as the search space expands significantly. As the size of datasets also grows, traditional approaches based on extensive search start to become prohibitively expensive in terms of computational resources and time, especially in data streaming scenarios. This paper describes an approach based on meta-learning that tackles two main challenges. The first is to predict key performance indicators of machine learning models. The second is to recommend the best algorithm/configuration for training a model for a given machine learning problem. When compared to a state-of-the-art method (AutoML), the proposed approach is up to 130x faster and only 4% worse in terms of average model quality. Hence, it is especially suited for scenarios in which models need to be updated regularly, such as in streaming scenarios with big data, in which some accuracy can be traded for a much shorter model training time.

2023

The Impact of Data Selection Strategies on Distributed Model Performance

Authors
Guimarães, M; Oliveira, F; Carneiro, D; Novais, P;

Publication
Ambient Intelligence - Software and Applications - 14th International Symposium on Ambient Intelligence, ISAmI 2023, Guimarães, Portugal, July 12-14, 2023

Abstract
Distributed Machine Learning, in which data and learning tasks are scattered across a cluster of computers, is one of the answers of the field to the challenges posed by Big Data. Still, in an era in which data abounds, decisions must still be made regarding which specific data to use on the training of the model, either because the amount of available data is simply too large, or because the training time or complexity of the model must be kept low. Typical approaches include, for example, selection based on data freshness. However, old data are not necessarily outdated and might still contain relevant patterns. Likewise, relying only on recent data may significantly decrease data diversity and representativity, and decrease model quality. The goal of this paper is to compare different heuristics for selecting data in a distributed Machine Learning scenario. Specifically, we ascertain whether selecting data based on their characteristics (meta-features), and optimizing for maximum diversity, improves model quality while, eventually, allowing to reduce model complexity. This will allow to develop more informed data selection strategies in distributed settings, in which the criteria are not only the location of the data or the state of each node in the cluster, but also include intrinsic and relevant characteristics of the data. © The Author(s), under exclusive license to Springer Nature Switzerland AG 2023.

2023

Block size, parallelism and predictive performance: finding the sweet spot in distributed learning

Authors
Oliveira, F; Carneiro, D; Guimaraes, M; Oliveira, O; Novais, P;

Publication
INTERNATIONAL JOURNAL OF PARALLEL EMERGENT AND DISTRIBUTED SYSTEMS

Abstract
As distributed and multi-organization Machine Learning emerges, new challenges must be solved, such as diverse and low-quality data or real-time delivery. In this paper, we use a distributed learning environment to analyze the relationship between block size, parallelism, and predictor quality. Specifically, the goal is to find the optimum block size and the best heuristic to create distributed Ensembles. We evaluated three different heuristics and five block sizes on four publicly available datasets. Results show that using fewer but better base models matches or outperforms a standard Random Forest, and that 32 MB is the best block size.