Cookies Policy
The website need some cookies and similar means to function. If you permit us, we will use those means to collect data on your visits for aggregated statistics to improve our service. Find out More
Accept Reject
  • Menu
Publications

Publications by Luís Filipe Teixeira

2021

Cervical Cancer Detection and Classification in Cytology Images Using a Hybrid Approach

Authors
Silva, EL; Sampaio, AF; Teixeira, LF; Vasconcelos, MJM;

Publication
ADVANCES IN VISUAL COMPUTING (ISVC 2021), PT II

Abstract
The high incidence of cervical cancer in women has prompted the research of automatic screening methods. This work focuses on two of the steps present in such systems, more precisely, the identification of cervical lesions and their respective classification. The development of automatic methods for these tasks is associated with some shortcomings, such as acquiring sufficient and representative clinical data. These limitations are addressed through a hybrid pipeline based on a deep learning model (RetinaNet) for the detection of abnormal regions, combined with random forest and SVM classifiers for their categorization, and complemented by the use of domain knowledge in its design. Additionally, the nuclei in each detected region are segmented, providing a set of nuclei-specific features whose impact on the classification result is also studied. Each module is individually assessed in addition to the complete system, with the latter achieving a precision, recall and F1 score of 0.04, 0.20 and 0.07, respectively. Despite the low precision, the system demonstrates potential as an analysis support tool with the capability of increasing the overall sensitivity of the human examination process.

2022

A Survey on Attention Mechanisms for Medical Applications: are we Moving Toward Better Algorithms?

Authors
Goncalves, T; Rio-Torto, I; Teixeira, LF; Cardoso, JS;

Publication
IEEE ACCESS

Abstract
The increasing popularity of attention mechanisms in deep learning algorithms for computer vision and natural language processing made these models attractive to other research domains. In healthcare, there is a strong need for tools that may improve the routines of the clinicians and the patients. Naturally, the use of attention-based algorithms for medical applications occurred smoothly. However, being healthcare a domain that depends on high-stake decisions, the scientific community must ponder if these high-performing algorithms fit the needs of medical applications. With this motto, this paper extensively reviews the use of attention mechanisms in machine learning methods (including Transformers) for several medical applications based on the types of tasks that may integrate several works pipelines of the medical domain. This work distinguishes itself from its predecessors by proposing a critical analysis of the claims and potentialities of attention mechanisms presented in the literature through an experimental case study on medical image classification with three different use cases. These experiments focus on the integrating process of attention mechanisms into established deep learning architectures, the analysis of their predictive power, and a visual assessment of their saliency maps generated by post-hoc explanation methods. This paper concludes with a critical analysis of the claims and potentialities presented in the literature about attention mechanisms and proposes future research lines in medical applications that may benefit from these frameworks.

2022

Classification of Facial Expressions Under Partial Occlusion for VR Games

Authors
Rodrigues, ASF; Lopes, JC; Lopes, RP; Teixeira, LF;

Publication
OPTIMIZATION, LEARNING ALGORITHMS AND APPLICATIONS, OL2A 2022

Abstract
Facial expressions are one of the most common way to externalize our emotions. However, the same emotion can have different effects on the same person and has different effects on different people. Based on this, we developed a system capable of detecting the facial expressions of a person in real-time, occluding the eyes (simulating the use of virtual reality glasses). To estimate the position of the eyes, in order to occlude them, Multi-task Cascade Convolutional Neural Networks (MTCNN) were used. A residual network, a VGG, and the combination of both models, were used to perform the classification of 7 different types of facial expressions (Angry, Disgust, Fear, Happy, Sad, Surprise, Neutral), classifying the occluded and non-occluded dataset. The combination of both models, achieved an accuracy of 64.9% for the occlusion dataset and 62.8% for no occlusion, using the FER-2013 dataset. The primary goal of this work was to evaluate the influence of occlusion, and the results show that the majority of the classification is done with the mouth and chin. Nevertheless, the results were far from the state-of-the-art, which is expect to be improved, mainly by adjusting the MTCNN.

2023

Deep learning-based human action recognition to leverage context awareness in collaborative assembly

Authors
Moutinho, D; Rocha, LF; Costa, CM; Teixeira, LF; Veiga, G;

Publication
ROBOTICS AND COMPUTER-INTEGRATED MANUFACTURING

Abstract
Human-Robot Collaboration is a critical component of Industry 4.0, contributing to a transition towards more flexible production systems that are quickly adjustable to changing production requirements. This paper aims to increase the natural collaboration level of a robotic engine assembly station by proposing a cognitive system powered by computer vision and deep learning to interpret implicit communication cues of the operator. The proposed system, which is based on a residual convolutional neural network with 34 layers and a long -short term memory recurrent neural network (ResNet-34 + LSTM), obtains assembly context through action recognition of the tasks performed by the operator. The assembly context was then integrated in a collaborative assembly plan capable of autonomously commanding the robot tasks. The proposed model showed a great performance, achieving an accuracy of 96.65% and a temporal mean intersection over union (mIoU) of 94.11% for the action recognition of the considered assembly. Moreover, a task-oriented evaluation showed that the proposed cognitive system was able to leverage the performed human action recognition to command the adequate robot actions with near-perfect accuracy. As such, the proposed system was considered as successful at increasing the natural collaboration level of the considered assembly station.

2023

GASTeN: Generative Adversarial Stress Test Networks

Authors
Cunha, L; Soares, C; Restivo, A; Teixeira, LF;

Publication
ADVANCES IN INTELLIGENT DATA ANALYSIS XXI, IDA 2023

Abstract
Concerns with the interpretability of ML models are growing as the technology is used in increasingly sensitive domains (e.g., health and public administration). Synthetic data can be used to understand models better, for instance, if the examples are generated close to the frontier between classes. However, data augmentation techniques, such as Generative Adversarial Networks (GAN), have been mostly used to generate training data that leads to better models. We propose a variation of GANs that, given a model, generates realistic data that is classified with low confidence by a given classifier. The generated examples can be used in order to gain insights on the frontier between classes. We empirically evaluate our approach on two well-known image classification benchmark datasets, MNIST and Fashion MNIST. Results show that the approach is able to generate images that are closer to the frontier when compared to the original ones, but still realistic. Manual inspection confirms that some of those images are confusing even for humans.

2023

MobileWeatherNet for LiDAR-Only Weather Estimation

Authors
da Silva, MP; Carneiro, D; Fernandes, J; Texeira, LF;

Publication
2023 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, IJCNN

Abstract
An autonomous vehicle relying on LiDAR data should be able to assess its limitations in real time without depending on external information or additional sensors. The point cloud generated by the sensor is subjected to significant degradation under adverse weather conditions (rain, fog, and snow), which limits the vehicle's visibility and performance. With this in mind, we show that point cloud data contains sufficient information to estimate the weather accurately and present MobileWeatherNet, a LiDAR-only convolutional neural network that uses the bird's-eye view 2D projection to extract point clouds' weather condition and improves state-of-the-art performance by 15% in terms of the balanced accuracy while reducing inference time by 63%. Moreover, this paper demonstrates that among common architectures, the use of the bird's eye view significantly enhances their performance without an increase in complexity. To the extent of our knowledge, this is the first approach that uses deep learning for weather estimation using point cloud data in the form of a bird's-eye-view projection.

  • 6
  • 10