2023
Autores
Neto P.C.; Caldeira E.; Cardoso J.S.; Sequeira A.F.;
Publicação
BIOSIG 2023 - Proceedings of the 22nd International Conference of the Biometrics Special Interest Group
Abstract
With the ever-growing complexity of deep learning models for face recognition, it becomes hard to deploy these systems in real life. Researchers have two options: 1) use smaller models; 2) compress their current models. Since the usage of smaller models might lead to concerning biases, compression gains relevance. However, compressing might be also responsible for an increase in the bias of the final model. We investigate the overall performance, the performance on each ethnicity subgroup and the racial bias of a State-of-the-Art quantization approach when used with synthetic and real data. This analysis provides a few more details on potential benefits of performing quantization with synthetic data, for instance, the reduction of biases on the majority of test scenarios. We tested five distinct architectures and three different training datasets. The models were evaluated on a fourth dataset which was collected to infer and compare the performance of face recognition models on different ethnicity.
2023
Autores
Ferreira, G; Teixeira, M; Belo, R; Silva, W; Cardoso, JS;
Publicação
2023 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, IJCNN
Abstract
The application of machine learning algorithms to predict the mechanism of action (MoA) of drugs can be highly valuable and enable the discovery of new uses for known molecules. The developed methods are usually evaluated with small subsets of MoAs with large support, leading to deceptively good generalization. However, these datasets may not accurately represent a practical use, due to the limited number of target MoAs. Accurate predictions for these rare drugs are important for drug discovery and should be a point of focus. In this work, we explore different training strategies to improve the performance of a well established deep learning model for rare drug MoA prediction. We explored transfer learning by first learning a model for common MoAs, and then using it to initialize the learning of another model for rarer MoAs. We also investigated the use of a cascaded methodology, in which results from an initial model are used as additional inputs to the model for rare MoAs. Finally, we proposed and tested an extension of Mixup data augmentation for multilabel classification. The baseline model showed an AUC of 73.2% for common MoAs and 62.4% for rarer classes. From the investigated methods, Mixup alone failed to improve the performance of a baseline classifier. Nonetheless, the other proposed methods outperformed the baseline for rare classes. Transfer Learning was preferred in predicting classes with less than 10 training samples, while the cascaded classifiers (with Mixup) showed better predictions for MoAs with more than 10 samples. However, the performance for rarer MoAs still lags behind the performance for frequent MoAs and is not sufficient for the reliable prediction of rare MoAs.
2023
Autores
Albuquerque, T; Fang, ML; Wiestler, B; Delbridge, C; Vasconcelos, MJM; Cardoso, JS; Schüffler, P;
Publicação
MEDICAL IMAGE COMPUTING AND COMPUTER ASSISTED INTERVENTION - MICCAI 2023 WORKSHOPS
Abstract
The most malignant tumors of the central nervous system are adult-type diffuse gliomas. Historically, glioma subtype classification has been based on morphological features. However, since 2016, WHO recognizes that molecular evaluation is critical for subtyping. Among molecular markers, the mutation status of IDH1 and the codeletion of 1p/19q are crucial for the precise diagnosis of these malignancies. In pathology laboratories, however, manual screening for those markers is time-consuming and susceptible to error. To overcome these limitations, we propose a novel multimodal biomarker classification method that integrates image features derived from brain magnetic resonance imaging and histopathological exams. The proposed model consists of two branches, the first branch takes as input a multi-scale Hematoxylin and Eosin whole slide image, and the second branch uses the pre-segmented region of interest from the magnetic resonance imaging. Both branches are based on convolutional neural networks. After passing the exams by the two embedding branches, the output feature vectors are concatenated, and a multi-layer perceptron is used to classify the glioma biomarkers as a multi-class problem. In this work, several fusion strategies were studied, including a cascade model with mid-fusion; a mid-fusion model, a late fusion model, and a mid-context fusion model. The models were tested using a publicly available data set from The Cancer Genome Atlas. Our cross-validated classification models achieved an area under the curve of 0.874, 0.863, and 0.815 for the proposed multimodal, magnetic resonance imaging, and Hematoxylin and Eosin stain slide images respectively, indicating our multimodal model outperforms its unimodal counterparts and the state-of-the-art glioma biomarker classification methods.
2023
Autores
Moutinho, D; Rocha, LF; Costa, CM; Teixeira, LF; Veiga, G;
Publicação
ROBOTICS AND COMPUTER-INTEGRATED MANUFACTURING
Abstract
Human-Robot Collaboration is a critical component of Industry 4.0, contributing to a transition towards more flexible production systems that are quickly adjustable to changing production requirements. This paper aims to increase the natural collaboration level of a robotic engine assembly station by proposing a cognitive system powered by computer vision and deep learning to interpret implicit communication cues of the operator. The proposed system, which is based on a residual convolutional neural network with 34 layers and a long -short term memory recurrent neural network (ResNet-34 + LSTM), obtains assembly context through action recognition of the tasks performed by the operator. The assembly context was then integrated in a collaborative assembly plan capable of autonomously commanding the robot tasks. The proposed model showed a great performance, achieving an accuracy of 96.65% and a temporal mean intersection over union (mIoU) of 94.11% for the action recognition of the considered assembly. Moreover, a task-oriented evaluation showed that the proposed cognitive system was able to leverage the performed human action recognition to command the adequate robot actions with near-perfect accuracy. As such, the proposed system was considered as successful at increasing the natural collaboration level of the considered assembly station.
2023
Autores
Cunha, L; Soares, C; Restivo, A; Teixeira, LF;
Publicação
ADVANCES IN INTELLIGENT DATA ANALYSIS XXI, IDA 2023
Abstract
Concerns with the interpretability of ML models are growing as the technology is used in increasingly sensitive domains (e.g., health and public administration). Synthetic data can be used to understand models better, for instance, if the examples are generated close to the frontier between classes. However, data augmentation techniques, such as Generative Adversarial Networks (GAN), have been mostly used to generate training data that leads to better models. We propose a variation of GANs that, given a model, generates realistic data that is classified with low confidence by a given classifier. The generated examples can be used in order to gain insights on the frontier between classes. We empirically evaluate our approach on two well-known image classification benchmark datasets, MNIST and Fashion MNIST. Results show that the approach is able to generate images that are closer to the frontier when compared to the original ones, but still realistic. Manual inspection confirms that some of those images are confusing even for humans.
2023
Autores
da Silva, MP; Carneiro, D; Fernandes, J; Texeira, LF;
Publicação
2023 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, IJCNN
Abstract
An autonomous vehicle relying on LiDAR data should be able to assess its limitations in real time without depending on external information or additional sensors. The point cloud generated by the sensor is subjected to significant degradation under adverse weather conditions (rain, fog, and snow), which limits the vehicle's visibility and performance. With this in mind, we show that point cloud data contains sufficient information to estimate the weather accurately and present MobileWeatherNet, a LiDAR-only convolutional neural network that uses the bird's-eye view 2D projection to extract point clouds' weather condition and improves state-of-the-art performance by 15% in terms of the balanced accuracy while reducing inference time by 63%. Moreover, this paper demonstrates that among common architectures, the use of the bird's eye view significantly enhances their performance without an increase in complexity. To the extent of our knowledge, this is the first approach that uses deep learning for weather estimation using point cloud data in the form of a bird's-eye-view projection.
The access to the final selection minute is only available to applicants.
Please check the confirmation e-mail of your application to obtain the access code.