Cookies
O website necessita de alguns cookies e outros recursos semelhantes para funcionar. Caso o permita, o INESC TEC irá utilizar cookies para recolher dados sobre as suas visitas, contribuindo, assim, para estatísticas agregadas que permitem melhorar o nosso serviço. Ver mais
Aceitar Rejeitar
  • Menu
Publicações

Publicações por LIAAD

2025

Simulating Biases for Interpretable Fairness in Offline and Online Classifiers

Autores
Inácio, R; Kokkinogenis, Z; Cerqueira, V; Soares, C;

Publicação
CoRR

Abstract

2025

Generating Large Semi-Synthetic Graphs of Any Size

Autores
Tuna, R; Soares, C;

Publicação
CoRR

Abstract

2025

MASTFM: Meta-learning and Data Augmentation to Stress Test Forecasting Models

Autores
Inácio, R; Cerqueira, V; Barandas, M; Soares, C;

Publicação
Machine Learning and Knowledge Discovery in Databases. Applied Data Science Track and Demo Track - European Conference, ECML PKDD 2025, Porto, Portugal, September 15-19, 2025, Proceedings, Part X

Abstract

2025

Mast: interpretable stress testing via meta-learning for forecasting model robustness evaluation

Autores
Inácio, R; Cerqueira, V; Barandas, M; Soares, C;

Publicação
Mach. Learn.

Abstract

2025

SPATA: Systematic Pattern Analysis for Detailed and Transparent Data Cards

Autores
Vitorino, J; Maia, E; Praça, I; Soares, C;

Publicação
CoRR

Abstract

2025

Evaluating Transfer Learning Methods on Real-World Data Streams: A Case Study in Financial Fraud Detection

Autores
Pereira, RR; Bono, J; Ferreira, HM; Ribeiro, P; Soares, C; Bizarro, P;

Publicação
Machine Learning and Knowledge Discovery in Databases. Applied Data Science Track - European Conference, ECML PKDD 2025, Porto, Portugal, September 15-19, 2025, Proceedings, Part IX

Abstract
When the available data for a target domain is limited, transfer learning (TL) methods leverage related data-rich source domains to train and evaluate models, before deploying them on the target domain. However, most TL methods assume fixed levels of labeled and unlabeled target data, which contrasts with real-world scenarios where both data and labels arrive progressively over time. As a result, evaluations based on these static assumptions may not reflect how methods perform in practice. To support a more realistic assessment of TL methods in dynamic settings, we propose an evaluation framework that (1) simulates varying data availability over time, (2) creates multiple domains via resampling of a given dataset and (3) introduces inter-domain variability through controlled transformations, e.g., including time-dependent covariate and concept shifts. These capabilities enable the systematic simulation of a large number of variants of the experiments, providing deeper insights into how algorithms may behave when deployed. We demonstrate the usefulness of the proposed framework by performing a case study on a proprietary real-world suite of card payment datasets. To support reproducibility, we also apply the framework on the publicly available Bank Account Fraud (BAF) dataset. By providing a methodology for evaluating TL methods over time and in different data availability conditions, our framework supports a better understanding of model behavior in real-world environments, which enables more informed decisions when deploying models in new domains. © 2025 Elsevier B.V., All rights reserved.

  • 11
  • 514