Cookies
O website necessita de alguns cookies e outros recursos semelhantes para funcionar. Caso o permita, o INESC TEC irá utilizar cookies para recolher dados sobre as suas visitas, contribuindo, assim, para estatísticas agregadas que permitem melhorar o nosso serviço. Ver mais
Aceitar Rejeitar
  • Menu
Publicações

Publicações por CTM

2024

108TiP CINDERELLA clinical trial: Using artificial intelligence-driven healthcare to enhance breast cancer locoregional treatment decisions

Autores
A. Pfob; E-A. Bonci; O. Kaidar-Person; M. Antunes; O. Ciani; H. Cruz; R. Di Micco; O.D. Gentilini; J. Heil; P. Kabata; M. Romariz; T. Gonçalves; H.G. Martins; L. Borsoi; M. Mika; N. Romem; T. Schinköthe; G. Silva; M. Bobowicz; M.J. Cardoso;

Publicação
ESMO Open

Abstract

2024

CINDERELLA Trial: validation of an artificial-intelligence cloud-based platform to improve the shared decision-making process and outcomes in breast cancer patients proposed for locoregional treatment

Autores
Eduard-Alexandru Bonci; Orit Kaidar-Person; Marilia Antunes; Oriana Ciani; Helena Cruz; Rosa Di Micco; Oreste Gentilini; Pedro Gouveia; Jörg Heil; Pawel Kabata; Nuno Freitas; Tiago Gonçalves; Miguel Romariz; Henrique Martins; Carlos Mavioso; Martin Mika; André Pfob; Timo Schinköthe; Giovani Silva; Maria-João Cardoso;

Publicação
European Journal of Surgical Oncology

Abstract

2024

CINDERELLA Clinical trial (NCT05196269): using artificial intelligence-driven healthcare to enhance breast cancer locoregional treatment decisions

Autores
Bonel, EA; Kaidar-Person, O; Antunes, M; Ciani, O; Cruz, H; Di Micco, R; Gentilini, O; Heil, J; Kabata, P; Romariz, M; Gonçalves, T; Martins, H; Borsoi, L; Mika, M; Pfob, A; Romem, N; Schinköthe, T; Silva, G; Senkus, E; Cardoso, MJ;

Publicação
ANNALS OF SURGICAL ONCOLOGY

Abstract

2024

Design and Usability Assessment of Multimodal Augmented Reality System for Gait Training

Autores
Pinheiro, C; Figueiredo, J; Pereira, T; Santos, CP;

Publicação
ROBOT 2023: SIXTH IBERIAN ROBOTICS CONFERENCE, VOL 2

Abstract
Biofeedback is a promising tool to complement conventional physical therapy by fostering active participation of neurologically impaired patients during treatment. This work aims at a user-centered design and usability assessment for different age groups of a novel wearable augmented reality application composed of a multimodal sensor network and corresponding control strategies for personalized biofeedback during gait training. The proposed solution includes wearable AR glasses that deliver visual cues controlled in real-time according to mediolateral center of mass position, sagittal ankle angle, or tibialis anterior muscle activity from inertial and EMG sensors. Control strategies include positive and negative reinforcement conditions and are based on the user's performance by comparing real-time sensor data with an automatically user-personalized threshold. The proposed solution allows ambulatory practice on daily scenarios, physiotherapists' involvement through a laptop screen, and contributes to further benchmark biofeedback regarding the type of sensor. Although old healthy adults with low academic degrees have a preference for guidance from an expert person, excellent usability scores (SUS scores: 81.25-96.87) were achieved with young and middle-aged healthy adults and one neurologically impaired patient.

2024

A Semantic-oriented Approach for Underwater Wireless Communications using Generative AI

Autores
Loureiro, JP; Mateus, A; Teixeira, B; Campos, R;

Publicação
Proceedings of the 2024 15th IFIP Wireless and Mobile Networking Conference, WMNC 2024

Abstract
Underwater wireless communications are crucial for supporting multiple maritime activities, such as environmental monitoring and offshore wind farms. However, the challenging underwater environment continues to pose obstacles to the development of long-range, broadband underwater wireless communication systems. State of the art solutions are limited to long range, narrowband acoustics and short range, broadband radio or optical communications. This precludes real-time wireless transmission of imagery over long distancesIn this paper, we propose SAGE, a semantic-oriented underwater communications approach to enable real-time wireless imagery transmission over noisy and narrowband channels. SAGE extracts semantically relevant information from images at the sender located underwater and generates a text description that is transmitted to the receiver at the surface, which in turn generates an image from the received text description. SAGE is evaluated using BLIP for image-to-text and Stable Diffusion for text-to-image, showing promising image similarity between the original and the generated images, and a significant reduction in latency up to a hundred-fold, encouraging further research in this area. © 2024 IFIP.

2024

<i>DeViL</i>: Decoding Vision features into Language

Autores
Dani, M; Rio Torto, I; Alaniz, S; Akata, Z;

Publicação
PATTERN RECOGNITION, DAGM GCPR 2023

Abstract
Post-hoc explanation methods have often been criticised for abstracting away the decision-making process of deep neural networks. In this work, we would like to provide natural language descriptions for what different layers of a vision backbone have learned. Our DeViL method generates textual descriptions of visual features at different layers of the network as well as highlights the attribution locations of learned concepts. We train a transformer network to translate individual image features of any vision layer into a prompt that a separate off-the-shelf language model decodes into natural language. By employing dropout both per-layer and per-spatial-location, our model can generalize training on image-text pairs to generate localized explanations. As it uses a pre-trained language model, our approach is fast to train and can be applied to any vision backbone. Moreover, DeViL can create open-vocabulary attribution maps corresponding to words or phrases even outside the training scope of the vision model. We demonstrate that DeViL generates textual descriptions relevant to the image content on CC3M, surpassing previous lightweight captioning models and attribution maps, uncovering the learned concepts of the vision backbone. Further, we analyze fine-grained descriptions of layers as well as specific spatial locations and show that DeViL outperforms the current state-of-the-art on the neuron-wise descriptions of the MILANNOTATIONS dataset.

  • 22
  • 317