Clinical Trials Directory

Trials / Unknown

UnknownNCT06031818

Usability and Clinical Effectiveness of an Interpretable Deep Learning Framework for Post-Hepatectomy Liver Failure Prediction

Usability and Clinical Effectiveness of an Interpretable Deep Learning Framework (VAE-MILP) Using Counterfactual Explanations and Layerwise Relevance Propagation Framework for Post-Hepatectomy Liver Failure Prediction

Status
Unknown
Phase
Study type
Observational
Enrollment
80 (estimated)
Sponsor
Maastricht University · Academic / Other
Sex
All
Age
Healthy volunteers
Not accepted

Summary

The goal of this in-silico clinical trial is to learn about the usability and clinical effectiveness of an interpretable deep learning framework (VAE-MLP) using counterfactual explanations and layerwise relevance propagation for prediction of post-hepatectomy liver failure (PHLF) in patients with hepatocellular carcinoma (HCC). The main questions it aims to answer are: * To investigate the usability of the VAE-MLP framework for explanation of the deep learning model. * To investigate the clinical effectiveness of VAE-MLP framework for prediction of post-hepatectomy liver failure in patients with hepatocellular carcinoma. In the usability trial the clinicians and radiologists will be shown the counterfactual explanations and layerwise relevance propagation (LRP) plots to evaluate the usability of the framework. In the clinical trial the clinicians and radiologists will make the prediction under two different conditions: with model explanation and without model explanation with a washout period of at least 14 days to evaluate the clinical effectiveness of the explanation framework.

Detailed description

Post-hepatectomy liver failure (PHLF) is a severe complication after liver resection. It is important to develop an interpretable model for predicting PHLF in order to facilitate effective collaboration with clinicians for decision-making. Two-dimensional shear wave elastography (2D-SWE) is a liver stiffness measurement (LSM) technology that was proven to be useful in liver fibrosis staging. Therefore 2D-SWE shows the potential value for liver function assessment and PHLF prediction. 2D-SWE images display color-coded tissue stiffness map of liver parenchyma, with red representing a solid tissue (higher stiffness) and blue representing a soft tissue (lower stiffness). Routine analysis of 2D-SWE fails to fully utilize all information available in the images and also suffers from inter-observer variance in choosing the optimal quantification region. Deep learning (DL) has demonstrated state-of-the-art performance on many medical imaging tasks such as classification or segmentation. However, despite significant progress in DL, the clinical translation of DL tools has so far been limited, partially due to a lack of interpretability of models, the so-called "black box" problem. Interpretability of DL systems is important for fostering clinical trust as well as timely correcting any faulty processes in the algorithms. Here, the investigators present a novel interpretable DL framework (VAE-MLP) which incorporates counterfactual analysis for the explanation of 2D medical images and LRP for the explanation of feature attributions of both medical images and clinical variables. The goal of this in-silico clinical trial is to learn about the usability and clinical effectiveness of an interpretable deep learning framework (VAE-MLP) using counterfactual explanations and layerwise relevance propagation for prediction of post-hepatectomy liver failure (PHLF) in patients with hepatocellular carcinoma. The main questions it aims to answer are: * To investigate the usability of the the interpretable deep learning framework (VAE-MLP) for explanation of the deep learning model. * To investigate the clinical effectiveness of the interpretable deep learning framework (VAE-MLP) for prediction of post-hepatectomy liver failure in patients with hepatocellular carcinoma. In the usability trial the clinicians and radiologists will be shown the counterfactual explanations and layerwise relevance propagation plots of 6 examples. The score of the Likert scale of a designed questionnaire is used to evaluate the usability of the framework. In the clinical trial the clinicians and radiologists will make the prediction under two different conditions: with model explanation and without model explanation with a washout period of at least 14 days. The accuracy, sensitivity and specificity is used to compare the clinical effectiveness of the explanation framework.

Conditions

Interventions

TypeNameDescription
OTHERThe explanation of deep learning framework (VAE-MLP) , including counterfactual explanations and layerwise relevance propagationThe radiologist and clinicians will be provided the model prediction results with the explanation of the model and they will fill in a questionnaire to evaluate the usability of the interpretable framework.
OTHERThe model predictionThe radiologist and clinicians will be provided the model prediction results without the explanation of the model and they will be asked to give their own prediction.
OTHERThe model prediction and the explanation of deep learning framework (VAE-MLP) , including counterfactual explanations and layerwise relevance propagationThe radiologist and clinicians will be provided the model prediction results with the explanation of the model and they will be asked to give their own prediction.

Timeline

Start date
2023-12-10
Primary completion
2024-02-28
Completion
2024-03-15
First posted
2023-09-11
Last updated
2024-02-06

Locations

1 site across 1 country: China

Source: ClinicalTrials.gov record NCT06031818. Inclusion in this directory is not an endorsement.