Source code for WORC.featureprocessing.StatisticalTestFeatures

#!/usr/bin/env python

# Copyright 2016-2023 Biomedical Imaging Group Rotterdam, Departments of
# Medical Informatics and Radiology, Erasmus MC, Rotterdam, The Netherlands
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
#     http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.

import matplotlib
matplotlib.use('agg')
import matplotlib.pyplot as plt

import os
import csv
import numpy as np
from scipy.stats import ttest_ind, ranksums, mannwhitneyu, chi2_contingency
import WORC.IOparser.config_io_classifier as config_io
from WORC.IOparser.file_io import load_features
from WORC.detectors.detectors import DebugDetector
from WORC.plotting.plot_pvalues_features import manhattan_importance


[docs]def StatisticalTestFeatures(features, patientinfo, config, output_csv=None, output_png=None, output_tex=None, plot_test='MWU', Bonferonni=True, fontsize='small', yspacing=1, threshold=0.05, verbose=True, label_type=None): """Perform several statistical tests on features, such as a student t-test. Parameters ---------- features: string, mandatory contains the paths to all .hdf5 feature files used. modalityname1=file1,file2,file3,... modalityname2=file1,... Thus, modalities names are always between a space and a equal sign, files are split by commas. We assume that the lists of files for each modality has the same length. Files on the same position on each list should belong to the same patient. patientinfo: string, mandatory Contains the path referring to a .txt file containing the patient label(s) and value(s) to be used for learning. See the Github Wiki for the format. config: string, mandatory path referring to a .ini file containing the parameters used for feature extraction. See the Github Wiki for the possible fields and their description. # TODO: outputs verbose: boolean, default True print final feature values and labels to command line or not. """ # Load variables from the config file config = config_io.load_config(config) if type(patientinfo) is list: patientinfo = ''.join(patientinfo) if type(config) is list: config = ''.join(config) if type(output_csv) is list: output_csv = ''.join(output_csv) if type(output_png) is list: output_png = ''.join(output_png) if type(output_tex) is list: output_tex = ''.join(output_tex) # Create output folder if required if not os.path.exists(os.path.dirname(output_csv)): os.makedirs(os.path.dirname(output_csv)) if label_type is None: label_type = config['Labels']['label_names'] # Read the features and classification data print("Reading features and label data.") label_data, image_features =\ load_features(features, patientinfo, label_type) # Extract feature labels and put values in an array feature_labels = image_features[0][1] feature_values = np.zeros([len(image_features), len(feature_labels)]) for num, x in enumerate(image_features): feature_values[num, :] = x[0] # ----------------------------------------------------------------------- # Perform statistical tests print("Performing statistical tests.") label_value = label_data['label'] label_name = label_data['label_name'] header = list() subheader = list() for i_name in label_name: header.append(str(i_name)) header.append('') header.append('') header.append('') header.append('') header.append('') subheader.append('Label') subheader.append('Ttest') subheader.append('Welch') subheader.append('Wilcoxon') subheader.append('Mann-Whitney') subheader.append('Chi2') subheader.append('') # Open the output_csv file if output_csv is not None: myfile = open(output_csv, 'w') wr = csv.writer(myfile, quoting=csv.QUOTE_ALL) wr.writerow(header) wr.writerow(subheader) savedict = dict() for i_class, i_name in zip(label_value, label_name): savedict[i_name[0]] = dict() pvalues = list() pvalueswelch = list() pvalueswil = list() pvaluesmw = list() pvalueschi2 = list() classlabels = i_class.ravel() for num, fl in enumerate(feature_labels): fv = feature_values[:, num] # Remove NaN values fv = fv[~np.isnan(fv)] class1 = [i for j, i in enumerate(fv) if classlabels[j] == 1] class2 = [i for j, i in enumerate(fv) if classlabels[j] == 0] pvalues.append(ttest_ind(class1, class2)[1]) pvalueswelch.append(ttest_ind(class1, class2, equal_var=False)[1]) pvalueswil.append(ranksums(class1, class2)[1]) try: pmwu = mannwhitneyu(class1, class2)[1] if pmwu == 0.0: print("[WORC Warning] Mann-Whitney U test resulted in a p-value of exactly 0.0, which is not valid. Replacing metric value by NaN.") pvaluesmw.append(np.nan) else: pvaluesmw.append(pmwu) except ValueError as e: print("[WORC Warning] " + str(e) + '. Replacing metric value by NaN.') pvaluesmw.append(np.nan) # Optional: perform chi2 test. Only do this when categorical, which we define as less than 20 options. unique_values = list(set(fv)) unique_values.sort() if len(unique_values) == 0: # All NaN print("[WORC Warning] " + fl + " has no value. Replacing chi2 metric value by NaN.") pvalueschi2.append(np.nan) elif len(unique_values) == 1: print("[WORC Warning] " + fl + " has only one value. Replacing chi2 metric value by NaN.") pvalueschi2.append(np.nan) elif len(unique_values) <= 20: class1_count = [class1.count(i) for i in unique_values] class2_count = [class2.count(i) for i in unique_values] obs = np.array([class1_count, class2_count]) try: _, p, _, _ = chi2_contingency(obs) pvalueschi2.append(p) except ValueError: print("[WORC Warning] " + fl + " has a zero element in table of frequencies. Replacing chi2 metric value by NaN.") pvalueschi2.append(np.nan) else: print("[WORC Warning] " + fl + " is no categorical variable. Replacing chi2 metric value by NaN.") pvalueschi2.append(np.nan) # Sort based on p-values: indices = np.argsort(np.asarray(pvaluesmw)) feature_labels_o = np.asarray(feature_labels)[indices].tolist() pvalues = np.asarray(pvalues)[indices].tolist() pvalueswelch = np.asarray(pvalueswelch)[indices].tolist() pvalueswil = np.asarray(pvalueswil)[indices].tolist() pvaluesmw = np.asarray(pvaluesmw)[indices].tolist() pvalueschi2 = np.asarray(pvalueschi2)[indices].tolist() savedict[i_name[0]]['ttest'] = pvalues savedict[i_name[0]]['welch'] = pvalueswelch savedict[i_name[0]]['wil'] = pvalueswil savedict[i_name[0]]['mw'] = pvaluesmw savedict[i_name[0]]['chi2'] = pvalueschi2 savedict[i_name[0]]['labels'] = feature_labels_o if output_csv is not None: for num in range(0, len(savedict[i_name[0]]['ttest'])): writelist = list() for i_name in savedict.keys(): labeldict = savedict[i_name] writelist.append(labeldict['labels'][num]) writelist.append(labeldict['ttest'][num]) writelist.append(labeldict['welch'][num]) writelist.append(labeldict['wil'][num]) writelist.append(labeldict['mw'][num]) writelist.append(labeldict['chi2'][num]) writelist.append('') wr.writerow(writelist) print("Saved data to CSV!") if output_png is not None or output_tex is not None: # Initialize objects objects_temp = labeldict['labels'] if plot_test == 'MWU': p_values_temp = labeldict['mw'] # remove the nan objects = list() p_values = list() for o, p in zip(objects_temp, p_values_temp): if not np.isnan(p): objects.append(o) p_values.append(p) # Debug defaults if DebugDetector().do_detection(): # No correction Bonferonni = False # Just select ~10 features sorted_p = p_values[:] sorted_p.sort() threshold = sorted_p[10] if Bonferonni: # Apply Bonferonni correction for multiple testing threshold = threshold / len(p_values) # Create labels labels = list() mapping = {0: 'Histogram', 1: 'Shape', 2: 'Orientation', 3: 'GLCM', 4: 'GLRLM', 5: 'GLSZM', 6: 'GLDM', 7: 'NGTDM', 8: 'Gabor', 9: 'Semantic', 10: 'DICOM', 11: 'LoG', 12: 'Vessel', 13: 'LBP', 14: 'Phase' } for o in objects: if 'hf_' in o.lower(): labels.append(0) elif 'sf_' in o.lower(): labels.append(1) elif 'of_' in o.lower(): labels.append(2) elif 'glcm_' in o or 'glcmms_' in o.lower(): labels.append(3) elif 'glrlm_' in o.lower(): labels.append(4) elif 'glszm_' in o.lower(): labels.append(5) elif 'gldm_' in o.lower(): labels.append(6) elif 'ngtdm_' in o.lower(): labels.append(7) elif 'gabor_' in o.lower(): labels.append(8) elif 'semf_' in o.lower(): labels.append(9) elif 'df_' in o.lower(): labels.append(10) elif 'logf_' in o.lower(): labels.append(11) elif 'vf_' in o.lower(): labels.append(12) elif 'lbp_' in o.lower(): labels.append(13) elif 'phasef_' in o.lower(): labels.append(14) else: raise KeyError(o) # Replace several labels objects = [o.replace('CalcFeatures_', '') for o in objects] objects = [o.replace('featureconverter_', '') for o in objects] objects = [o.replace('PREDICT_', '') for o in objects] objects = [o.replace('PyRadiomics_', '') for o in objects] objects = [o.replace('Pyradiomics_', '') for o in objects] objects = [o.replace('predict_', '') for o in objects] objects = [o.replace('pyradiomics_', '') for o in objects] objects = [o.replace('_predict', '') for o in objects] objects = [o.replace('_pyradiomics', '') for o in objects] objects = [o.replace('original_', '') for o in objects] objects = [o.replace('train_', '') for o in objects] objects = [o.replace('test_', '') for o in objects] objects = [o.replace('1_0_', '') for o in objects] objects = [o.replace('hf_', '') for o in objects] objects = [o.replace('sf_', '') for o in objects] objects = [o.replace('of_', '') for o in objects] objects = [o.replace('GLCM_', '') for o in objects] objects = [o.replace('GLCMMS_', '') for o in objects] objects = [o.replace('GLRLM_', '') for o in objects] objects = [o.replace('GLSZM_', '') for o in objects] objects = [o.replace('GLDM_', '') for o in objects] objects = [o.replace('NGTDM_', '') for o in objects] objects = [o.replace('Gabor_', '') for o in objects] objects = [o.replace('semf_', '') for o in objects] objects = [o.replace('df_', '') for o in objects] objects = [o.replace('logf_', '') for o in objects] objects = [o.replace('vf_', '') for o in objects] objects = [o.replace('Frangi_', '') for o in objects] objects = [o.replace('LBP_', '') for o in objects] objects = [o.replace('phasef_', '') for o in objects] objects = [o.replace('tf_', '') for o in objects] objects = [o.replace('_CT_0', '') for o in objects] objects = [o.replace('_MR_0', '') for o in objects] objects = [o.replace('CT_0', '') for o in objects] objects = [o.replace('MR_0', '') for o in objects] # Sort based on labels sort_indices = np.argsort(np.asarray(labels)) p_values = [p_values[i] for i in sort_indices] labels = [labels[i] for i in sort_indices] objects = [objects[i] for i in sort_indices] # Make manhattan plot manhattan_importance(values=p_values, labels=labels, output_png=output_png, feature_labels=objects, threshold_annotated=threshold, mapping=mapping, output_tex=output_tex) return savedict