TY - JOUR
T1 - Ten simple rules for reporting machine learning methods implementation and evaluation on biomedical data
AU - Seghier, Mohamed L.
N1 - Funding Information:
This work was funded by Khalifa University.
Publisher Copyright:
© 2021 Wiley Periodicals LLC.
PY - 2022/1
Y1 - 2022/1
N2 - There is a huge discrepancy in how researchers implement, evaluate, and report the performance of a machine learning method for classification or segmentation of biomedical data. Poor reporting and inadequate inferences are, however, not unusual to see in current literature. More specifically, vague aims and scope, missing details about the data, ambiguous preprocessing procedures, lack of clarity regarding the method's implementation, poor validation and testing, invalid comparisons between methods, and the absence of a clear rationale for performance metric choices are making it difficult to draw the right conclusions from many studies in the field. This report suggests 10 guidelines and principles that should be followed when reporting the implementation of a method and the evaluation of its performance in order to make the study transparent, interpretable, replicable, and useful. All stages of data processing and method's performance evaluation should be clearly described, and parameters and metric choices must be justified in order to aid readers in appreciating the performance of the method or in comparing it with other relevant methods. We feel that these guidelines are important for clear scientific communication in the field of biomedical data processing.
AB - There is a huge discrepancy in how researchers implement, evaluate, and report the performance of a machine learning method for classification or segmentation of biomedical data. Poor reporting and inadequate inferences are, however, not unusual to see in current literature. More specifically, vague aims and scope, missing details about the data, ambiguous preprocessing procedures, lack of clarity regarding the method's implementation, poor validation and testing, invalid comparisons between methods, and the absence of a clear rationale for performance metric choices are making it difficult to draw the right conclusions from many studies in the field. This report suggests 10 guidelines and principles that should be followed when reporting the implementation of a method and the evaluation of its performance in order to make the study transparent, interpretable, replicable, and useful. All stages of data processing and method's performance evaluation should be clearly described, and parameters and metric choices must be justified in order to aid readers in appreciating the performance of the method or in comparing it with other relevant methods. We feel that these guidelines are important for clear scientific communication in the field of biomedical data processing.
UR - http://www.scopus.com/inward/record.url?scp=85118534163&partnerID=8YFLogxK
U2 - 10.1002/ima.22674
DO - 10.1002/ima.22674
M3 - Article
AN - SCOPUS:85118534163
SN - 0899-9457
VL - 32
SP - 5
EP - 11
JO - International Journal of Imaging Systems and Technology
JF - International Journal of Imaging Systems and Technology
IS - 1
ER -