Performance and Agreement When Annotating Chest X-ray Text Reports - A Preliminary Step in the Development of a Deep Learning-Based Prioritization and Detection System

Publikation: Bidrag til tidsskriftTidsskriftartikelForskningfagfællebedømt

Standard

Performance and Agreement When Annotating Chest X-ray Text Reports - A Preliminary Step in the Development of a Deep Learning-Based Prioritization and Detection System. / Li, Dana; Pehrson, Lea Marie; Bonnevie, Rasmus; Fraccaro, Marco; Thrane, Jakob; Tøttrup, Lea; Lauridsen, Carsten Ammitzbøl; Butt Balaganeshan, Sedrah; Jankovic, Jelena; Andersen, Tobias Thostrup; Mayar, Alyas; Hansen, Kristoffer Lindskov; Carlsen, Jonathan Frederik; Darkner, Sune; Nielsen, Michael Bachmann.

I: Diagnostics, Bind 13, Nr. 6, 2023.

Publikation: Bidrag til tidsskriftTidsskriftartikelForskningfagfællebedømt

Harvard

Li, D, Pehrson, LM, Bonnevie, R, Fraccaro, M, Thrane, J, Tøttrup, L, Lauridsen, CA, Butt Balaganeshan, S, Jankovic, J, Andersen, TT, Mayar, A, Hansen, KL, Carlsen, JF, Darkner, S & Nielsen, MB 2023, 'Performance and Agreement When Annotating Chest X-ray Text Reports - A Preliminary Step in the Development of a Deep Learning-Based Prioritization and Detection System', Diagnostics, bind 13, nr. 6. https://doi.org/10.3390/diagnostics13061070

APA

Li, D., Pehrson, L. M., Bonnevie, R., Fraccaro, M., Thrane, J., Tøttrup, L., Lauridsen, C. A., Butt Balaganeshan, S., Jankovic, J., Andersen, T. T., Mayar, A., Hansen, K. L., Carlsen, J. F., Darkner, S., & Nielsen, M. B. (2023). Performance and Agreement When Annotating Chest X-ray Text Reports - A Preliminary Step in the Development of a Deep Learning-Based Prioritization and Detection System. Diagnostics, 13(6). https://doi.org/10.3390/diagnostics13061070

Vancouver

Li D, Pehrson LM, Bonnevie R, Fraccaro M, Thrane J, Tøttrup L o.a. Performance and Agreement When Annotating Chest X-ray Text Reports - A Preliminary Step in the Development of a Deep Learning-Based Prioritization and Detection System. Diagnostics. 2023;13(6). https://doi.org/10.3390/diagnostics13061070

Author

Li, Dana ; Pehrson, Lea Marie ; Bonnevie, Rasmus ; Fraccaro, Marco ; Thrane, Jakob ; Tøttrup, Lea ; Lauridsen, Carsten Ammitzbøl ; Butt Balaganeshan, Sedrah ; Jankovic, Jelena ; Andersen, Tobias Thostrup ; Mayar, Alyas ; Hansen, Kristoffer Lindskov ; Carlsen, Jonathan Frederik ; Darkner, Sune ; Nielsen, Michael Bachmann. / Performance and Agreement When Annotating Chest X-ray Text Reports - A Preliminary Step in the Development of a Deep Learning-Based Prioritization and Detection System. I: Diagnostics. 2023 ; Bind 13, Nr. 6.

Bibtex

@article{a7e1b319f4be48e2b21ee5dcab0f1969,
title = "Performance and Agreement When Annotating Chest X-ray Text Reports - A Preliminary Step in the Development of a Deep Learning-Based Prioritization and Detection System",
abstract = "A chest X-ray report is a communicative tool and can be used as data for developing artificial intelligence-based decision support systems. For both, consistent understanding and labeling is important. Our aim was to investigate how readers would comprehend and annotate 200 chest X-ray reports. Reports written between 1 January 2015 and 11 March 2022 were selected based on search words. Annotators included three board-certified radiologists, two trained radiologists (physicians), two radiographers (radiological technicians), a non-radiological physician, and a medical student. Consensus labels by two or more of the experienced radiologists were considered {"}gold standard{"}. Matthew's correlation coefficient (MCC) was calculated to assess annotation performance, and descriptive statistics were used to assess agreement between individual annotators and labels. The intermediate radiologist had the best correlation to {"}gold standard{"} (MCC 0.77). This was followed by the novice radiologist and medical student (MCC 0.71 for both), the novice radiographer (MCC 0.65), non-radiological physician (MCC 0.64), and experienced radiographer (MCC 0.57). Our findings showed that for developing an artificial intelligence-based support system, if trained radiologists are not available, annotations from non-radiological annotators with basic and general knowledge may be more aligned with radiologists compared to annotations from sub-specialized medical staff, if their sub-specialization is outside of diagnostic radiology.",
author = "Dana Li and Pehrson, {Lea Marie} and Rasmus Bonnevie and Marco Fraccaro and Jakob Thrane and Lea T{\o}ttrup and Lauridsen, {Carsten Ammitzb{\o}l} and {Butt Balaganeshan}, Sedrah and Jelena Jankovic and Andersen, {Tobias Thostrup} and Alyas Mayar and Hansen, {Kristoffer Lindskov} and Carlsen, {Jonathan Frederik} and Sune Darkner and Nielsen, {Michael Bachmann}",
year = "2023",
doi = "10.3390/diagnostics13061070",
language = "English",
volume = "13",
journal = "Diagnostics",
issn = "2075-4418",
publisher = "MDPI AG",
number = "6",

}

RIS

TY - JOUR

T1 - Performance and Agreement When Annotating Chest X-ray Text Reports - A Preliminary Step in the Development of a Deep Learning-Based Prioritization and Detection System

AU - Li, Dana

AU - Pehrson, Lea Marie

AU - Bonnevie, Rasmus

AU - Fraccaro, Marco

AU - Thrane, Jakob

AU - Tøttrup, Lea

AU - Lauridsen, Carsten Ammitzbøl

AU - Butt Balaganeshan, Sedrah

AU - Jankovic, Jelena

AU - Andersen, Tobias Thostrup

AU - Mayar, Alyas

AU - Hansen, Kristoffer Lindskov

AU - Carlsen, Jonathan Frederik

AU - Darkner, Sune

AU - Nielsen, Michael Bachmann

PY - 2023

Y1 - 2023

N2 - A chest X-ray report is a communicative tool and can be used as data for developing artificial intelligence-based decision support systems. For both, consistent understanding and labeling is important. Our aim was to investigate how readers would comprehend and annotate 200 chest X-ray reports. Reports written between 1 January 2015 and 11 March 2022 were selected based on search words. Annotators included three board-certified radiologists, two trained radiologists (physicians), two radiographers (radiological technicians), a non-radiological physician, and a medical student. Consensus labels by two or more of the experienced radiologists were considered "gold standard". Matthew's correlation coefficient (MCC) was calculated to assess annotation performance, and descriptive statistics were used to assess agreement between individual annotators and labels. The intermediate radiologist had the best correlation to "gold standard" (MCC 0.77). This was followed by the novice radiologist and medical student (MCC 0.71 for both), the novice radiographer (MCC 0.65), non-radiological physician (MCC 0.64), and experienced radiographer (MCC 0.57). Our findings showed that for developing an artificial intelligence-based support system, if trained radiologists are not available, annotations from non-radiological annotators with basic and general knowledge may be more aligned with radiologists compared to annotations from sub-specialized medical staff, if their sub-specialization is outside of diagnostic radiology.

AB - A chest X-ray report is a communicative tool and can be used as data for developing artificial intelligence-based decision support systems. For both, consistent understanding and labeling is important. Our aim was to investigate how readers would comprehend and annotate 200 chest X-ray reports. Reports written between 1 January 2015 and 11 March 2022 were selected based on search words. Annotators included three board-certified radiologists, two trained radiologists (physicians), two radiographers (radiological technicians), a non-radiological physician, and a medical student. Consensus labels by two or more of the experienced radiologists were considered "gold standard". Matthew's correlation coefficient (MCC) was calculated to assess annotation performance, and descriptive statistics were used to assess agreement between individual annotators and labels. The intermediate radiologist had the best correlation to "gold standard" (MCC 0.77). This was followed by the novice radiologist and medical student (MCC 0.71 for both), the novice radiographer (MCC 0.65), non-radiological physician (MCC 0.64), and experienced radiographer (MCC 0.57). Our findings showed that for developing an artificial intelligence-based support system, if trained radiologists are not available, annotations from non-radiological annotators with basic and general knowledge may be more aligned with radiologists compared to annotations from sub-specialized medical staff, if their sub-specialization is outside of diagnostic radiology.

U2 - 10.3390/diagnostics13061070

DO - 10.3390/diagnostics13061070

M3 - Journal article

C2 - 36980376

VL - 13

JO - Diagnostics

JF - Diagnostics

SN - 2075-4418

IS - 6

ER -

ID: 342673761