Receive a weekly summary and discussion of the top papers of the week by leading researchers in the field.

In IEEE transactions on artificial intelligence

Amidst the ongoing pandemic, the assessment of computed tomography (CT) images for COVID-19 presence can exceed the workload capacity of radiologists. Several studies addressed this issue by automating COVID-19 classification and grading from CT images with convolutional neural networks (CNNs). Many of these studies reported initial results of algorithms that were assembled from commonly used components. However, the choice of the components of these algorithms was often pragmatic rather than systematic and systems were not compared to each other across papers in a fair manner. We systematically investigated the effectiveness of using 3-D CNNs instead of 2-D CNNs for seven commonly used architectures, including DenseNet, Inception, and ResNet variants. For the architecture that performed best, we furthermore investigated the effect of initializing the network with pretrained weights, providing automatically computed lesion maps as additional network input, and predicting a continuous instead of a categorical output. A 3-D DenseNet-201 with these components achieved an area under the receiver operating characteristic curve of 0.930 on our test set of 105 CT scans and an AUC of 0.919 on a publicly available set of 742 CT scans, a substantial improvement in comparison with a previously published 2-D CNN. This article provides insights into the performance benefits of various components for COVID-19 classification and grading systems. We have created a challenge on to allow for a fair comparison between the results of this and future research.

de Vente Coen, Boulogne Luuk H, Venkadesh Kiran Vaidhya, Sital Cheryl, Lessmann Nikolas, Jacobs Colin, Sanchez Clara I, van Ginneken Bram


** CO-RADS, 3-D convolutional neural network (CNN), COVID-19, deep learning, medical imaging**