Computer aided diagnosis systems are used to assist radiologists in their decision making. The sensitivity of these systems is hindered by the complexity of the structures inside the lungs. Several systems and methods have been proposed to detect and classify lung nodules, but all of them have their strengths and weaknesses. One way to overcome the weaknesses is to combine multiple systems. Systems based on handcrafted features capture a limited set of characteristics from the image, while deep learning based classifiers can deal with a wider range of structures. In this work, several ways to combine a handcrafted feature based classifier with four convolutional neural network are explored. The systems were combined merging the probabilities assigned to the detections in several ways. Support-vector machine, multilayer perceptron and random forest classifiers were used to combine the selected classifiers. The LUNA16 Challenge was used to evaluate the performance of the resulting hybrid systems. In all cases, the hybrid systems outperformed the individual systems. Although the average of sensitivities are similar for most of the combinations, the best hybrid system achieves a gain of 35 extra nodules at 4 FP per scan.
Convolutional neural networks are known to require large amounts of data to achieve optimal performance. In addition, data is commonly computationally augmented using a variety of geometric and intensity transformations to further extent the set of training samples. In medical imaging, annotated data is often scarce or costly to obtain, and there is considerable interest in methods to reduce the amount of data needed. In this work, we investigate the relative benefit of increasing the amount of original data, with respect to computationally augmenting the amount of training samples, for the case of false positive reduction of lung nodules candidates. To this end, we have implemented a previously published topology for classification, shown to achieve state of the art results on the publicly available Luna16 dataset. Numerous models were trained using different amounts of unique training samples and different degrees of data augmentation involving rotations and translations, and the performance was compared. Results indicate that in general, better performance is achieved when increasing the amount of data, or augmenting the data more extensively, as expected. Surprisingly however, we observed that after reaching a certain amount of unique training samples, data augmentation leads to significantly better performance compared to adding the same number of new samples to the training dataset. We hypothesize that the augmentation has aided in learning more general {rotation and translation invariant-features, leading to improved performance on unseen data. Future experiments include more detailed characterization of this behavior, and relating this to the topology and amount of parameters to be trained.
Access to the requested content is limited to institutions that have purchased or subscribe to SPIE eBooks.
You are receiving this notice because your organization may not have SPIE eBooks access.*
*Shibboleth/Open Athens users─please
sign in
to access your institution's subscriptions.
To obtain this item, you may purchase the complete book in print or electronic format on
SPIE.org.
INSTITUTIONAL Select your institution to access the SPIE Digital Library.
PERSONAL Sign in with your SPIE account to access your personal subscriptions or to use specific features such as save to my library, sign up for alerts, save searches, etc.