High-Resolution Breast Cancer Screening with Multi-View Deep Convolutional Neural Networks

Size: px
Start display at page:

Download "High-Resolution Breast Cancer Screening with Multi-View Deep Convolutional Neural Networks"

Transcription

1 1 High-Resolution Breast Cancer Screening with Multi-View Deep Convolutional Neural Networks Krzysztof J. Geras 1, Stacey Wolfson 3, Yiqiu Shen 1, S. Gene Kim 3,4, Linda Moy 3,4, Kyunghyun Cho 1,2,5 arxiv: v2 [cs.cv] 6 Nov 2017 Abstract Recent advances in deep learning for natural images has prompted a surge of interest in applying similar techniques to medical images. Most of the initial attempts focused on replacing the input of a deep convolutional neural network with a medical image, which does not take into consideration the fundamental differences between these two types of images. Specifically, fine details are necessary for detection in medical images, unlike in natural images where coarse structures matter. This difference makes it inadequate to use the existing network architectures developed for natural images, because they work on an heavily downsampled image to reduce the memory requirements. This hides details necessary to make accurate predictions. Additionally, a single exam in medical imaging often comes with a set of views which must be fused in order to reach a correct conclusion. In our work, we propose to use a multi-view deep convolutional neural network that handles a set of high-resolution medical images. We evaluate it on large-scale mammographybased breast cancer screening (BI-RADS prediction) using 886 thousand images. We focus on investigating the impact of training set size and image size on the prediction accuracy. Our results highlight that performance increases with the size of training set, and that the best performance can only be achieved using the original resolution. This suggests that medical imaging research using deep learning must utilize as much data as possible with the least amount of potentially harmful preprocessing. Index Terms breast cancer screening, deep convolutional neural networks, deep learning, machine learning, mammography Breast cancer is the second leading cause of death among women in the United States. It is estimated that 232,000 women were diagnosed with breast cancer and approximately 40,000 died from the disease in 2015 [1]. Screening mammography is the main imaging test used to detect occult breast cancer. Multiple randomized clinical trials have shown a 30% reduction in mortality in asymptomatic women who were undergoing screening mammography [2], [3]. Although mammography is the only imaging test that reduced breast cancer mortality [2], [3], [4], the appropriate screening interval for mammograms has been the subject of public debate with different professional societies offering varying guidelines for mammographic screening [2], [3], [4], [5]. In particular, there has been public discussion regarding the potential harms of screening. These harms include false positive recalls and false 1 Center for Data Science, New York University, 60 5th Ave, New York, NY Courant Institute of Mathematical Sciences, New York University, 251 Mercer St, New York, NY Center for Biomedical Imaging, Radiology, NYU School of Medicine, 660 1st Avenue, New York, NY Perlmutter Cancer Center, NYU Langone Medical Center, 160 E 34th St, New York, NY CIFAR Azrieli Global Scholar This work has been submitted to the IEEE for possible publication. Copyright may be transferred without notice, after which this version may no longer be accessible. positive biopsies as well as anxiety caused by recall for diagnostic testing after a screening exam. Overall, the recall rate following a screening mammogram is between 10-15%. This equates to about 3.3 to 4.5 million callback exams for additional testing [6]. The vast majority of the women asked to return following an inconclusive mammogram undergo another mammogram and/or ultrasound for clarification. Most of these false positive findings are found to represent normal breast tissue. Only 1% to 2% of women who have an abnormal screening mammogram are recommended to undergo a biopsy. Only 20-40% of these biopsies yield a diagnosis of cancer [7]. In 2014, over 39 million screening and diagnostic mammography exams were performed in the US. Therefore, in addition to the anxiety from undergoing a false positive mammogram, there are significant costs associated with unnecessary follow ups and biopsies. Clearly, there is an unmet need to shift the balance of routine breast cancer screening towards more benefit and less harm. MAMMOGRAPHY-BASED BREAST CANCER SCREENING Screening mammography has been recognized as one of the leading contributors to reducing breast cancer mortality by 30% since the 1980s. Despite such successes, breast cancer remains the second leading cause of cancer death in women and the leading cause of death in women aged between 45 and 55 [1]. This fact illustrates key limitations in the sensitivity of current approaches to breast cancer screening largely based on mammography. The specificity of the screening mammography is also limited as the continued large fraction of false positives often leads to unnecessary follow up procedures including biopsies that cause considerable pain, anxiety and cost. It is therefore necessary to improve diagnostic accuracy of mammography in terms of both sensitivity for early detection of breast cancer and specificity to keep the recall rate low. Breast Cancer Screening as a Deep Learning Task Deep learning has recently seen enormous success in challenging problems such as object recognition in natural images, automatic speech recognition and machine translation [8]. This success has prompted a surge of interest in applying deep convolutional networks (DCN) to medical imaging. Many recent studies have shown the potential of applying such networks to medical imaging, including breast screening mammography; however, without investigating the fundamental differences between medical and natural images and their impact on the design choices and performance of proposed models. For

2 2 instance, many recent works have either significantly downscaled a whole image or focused on classifying a small region of interest. This might be detrimental to performance of such models given the well-known dependency of breast cancer screening on fine details, such as the existence of a cluster of microcalcifications, as well as global structures, such as the symmetry between two breasts. Furthermore, the potential of DCNs has only been assessed in limited settings of small data sets often consisting of less than 1,000 images, while the success of such networks in natural object recognition is largely attributed to the availability of more than 1M annotated images. This further hinders our understanding of the true potential of DCNs in medical imaging, particularly in breast cancer screening. In this work, we conduct an investigation into analyzing and understanding fundamental properties of deep convolutional networks in the context of breast cancer screening. We start by building a large-scale data set of 201,698 screening mammographic exams (886,437 images) collected at multiple sites of our institution. We develop a novel DCN that is able to handle multiple views of screening mammography and to utilize large high-resolution images without downscaling. We refer to this DCN as a multi-view deep convolutional network (MV-DCN). Our network learns to predict the assessment of a radiologist, classifying an incoming example as BI-RADS 0 ( incomplete ), BI-RADS 1 ( normal ) or BI-RADS 2 ( benign finding ). We study the impact of the data set size and image resolution on the screening performance of the proposed MV-DCN, which would serve as a de facto guideline for optimizing future deep neural networks for medical imaging. We further investigate the potential of the proposed MV-DCN by visualizing predictions made by MV-DCN. HIGH-RESOLUTION MULTI-VIEW DEEP CONVOLUTIONAL NEURAL NETWORKS Deep Convolutional Neural Network A deep convolutional neural network [9], [10] is a classifier that takes an image x as input, often with multiple channels corresponding to different colors (e.g., RGB), and outputs the conditional probability distribution over the categories p(y x). This is done by a series of nonlinear functions that gradually transform the input pixel-level image. A major property of the deep convolutional network, which distinguishes it from a multi-layer perceptron, is that it heavily relies on convolutional and pooling layers, which make the network invariant to local translation of visual features in the input. Multi-View Deep Convolutional Neural Network Unlike an object recognition task with natural images, an exam in medical imaging often comes with a set of views. For instance, it is standard in screening mammography to obtain cranial caudal (CC) and mediolateral oblique (MLO) views for each breast of a patient, resulting in a set of four images. We will refer to them as L-CC, R-CC, L-MLO and R-MLO (Figure 1). There is a rich literature in building a deep neural network for multi-view data. Most of them fall into one of two families. First, there are works on unsupervised feature extraction from multiple views using a variant of deep autoencoders [11], [12], [13]. They usually train a multi-view deep neural network with unlabeled examples, and use the output of such a network as a feature extractor, followed by a standard classifier. On the other hand, [14] proposed to build a multi-view deep convolutional network directly for classification. We propose a variant of MV-DCN which was motivated by [14]. This MV-DCN computes the output in two stages. In the first stage, a number of convolutional and pooling layers is separately applied to each of the views. We denote such viewspecific representation by h v, where v refers to the index of the view. These view-specific representations are concatenated to form a vector, [h L CC, h R CC, h L MLO, h R MLO ], which is an input to the second stage - a fully connected layer followed by a softmax layer producing output distribution p(y x). The whole network is trained jointly by stochastic gradient descent with backpropagation [15]. Furthermore, we employ a number of regularization techniques to avoid the behavior of overfitting due to the relatively small size of training dataset, such as data augmentation by random cropping [16] and dropout [17]. These will be described later in detail. High-Resolution Convolutional Neural Network It is common in object recognition and detection in natural images to heavily downscale an original high-resolution image. For instance, the input to the network of the best performer in ImageNet Challenge 2015 (classification task) was an image downscaled to [18]. This is often done to improve the computational efficiency, both in terms of computation and memory, and also because no significant improvement has been observed with higher-resolution images. It reflects an inherent property of natural images, in which the objects of interest are usually presented in relatively larger portions than other objects and what matter most are their macro-structures, such as shapes, colors and other global descriptors. However, downscaling of an input image is not desirable in the case of medical images, and in particular for early-stage screening based on breast mammography. Often a cue for diagnosis is a subtle finding which may be identified only at the original resolution. In order to address the computational issues of handling fullresolution images, we propose to use aggressive convolution and pooling layers. First, we use convolution layers with strides larger than one in the first two convolutional layers. Also, the first pooling layer has a larger stride than the other pooling layers. Thereby, we greatly reduce the size of feature maps early in the network. Although this aggressive convolution and pooling loses some spatial information, the parameters of the network are adjusted to minimize this information loss during training. This is unlike downscaling of the input, which loses information unconditionally. Second, we take the average of output vectors from the last feature map instead of concatenating them [19], which has been a more common practice [16], [20]. This drastically reduces the dimensionality of the view-specific vector without much, if any, performance degradation [21]. Using both of these

3 3 right mediolateral olique (R-MLO) left mediolateral olique (L-MLO) right cranio-caudal (R-CC) left cranio-caudal (L-CC) Fig. 1. The four standard views used in our experiments. approaches, we are able to build an MV-DCN that takes four pixels images (one per view) as input without any downscaling. RELATED WORK Let us briefly review recent deep learning based approaches to breast mammography, summarized in Table I. a) Multi-Stage vs. End-to-End Approaches.: Traditionally disjoint breast cancer screening or lesion detection is done in three stages: detection, analysis and final assessment/management. In the first stage, a breast mammography image is segmented into different types of regions, such as foreground (breast) and background. Within the segmented region of breast, the second stage focuses on extracting a set of regions of interest (ROI) that will be examined in more detail. In the third stage, each of those ROI s is determined to be a malignant lesion or not. The outcome of the third stage is used to make the final decision on a given case consisting of multiple views. Most of the recent research on applying deep learning to breast mammography have focused on replacing one or more stages in this existing multi-stage pipeline; for instance, mass detection [22], [23], [24]. In their work, a deep neural network is trained to determine whether a small patch is a mass. Others have focused on training a deep neural network for classifying a small region of interest into one of a few categories, assuming an existing mass detection system [25], [26], [27], [28]. On the other hand, a small number of research groups have considered replacing the whole multi-stage approach with a single, or a series of, trainable machine learning algorithms. Kooi et al. [29] proposed to use a random forest classifier for mass detection followed by a DCN that classifies each detected mass. A similar approach was proposed by Becker et al. [30]. Akselrod-Ballin et al. [31] further proposed to use deep convolutional networks for both mass detection and classification, potentially enabling end-to-end training. Two groups [32], [33] went even further by proposing a single deep convolutional network that classifies a whole image, or a set of multiple views. The work by Carneiro et al. [33] is closest to our approach in this paper. In both of the works, a single deep convolutional neural network takes as input a set of multiple views of an exam and predicts its BI-RADS label. b) Data Size.: Although it is recognized that one of the driving forces behind the success of deep learning is the availability of large scale data, it has not been exploited when applying deep learning to mammography. As evident in Table I, most of the recent works use less than 1,000 images for both training and testing. To avoid the issue of small training data, most of the earlier works resorted to training with many small patches, or ROI s, avoiding end-toend training. One exception is [33] in which Carneiro et al. use the whole image, however, with the deep convolutional network pretrained for object recognition in natural images. Unlike these earlier approaches, we use a large-scale data set of an unprecedented size, consisting of 886,437 images. This allows us to carefully study the impact of the size of training data set. c) Natural vs. Controlled Distribution.: Breast screening is aimed at a general population rather than a selected group of patients. This implies that the distribution of the screening outcome is heavily skewed toward normal (BI-RADS 1). In our training set which closely follows a general population distribution, approximately 46% of the cases were assigned BI-RADS 1 ( normal ), while 41% were assigned BI-RADS 2 ( benign finding ) and 13% BI-RADS 0 ( incomplete ). This is in contrast to two widely-used, publicly available datasets, INBreast [34], DDSM [35], [36] and other curated small-scale datasets from recent literature (see those in Table I). These datasets are often constructed to include approximately the same proportions of normal and abnormal cases, resulting in, what we refer to as, a controlled distribution of outcomes which differs from a natural distribution. For instance, IN- Breast has approximately achieved a balance between benign and malignant cases. This type of artificial balancing, or equivalently upsampling of malignant cases, may bias a model to more often predict a given case as malignant and require a

4 4 TABLE I PREVIOUS WORKS ON DEEP LEARNING FOR BREAST MAMMOGRAPHY. task ref. E2E #images image MV input dist. size 829k (57k) IMG N [33] 680 ( 340) IMG C [32] 410 (CV) IMG C [31] 850 ( 170) IMG C [25] 607 (CV) ROI C [29] 44, ROI N (18,000) [26] 1820 (182) ROI C [27] 736 ( 300) ROI C [28] 1606 ( 378) ROI N [22] 116 (CV) ROI C [23] 410 (CV) ROI C [24] 2500 (250) ROI C [37] 1000 (204) N/A ROI C [38] 1410 (N/A) N/A ROI C BI-RADS lesion mass MC When more than one data set was used, we list the size of the largest data set. denotes this paper. The table should be read with the following footnotes. The target task; BI-RADS: BI-RADS prediction, lesion: lesion classification (benign vs. malignant), mass: mass detection, and MC: microcalcification detection. Whether the proposed system is trainable end-to-end. For instance, a system that requires an external system for extracting regions of interest (ROI) is not end-to-end, while a system that uses convolutional networks for both ROI extraction and lesion classification is. In the parentheses there is the number of test images or CV if cross-validation is used. denotes the size of the input image to a deep neural network. Whether multiple views per one exam are utilized. Whether the data reflects natural distribution (N) or controlled distribution (C). Whether the input to a deep neural network is a whole image (IMG) or a small subset (ROI). Did not use images as input to the learning algorithm. recall more often than necessary. Unlike these earlier works, in this paper, we use the full data without artificial balancing of outcomes to ensure that any trained deep convolutional network will closely reflect the natural distribution of outcomes. Collection DATA This is a Health Insurance Portability and Accountability (HIPAA)-compliant, retrospective study approved by our Institutional Review Board. Consecutive screening mammograms for 129,208 patients aged 1 between 19 and 99 (mean: 57.2, std: 11.6) collected within seven years ( ) at five imaging sites affiliated with New York University School of Medicine were used in this study. These imaging centers are located in the New York City metropolitan area (a large academic center and two large ambulatory care practices), where, altogether, over 70,000 mammograms are performed annually. The ethnic makeup of the patient cohort for this study reflects the population pool in NYC, which is 50% Caucasian, 30% African American, 5% Asian and 15% Hispanic. Data Statistics There are altogether 129,208 patients, 201,698 exams and 886,437 images in the data set. We divide the data into training, validation and test sets in the following manner. First we sort data for all patients according to the date of the last 1 When more than one exam for a patient was in the data set, we included the ages of that patient at the time of all exams to compute the values above. visit of the patient. We use the first 80% of the patients in this order as the training data, the next 10% as the validation data and the last 10% as the test data. For each patient in the test set, we evaluate our model s performance in predicting only the label for the last visit of each patient. This reflects best the way the model would be used once deployed. TABLE II DISTRIBUTION OF DATA ASSOCIATED WITH DIFFERENT BI-RADS IN TRAINING, VALIDATION AND TEST DATA. EACH CELL IN THE TABLE HAS THE FOLLOWING FORMAT: NUMBER OF EXAMS / NUMBER OF IMAGES. BI-RADS 0 BI-RADS 1 BI-RADS 2 Training / / / Validation 2634 / / / Test 1341 / / / Data preprocessing and augmentation We normalize the images in the following way. For each image we compute the mean, µ, and the standard deviation, σ, of its pixels. We then subtract µ from each pixel and divide each pixel by σ. Additionally, we flip horizontally the images of R-CC and R-MLO views so that the breast is always on the same side of the image. Since the images vary in size, we crop all of them to the size of pixels. The position of the crop is determined in the following manner. First, the crop area is placed leftmost on the horizontal axis and centrally on the vertical axis. Then noise is added to this position. Let us denote the number of pixels between the top border of the crop area and the top border of the image by b top and analogously define b bottom and b right. We draw a number, t vertical from a uniform distribution U( min(b top, 100), min(b bottom, 100)) and t horizontal from U(0, min(b right, 100)). Finally we translate the position of the crop area by t horizontal pixels horizontally and t vertical pixels vertically. During training this noise is sampled independently every time an image is used. During validation there is no noise added to the position of the crop area. At test time, we feed ten sets of four randomly cropped views to the network. The final prediction is made by averaging predictions for all crops. The aim of this averaging is twofold; first, to use information from outside the center of the image while keeping the size of the input fixed and second, to make prediction of the network more stable. A small fraction of data contains more than one image per view. For such cases one image per view is sampled randomly uniformly each time an exam is used during training and testing. During validation the image with the earliest time stamp is always used. Evaluation Metrics SETTINGS When there are two classes the most frequently applied performance metric is the AUC (area under the ROC curve). However, since there are three classes in our learning task, we cannot apply this metric directly. Instead we compute three AUCs, each time treating one of the three classes as a positive class and the remaining two as negative. We use the macro average of the three AUCs, abbreviated as macauc, as the main performance metric in this work.

5 5 layer kernel size stride #maps repetition global average pooling 256 convolution max pooling convolution max pooling convolution max pooling convolution convolution max pooling convolution input 1 Fig. 2. Description of one deep convolutional network column for a single view. It transforms the input view (a gray-scale image) into a 256-dimensional vector. Classifier p(y x) Fully connected layer (1024 hidden units) Concatenation (256 4 dim) DCN DCN DCN DCN L-CC R-CC L-MLO R-MLO Fig. 3. An overview of the proposed multi-view deep convolutional network. DCN refers to the convolutional network network column from Figure 2. The arrow indicates the direction of information flow. Unlike other widely used nonlinear classifiers, such as a support vector machine or a random forest, a deep convolutional neural network outputs a proper conditional distribution p(y x). It allows us to compute the network s confidence in its prediction by computing the entropy of this distribution, i.e., H(y x) = y C p(y x) log p(y x), (1) where y iterates over all possible classes C. The larger the entropy, the less confident the network is about its prediction. Based on H, we can quantify the change in accuracy (measured by AUC) with respect to the network s confidence. We refer to AUC computed with the most confident 30% of test examples as HC-AUC and we define HC-macAUC to be the average of the three HC-AUC s. Model Setup The overall architecture of our network is shown in Figure 3. Each column corresponding to a different view has an architecture described in Figure 2. We apply the rectifier function after convolutional layers. In addition to augmenting the data by cropping the images at random positions, we regularize the network in three ways. First, we tie the weights in the corresponding columns, i.e., the parameters of the columns processing L-CC and R-CC views are shared as are those of the columns processing L-MLO and R-MLO views. Second, we add Gaussian noise to the input (with the mean of zero and the standard deviation of 0.01). Third, we apply dropout (with a rate of 0.2) after the fully connected layer. We turn off the input noise and dropout during validation and testing. The parameters of the network are initialized using the recipe of Glorot & Bengio [39] and learned using the Adam algorithm [40] with the initial learning rate of Due to the memory limitations of our hardware, the mini-batch size is set to four. We train the network for up to 100 epochs. After each training epoch we compute the macauc on the validation set. We report the test error of the model which achieved the lowest macauc on the validation set. Effect of Scale QUANTITATIVE RESULTS ANALYSIS First, we validate our earlier claim on the need of largescale data for harnessing the most out of deep convolutional neural networks. We train separate networks on the training sets of different sizes; 100%, 50%, 20% and 10%, 5%, 2% and 1% of the original training set. In Table III, we observe that the classification performance improves as the number of training examples increases. This shows the importance of using a large training set. This is consistent with observations made in many other fields such as computer vision, natural language processing and speech recognition [8]. Effect of Resolution We then investigate the effect of resolution of input images. Using the full training set, we train networks with varying input resolutions; scaling both dimensions of the input by 1/8, 1/4 and 1/2. We use bicubic interpolation to downscale the input. When the input resolution is significantly smaller than the original some convolutional layers in the later stages cannot be applied because the size of the feature maps becomes smaller than the size of a convolutional kernel. In that case, we simply skip the remaining layers until the global average pooling. As shown in Table IV, we already see a drop in performance when each dimension of the input was downscaled by half. Further degradation of performance was observed with more aggressive downscaling. Confidence We measure confidence of predictions in terms of the entropy of the output distribution (cf. Equation 1). As shown TABLE III THE EFFECT OF CHANGING THE FRACTION OF THE TRAINING DATA USED. INCREASING THE AMOUNT OF DATA YIELDS BETTER RESULTS. fraction 1% 2% 5% 10% 20% 50% 100% 0 vs. others vs. others vs. others macauc HC-macAUC TABLE IV THE EFFECT OF DECREASING THE RESOLUTION OF THE IMAGE. scale 1/8 1/4 1/2 1 0 vs. others vs. others vs. others macauc HC-macAUC

6 6 true positive rate BI-RADS 0 BI-RADS 1 BI-RADS false positive rate true positive rate BI-RADS 0 BI-RADS 1 BI-RADS false positive rate Fig. 4. ROCs computed with all test data (left) and ROCs computed with test data which the network was confident about (right). ROCs for BI-RADS 1 and BI-RADS 2 classes improve a lot for confident examples while BI-RADS 0 remain similar. in Table V, confident predictions of the proposed model are more accurate. This phenomenon was apparent in in all the experiments (see Table III, Table IV and Figure 4). TABLE V THE ACCURACY (MACAUC) USING THE TOP-P % VALIDATION EXAMPLES ACCORDING TO ENTROPY (BALANCED ACCORDING TO THE PRIOR DISTRIBUTION). THE LOWER THE P, THE MORE CONFIDENT PREDICTIONS. WHEN P = 30%, WE REFER TO THE MACAUC AS A HIGH-CONFIDENCE MACAUC (HC-MACAUC). WHEN REPORTING HC-AUC ON THE TEST SET, WE USE THE CUT-OFF ENTROPY VALUES COMPUTED ON THE VALIDATION SET. P 10% 20% 30% 50% 100% HC-macAUC VISUALIZATION A flip side of high effectiveness of a deep convolutional neural network is the difficulty in interpreting its internal processing. Only recently there have been some efforts on visualizing deep convolutional neural networks for computer vision [41], [42]. These recent approaches however are not computationally efficient and are not easy to apply to medical images due to a number of reasons, including the need for training with a large data set [41] and the availability of good image statistics [42]. Instead, we propose a simpler visualization technique in this paper that does not require any further training. We look at the sensitivity of the network s output to the perturbation of each input pixel. The network outputs the conditional distribution over all the categories, and we can measure the entropy (or confidence) of the predictive distribution H(y x). We can use standard backpropagation to compute H x for the pixel (i, j) of the v-th view. Those input pixels v ij that influence the confidence of the network will have high values, and those that do not contribute much will have low values ( 0). We show two examples of such visualization for patients which were confirmed by a follow-up examination to have breast cancer in Figure 5. CONCLUSION In this paper we have made a first step towards end-to-end large scale training of multi-view deep convolutional networks for breast cancer screening. We have shown experimentally that it is essential to keep the images at high-resolution. We expect this to hold for other learning tasks with medical images where fine details determine the outcome. We also demonstrated it is necessary to use a large number of exams. Although we used the largest breast cancer screening data set ever reported in literature, the performance of our model has not saturated and is expected to improve with more data. Our network s performance was lowest on differentiating BI-RADS 0 from the other classes. We attribute it to the issue of noise in BI-RADS labeling. Doctors often disagree on how a particular exam should be classified [3] and in fact, less than 1% of the screening population has cancer [2], [5]. To alleviate this problem we can instead use the information on whether a person actually went on to develop breast cancer in the future as a label. It is also worth noting that, because of limited computational resources, we had to heavily rely on our experience in the choice of learning hyperparameters. We did not perform a systematic search for optimal hyperparameters, which often has a great impact on the performance of a neural network in limited data scenarios [43], [44]. The methods we used in this work are powerful and our results can be improved simply by the means of applying more computational resources without significantly changing the methodology. ACKNOWLEDGMENT We would like to thank Jure Žbontar, Yann LeCun, Pablo Sprechmann, Cem Deniz, Jingyi Su, Nan Wu and Masha Zorin

7 7 (a) BI-RADS 0 (b) BI-RADS 2 Fig. 5. Examples of visualizations of decisions made by our model. On the left in both (a) and (b) panels there are images of breast with arrows indicating possible suspicious findings. On the right in both (a) and (b) panels there are the same images as on the left in the corresponding panels with regions of the images (highlighted in red) which influence confidence of predictions of our neural network. Please note that our visualization highlights parts of the image that are relevant for all classes (BI-RADS 0, BI-RADS 1, and BI-RADS 2) and that those highlighted areas include locations indicated in the images on the left. Panel (a) shows the right breast of a 61 years old patient who was assigned BI-RADS 0 in her screening mammography. Biopsy confirmed that the finding indicated in the image by the right arrow was invasive ductal carcinoma. Panel (b) shows the left breast of a 62 years old patient. She had a prior breast surgery and a biopsy marker in one of her breasts as indicated by the blue arrows. Our neural network correctly and confidently predicted that the artifacts were benign and indicated scar markers and a biopsy clip. for insightful comments on this work. We would also like to thank Joe Katsnelson and Mario Videna for their efforts in supporting our computing environment. R EFERENCES [1] R. L. Siegel, K. D. Miller, and A. Jemal, Cancer statistics, 2015, CA: a cancer journal for clinicians, vol. 65, no. 1, pp. 5 29, [2] S. W. Duffy, L. Tabar, H. H. Chen, M. Holmqvist, M. F. Yen, S. Abdsalah, B. Epstein, E. Frodis, E. Ljungberg, C. Hedborg-Melander, A. Sundbom, M. Tholin, M. Wiege, A. Akerlund, H. M. Wu, T. S. Tung, Y. H. Chiu, C. P. Chiu, C. C. Huang, R. A. Smith, M. Rosen, M. Stenbeck, and L. Holmberg, The impact of organized mammography service screening on breast carcinoma mortality in seven swedish counties, Cancer, vol. 95, no. 3, pp , [3] D. B. Kopans, Beyond randomized controlled trials: organized mammographic screening substantially reduces breast carcinoma mortality, Cancer, vol. 94, no. 2, pp ; author reply 581 3, [Online]. Available: [4] S. W. Duffy, L. Tabar, and R. A. Smith, The mammographic screening trials: commentary on the recent work by Olsen and Gotzsche, CA Cancer J Clin, vol. 52, no. 2, pp , [Online]. Available: [5] L. Tabar, B. Vitak, H. H. Chen, M. F. Yen, S. W. Duffy, and R. A. Smith, Beyond randomized controlled trials: organized mammographic screening substantially reduces breast carcinoma mortality, Cancer, vol. 91, no. 9, pp , [Online]. Available: [6] A. N. Tosteson, D. G. Fryback, C. S. Hammond, L. G. Hanna, M. R. Grove, M. Brown, Q. Wang, K. Lindfors, and E. D. Pisano, Consequences of false-positive screening mammograms, JAMA internal medicine, vol. 174, no. 6, pp , [7] D. B. Kopans, An open letter to panels that are deciding guidelines for breast cancer screening, Breast Cancer Res Treat, vol. 151, no. 1, pp , [Online]. Available: http: // [8] Y. LeCun, Y. Bengio, and G. Hinton, Deep learning, Nature, vol. 521, no. 7553, pp , [9] Y. LeCun, B. Boser, J. S. Denker, D. Henderson, R. E. Howard, W. Hubbard, and L. D. Jackel, Backpropagation applied to handwritten zip code recognition, Neural computation, vol. 1, no. 4, pp , [10] Y. LeCun, L. Bottou, Y. Bengio, and P. Haffner, Gradient-based learning applied to document recognition, Proceedings of the IEEE, vol. 86, no. 11, pp , [11] J. Ngiam, A. Khosla, M. Kim, J. Nam, H. Lee, and A. Y. Ng, Multimodal deep learning, in International conference on machine learning, 2011, pp [12] N. Srivastava and R. Salakhutdinov, Multimodal learning with deep boltzmann machines, in Advances in neural information processing systems, 2012, pp [13] W. Wang, R. Arora, K. Livescu, and J. Bilmes, On deep multiview representation learning, in International Conference on Machine Learning, 2015, pp [14] H. Su, S. Maji, E. Kalogerakis, and E. Learned-Miller, Multi-view convolutional neural networks for 3d shape recognition, in Proceedings of the IEEE International Conference on Computer Vision, 2015, pp [15] D. E. Rumelhart, G. E. Hinton, and R. J. Williams, Learning internal representations by error-propagation, in Parallel Distributed Processing: Explorations in the Microstructure of Cognition. Volume 1. MIT Press, Cambridge, MA, 1986, vol. 1, no. 6088, pp [16] A. Krizhevsky, I. Sutskever, and G. E. Hinton, ImageNet classification with deep convolutional neural networks, in Advances in neural information processing systems, 2012, pp [17] N. Srivastava, G. E. Hinton, A. Krizhevsky, I. Sutskever, and R. Salakhutdinov, Dropout: a simple way to prevent neural networks from overfitting. Journal of Machine Learning Research, vol. 15, no. 1, pp , [18] K. He, X. Zhang, S. Ren, and J. Sun, Deep residual learning for image recognition, in IEEE Conference on Computer Vision and Pattern Recognition, [19] M. Lin, Q. Chen, and S. Yan, Network in network, in International Conference on Learning Representations, [20] K. Simonyan and A. Zisserman, Very deep convolutional networks for large-scale image recognition, in International Conference on Learning Representation, [21] C. Szegedy, W. Liu, Y. Jia, P. Sermanet, S. Reed, D. Anguelov, D. Erhan, V. Vanhoucke, and A. Rabinovich, Going deeper with convolutions, in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2015, pp [22] I. Domingues and J. S. Cardoso, Mass detection on mammogram images: a first assessment of deep learning techniques, [23] N. Dhungel, G. Carneiro, and A. P. Bradley, Automated mass detection in mammograms using cascaded deep learning and random forests, in International Conference on Digital Image Computing: Techniques and Applications. IEEE, 2015, pp [24] M. G. Ertosun and D. L. Rubin, Probabilistic visual search for masses within mammography images using deep learning, in IEEE International Conference on Bioinformatics and Biomedicine. IEEE, 2015, pp

8 8 [25] B. Q. Huynh, H. Li, and M. L. Giger, Digital mammographic tumor classification using transfer learning from deep convolutional neural networks, Journal of Medical Imaging, vol. 3, no. 3, pp , [26] D. Lévy and A. Jain, Breast mass classification from mammograms using deep convolutional neural networks, arxiv: , [27] J. Arevalo, F. A. González, R. Ramos-Pollán, J. L. Oliveira, and M. A. G. Lopez, Representation learning for mammography mass lesion classification with convolutional neural networks, Computer methods and programs in biomedicine, vol. 127, pp , [28] J.-J. Mordang, T. Janssen, A. Bria, T. Kooi, A. Gubern-Mérida, and N. Karssemeijer, Automatic microcalcification detection in multivendor mammography using convolutional neural networks, in International Workshop on Digital Mammography. Springer, 2016, pp [29] T. Kooi, G. Litjens, B. van Ginneken, A. Gubern-Mérida, C. I. Sánchez, R. Mann, A. den Heeten, and N. Karssemeijer, Large scale deep learning for computer aided detection of mammographic lesions, Medical image analysis, vol. 35, pp , [30] A. S. Becker, M. Marcon, S. Ghafoor, M. C. Wurnig, T. Frauenfelder, and A. Boss, Deep learning in mammography: Diagnostic accuracy of a multipurpose image analysis software in the detection of breast cancer. Investigative Radiology, [31] A. Akselrod-Ballin, L. Karlinsky, S. Alpert, S. Hasoul, R. Ben-Ari, and E. Barkan, A region based convolutional network for tumor detection and classification in breast mammography, in International Workshop on Large-Scale Annotation of Biomedical Data and Expert Label Synthesis. Springer, 2016, pp [32] W. Zhu, Q. Lou, Y. S. Vang, and X. Xie, Deep multi-instance networks with sparse label assignment for whole mammogram classification, arxiv: , [33] G. Carneiro, J. Nascimento, and A. P. Bradley, Unregistered multiview mammogram analysis with pre-trained deep learning models, in International Conference on Medical Image Computing and Computer- Assisted Intervention. Springer, 2015, pp [34] I. C. Moreira, I. Amaral, I. Domingues, A. Cardoso, M. J. Cardoso, and J. S. Cardoso, Inbreast: toward a full-field digital mammographic database, Academic radiology, vol. 19, no. 2, pp , [35] K. Bowyer, D. Kopans, W. Kegelmeyer, R. Moore, M. Sallam, K. Chang, and K. Woods, The digital database for screening mammography, in Third international workshop on digital mammography, vol. 58, 1996, p. 27. [36] M. Heath, K. Bowyer, D. Kopans, P. Kegelmeyer Jr, R. Moore, K. Chang, and S. Munishkumaran, Current status of the digital database for screening mammography, in Digital mammography. Springer, 1998, pp [37] J. Wang, X. Yang, H. Cai, W. Tan, C. Jin, and L. Li, Discrimination of breast cancer with microcalcifications on mammography by deep learning, Scientific reports, vol. 6, p , [38] A. J. Bekker, H. Greenspan, and J. Goldberger, A multi-view deep learning architecture for classification of breast microcalcifications, in IEEE International Symposium on Biomedical Imaging, 2016, pp [39] X. Glorot and Y. Bengio, Understanding the difficulty of training deep feedforward neural networks. in International Conference on Artificial Intelligence and Statistics, [40] D. Kingma and J. Ba, Adam: A method for stochastic optimization, in International Conference on Learning Representation, [41] M. D. Zeiler and R. Fergus, Visualizing and understanding convolutional networks, in European Conference on Computer Vision. Springer, 2014, pp [42] J. Yosinski, J. Clune, A. Nguyen, T. Fuchs, and H. Lipson, Understanding neural networks through deep visualization, arxiv: , [43] J. S. Bergstra, R. Bardenet, Y. Bengio, and B. Kégl, Algorithms for hyper-parameter optimization, in Advances in Neural Information Processing Systems 24, 2011, pp [44] J. Snoek, O. Rippel, K. Swersky, R. Kiros, N. Satish, N. Sundaram, M. M. A. Patwary, M. Prabhat, and R. P. Adams, Scalable bayesian optimization using deep neural networks, in International Conference on Machine Learning, 2015.

High-Resolution Breast Cancer Screening with Multi-View Deep Convolutional Neural Networks

High-Resolution Breast Cancer Screening with Multi-View Deep Convolutional Neural Networks High-Resolution Breast Cancer Screening with Multi-View Deep Convolutional Neural Networks Krzysztof J. Geras Joint work with Kyunghyun Cho, Linda Moy, Gene Kim, Stacey Wolfson and Artie Shen. GTC 2017

More information

arxiv: v1 [cs.cv] 21 Jul 2017

arxiv: v1 [cs.cv] 21 Jul 2017 A Multi-Scale CNN and Curriculum Learning Strategy for Mammogram Classification William Lotter 1,2, Greg Sorensen 2, and David Cox 1,2 1 Harvard University, Cambridge MA, USA 2 DeepHealth Inc., Cambridge

More information

arxiv: v2 [cs.cv] 8 Mar 2018

arxiv: v2 [cs.cv] 8 Mar 2018 Automated soft tissue lesion detection and segmentation in digital mammography using a u-net deep learning network Timothy de Moor a, Alejandro Rodriguez-Ruiz a, Albert Gubern Mérida a, Ritse Mann a, and

More information

CSE Introduction to High-Perfomance Deep Learning ImageNet & VGG. Jihyung Kil

CSE Introduction to High-Perfomance Deep Learning ImageNet & VGG. Jihyung Kil CSE 5194.01 - Introduction to High-Perfomance Deep Learning ImageNet & VGG Jihyung Kil ImageNet Classification with Deep Convolutional Neural Networks Alex Krizhevsky, Ilya Sutskever, Geoffrey E. Hinton,

More information

Understanding Convolutional Neural

Understanding Convolutional Neural Understanding Convolutional Neural Networks Understanding Convolutional Neural Networks David Stutz July 24th, 2014 David Stutz July 24th, 2014 0/53 1/53 Table of Contents - Table of Contents 1 Motivation

More information

Investigating the performance of a CAD x scheme for mammography in specific BIRADS categories

Investigating the performance of a CAD x scheme for mammography in specific BIRADS categories Investigating the performance of a CAD x scheme for mammography in specific BIRADS categories Andreadis I., Nikita K. Department of Electrical and Computer Engineering National Technical University of

More information

Mammogram Analysis: Tumor Classification

Mammogram Analysis: Tumor Classification Mammogram Analysis: Tumor Classification Literature Survey Report Geethapriya Raghavan geeragh@mail.utexas.edu EE 381K - Multidimensional Digital Signal Processing Spring 2005 Abstract Breast cancer is

More information

A convolutional neural network to classify American Sign Language fingerspelling from depth and colour images

A convolutional neural network to classify American Sign Language fingerspelling from depth and colour images A convolutional neural network to classify American Sign Language fingerspelling from depth and colour images Ameen, SA and Vadera, S http://dx.doi.org/10.1111/exsy.12197 Title Authors Type URL A convolutional

More information

Deep learning and non-negative matrix factorization in recognition of mammograms

Deep learning and non-negative matrix factorization in recognition of mammograms Deep learning and non-negative matrix factorization in recognition of mammograms Bartosz Swiderski Faculty of Applied Informatics and Mathematics Warsaw University of Life Sciences, Warsaw, Poland bartosz_swiderski@sggw.pl

More information

arxiv: v1 [stat.ml] 23 Jan 2017

arxiv: v1 [stat.ml] 23 Jan 2017 Learning what to look in chest X-rays with a recurrent visual attention model arxiv:1701.06452v1 [stat.ml] 23 Jan 2017 Petros-Pavlos Ypsilantis Department of Biomedical Engineering King s College London

More information

Mammogram Analysis: Tumor Classification

Mammogram Analysis: Tumor Classification Mammogram Analysis: Tumor Classification Term Project Report Geethapriya Raghavan geeragh@mail.utexas.edu EE 381K - Multidimensional Digital Signal Processing Spring 2005 Abstract Breast cancer is the

More information

CS-E Deep Learning Session 4: Convolutional Networks

CS-E Deep Learning Session 4: Convolutional Networks CS-E4050 - Deep Learning Session 4: Convolutional Networks Jyri Kivinen Aalto University 23 September 2015 Credits: Thanks to Tapani Raiko for slides material. CS-E4050 - Deep Learning Session 4: Convolutional

More information

arxiv: v1 [cs.cv] 18 Dec 2016

arxiv: v1 [cs.cv] 18 Dec 2016 Deep Multi-instance Networks with Sparse Label Assignment for Whole Mammogram Classification Wentao Zhu, Qi Lou, Yeeleng Scott Vang, and Xiaohui Xie {wentaoz1, xhx}@ics.uci.edu, {qlou, ysvang}@uci.edu

More information

Deep-Learning Based Semantic Labeling for 2D Mammography & Comparison of Complexity for Machine Learning Tasks

Deep-Learning Based Semantic Labeling for 2D Mammography & Comparison of Complexity for Machine Learning Tasks Deep-Learning Based Semantic Labeling for 2D Mammography & Comparison of Complexity for Machine Learning Tasks Paul H. Yi, MD, Abigail Lin, BSE, Jinchi Wei, BSE, Haris I. Sair, MD, Ferdinand K. Hui, MD,

More information

A HMM-based Pre-training Approach for Sequential Data

A HMM-based Pre-training Approach for Sequential Data A HMM-based Pre-training Approach for Sequential Data Luca Pasa 1, Alberto Testolin 2, Alessandro Sperduti 1 1- Department of Mathematics 2- Department of Developmental Psychology and Socialisation University

More information

Digital breast tomosynthesis

Digital breast tomosynthesis GE Healthcare Digital breast tomosynthesis Daniel B. Kopans, M.D., F.A.C.R. Professor of Radiology Harvard Medical School Senior Radiologist - Breast Imaging Division Massachusetts General Hospital Since

More information

Automated Mass Detection from Mammograms using Deep Learning and Random Forest

Automated Mass Detection from Mammograms using Deep Learning and Random Forest Automated Mass Detection from Mammograms using Deep Learning and Random Forest Neeraj Dhungel 1 Gustavo Carneiro 1 Andrew P. Bradley 2 1 ACVT, University of Adelaide, Australia 2 University of Queensland,

More information

Expert identification of visual primitives used by CNNs during mammogram classification

Expert identification of visual primitives used by CNNs during mammogram classification Expert identification of visual primitives used by CNNs during mammogram classification Jimmy Wu a, Diondra Peck b, Scott Hsieh c, Vandana Dialani, MD d, Constance D. Lehman, MD e, Bolei Zhou a, Vasilis

More information

Classification of breast cancer histology images using transfer learning

Classification of breast cancer histology images using transfer learning Classification of breast cancer histology images using transfer learning Sulaiman Vesal 1 ( ), Nishant Ravikumar 1, AmirAbbas Davari 1, Stephan Ellmann 2, Andreas Maier 1 1 Pattern Recognition Lab, Friedrich-Alexander-Universität

More information

COMPARATIVE STUDY ON FEATURE EXTRACTION METHOD FOR BREAST CANCER CLASSIFICATION

COMPARATIVE STUDY ON FEATURE EXTRACTION METHOD FOR BREAST CANCER CLASSIFICATION COMPARATIVE STUDY ON FEATURE EXTRACTION METHOD FOR BREAST CANCER CLASSIFICATION 1 R.NITHYA, 2 B.SANTHI 1 Asstt Prof., School of Computing, SASTRA University, Thanjavur, Tamilnadu, India-613402 2 Prof.,

More information

NMF-Density: NMF-Based Breast Density Classifier

NMF-Density: NMF-Based Breast Density Classifier NMF-Density: NMF-Based Breast Density Classifier Lahouari Ghouti and Abdullah H. Owaidh King Fahd University of Petroleum and Minerals - Department of Information and Computer Science. KFUPM Box 1128.

More information

FULLY AUTOMATED CLASSIFICATION OF MAMMOGRAMS USING DEEP RESIDUAL NEURAL NETWORKS. Gustavo Carneiro

FULLY AUTOMATED CLASSIFICATION OF MAMMOGRAMS USING DEEP RESIDUAL NEURAL NETWORKS. Gustavo Carneiro FULLY AUTOMATED CLASSIFICATION OF MAMMOGRAMS USING DEEP RESIDUAL NEURAL NETWORKS Neeraj Dhungel Gustavo Carneiro Andrew P. Bradley? Electrical and Computer Engineering, The University of British Columbia,

More information

Automated diagnosis of pneumothorax using an ensemble of convolutional neural networks with multi-sized chest radiography images

Automated diagnosis of pneumothorax using an ensemble of convolutional neural networks with multi-sized chest radiography images Automated diagnosis of pneumothorax using an ensemble of convolutional neural networks with multi-sized chest radiography images Tae Joon Jun, Dohyeun Kim, and Daeyoung Kim School of Computing, KAIST,

More information

Convolutional Neural Networks (CNN)

Convolutional Neural Networks (CNN) Convolutional Neural Networks (CNN) Algorithm and Some Applications in Computer Vision Luo Hengliang Institute of Automation June 10, 2014 Luo Hengliang (Institute of Automation) Convolutional Neural Networks

More information

Lung Nodule Segmentation Using 3D Convolutional Neural Networks

Lung Nodule Segmentation Using 3D Convolutional Neural Networks Lung Nodule Segmentation Using 3D Convolutional Neural Networks Research paper Business Analytics Bernard Bronmans Master Business Analytics VU University, Amsterdam Evert Haasdijk Supervisor VU University,

More information

Deep CNNs for Diabetic Retinopathy Detection

Deep CNNs for Diabetic Retinopathy Detection Deep CNNs for Diabetic Retinopathy Detection Alex Tamkin Stanford University atamkin@stanford.edu Iain Usiri Stanford University iusiri@stanford.edu Chala Fufa Stanford University cfufa@stanford.edu 1

More information

Multi-attention Guided Activation Propagation in CNNs

Multi-attention Guided Activation Propagation in CNNs Multi-attention Guided Activation Propagation in CNNs Xiangteng He and Yuxin Peng (B) Institute of Computer Science and Technology, Peking University, Beijing, China pengyuxin@pku.edu.cn Abstract. CNNs

More information

COMP9444 Neural Networks and Deep Learning 5. Convolutional Networks

COMP9444 Neural Networks and Deep Learning 5. Convolutional Networks COMP9444 Neural Networks and Deep Learning 5. Convolutional Networks Textbook, Sections 6.2.2, 6.3, 7.9, 7.11-7.13, 9.1-9.5 COMP9444 17s2 Convolutional Networks 1 Outline Geometry of Hidden Unit Activations

More information

Differentiating Tumor and Edema in Brain Magnetic Resonance Images Using a Convolutional Neural Network

Differentiating Tumor and Edema in Brain Magnetic Resonance Images Using a Convolutional Neural Network Original Article Differentiating Tumor and Edema in Brain Magnetic Resonance Images Using a Convolutional Neural Network Aida Allahverdi 1, Siavash Akbarzadeh 1, Alireza Khorrami Moghaddam 2, Armin Allahverdy

More information

arxiv: v2 [cs.cv] 19 Dec 2017

arxiv: v2 [cs.cv] 19 Dec 2017 An Ensemble of Deep Convolutional Neural Networks for Alzheimer s Disease Detection and Classification arxiv:1712.01675v2 [cs.cv] 19 Dec 2017 Jyoti Islam Department of Computer Science Georgia State University

More information

AN ALGORITHM FOR EARLY BREAST CANCER DETECTION IN MAMMOGRAMS

AN ALGORITHM FOR EARLY BREAST CANCER DETECTION IN MAMMOGRAMS AN ALGORITHM FOR EARLY BREAST CANCER DETECTION IN MAMMOGRAMS Isaac N. Bankman', William A. Christens-Barryl, Irving N. Weinberg2, Dong W. Kim3, Ralph D. Semmell, and William R. Brody2 The Johns Hopkins

More information

Using Deep Convolutional Neural Networks to Predict Semantic Features of Lesions in Mammograms

Using Deep Convolutional Neural Networks to Predict Semantic Features of Lesions in Mammograms Using Deep Convolutional Neural Networks to Predict Semantic Features of Lesions in Mammograms Vibhu Agarwal Stanford University vibhua@stanford.edu Clayton Carson Stanford University carsoncz@stanford.edu

More information

A Deep Learning Approach for Breast Cancer Mass Detection

A Deep Learning Approach for Breast Cancer Mass Detection A Deep Learning Approach for Breast Cancer Mass Detection Wael E.Fathy 1, Amr S. Ghoneim 2 Teaching Assistant 1, Assistant Professor 2 Department of Computer Science, Faculty of Computers and Information

More information

arxiv: v3 [stat.ml] 27 Mar 2018

arxiv: v3 [stat.ml] 27 Mar 2018 ATTACKING THE MADRY DEFENSE MODEL WITH L 1 -BASED ADVERSARIAL EXAMPLES Yash Sharma 1 and Pin-Yu Chen 2 1 The Cooper Union, New York, NY 10003, USA 2 IBM Research, Yorktown Heights, NY 10598, USA sharma2@cooper.edu,

More information

DeepMiner: Discovering Interpretable Representations for Mammogram Classification and Explanation

DeepMiner: Discovering Interpretable Representations for Mammogram Classification and Explanation DeepMiner: Discovering Interpretable Representations for Mammogram Classification and Explanation Jimmy Wu 1, Bolei Zhou 1, Diondra Peck 2, Scott Hsieh 3, Vandana Dialani, MD 4 Lester Mackey 5, and Genevieve

More information

arxiv: v1 [cs.cv] 15 Aug 2018

arxiv: v1 [cs.cv] 15 Aug 2018 Ensemble of Convolutional Neural Networks for Dermoscopic Images Classification Tomáš Majtner 1, Buda Bajić 2, Sule Yildirim 3, Jon Yngve Hardeberg 3, Joakim Lindblad 4,5 and Nataša Sladoje 4,5 arxiv:1808.05071v1

More information

arxiv: v2 [cs.cv] 22 Mar 2018

arxiv: v2 [cs.cv] 22 Mar 2018 Deep saliency: What is learnt by a deep network about saliency? Sen He 1 Nicolas Pugeault 1 arxiv:1801.04261v2 [cs.cv] 22 Mar 2018 Abstract Deep convolutional neural networks have achieved impressive performance

More information

Building an Ensemble System for Diagnosing Masses in Mammograms

Building an Ensemble System for Diagnosing Masses in Mammograms Building an Ensemble System for Diagnosing Masses in Mammograms Yu Zhang, Noriko Tomuro, Jacob Furst, Daniela Stan Raicu College of Computing and Digital Media DePaul University, Chicago, IL 60604, USA

More information

Classification of benign and malignant masses in breast mammograms

Classification of benign and malignant masses in breast mammograms Classification of benign and malignant masses in breast mammograms A. Šerifović-Trbalić*, A. Trbalić**, D. Demirović*, N. Prljača* and P.C. Cattin*** * Faculty of Electrical Engineering, University of

More information

Less is More: Culling the Training Set to Improve Robustness of Deep Neural Networks

Less is More: Culling the Training Set to Improve Robustness of Deep Neural Networks Less is More: Culling the Training Set to Improve Robustness of Deep Neural Networks Yongshuai Liu, Jiyu Chen, and Hao Chen University of California, Davis {yshliu, jiych, chen}@ucdavis.edu Abstract. Deep

More information

arxiv: v2 [cs.cv] 7 Jun 2018

arxiv: v2 [cs.cv] 7 Jun 2018 Deep supervision with additional labels for retinal vessel segmentation task Yishuo Zhang and Albert C.S. Chung Lo Kwee-Seong Medical Image Analysis Laboratory, Department of Computer Science and Engineering,

More information

arxiv: v1 [cs.cv] 17 Aug 2017

arxiv: v1 [cs.cv] 17 Aug 2017 Deep Learning for Medical Image Analysis Mina Rezaei, Haojin Yang, Christoph Meinel Hasso Plattner Institute, Prof.Dr.Helmert-Strae 2-3, 14482 Potsdam, Germany {mina.rezaei,haojin.yang,christoph.meinel}@hpi.de

More information

arxiv: v2 [cs.lg] 1 Jun 2018

arxiv: v2 [cs.lg] 1 Jun 2018 Shagun Sodhani 1 * Vardaan Pahuja 1 * arxiv:1805.11016v2 [cs.lg] 1 Jun 2018 Abstract Self-play (Sukhbaatar et al., 2017) is an unsupervised training procedure which enables the reinforcement learning agents

More information

Do Deep Neural Networks Suffer from Crowding?

Do Deep Neural Networks Suffer from Crowding? Do Deep Neural Networks Suffer from Crowding? Anna Volokitin Gemma Roig ι Tomaso Poggio voanna@vision.ee.ethz.ch gemmar@mit.edu tp@csail.mit.edu Center for Brains, Minds and Machines, Massachusetts Institute

More information

A CONVOLUTION NEURAL NETWORK ALGORITHM FOR BRAIN TUMOR IMAGE SEGMENTATION

A CONVOLUTION NEURAL NETWORK ALGORITHM FOR BRAIN TUMOR IMAGE SEGMENTATION A CONVOLUTION NEURAL NETWORK ALGORITHM FOR BRAIN TUMOR IMAGE SEGMENTATION 1 Priya K, 2 Dr.O. Saraniya 1 PG Scholar, 2 Assistant Professor Department Of ECE Government College of Technology, Coimbatore,

More information

Investigation of multiorientation and multiresolution features for microcalcifications classification in mammograms

Investigation of multiorientation and multiresolution features for microcalcifications classification in mammograms Investigation of multiorientation and multiresolution features for microcalcifications classification in mammograms Aqilah Baseri Huddin, Brian W.-H. Ng, Derek Abbott 3 School of Electrical and Electronic

More information

Convolutional capsule network for classification of breast cancer histology images

Convolutional capsule network for classification of breast cancer histology images Convolutional capsule network for classification of breast cancer histology images Tomas Iesmantas 1 and Robertas Alzbutas 1 1 Kaunas University of Technology, K. Donelaičio g. 73, Kaunas 44249 tomas.iesmantas@ktu.lt

More information

arxiv: v1 [cs.cv] 26 Feb 2018

arxiv: v1 [cs.cv] 26 Feb 2018 Classification of breast cancer histology images using transfer learning Sulaiman Vesal 1, Nishant Ravikumar 1, AmirAbbas Davari 1, Stephan Ellmann 2, Andreas Maier 1 arxiv:1802.09424v1 [cs.cv] 26 Feb

More information

arxiv: v1 [cs.ai] 28 Nov 2017

arxiv: v1 [cs.ai] 28 Nov 2017 : a better way of the parameters of a Deep Neural Network arxiv:1711.10177v1 [cs.ai] 28 Nov 2017 Guglielmo Montone Laboratoire Psychologie de la Perception Université Paris Descartes, Paris montone.guglielmo@gmail.com

More information

Skin cancer reorganization and classification with deep neural network

Skin cancer reorganization and classification with deep neural network Skin cancer reorganization and classification with deep neural network Hao Chang 1 1. Department of Genetics, Yale University School of Medicine 2. Email: changhao86@gmail.com Abstract As one kind of skin

More information

DEEP LEARNING-BASED ASSESSMENT OF TUMOR-ASSOCIATED STROMA FOR DIAGNOSING BREAST CANCER IN HISTOPATHOLOGY IMAGES

DEEP LEARNING-BASED ASSESSMENT OF TUMOR-ASSOCIATED STROMA FOR DIAGNOSING BREAST CANCER IN HISTOPATHOLOGY IMAGES DEEP LEARNING-BASED ASSESSMENT OF TUMOR-ASSOCIATED STROMA FOR DIAGNOSING BREAST CANCER IN HISTOPATHOLOGY IMAGES Babak Ehteshami Bejnordi, Jimmy Lin, Ben Glass, Maeve Mullooly, Gretchen L Gierach, Mark

More information

arxiv: v1 [cs.cv] 30 May 2018

arxiv: v1 [cs.cv] 30 May 2018 A Robust and Effective Approach Towards Accurate Metastasis Detection and pn-stage Classification in Breast Cancer Byungjae Lee and Kyunghyun Paeng Lunit inc., Seoul, South Korea {jaylee,khpaeng}@lunit.io

More information

Y-Net: Joint Segmentation and Classification for Diagnosis of Breast Biopsy Images

Y-Net: Joint Segmentation and Classification for Diagnosis of Breast Biopsy Images Y-Net: Joint Segmentation and Classification for Diagnosis of Breast Biopsy Images Sachin Mehta 1, Ezgi Mercan 1, Jamen Bartlett 2, Donald Weaver 2, Joann G. Elmore 1, and Linda Shapiro 1 1 University

More information

HHS Public Access Author manuscript Med Image Comput Comput Assist Interv. Author manuscript; available in PMC 2018 January 04.

HHS Public Access Author manuscript Med Image Comput Comput Assist Interv. Author manuscript; available in PMC 2018 January 04. Discriminative Localization in CNNs for Weakly-Supervised Segmentation of Pulmonary Nodules Xinyang Feng 1, Jie Yang 1, Andrew F. Laine 1, and Elsa D. Angelini 1,2 1 Department of Biomedical Engineering,

More information

Mammographic Breast Density Classification by a Deep Learning Approach

Mammographic Breast Density Classification by a Deep Learning Approach Mammographic Breast Density Classification by a Deep Learning Approach Aly Mohamed, PhD Robert Nishikawa, PhD Wendie A. Berg, MD, PhD David Gur, ScD Shandong Wu, PhD Department of Radiology, University

More information

Sign Language Recognition using Convolutional Neural Networks

Sign Language Recognition using Convolutional Neural Networks Sign Language Recognition using Convolutional Neural Networks Lionel Pigou, Sander Dieleman, Pieter-Jan Kindermans, Benjamin Schrauwen Ghent University, ELIS, Belgium Abstract. There is an undeniable communication

More information

An Artificial Neural Network Architecture Based on Context Transformations in Cortical Minicolumns

An Artificial Neural Network Architecture Based on Context Transformations in Cortical Minicolumns An Artificial Neural Network Architecture Based on Context Transformations in Cortical Minicolumns 1. Introduction Vasily Morzhakov, Alexey Redozubov morzhakovva@gmail.com, galdrd@gmail.com Abstract Cortical

More information

Automated detection of masses on whole breast volume ultrasound scanner: false positive reduction using deep convolutional neural network

Automated detection of masses on whole breast volume ultrasound scanner: false positive reduction using deep convolutional neural network Automated detection of masses on whole breast volume ultrasound scanner: false positive reduction using deep convolutional neural network Yuya Hiramatsu a, Chisako Muramatsu* a, Hironobu Kobayashi b, Takeshi

More information

Mammography is a most effective imaging modality in early breast cancer detection. The radiographs are searched for signs of abnormality by expert

Mammography is a most effective imaging modality in early breast cancer detection. The radiographs are searched for signs of abnormality by expert Abstract Methodologies for early detection of breast cancer still remain an open problem in the Research community. Breast cancer continues to be a significant problem in the contemporary world. Nearly

More information

Classification of Mammograms using Gray-level Co-occurrence Matrix and Support Vector Machine Classifier

Classification of Mammograms using Gray-level Co-occurrence Matrix and Support Vector Machine Classifier Classification of Mammograms using Gray-level Co-occurrence Matrix and Support Vector Machine Classifier P.Samyuktha,Vasavi College of engineering,cse dept. D.Sriharsha, IDD, Comp. Sc. & Engg., IIT (BHU),

More information

Mammographic Cancer Detection and Classification Using Bi Clustering and Supervised Classifier

Mammographic Cancer Detection and Classification Using Bi Clustering and Supervised Classifier Mammographic Cancer Detection and Classification Using Bi Clustering and Supervised Classifier R.Pavitha 1, Ms T.Joyce Selva Hephzibah M.Tech. 2 PG Scholar, Department of ECE, Indus College of Engineering,

More information

Fully Convolutional Network-based Multi-Task Learning for Rectum and Rectal Cancer Segmentation

Fully Convolutional Network-based Multi-Task Learning for Rectum and Rectal Cancer Segmentation Fully Convolutional Network-based Multi-Task Learning for Rectum and Rectal Cancer Segmentation Joohyung Lee 1, Ji Eun Oh 1, Min Ju Kim 2, Bo Yun Hur 2, Sun Ah Cho 1 1, 2*, and Dae Kyung Sohn 1 Innovative

More information

arxiv: v2 [cs.cv] 3 Jun 2018

arxiv: v2 [cs.cv] 3 Jun 2018 S4ND: Single-Shot Single-Scale Lung Nodule Detection Naji Khosravan and Ulas Bagci Center for Research in Computer Vision (CRCV), School of Computer Science, University of Central Florida, Orlando, FL.

More information

Using Deep Convolutional Networks for Gesture Recognition in American Sign Language

Using Deep Convolutional Networks for Gesture Recognition in American Sign Language Using Deep Convolutional Networks for Gesture Recognition in American Sign Language Abstract In the realm of multimodal communication, sign language is, and continues to be, one of the most understudied

More information

Neural Network Based Technique to Locate and Classify Microcalcifications in Digital Mammograms

Neural Network Based Technique to Locate and Classify Microcalcifications in Digital Mammograms Neural Network Based Technique to Locate and Classify Microcalcifications in Digital Mammograms Author Verma, Brijesh Published 1998 Conference Title 1998 IEEE World Congress on Computational Intelligence

More information

arxiv: v3 [stat.ml] 28 Oct 2017

arxiv: v3 [stat.ml] 28 Oct 2017 Interpretable Deep Learning applied to Plant Stress Phenotyping arxiv:1710.08619v3 [stat.ml] 28 Oct 2017 Sambuddha Ghosal sghosal@iastate.edu Asheesh K. Singh singhak@iastate.edu Arti Singh arti@iastate.edu

More information

Computer aided detection of clusters of microcalcifications on full field digital mammograms

Computer aided detection of clusters of microcalcifications on full field digital mammograms Computer aided detection of clusters of microcalcifications on full field digital mammograms Jun Ge, a Berkman Sahiner, Lubomir M. Hadjiiski, Heang-Ping Chan, Jun Wei, Mark A. Helvie, and Chuan Zhou Department

More information

3D Deep Learning for Multi-modal Imaging-Guided Survival Time Prediction of Brain Tumor Patients

3D Deep Learning for Multi-modal Imaging-Guided Survival Time Prediction of Brain Tumor Patients 3D Deep Learning for Multi-modal Imaging-Guided Survival Time Prediction of Brain Tumor Patients Dong Nie 1,2, Han Zhang 1, Ehsan Adeli 1, Luyan Liu 1, and Dinggang Shen 1(B) 1 Department of Radiology

More information

ITERATIVELY TRAINING CLASSIFIERS FOR CIRCULATING TUMOR CELL DETECTION

ITERATIVELY TRAINING CLASSIFIERS FOR CIRCULATING TUMOR CELL DETECTION ITERATIVELY TRAINING CLASSIFIERS FOR CIRCULATING TUMOR CELL DETECTION Yunxiang Mao 1, Zhaozheng Yin 1, Joseph M. Schober 2 1 Missouri University of Science and Technology 2 Southern Illinois University

More information

DIAGNOSTIC CLASSIFICATION OF LUNG NODULES USING 3D NEURAL NETWORKS

DIAGNOSTIC CLASSIFICATION OF LUNG NODULES USING 3D NEURAL NETWORKS DIAGNOSTIC CLASSIFICATION OF LUNG NODULES USING 3D NEURAL NETWORKS Raunak Dey Zhongjie Lu Yi Hong Department of Computer Science, University of Georgia, Athens, GA, USA First Affiliated Hospital, School

More information

Since its introduction in 2000, digital mammography has become

Since its introduction in 2000, digital mammography has become Review Article Smith A, PhD email : Andrew.smith@hologic.com Since its introduction in 2000, digital mammography has become an accepted standard of care in breast cancer screening and has paved the way

More information

arxiv: v1 [cs.lg] 4 Feb 2019

arxiv: v1 [cs.lg] 4 Feb 2019 Machine Learning for Seizure Type Classification: Setting the benchmark Subhrajit Roy [000 0002 6072 5500], Umar Asif [0000 0001 5209 7084], Jianbin Tang [0000 0001 5440 0796], and Stefan Harrer [0000

More information

arxiv: v1 [cs.cv] 9 Oct 2018

arxiv: v1 [cs.cv] 9 Oct 2018 Automatic Segmentation of Thoracic Aorta Segments in Low-Dose Chest CT Julia M. H. Noothout a, Bob D. de Vos a, Jelmer M. Wolterink a, Ivana Išgum a a Image Sciences Institute, University Medical Center

More information

arxiv: v2 [cs.cv] 29 Jan 2019

arxiv: v2 [cs.cv] 29 Jan 2019 Comparison of Deep Learning Approaches for Multi-Label Chest X-Ray Classification Ivo M. Baltruschat 1,2,, Hannes Nickisch 3, Michael Grass 3, Tobias Knopp 1,2, and Axel Saalbach 3 arxiv:1803.02315v2 [cs.cv]

More information

Efficient Deep Model Selection

Efficient Deep Model Selection Efficient Deep Model Selection Jose Alvarez Researcher Data61, CSIRO, Australia GTC, May 9 th 2017 www.josemalvarez.net conv1 conv2 conv3 conv4 conv5 conv6 conv7 conv8 softmax prediction???????? Num Classes

More information

Task 1: Machine Learning with Spike-Timing-Dependent Plasticity (STDP)

Task 1: Machine Learning with Spike-Timing-Dependent Plasticity (STDP) DARPA Report Task1 for Year 1 (Q1-Q4) Task 1: Machine Learning with Spike-Timing-Dependent Plasticity (STDP) 1. Shortcomings of the deep learning approach to artificial intelligence It has been established

More information

arxiv: v2 [cs.lg] 30 Oct 2013

arxiv: v2 [cs.lg] 30 Oct 2013 Prediction of breast cancer recurrence using Classification Restricted Boltzmann Machine with Dropping arxiv:1308.6324v2 [cs.lg] 30 Oct 2013 Jakub M. Tomczak Wrocław University of Technology Wrocław, Poland

More information

CLASSIFICATION OF ABNORMALITY IN B -MASS BY ARCHITECTURAL DISTORTION

CLASSIFICATION OF ABNORMALITY IN B -MASS BY ARCHITECTURAL DISTORTION CLASSIFICATION OF ABNORMALITY IN B -MASS BY ARCHITECTURAL DISTORTION #1 Venmathi.A.R., * 2 D.C.Jullie Josphine #1.Dept of ECE, Kings Engineering College * 2. Dept of CSE,Kings Engineering college Abstract-The

More information

Personalized Colorectal Cancer Survivability Prediction with Machine Learning Methods*

Personalized Colorectal Cancer Survivability Prediction with Machine Learning Methods* Personalized Colorectal Cancer Survivability Prediction with Machine Learning Methods* 1 st Samuel Li Princeton University Princeton, NJ seli@princeton.edu 2 nd Talayeh Razzaghi New Mexico State University

More information

arxiv: v3 [cs.cv] 9 Nov 2017

arxiv: v3 [cs.cv] 9 Nov 2017 Detecting and classifying lesions in mammograms with Deep Learning Dezső Ribli 1*, Anna Horváth 2, Zsuzsa Unger 3, Péter Pollner 4, and István Csabai 1 1 Department of Physics of Complex Systems, Eötvös

More information

The Impact of Visual Saliency Prediction in Image Classification

The Impact of Visual Saliency Prediction in Image Classification Dublin City University Insight Centre for Data Analytics Universitat Politecnica de Catalunya Escola Tècnica Superior d Enginyeria de Telecomunicacions de Barcelona Eric Arazo Sánchez The Impact of Visual

More information

TWO HANDED SIGN LANGUAGE RECOGNITION SYSTEM USING IMAGE PROCESSING

TWO HANDED SIGN LANGUAGE RECOGNITION SYSTEM USING IMAGE PROCESSING 134 TWO HANDED SIGN LANGUAGE RECOGNITION SYSTEM USING IMAGE PROCESSING H.F.S.M.Fonseka 1, J.T.Jonathan 2, P.Sabeshan 3 and M.B.Dissanayaka 4 1 Department of Electrical And Electronic Engineering, Faculty

More information

Quantifying Radiographic Knee Osteoarthritis Severity using Deep Convolutional Neural Networks

Quantifying Radiographic Knee Osteoarthritis Severity using Deep Convolutional Neural Networks Quantifying Radiographic Knee Osteoarthritis Severity using Deep Convolutional Neural Networks Joseph Antony, Kevin McGuinness, Noel E O Connor, Kieran Moran Insight Centre for Data Analytics, Dublin City

More information

Research Article. Automated grading of diabetic retinopathy stages in fundus images using SVM classifer

Research Article. Automated grading of diabetic retinopathy stages in fundus images using SVM classifer Available online www.jocpr.com Journal of Chemical and Pharmaceutical Research, 2016, 8(1):537-541 Research Article ISSN : 0975-7384 CODEN(USA) : JCPRC5 Automated grading of diabetic retinopathy stages

More information

Recurrent Fully Convolutional Neural Networks for Multi-slice MRI Cardiac Segmentation

Recurrent Fully Convolutional Neural Networks for Multi-slice MRI Cardiac Segmentation Recurrent Fully Convolutional Neural Networks for Multi-slice MRI Cardiac Segmentation Rudra P K Poudel, Pablo Lamata and Giovanni Montana Department of Biomedical Engineering, King s College London, SE1

More information

Comparison of Two Approaches for Direct Food Calorie Estimation

Comparison of Two Approaches for Direct Food Calorie Estimation Comparison of Two Approaches for Direct Food Calorie Estimation Takumi Ege and Keiji Yanai Department of Informatics, The University of Electro-Communications, Tokyo 1-5-1 Chofugaoka, Chofu-shi, Tokyo

More information

Radiology Rounds A Newsletter for Referring Physicians Massachusetts General Hospital Department of Radiology

Radiology Rounds A Newsletter for Referring Physicians Massachusetts General Hospital Department of Radiology Radiology Rounds A Newsletter for Referring Physicians Massachusetts General Hospital Department of Radiology Special Issue Response to the Recent US Preventive Services Task Force Recommendations for

More information

Automatic Prostate Cancer Classification using Deep Learning. Ida Arvidsson Centre for Mathematical Sciences, Lund University, Sweden

Automatic Prostate Cancer Classification using Deep Learning. Ida Arvidsson Centre for Mathematical Sciences, Lund University, Sweden Automatic Prostate Cancer Classification using Deep Learning Ida Arvidsson Centre for Mathematical Sciences, Lund University, Sweden Outline Autoencoders, theory Motivation, background and goal for prostate

More information

A Novel Approach to Breast Ultrasound Image Segmentation Based on the Characteristics of Breast Tissue and Particle Swarm Optimization

A Novel Approach to Breast Ultrasound Image Segmentation Based on the Characteristics of Breast Tissue and Particle Swarm Optimization A Novel Approach to Breast Ultrasound Image Segmentation Based on the Characteristics of Breast Tissue and Particle Swarm Optimization Yanhui Guo,, H.D. Cheng,, Jiawei Tian 3, Yingtao Zhang School of Computer

More information

PMR5406 Redes Neurais e Lógica Fuzzy. Aula 5 Alguns Exemplos

PMR5406 Redes Neurais e Lógica Fuzzy. Aula 5 Alguns Exemplos PMR5406 Redes Neurais e Lógica Fuzzy Aula 5 Alguns Exemplos APPLICATIONS Two examples of real life applications of neural networks for pattern classification: RBF networks for face recognition FF networks

More information

MAMMOGRAM AND TOMOSYNTHESIS CLASSIFICATION USING CONVOLUTIONAL NEURAL NETWORKS

MAMMOGRAM AND TOMOSYNTHESIS CLASSIFICATION USING CONVOLUTIONAL NEURAL NETWORKS University of Kentucky UKnowledge Theses and Dissertations--Computer Science Computer Science 2017 MAMMOGRAM AND TOMOSYNTHESIS CLASSIFICATION USING CONVOLUTIONAL NEURAL NETWORKS Xiaofei Zhang University

More information

Automatic Classification of Breast Masses for Diagnosis of Breast Cancer in Digital Mammograms using Neural Network

Automatic Classification of Breast Masses for Diagnosis of Breast Cancer in Digital Mammograms using Neural Network IJSTE - International Journal of Science Technology & Engineering Volume 1 Issue 11 May 2015 ISSN (online): 2349-784X Automatic Classification of Breast Masses for Diagnosis of Breast Cancer in Digital

More information

arxiv: v1 [cs.cv] 13 Jul 2018

arxiv: v1 [cs.cv] 13 Jul 2018 Multi-Scale Convolutional-Stack Aggregation for Robust White Matter Hyperintensities Segmentation Hongwei Li 1, Jianguo Zhang 3, Mark Muehlau 2, Jan Kirschke 2, and Bjoern Menze 1 arxiv:1807.05153v1 [cs.cv]

More information

Design of Palm Acupuncture Points Indicator

Design of Palm Acupuncture Points Indicator Design of Palm Acupuncture Points Indicator Wen-Yuan Chen, Shih-Yen Huang and Jian-Shie Lin Abstract The acupuncture points are given acupuncture or acupressure so to stimulate the meridians on each corresponding

More information

Computer Aided Detection of Abnormalities in Mammograms

Computer Aided Detection of Abnormalities in Mammograms Computer Aided Detection of Abnormalities in Mammograms I W Hutt 1, S M Astley 1 & C R M Boggis 2 1: Wolfson Image Analysis Unit, Dept of Medical Biophysics, Manchester University, Oxford. Rd, Manchester,

More information

Flexible, High Performance Convolutional Neural Networks for Image Classification

Flexible, High Performance Convolutional Neural Networks for Image Classification Flexible, High Performance Convolutional Neural Networks for Image Classification Dan C. Cireşan, Ueli Meier, Jonathan Masci, Luca M. Gambardella, Jürgen Schmidhuber IDSIA, USI and SUPSI Manno-Lugano,

More information

Sleep Staging with Deep Learning: A convolutional model

Sleep Staging with Deep Learning: A convolutional model Sleep Staging with Deep Learning: A convolutional model Isaac Ferna ndez-varela1, Dimitrios Athanasakis2, Samuel Parsons3 Elena Herna ndez-pereira1, and Vicente Moret-Bonillo1 1- Universidade da Corun

More information

Discovery of Rare Phenotypes in Cellular Images Using Weakly Supervised Deep Learning

Discovery of Rare Phenotypes in Cellular Images Using Weakly Supervised Deep Learning Discovery of Rare Phenotypes in Cellular Images Using Weakly Supervised Deep Learning Heba Sailem *1, Mar Arias Garcia 2, Chris Bakal 2, Andrew Zisserman 1, and Jens Rittscher 1 1 Department of Engineering

More information

Retinopathy Net. Alberto Benavides Robert Dadashi Neel Vadoothker

Retinopathy Net. Alberto Benavides Robert Dadashi Neel Vadoothker Retinopathy Net Alberto Benavides Robert Dadashi Neel Vadoothker Motivation We were interested in applying deep learning techniques to the field of medical imaging Field holds a lot of promise and can

More information