Hybrid Deep Learning for Reflectance Confocal Microscopy Skin Images

Reflectance Confocal Microscopy (RCM) is used for evaluation of human skin disorders and the effects of skin treatments by imaging the skin layers at different depths. Traditionally, clinical experts manually categorize the images captured into different skin layers. This time-consuming labeling task impedes the convenient analysis of skin image datasets. In recent automated image recognition tasks, deep learning with convolutional neural nets (CNN) has achieved remarkable results. However in many clinical settings, training data is often limited and insufficient for CNN training. For recognition of RCM skin images, we demonstrate that a CNN trained on a moderate size dataset leads to low accuracy. We introduce a hybrid deep learning approach which uses traditional texton-based feature vectors as input to train a deep neural network. This hybrid method uses fixed filters in the input layer instead of tuned filters, yet superior performance is achieved. Our dataset consists of 1500 images from 15 RCM stacks belonging to six different categories of skin layers. We show that our hybrid deep learning approach performs with a test accuracy of 82% compared with 51% for CNN. We also compare the results with additional proposed methods for RCM image recognition and show improved accuracy. This work was presented in ICPR 2016 as an oral presentation.

We then extend the experiments by using pre-trained CNN as feature extractors and fine-tune the pre-trained CNNs. This work is yet to be published, more details will be made available soon.



Hybrid Deep Learning for Reflectance Confocal Microscopy Skin Images
Parneet Kaur, Kristin Dana, Oana Cula, Catherine Mack
International Conference on Pattern Recognition (ICPR), 2016

paper | abstract | bibtex

  title={Hybrid deep learning for Reflectance Confocal Microscopy skin images},
  author={Kaur, Parneet and Dana, Kristin J and Cula, Gabriela Oana and Mack, M Catherine},
  booktitle={Pattern Recognition (ICPR), 2016 23rd International Conference on},

Written on December 8, 2016