Rui Li1 , Pelz J1 , Haake AR1
1Golisano College of Computing and Information Science, Rochester Institute of Technology, Rochester, NY. Corresponding Author: Rui Li, Golisano College of Computing and Information Science, Rochester Institute of Technology, Rochester, NY, Tel: +1 585-475-7203; E-Mail: [email protected] Received Date: 18 Apr 2016 Accepted Date: 17 May 2016 Published Date: 21 Jul 2016 Copyright © 2016 Li R Citation: Li R, Pelz J and Haake AR. (2016). Using Experts’ Perceptual Skill for Dermatological Image Segmentation. M J Derm. 1(1): 007. ABSTRACT
There is a growing reliance on imaging equipment in medical domain, hence medical experts’ specialized visual perceptual capability becomes the key of their superior performance. In this paper, we propose a principled generative model to detect and segment out dermatological lesions by exploiting the experts’ perceptual expertise represented by their patterned eye movement behaviors during examining and diagnosing dermatological images. The image superpixels’ diagnostic significance levels are inferred based on the correlations between their appearances and the spatial structures of the experts’ signature eye movement patterns. In this process, the global relationships between the superpixels are also manifested by the spans of the signature eye movement patterns. Our model takes into account these dependencies between experts’ perceptual skill and image properties to generate a holistic understanding of cluttered dermatological images. A Gibbs sampler is derived to use the generative model’s structure to estimate the diagnostic significance and lesion spatial distributions from superpixel-based representation of dermatological images and experts’ signature eye movement patterns. We demonstrate the effectiveness of our approach on a set of dermatological images on which dermatologists’ eye movements are recorded. It suggests that the integration of experts’ perceptual skill and dermatological images is able to greatly improve medical image understanding and retrieval. KEYWORDS
Perceptual Expertise; Dermatological Image Understanding; Probabilistic Modeling; Gibbs Sampling; Eye Tracking Experiments; Superpixels. INTRODUCTION
Image understanding in knowledge-rich domains is particularly challenging, because experts’ domain knowledge and perceptual expertise are demanded to transform image pixels into meaningful contents [1-3]. This motivates using active learning methods to incorporate experts’ specialized capability into this process in order to improve the segmentation performance [4-6)]. However, traditional knowledge acquisition used by active learning methods, such as manual markings, annotations and verbal reports, poses series of significant problems [7, 2]. Because tacit (implicit) knowledge as an integral part of expertise is not consciously accessible to experts, it is difficult for them to identify exactly the diagnostic reasoning processes involved in decision-making, On the other hand, empirical studies suggest that eye movements, as both direct input and measurable output of real time signal processing in the brain, provide us an effective and reliable measure of both human cognitive processing and perceptual skill [8, 9]. Recently, studies try to incorporate perceptual skill into image understanding approaches [4, 7, 2, 6, 10]. Experts’ eye movement data are projected into image feature space to evaluate feature saliency by weighting local features close to each fixation. Salient image features are then mapped back to spatial space in order to highlight regions of interest and at- tention selection [4]. Furthermore, a conceptual framework is developed to measure image feature relevance based on corresponding patterned eye movement deployments defined by pair-wise comparison between multiple experts’ eye movement data [7, 2]. In an image retrieval study [10], images are segmented generically based on image features first. In the later step of image matching, the similarity measures between query image segments and candidate images are weighted by subjects’ fixation data. There are significant limits associated with these approaches. Some of them treat eye movements as a static process by solely using fixation locations without taking the dynamic nature into account [11]. Other studies apply dynamic models to capture sequential information of eye movements, but the cardinality is heuristic and the eye movement descriptions are limited. Since these studies directly use the observed eye movement data, which are noisy and inconsistent, to evaluate image features, their methods’ reliability and robustness are undermined. To address these limitation, we adopt Signature Patterns to represent experts’ perceptual skill [12-14]. Signature Patterns summarize a set of eye movement patterns consistently exhibited by dermatologists while examining various dermatological images in Fig. 1. Signature Patterns encode rich information of experts’ eye movements including fixation locations, fixation durations, and saccade amplitude. Recent eye tracking studies suggest that all these descriptions characterize important diagnostic reasoning processes [15, 16, 9]. Based on these eye movement descriptions, Signature Patterns shed light on both diagnostic significance of image regions and their relative spatial structures. We believe that the patterned behaviors underlying observed human behavioral data is a critical intermediate component to advance domain-specific image understanding. From this perspective, we utilize the spatialtemporal properties discovered in experts’ eye movement data to guide medical image segmentation. In particular, we propose a principled generative model to combine experts’ Signature Patterns with image properties in order to improve image segmentation by estimating diagnostic significance and varying visual spatial structures of the medical images. Perceptual Expertise Based Segmentation
Instead of directly evaluating image features with noisy unstable fixations, we formulated this problem into a probabilistic inference one: given the image features and Signature Patterns from experts, we are inferring diagnostic significance of image regions and their spatial structures [17]. Our method simultaneously grouped dense image features into spatially coherent segments via color and texture cues, and refined these partitions using Signature Patterns, as depicted in the probabilistic generative model in Figure 2.
p(s(i)|on , Π) = Multi(s(i)| Π on ) -------------> (3) Where Π on is regularized by a Dirichlet prior Π on Dir(α). After we independently selected a diagnostic significance level for each superpixel i, we sampled features using parameters determined by s (i) :
p(x(i)nm|s(i) , θ(i)) = Multi(x(i)nm|θ(i)s(i) ) ---------------- >(4) We used tn to denote spatial structures of Signature Patterns’ eye movement units which are scale-normalized and centered, as shown in Fig. 2. Depend on the perceptual category of an image on , tneither parameterizes a unimodal Gaussian distribution for Concentrating and Clutter patterns, or a bimodal Gaussian distribution for Switching pattern. tn is sampled from a normal-inverse-Wishard distribution parameterized by λ = {μ, Λ }. Furthermore, θ(e) specifies tn’s positions relative to an image-specific reference transformation. These transformations are independently sampled from some common prior distribution. We assigned each perceptual category unimodal Gaussian transformation densities:
p(θ(e)|λ , on) = N(θ(e)|ςon, γon) -------------->(5) Where λ = {ς,γ}, and these transformation distribution are regularized by conjugate normal-inverse-wishart prior λ NIW(R). We then modeled the eye movement units of Signature Patterns as a Gibbs distribution:
p(x(e)nm|s(i), tn,θ(e)) ∝ exp(βf(x(i)nm, s(i),tn,θ(e)))------------->(6) Where f denotes the correlation between spatial structures of Signature Patterns and diagnostic significance. And ß is a free parameter to control the overall reliability of Signature Patterns. A Gibbs sampler was developed to learn the parameters of our model. We assumed that hyperparameter values are fixed. We also employed conjugate prior distributions for the diagnostic significance and spatial structure densities. These allowed us to learn the parameters by iteratively sampling significance s(i) using likelihoods which analytically marginalize significance probabilities πand parameters θ(e) This Rao-Blackwellized Gibbs sampler approximates the posterior distribution p(s(i)|X(i) , on) by sampling from it. This posterior can then be used to estimate the underlying parameters. Experimental Evaluation Eye tracking experiments were conducted to collect eye movement sequences from eleven board-certified dermatologists while they were examining and diagnosing a set of 70 dermatological images, each representing a different diagnosis. A SMI (SensoMotoric Instruments) eye tracking apparatus, which was running at 50 Hz sampling rate and has reported accuracy of 0.50 visual angle, was applied to display the stimuli at a resolution of 1680 by 1050 pixels. The subjects viewed the medical images binocularly at a distance of about 60 cm. They were instructed to examine the medical images and make a diagnosis. The experiment started with a 9-point calibration and the calibration was validated after every 10 images. Based on the inferred Signature Patterns, we ran our model to do segmentation. We compared our results to the normalized cuts spectral clustering method (NCuts), and a Guassian mixture model (GMM) based on the same set of color and texture features as used in our model. Our model consistently detected variability in the lesion spatial distributions, and capture both primary and secondary abnormalities. In contrast, NCuts tends to generate regions of equal size, which produces distortions. If the lesion scales are comparable with the body parts’ in the images, NCuts is relatively effective (4th row and 6th row, 5th column in Fig. 3). Otherwise, it tends to fail. Without incorporating domain-specific expertise, GMM discards the dependencies among the superpixels at diagnostic level. So it fails to either differentiate the regions with similar appearance (3rd row and 6th column in Fig. 3).
Fig.3: Qualitative comparison of dermatological image segmentation for each perceptual categories. Column 1: actual photographs. Column 2: Signature Patterns. Column 3 and 4: Our algorithm (red=primary abnormality). Column 5: NCuts spectral clustering method. Column 6: Gaussian mixture model. Best viewed in color
Using Experts’ Perceptual Expertise for Dermatological Image Segmentation
To quantitatively evaluate results, we measured overlap with dermatologists’ segments via the Rand index in Fig. 5. Although GMM performs well for some images with unambiguous features, our method was often substantially better. Signature Patterns enabled us to improve segmentation performance without learning images layers. The perceptual skill encoded in Signature Patterns helped to capture global spatial structures of lesions among dermatological images. Note that superpixels being fixated on were not necessarily to be diagnostically significant, and those without fixations on them may be equivalently important as parts of primary abnormalities. Our model captured this subtlety by inferring the correlations between Signature Patterns and superpixel properties as well as taking Signature Patterns’ spatial transformation across different images into consideration.
We developed a principled generative model for dermatological image segmentation which integrates the extracted experts’ perceptual expertise represented by their patterned eye movement behaviors and image superpixel properties to evaluate diagnostic significance of image regions. Our experiments show that the effective representation of visual perceptual expertise allows better domain-specific image segmentation. This provides a promising starting point for jointly performing experts’ perceptual skill eval uation and image understanding. We believe that such integration will be able to greatly improve medical training and medical image retrieval.