Next Article in Journal
Comparison of Deep-Learning and Conventional Machine-Learning Methods for the Automatic Recognition of the Hepatocellular Carcinoma Areas from Ultrasound Images
Previous Article in Journal
Landmark-Based Inertial Navigation System for Autonomous Navigation of Missile Platform
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Simple Shading Correction Method for Brightfield Whole Slide Imaging

1
Gwangju Institute of Science and Technology, the School of Mechanical Engineering, Gwangju 61005, Korea
2
Korea Photonics Technology Institute, Intelligent Photonic Sensor Research Center, Gwangju 61007, Korea
3
College of Medicine, Dankook University, Cheonan-si 31116, Korea
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Sensors 2020, 20(11), 3084; https://doi.org/10.3390/s20113084
Submission received: 12 March 2020 / Revised: 17 May 2020 / Accepted: 27 May 2020 / Published: 29 May 2020
(This article belongs to the Section Biomedical Sensors)

Abstract

:
Whole slide imaging (WSI) refers to the process of creating a high-resolution digital image of a whole slide. Since digital images are typically produced by stitching image sequences acquired from different fields of view, the visual quality of the images can be degraded owing to shading distortion, which produces black plaid patterns on the images. A shading correction method for brightfield WSI is presented, which is simple but robust not only against typical image artifacts caused by specks of dust and bubbles, but also against fixed-pattern noise, or spatial variations in pixel values under uniform illumination. The proposed method comprises primarily of two steps. The first step constructs candidates of a shading distortion model from a stack of input image sequences. The second step selects the optimal model from the candidates. The proposed method was compared experimentally with two previous state-of-the-art methods, regularized energy minimization (CIDRE) and background and shading correction (BaSiC) and showed better correction scores, as smooth operations and constraints were not imposed when estimating the shading distortion. The correction scores, averaged over 40 image collections, were as follows: proposed method, 0.39 ± 0.099; CIDRE method, 0.67 ± 0.047; BaSiC method, 0.55 ± 0.038. Based on the quantitative evaluations, we can confirm that the proposed method can correct not only shading distortion, but also fixed-pattern noise, compared with the two previous state-of-the-art methods.

1. Introduction

Whole slide imaging (WSI), which refers to scanning a whole slide and creating a single high-resolution digital image, is a relatively new technique that solves the narrow field of view (FOV) problem of traditional optical microscopes [1,2,3]. Since the introduction of the automated WSI device in 1999 [4], the WSI devices can acquire faster and more accurate high-quality digital images of the whole slide due to its recent advances in hardware and software technologies [1,5]. These advances have made it possible for pathologists to convert from viewing slides from a microscope to a computer monitor that displays high-resolution whole-slide images [3]. This transfer extends the field of applications of WSI devices to both diagnostic and histopathologic education and proficiency tests [6,7,8,9].
Digital slide images are typically generated by stitching image sequences acquired from different FOVs of a sample into a single whole-slide image. However, shading distortion, caused by uneven illumination [10], can produce black plaid patterns when stitching image sequences acquired to cover the whole slide, which degrades the visual quality of the digital images [11,12,13]. Moreover, the black plaid patterns on the images adversely affect subsequent image analysis, such as segmentation and quantification [14].
Previous shading correction methods can be classified into two categories: prospective and retrospective [14,15]. The prospective approach, e.g., the calib–zero [16] and empty–zero methods [17], estimates shading distortion from a set of image sequences for a reference slide obtained at various locations. Many commercial WSI devices provide the approach as a built-in function; however, acquiring image sequences of the reference slide whenever the brightness of light source changes requires additional effort, which is often prohibitive for many bioimaging researchers [15]. In contrast to the prospective approach, the retrospective approach estimates the shading distortion directly from actual image sequences that are acquired to generate the whole-slide image without the requirement of additional reference images.
The corrected intensity distributions using regularized energy minimization (CIDRE) [18] and background and shading correction (BaSiC) [15], which are categorized as the prospective approach, have been recently acknowledged as state-of-the-art methods, as they have achieved the best performances among 14 prospective and retrospective methods, as reported by reference [15]. The two methods assume smooth shading distortions. The CIDRE method estimates the shading distortion by minimizing a regularized energy function comprising a smoothness term, which enforces a smoothed shading distortion model. By contrast, the BaSiC method imposes sparse constraints on the Fourier-transformed shading distortion model to enforce smoothness. However, the two methods cannot accurately model the shading distortion of real images with fixed-pattern noise, which are a spatial variation in pixel values under uniform illumination [19], as the assumption suppresses spatial variations when estimating the shading distortion. The fixed-pattern noises should be corrected, as they resulted in the image quality degradation [20], which is especially visible when the high-resolution whole-slide images are zoomed in on the monitor screen.
We herein propose a shading correction method, which is straightforward, but robust against not only typical image artifacts caused by dust, scratches and bubbles, but also fixed-pattern noise that degrades the image quality. The proposed method comprises two steps. The first step reconstructs candidates for the shading distortion model by sorting intensities at each pixel axis from a stack of actual image sequences, and the second step selects the optimal shading distortion model, which does not include the typical image artifact, from the candidates by measuring the degree of smoothness. The proposed method removes the fixed-pattern noises in the process of correcting shading distortion, as the estimated flat-field distortion model contained information on the fixed-pattern noises. Therefore, the proposed method does not perform additional noise removal methods. We compared the proposed method with the two state-of-the-art methods, CIDRE [18] and BaSiC [15], for a quantitative evaluation. In the experiments, the proposed method showed better correction scores on diverse imaging conditions with fixed-pattern noise, as we did not impose additional smooth operations and constraints.

2. Experimental Setup

The schematics of the setup for the self-developed WSI device is shown in Figure 1. The device comprised four modules: a Köhler illumination to project brightfield light onto the sample, a motorized XY stage to move the sample along the x and y axes, a piezo stage to adjust the focus of the objective lens and a camera module to acquire sample images. We established Köhler illumination using a light source, two lenses (L1 and L2) and a condenser lens (CL) to generate an even illumination and ensure that the light source was not visible in the acquired images. The light source was a white light-emitting diode (LED) (MNWHL4, Thorlabs, Newton, NJ, USA). A 40 × 0.75NA objective lens (MRH00401, Nikon, Tokyo, Japan) and a tube lens (TTL200, Thorlabs, Newton, NJ, USA) were located between the condenser lens and the camera to project a magnified image of the sample on the sensor plane of the camera.
The image of the sample was acquired using a camera (CP80-4-C500, Optronis, Kehl, Germany) with a pixel size of 7 μm. The piezo stage (PD72Z4CAQ, Physik Instrumente (PI), Karlsruhe, Germany), which adjusts the focal plane of the objective lens, was used for the camera to acquire the focused image; the motorized XY stage (SCAN 75 × 50, Märzhäuser Wetzlar, Germany), which moves the slide along the x and y axes, was used to scan the whole slide. The image size of the camera used in the experiments was 2304 × 1720 pixels, and the FOV was 400 µm × 300 μm. The camera, motorized XY stage, and piezo stage were controlled by a self-developed software to generate digital slide images automatically, and the software was executed on an Intel Core i7-4790 CPU with 32 GB of memory.

3. Methods

The image formation process, which relates an image sequence Y = [ Y 1 , , Y n ] acquired from the device and its corrected image sequence X = [ X 1 , , X n ] , can be approximated as a linear function [21], as follows:
Y i ( k ) = X i ( k ) × F ( k ) + D ( k ) ,
where F ( k ) denotes a flat-field distortion at the k t h pixel caused by uneven illumination [17,22]; D ( k ) denotes dark-field distortion, which occurs if no light is incident [23]. Because the dark-field distortion is typically nearly uniform, varying by only a few intensity values [18], the proposed method assumes D ( k ) = 0 .
Therefore, the corrected image can be easily calculated by reversing the image formation process (Equation (2)) as follows:
X ^ i ( k ) = Y i ( k ) F ^ ( k ) ,
where F ^ ( k ) denotes the estimate of the actual flat-field distortion F ( k ) . Because the actual distortion could not be determined precisely in practice, we estimated the flat-field distortion F ^ , in which the corrected image X ^ i optimally estimated X i from Y i [15,18,21].
Figure 2. shows the flow of the proposed method that estimates the flat-field distortion and corrects the shading distortion. The proposed method comprises primarily of two main steps. The first step constructs candidates for flat-field distortion (Figure 2b) from a stack of input sequences (Figure 2a). Using the brightfield WSI device allows the background pixels, which does not contain objects such as cells to block the illumination light, to have higher values than the object pixels. Based on this assumption, the proposed method constructs candidates by separately sorting the intensity values of red, blue and green channels at each pixel axis and then merging those three channels into a single RGB image. Figure 2b shows the candidates constructed from the image sequences in Figure 2a. The highest intensity values of three channels (the brightest pixel values) at each pixel axis are placed at the first candidate, whereas the darkest pixel values are placed at the last (nth) candidate.
In the second step of the proposed method, the optimal flat-field distortion among the candidates is selected. The candidate constructed with the highest values at each pixel axis can be considered as the optimal flat-field distortion. However, if typical artifacts, such as specks of dust, scratches and bubbles appear on the microscopic slide, they are reflected in the candidates constructed with higher values, as shown in Figure 3. Figure 3a shows more detailed (enlarged) images for 1, 2 and 3 of Figure 2b, while Figure 3b shows the intensity profile for a red channel on a dotted line segment. Therefore, estimating an optimal flat-field distortion robust to typical image artifacts can be considered as a problem of selecting a candidate with minimum artifacts. After estimating the flat-field distortion F ^ , input image sequences are corrected using Equation (2). Figure 2d shows result image sequences that the input image sequences (Figure 2a) are corrected by the estimated flat-field distortion (Figure 2c).
Candidates with more artifacts generally have more high-frequency components than those with no artifacts, as shown in Figure 3. Hence, the proposed method measures the degree of smoothness for each candidate. We define the degree of smoothness as the sum of the local coefficient of variation (LCoV), as the LCoV is used to determine whether a given pixel belongs to a high-frequency region or a homogenous region [24,25], as follows:
S R i = k σ R i ( k ) μ R i ( k ) ,
where σ R i ( k ) and μ R i ( k ) denote the local standard deviation and mean for a red channel at the k th pixel, respectively. A sliding window is used to compute the local standard deviation and the mean and the user determines the size of the sliding window. In the experiments, we set the size of the sliding window to 5 × 5 pixels. R in Equation (3) represents a red channel. The proposed method separately selects the optimal flat-field distortion model for each channel and then merges the distortion model for the abovementioned three channels into a single RGB distortion model. Figure 2c shows a graph (right of Figure 2c) representing the degree of smoothness of the candidates (Figure 2b) for each channel and the color shading model (left of Figure 2c) for the input images (Figure 2a).

4. Results

In the experiments, we scanned various samples with different staining qualities and under varying imaging conditions using a self-developed WSI device to confirm the feasibility of the proposed shading correction method Figure 4. shows examples of images before (Figure 4a) and after (Figure 4c) applying the shading correction for each image tile. Here, the image tile is one of the image sequences acquired by the camera while translating the motorized XY stage to cover the whole slide. Figure 4a shows the image tiles for various samples. Figure 4b shows the flat-field distortion models estimated by the proposed method, whereas Figure 4c shows the result images corrected by each distortion model. Each sample used in the experiments exhibited different cell types and densities as well as different flat-field distortions. Figure 5 shows examples of the shading correction results for the whole-slide images; Figure 5a,b show the whole-slide images before and after applying the shading correction, respectively. We used the microscopy image stitching tool available in ImageJ [26], which stitches partially overlapped image sequences, to generate whole-slide images.
The proposed method was compared with two state-of-the-art methods, CIDRE [18] and BaSiC [15], for a quantitative evaluation. To assess the quality of the correction, we quantified the error of the shading-corrected image X ^ i with a correction score [15] Γ as follows:
Γ = i | X ^ i X i true | i | Y i X i true | ,
where X i true denotes the ground-truth image. The ground-truth images in the experiments were obtained based on the empty-zero method [17], which estimated the shading distortion as the average of images of an empty slide obtained at various locations. The correction score Γ is in the interval [ 0 , ) , where 0 indicates perfect correction, 1 indicates the same amount of error as the uncorrected image and >1 indicates a greater disagreement. Figure 6 shows the correction scores of the proposed method and the two previous methods, CIDRE and BaSiC. The correction scores, averaged over 40 image collections, are as follows: the proposed method, 0.39 ± 0.099; the CIDRE method, 0.67 ± 0.047; the BaSiC method, 0.55 ± 0.038. Each symbol in Figure 6 indicates one of 40 image collections and each image collection consists of 100 images.
To verify the minimum number of images to obtain the correction score given in Figure 6, we evaluated the performance of the proposed method as a function of the number of images. Figure 7 shows a graph representing correction scores according to the number of images used to model the shading distortion. The x-axis of the graph represents the number of images, randomly selected from each collection. The y-axis represents the correction scores, averaged over five different sets of images randomly selected from each component. In Figure 7, each curve is the average score of the five different sets of images, while the shaded area represents the standard deviation. As shown in Figure 7, the proposed method obtained the correction score given in Figure 6 if more than 30 images are given to model the shading distortion.
The proposed method outperformed the two previous state-of-the-art methods because it could estimate the flat-field distortion that was robust to the fixed-pattern noise. The statistical test was performed for comparison between the means using the Wilcoxon signed-rank test, used in BaSiC [15]. It showed a significant difference between the proposed method and the two methods (n = 100, p-value = 1.819 × 10−12). Figure 8 shows an example of the fixed-pattern noise, which is a spatial variation in pixel values under uniform illumination, where Figure 8b shows an image in which a specific area of the input image is enlarged, whereas Figure 8c c shows the intensity profile for the red channel on the dotted line segment of Figure 8b. Figure 9 shows the resulting images that were shading-corrected by the two previous methods and the proposed method for the dotted box in Figure 8a. As shown in Figure 9, the results of the shading correction by the two previous methods still exhibited fixed-pattern noise, whereas it was reduced by the proposed method.
Since the shading distortion was greatly affected by the brightness of illumination, we evaluated the performance of shading correction according to the brightness of illumination. The brightness of the illumination was divided into three levels: 100, 150 and 200. Each level represents the average intensity of the images of an empty slide obtained at various locations. We acquired 40 image collections for each level, and the performance of the previous methods and proposed method was evaluated and compared for each level. Figure 10 shows an example of an image acquired at each level, and Table 1 shows the mean and variance of correction scores evaluated for each level. The experiments confirmed that the proposed method showed better correction scores than the previous methods regardless of the brightness of the illumination.
The fundamental difference between the previous methods and the proposed method is that the former assumes the smoothness of a flat-field distortion. The proposed method does not enforce the smoothness, which can eliminate information for fixed-pattern noise when estimating the flat-field distortion, whereas the previous methods assume smooth distortions. Figure 11 shows the flat-field distortion models estimated by the two previous methods (Figure 11a,b) and the proposed method (Figure 11c). The result images in Figure 1 were generated by the distortion models shown in Figure 11. As shown in Figure 11a,b,e,f, the flat-field distortion models of the two previous methods were smooth, which caused fixed-pattern noise to appear in the shading-corrected images of Figure 9a,b, as the input image sequences included the fixed-pattern noise, as shown in Figure 8. Meanwhile, the proposed method reduced the fixed-pattern noise as the flat-field distortion model estimated by the proposed method contained the necessary information, as shown in Figure 11c,g. Therefore, the proposed method can reduce fixed-pattern noise compared with the two previous state-of-the-art methods.
Because the images that were shading corrected by the previous methods (Figure 9a,b) resembled those corrupted by noise, we additionally evaluated the quality of the shading-corrected images using a peak signal-to-ratio (PSNR). Figure 2 shows the PSNRs of the proposed and two previous methods. The higher the PSNR, the better is the quality of the shading-corrected image. The PSNRs, averaged over 40 image collections, are as follows: the proposed method, 40.80 ± 0.85 dB; the CIDRE method, 36.32 ± 1.09 dB; the BaSiC method, 37.83 ± 0.66 dB, respectively. Based on the quantitative evaluations with the correction scores and PSNR, we can confirm that the proposed method can correct not only shading distortion, but also fixed-pattern noise, compared with the two previous methods, as shown in Figure 6 and Figure 12.
To confirm that the proposed method is simple, but robust compared with the previous two methods, we evaluated the proposed method and the two previous methods in terms of the processing time. Since the processing time depends on the two factors, the number of images and the image size, we measured the processing time according to the two factors. Table 2 shows the processing time according to the image size and the number of images. The BaSiC code and the CIDRE code, written in Matlab, were downloaded from their official GitHub website at https://github.com/marrlab/BaSiC and https://github.com/smithk/cidre, respectively. The proposed method was also implemented using Matlab to measure the processing time in the same condition. As shown in Table 2, the proposed method showed the faster processing time, even if the CIDRE was almost constant regardless of the image size and the number of input images, because the processing time of the proposed method increased slowly according to the number of images and image sizes.

5. Conclusions

A simple shading correction method for brightfield WSI proposed herein, which was robust to not only typical image artifacts caused by specks of dust and bubbles, but also to fixed-pattern noise, which were spatial variations in pixel values under uniform illumination. The proposed method constructed candidates for the shading distortion model by sorting intensities at each pixel axis from a stack of input image sequences. This was based on the assumption that the brightfield WSI allowed the background pixels—which did not contain objects such as cells to block the illumination light—to have higher values than the object pixels. Next, the optimal shading distortion model that did not include the typical image artifact was selected from among the candidates, based on the local coefficient of variance. The proposed method did not enforce the smoothness, which could eliminate fixed-pattern noise information when estimating the shading distortion, whereas the two previous state-of-the-art methods assumed smooth distortions. Experimentally, we achieved better correction scores compared with the two previous methods under diverse imaging conditions with fixed-pattern noise, as the proposed method could correct not only shading distortion, but also fixed-pattern noise. However, the proposed method could not correct the shading distortion of the image sequences acquired by darkfield microscopy or fluorescence microscopy. In future studies, we will modify the proposed method to be applicable to non-brightfield WSI.
We searched other benchmark datasets to evaluate the proposed method but have not found datasets suitable for brightfield whole slide imaging. Therefore, we opened the image collections used in the experiments to GitHub at https://github.com/pair-kopti/Shading-correction. We expect that the dataset will be used to evaluate the novel shading correction method for brightfield whole slide imaging.

Author Contributions

Conceptualization, A.P. and J.B.E.; software, Y.-O.T.; validation, A.P and J.E.; formal analysis, Y.-O.T, J.C. and J.E.; investigation A.P. and H.-S.K.; writing—original draft preparation, Y.-O.T., J.C. and A.P.; writing—review and editing, A.P., H.-S.K. and J.B.E.; project administration, J.B.E. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Technology Innovation Program (20002831, Development of a novel imaging platform for the screening of cervical cancer fast speed digital scanner and AI-engined 3D image reconstruction algorithm) funded by the Ministry of Trade, Industry & Energy (MOTIE, Korea).

Conflicts of Interest

The authors declare that there are no conflict of interest related to this article.

References

  1. Farahani, N.; Parwani, A.V.; Pantanowitz, L. Whole slide imaging in pathology: Advantages, limitations, and emerging perspectives. Pathol. Lab. Med. Int. 2015, 7, 23–33. [Google Scholar]
  2. Kalinski, T.; Zwönitzer, R.; Sel, S.; Evert, M.; Guenther, T.; Hofmann, H.; Bernarding, J.; Roessner, A. Virtual 3D microscopy using multiplane whole slide images in diagnostic pathology. Am. J. Clin. Pathol. 2008, 130, 259–264. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. Zarella, M.D.; Bowman, D.; Aeffner, F.; Farahani, N.; Xthona, A.; Absar, S.F.; Parwani, A.; Bui, M.; Hartman, D.J. A practical guide to whole slide imaging: A white paper from the digital pathology association. Arch. Pathol. Lab. Med. 2019, 143, 222–234. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  4. Ho, J.; Parwani, A.V.; Jukic, D.M.; Yagi, Y.; Anthony, L.; Gilbertson, J.R. Use of whole slide imaging in surgical pathology quality assurance: Design and pilot validation studies. Hum. Pathol. 2006, 37, 322–331. [Google Scholar] [PubMed]
  5. Guo, C.; Bian, Z.; Jiang, S.; Murphy, M.; Zhu, J.; Wang, R.; Song, R.; Shao, X.; Zhang, Y.; Zheng, G. OpenWSI: A low-cost, high-throughput whole slide imaging system via single-frame autofocusing and open-source hardware. Opt. Lett. 2020, 45, 260–263. [Google Scholar] [CrossRef] [Green Version]
  6. Hanna, M.G.; Reuter, V.E.; Hameed, M.R.; Tan, L.K.; Chiang, S.; Sigel, C.; Hollmann, T.; Giri, D.; Samboy, J.; Moradel, C.; et al. Whole slide imaging equivalency and efficiency study: Experience at a large academic center. Mod. Pathol. 2019, 32, 916–928. [Google Scholar] [CrossRef]
  7. Lee, S.H. Virtual microscopy: Past, present and future. Pathology 2011, 43, S4. [Google Scholar]
  8. Saco, A.; Bombi, J.A.; Garcia, A.; Ramírez, J.; Ordi, J. Current status of whole-slide imaging in education. Pathobiology 2016, 83, 79–88. [Google Scholar]
  9. Stewart, J., III; Miyazaki, K.; Bevans-Wilkins, K.; Ye, C.; Kurtycz, D.F.; Selvaggi, S.M. Virtual microscopy for cytology proficiency testing: Are we there yet? Cancer Cytopathol. 2007, 111, 203–209. [Google Scholar] [CrossRef]
  10. Goldman, D.B. Vignette and exposure calibration and compensation. IEEE Trans. Pattern. Anal. 2010, 32, 2276–2288. [Google Scholar]
  11. Chernavskaia, O.; Guo, S.; Meyer, T.; Vogler, N.; Akimov, D.; Heuke, S.; Heintzmann, R.; Bocklitz, T.; Popp, J. Correction of mosaicking artifacts in multimodal images caused by uneven illumination. J. Chemometr 2017, 31, e2901. [Google Scholar] [CrossRef]
  12. Kask, P.; Palo, K.; Hinnah, C.; Pommerencke, T. Flat field correction for high-throughput imaging of fluorescent samples. J. Microsc. 2016, 263, 328–340. [Google Scholar] [CrossRef] [PubMed]
  13. Piccinini, F.; Lucarelli, E.; Gherardi, A.; Bevilacqua, A. Multi-image based method to correct vignetting effect in light microscopy images. J. Microsc. 2012, 248, 6–22. [Google Scholar] [CrossRef] [PubMed]
  14. Tomaževič, D.; Likar, B.; Pernuš, F. Comparative evaluation of retrospective shading correction methods. J. Microsc. 2002, 208, 212–223. [Google Scholar] [CrossRef]
  15. Peng, T.; Thorn, K.; Schroeder, T.; Wang, L.; Theis, F.J.; Marr, C.; Navab, N. A BaSiC tool for background and shading correction of optical microscopy images. Nat. Commun. 2017, 8, 14836. [Google Scholar] [CrossRef] [PubMed]
  16. Varga, V.S.; Bocsi, J.; Sipos, F.; Csendes, G.; Tulassay, Z.; Molnár, B. Scanning fluorescent microscopy is an alternative for quantitative fluorescent cell analysis. Cytometry A 2004, 60, 53–62. [Google Scholar] [CrossRef] [PubMed]
  17. Young, I.T. Shading correction: Compensation for illumination and sensor inhomogeneities. Curr. Protoc. Cytom. 2000, 14, 2.11. [Google Scholar] [CrossRef]
  18. Smith, K.; Li, Y.; Piccinini, F.; Csucs, G.; Balazs, C.; Bevilacqua, A.; Horvath, P. CIDRE: An illumination-correction method for optical microscopy. Nat. Methods 2015, 12, 404. [Google Scholar] [CrossRef]
  19. Konnik, M.; Welsh, J. High-level numerical simulations of noise in CCD and CMOS photosensors: Review and tutorial. arXiv 2014, arXiv:1412.4031. Available online: https://arxiv.org/abs/1412.4031 (accessed on 28 May 2020).
  20. Lim, S.; Gamal, A. Gain fixed-pattern-noise correction via optical flow. IEEE Trans. Circuits Syst. I Reg. Papers. 2004, 51, 779–786. [Google Scholar] [CrossRef] [Green Version]
  21. Likar, B.; Maintz, J.A.; Viergever, M.A.; Pernus, F. Retrospective shading correction based on entropy minimization. J. Microsc. 2000, 197, 285–295. [Google Scholar] [CrossRef] [PubMed]
  22. Leong, F.W.; Brady, M.; McGee, J.O.D. Correction of uneven illumination (vignetting) in digital microscopy images. J. Clin. Pathol. 2003, 56, 619–621. [Google Scholar] [CrossRef] [PubMed]
  23. Irie, K.; McKinnon, A.; Unsworth, K.; Woodhead, I. A model for measurement of noise in CCD digital-video cameras. Meas. Sci. Technol. 2008, 19, 045207. [Google Scholar]
  24. Bovolo, F.; Bruzzone, L. A detail-preserving scale-driven approach to change detection in multitemporal SAR images. IEEE Trans. Geosci. Remote 2005, 43, 2963–2972. [Google Scholar] [CrossRef]
  25. Huang, S.; Huang, W.; Zhang, T. A New SAR image segmentation algorithm for the detection of target and shadow regions. Sci. Rep. 2016, 6, 38596. [Google Scholar] [CrossRef] [PubMed]
  26. Chalfoun, J.; Majurski, M.; Blattner, T.; Bhadriraju, K.; Keyrouz, W.; Bajcsy, P.; Brady, M. Mist: Accurate and scalable microscopy image stitching tool with stage modeling and error minimization. Sci. Rep. 2017, 7, 4988. [Google Scholar] [CrossRef] [PubMed] [Green Version]
Figure 1. Schematics of the experimental setup. Light-emitting diode (LED): light source, L: lens, M: mirror, CL: condenser lens, OBJ: objective lens, PS: piezo stage, TL: tube lens and CAM: camera.
Figure 1. Schematics of the experimental setup. Light-emitting diode (LED): light source, L: lens, M: mirror, CL: condenser lens, OBJ: objective lens, PS: piezo stage, TL: tube lens and CAM: camera.
Sensors 20 03084 g001
Figure 2. Flow of the proposed method: (a) input image sequence, (b) candidates of flat-field distortion, (c) selected flat-field distortion and (d) corrected image sequences.
Figure 2. Flow of the proposed method: (a) input image sequence, (b) candidates of flat-field distortion, (c) selected flat-field distortion and (d) corrected image sequences.
Sensors 20 03084 g002
Figure 3. Example of artifacts reflected on candidates constructed with higher values: (a) enlarged images for 1, 2 and 3 of Figure 2, and (b) intensity profiles for a red channel on a dotted line.
Figure 3. Example of artifacts reflected on candidates constructed with higher values: (a) enlarged images for 1, 2 and 3 of Figure 2, and (b) intensity profiles for a red channel on a dotted line.
Sensors 20 03084 g003
Figure 4. Examples of image tiles with shading corrected by the proposed method: (a) input image tiles, (b) flat-field distortion models and (c) shading-corrected images.
Figure 4. Examples of image tiles with shading corrected by the proposed method: (a) input image tiles, (b) flat-field distortion models and (c) shading-corrected images.
Sensors 20 03084 g004
Figure 5. Examples of whole-slide images, with shading corrected by the proposed method: (a) whole-slide images before applying shading correction and (b) whole-slide images after applying shading correction.
Figure 5. Examples of whole-slide images, with shading corrected by the proposed method: (a) whole-slide images before applying shading correction and (b) whole-slide images after applying shading correction.
Sensors 20 03084 g005
Figure 6. Correction scores of regularized energy minimization (CIDRE), BaSiC and the proposed method. Each symbol represents one of 40 image collections.
Figure 6. Correction scores of regularized energy minimization (CIDRE), BaSiC and the proposed method. Each symbol represents one of 40 image collections.
Sensors 20 03084 g006
Figure 7. Comparison of gamma scores for number of input images.
Figure 7. Comparison of gamma scores for number of input images.
Sensors 20 03084 g007
Figure 8. Example of fixed-pattern noise: (a) input image tile, (b) enlarged image and (c) intensity profile for the red channel on dotted line segment of (b).
Figure 8. Example of fixed-pattern noise: (a) input image tile, (b) enlarged image and (c) intensity profile for the red channel on dotted line segment of (b).
Sensors 20 03084 g008
Figure 9. Shading correction results of two previous methods and the proposed method: (a) result image corrected by CIDRE method, (b) result image corrected by BaSiC method, (c) result image corrected by the proposed method, (d) ground-truth image, (e) intensity profile on dotted line of (a), (f) intensity profile on dotted line of (b), (g) intensity profile on dotted line of (c) and (h) intensity profile for the red channel on dotted line of (d).
Figure 9. Shading correction results of two previous methods and the proposed method: (a) result image corrected by CIDRE method, (b) result image corrected by BaSiC method, (c) result image corrected by the proposed method, (d) ground-truth image, (e) intensity profile on dotted line of (a), (f) intensity profile on dotted line of (b), (g) intensity profile on dotted line of (c) and (h) intensity profile for the red channel on dotted line of (d).
Sensors 20 03084 g009
Figure 10. Example of image acquired at each brightness level. (a) 100, (b) 150, and (c) 200.
Figure 10. Example of image acquired at each brightness level. (a) 100, (b) 150, and (c) 200.
Sensors 20 03084 g010
Figure 11. Flat-field distortion model estimated by two previous methods and the proposed method: (a) distortion model estimated by CIDRE method; (b) distortion model estimated by BaSiC method; (c) distortion model estimated by the proposed method; (d) ground-truth model; (e) intensity profile on dotted line of (a), (f) intensity profile on dotted line of (b), (g) intensity profile on dotted line of (c) and (h) intensity profile on dotted line of (d).
Figure 11. Flat-field distortion model estimated by two previous methods and the proposed method: (a) distortion model estimated by CIDRE method; (b) distortion model estimated by BaSiC method; (c) distortion model estimated by the proposed method; (d) ground-truth model; (e) intensity profile on dotted line of (a), (f) intensity profile on dotted line of (b), (g) intensity profile on dotted line of (c) and (h) intensity profile on dotted line of (d).
Sensors 20 03084 g011
Figure 12. Peak signal-to-noise ratio (PSNR) of CIDRE, BaSiC and the proposed method.
Figure 12. Peak signal-to-noise ratio (PSNR) of CIDRE, BaSiC and the proposed method.
Sensors 20 03084 g012
Table 1. Comparison of correction scores according to the brightness of the illumination.
Table 1. Comparison of correction scores according to the brightness of the illumination.
Brightness of IlluminationCIDREBaSiCProposed Method
1000.6907 ± 0.04790.5909 ± 0.02700.4261 ± 0.0098
1500.6667 ± 0.10350.5322 ± 0.03320.3034 ± 0.0087
2000.6843 ± 0.18340.5171 ± 0.04170.3619 ± 0.1194
Table 2. Processing time (s) according to image size and number of images.
Table 2. Processing time (s) according to image size and number of images.
Method No. of Images100200300400
Image Size
CIDRE576 × 430191.66113.8690.56153.11
1152 × 860205.498.82120.62154.97
2304 × 1720231.03133.22142.44248.76
BaSiC576 × 430343.96522.44395.56520.59
1152 × 8601446.132226.991629.792388.19
2304 × 17203213.244765.283561.355320.58
Proposed
Method
576 × 43012.5218.625.0862.77
1152 × 86029.7644.3660.13152.21
2304 × 172099.27150.51201.99524.71

Share and Cite

MDPI and ACS Style

Tak, Y.-O.; Park, A.; Choi, J.; Eom, J.; Kwon, H.-S.; Eom, J.B. Simple Shading Correction Method for Brightfield Whole Slide Imaging. Sensors 2020, 20, 3084. https://doi.org/10.3390/s20113084

AMA Style

Tak Y-O, Park A, Choi J, Eom J, Kwon H-S, Eom JB. Simple Shading Correction Method for Brightfield Whole Slide Imaging. Sensors. 2020; 20(11):3084. https://doi.org/10.3390/s20113084

Chicago/Turabian Style

Tak, Yoon-Oh, Anjin Park, Janghoon Choi, Jonghyun Eom, Hyuk-Sang Kwon, and Joo Beom Eom. 2020. "Simple Shading Correction Method for Brightfield Whole Slide Imaging" Sensors 20, no. 11: 3084. https://doi.org/10.3390/s20113084

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop