Open Access
16 May 2022 Unsupervised super-resolution reconstruction of hyperspectral histology images for whole-slide imaging
Author Affiliations +
Abstract

Significance: Hyperspectral imaging (HSI) provides rich spectral information for improved histopathological cancer detection. However, acquiring high-resolution HSI data for whole-slide imaging (WSI) can be time-consuming and requires a huge amount of storage space.

Aim: WSI using a color camera can be achieved with fast speed, high image resolution, and excellent image quality due to the established techniques. We aim to develop an RGB-guided unsupervised hyperspectral super-resolution reconstruction method that is hypothesized to improve image quality while maintaining the spectral characteristics.

Approach: High-resolution hyperspectral images of 32 histologic slides were obtained via automated WSI. High-resolution RGB histology images were registered to the hyperspectral images for RGB guidance. An unsupervised super-resolution network was trained to take the downsampled low-resolution hyperspectral patches (LR-HSI) and high-resolution RGB patches (HR-RGB) as inputs to reconstruct high-resolution hyperspectral patches (HR-HSI). Then, an Inception-based network was trained with the HR-RGB, original HR-HSI, and generated HR-HSI, respectively, for whole-slide histopathological cancer detection.

Results: Our super-resolution reconstruction network generated high-resolution hyperspectral images with well-maintained spectral characteristics and improved image quality. Image classification using the original hyperspectral data outperformed RGB because of the extra spectral information. The generated hyperspectral image patches further improved the results.

Conclusions: The proposed method potentially reduces image acquisition time, saves storage space without compromising image quality, and improves the image classification performance.

1.

Introduction

Computer-aided pathology (CAP) is an active research area that aims to improve the reproducibility and objectivity of pathological diagnosis and save time in routine examination.1 Machine learning techniques, particularly deep learning models, have played an essential role in CAP.2 Various studies of deep learning-based cancer detection in whole-slide digitized histology images have been investigated, and most of them were carried out in RGB images. Hyperspectral imaging (HSI) is a noncontact and label-free imaging modality that has emerged in the medical imaging field. It captures the spatial and spectral information of the imaged tissue, revealing the chemical composition and morphological features in a single image modality and thus offering more fine features that are potentially useful for image segmentation and classification. Studies have proven the usefulness of HSI in microscopy applications.315 One of the promising applications for this technology is employing HSI for whole-slide imaging (WSI) to aid the histopathological cancer detection of tissue samples because HSI not only provides a reproducible and quantitative diagnosis of the slides but also improves the classification results compared with RGB.3,5 Although various steps such as fixation and embedding during the preparation of the histological samples may alter a few tissue features, such as the texture and the water content, lots of important molecules such as proteins are preserved in the tissue slides. The light absorption, scattering, and autofluorescence of these molecular components all contribute to the spectral characteristics in hyperspectral histologic images,6,16 which provide a different perspective from the traditional gross-level reflectance HSI. Even for the dyes, HSI can expand the three-channel color information into a much wider spectral dimension, which might increase the discriminability. Nakaya et al.17 used a support vector machine classifier and the average spectra of nuclei extracted from hyperspectral images for colon cancer detection. Ishikawa et al.18 proposed a pattern recognition method named hyperspectral analysis of histopathological slides based on stain spectrum to process the spectra of the histologic hyperspectral images for pancreatic tumor nuclei detection. Ortega et al.3 implemented automatic breast cancer cell detection in hyperspectral histologic images with an average testing AUC of 0.90. The comparison between the classification results using HSI and RGB suggests that HSI outperforms RGB. Our previous studies investigated the feasibility and usefulness of HSI for head and neck squamous cell carcinoma (SCC) nuclei detection in histologic slides.4,19 The comparison between HSI, HSI-synthesized RGB, and RGB indicates that the extra spectral information from HSI can improve the outcome. We also investigated whole-image histopathological cancer detection and proved the advantage of using HSI for head and neck cancers.5,20

There are a few difficulties when applying hyperspectral microscopy in real clinical situations. First, the size of a hyperspectral image file is large. For example, one hyperspectral image with a dimension of 3600 pixels × 2048 pixels × 150 bands saved in single precision requires about 4 GB of storage space. A whole-slide hyperspectral image acquired at a low magnification easily exceeds 100 GB. In some applications such as hematology, high magnification is necessary, which may put an extreme burden on data storage devices. Current limitations on data storage makes it difficult to establish a comprehensive hyperspectral microscopic dataset, which is essential for a thorough study of deep learning methods. Second, the acquisition of hyperspectral images can be time-consuming. Most hyperspectral microscopy studies used push-broom or spectral-scanning systems to obtain a sufficient spectral and spatial resolution. The scanning process of each image can take several seconds, which greatly increases the acquisition time of a whole-slide image. Even though a snapshot hyperspectral camera speeds up the acquisition, its intrinsic low spatial resolution would not meet the requirement of some applications in which fine structures in the slides need to be seen. Third, it is more difficult to implement autofocusing with a hyperspectral camera than a color camera. During the scanning of the histological slides, the change of slide thickness can result in unfocused images, and cellular components from different layers can become blurry. State-of-the-art microscopes and whole-slide scanners are integrated with autofocusing and extended focal imaging (EFI) algorithms to solve these problems; they are based on the acquisition of a stack of RGB images along the Z axis. For example, in Figs. 1(a) and (b), the green and blue arrows point at two regions where tissues with different thicknesses cannot be in focus simultaneously. With the EFI algorithm, a stack of images is acquired at different Z positions and then flattened to make an evenly focused image, as shown in Fig. 1(c). Although it is possible to apply such image enhancement algorithms to HSI, it would further prolong the scanning time tremendously. Finally, some bands in a hyperspectral image, especially the first and last several bands, can have large amounts of noise due to the low sensitivity. Both the unsatisfied focus and noise can result in low image quality and missing spatial details, i.e., high-frequency information, as shown in Figs. 1(d)1(f). On the other hand, whole-slide scanning with a color camera has become routine in the field of pathology. Digital histology images that are acquired with high-quality color cameras have very high image resolution and contrast. Making use of such advantageous features with HSI potentially improves the image quality of hyperspectral images.

Fig. 1

Incidences of low image quality of hyperspectral images. (a) and (b) Two images showing how tissues with different thicknesses in the same image cannot be in focus at the same time. (c) RGB histology image acquired with EFI function integrated with the state-of-the-art microscope. (d) High-resolution digital histology image of a thyroid slide with high image quality. (e) HSI-synthesized RGB image of the same region as (d) but with lower contrast because of the unsatisfying focusing during the slide scanning without autofocusing. (f) The first band from the hyperspectral image of the same region as (d) with a lot of noise due to low sensitivity.

JBO_27_5_056502_f001.png

Hyperspectral image super-resolution reconstruction,2123 also known as HSI pansharpening24 or HSI spatial resolution enhancement,25 is a technique that uses the spatial information from a high-resolution panchromatic image (PAN) and the color information from a low-resolution spectral image to generate a high-resolution spectral image. Different from the concept of “super-resolution imaging” that aims to increase the optical resolution of an imaging system, super-resolution image reconstruction uses various digital image processing techniques to generate images with higher resolutions than the original ones. It has been developed mainly for remote sensing applications. Previously proposed HSI super-resolution reconstruction methods, such as component substitution, multiresolution analysis, hybrid methods, and model-based methods, either result in spectral distortion, generate blurry results, or add complexity to implementation.24,26 In recent years, deep learning algorithms have been explored for the super-resolution reconstruction of hyperspectral images. Masi et al.27 adopted a three-layer convolutional neural network (CNN) named PNN to reconstruct high-resolution multispectral images from the stack of interpolated low-resolution multispectral images and high-resolution PAN. However, the network had a limited learning ability due to the shallow architecture. Yang et al.28 proposed a ResNet-based architecture named PanNet that did not fully exploit the spatial information due to the use of the high-pass filter. Yao et al.29 implemented pixelwise regression for hyperspectral pansharpening using a U-Net. Zheng et al.26 developed a two-part framework that first enhanced the spatial resolution in hyperspectral images through contrast limited adaptive histogram equalization and then used a deep residual neural network to further boost the fusion accuracy.

Since many microscopes and whole-slide scanners can have more than one camera that share the same field of view (FOV), applying the super-resolution reconstruction technique for hyperspectral microscopy and WSI is feasible. Considering the easy acquisition and wide utilization of high-quality RGB histology images, we propose using high-resolution RGB images to guide the super-resolution reconstruction of high-resolution hyperspectral images. Dey et al.30 adopted a linear degradation model to super-resolve imaging mass spectrometry with hematoxylin and eosin (H&E)-stained histology images, which neglected the nonlinearity in spectral responses and resulted in slightly noisy images. Jiang et al.25 used the matrix factorization method to fuse high-resolution RGB histology images and low-resolution hyperspectral microscopic images acquired with a push-broom camera, but the method was only tested for spatial enhancement scales of 3× and 4×. Moreover, none of these studies have investigated deep learning algorithms. Despite a wide variety of models proposed for super-resolution reconstruction in remote sensing, most of them were based on supervised learning methods.2629,31,32 Using this technique, the reconstructed high-resolution hyperspectral image might inherit the noise from the reference hyperspectral image. In addition, the goal of many previous studies was to reconstruct images as close as possible to the real high-resolution spectral images, but what if we can generate images with improved image quality?

Ideally, the super-resolution reconstruction of hyperspectral microscopic images should take full advantage of the superb image quality of digital histology images while maintaining the spectral features that are critical for image classification. Therefore, in this work, we develop a simple yet effective unsupervised super-resolution reconstruction network that fuses the spatial information from the high-resolution RGB images and the spectral information from the low-resolution hyperspectral images. With the proposed method, it is possible to save the acquisition time and storage space for hyperspectral images, as well as compensate for the low quality of some bands in the hyperspectral images, thus promoting the application of HSI for WSI and automatic histopathological cancer detection.

2.

Methods

2.1.

Histologic Slides and Hyperspectral Data Acquisition

In this work, we utilized 32 H&E-stained histologic slides from 16 patients with head and neck SCC.33,34 Each slide was from either tumor (T) or normal (N) tissue, as confirmed and annotated by board-certified pathologists. The tissue was resected during a routine surgery, after which the specimen was inked, formalin-fixed, and paraffin-embedded. The top section of each specimen was obtained using a microtome and stained with H&E. High-resolution digital histology images of the slides were obtained using a whole-slide scanner with a 40× objective lens right after the preparation of the slides.

For the acquisition of hyperspectral images, we used the hyperspectral microscopic imaging system that has been reported in our previous works4,5,20 to scan the slides. The dimensions of the hyperspectral images were 2000  pixels×2000  pixels×87  bands, covering a wavelength range of 470 to 720 nm. With the objective lens of 10× magnification, the FOV of the hyperspectral camera was about 1113  μm×1113  μm. A motorized stage was automated to move the slide along the horizontal and vertical directions with a step size of 1 mm, and one image was acquired at each step. The two adjacent images had an overlap area of about 113  μm wide (203 pixels). During the automatic scanning of the slide, the hyperspectral camera always waited for a few seconds after the stage motion to start image acquistion, giving us adequate time to focus the camera by manually adjusting the fine focusing knob without interrupting the scanning. Figure 2 shows an illustration of the scanning of an entire histologic slide. Each square in the figure indicates one hyperspectral image with a spatial size of 2000×2000  pixels, and every two adjacent images overlap 203 pixels with a scanning step size of 1 mm.

Fig. 2

Whole-slide scanning with a step size of 1 mm. The gray grids are the positions where no image was acquired because minimal tissue was detected in the FOV.

JBO_27_5_056502_f002.png

2.2.

Data Preprocessing

Before the scanning of each slide, a white reference hyperspectral image was obtained at a blank area on the slide, and a dark current image was acquired automatically by the camera with the camera shutter closed. Afterward, each raw hyperspectral image was calibrated with the white reference and dark current images to obtain the transmittance of the tissue:

Eq. (1)

Transmittance(λ)=IRaw(λ)IDark(λ)IWhite(λ)IDark(λ),
where Transmittance(λ) is the wavelength-dependent transmittance; Iraw(λ) is the intensity value for wavelength λ in the raw hyperspectral image; and Iwhite(λ) and Idark(λ) are the intensity values for wavelength λ in the white reference and dark current images, respectively.

For each calibrated hyperspectral image, we synthesized an RGB image using a customized HSI-to-RGB transformation5 and a grayscale image by calculating the average of all 87 bands. The HSI-synthesized RGB images were generated for a better visualization of the FOV of hyperspectral images but not used for the training or validation of any network. The grayscale images were later used for image registration because registration algorithms cannot be directly applied on the 87-band hyperspectral image. The high-resolution RGB images that were used to guide the super-resolution reconstruction were cropped from digital histology images, which were originally imaged with a whole-slide scanner at 40× objective magnification, as stated in the previous section. We looked at each HSI-synthesized RGB image, found the corresponding region in the whole-slide digital histology image, and cropped the high-resolution RGB image with a slightly larger FOV than the HSI-synthesized RGB image. Then, all high-resolution RGB images were registered to their matching hyperspectral images (average grayscale images) using affine registration with the Oriented FAST and Rotated BRIEF (ORB) feature detector35 from the OpenCV package and were resized to a spatial size of 2000×2000  pixels. The registration achieved a pixel-to-pixel alignment between the high-resolution RGB images and the high-resolution hyperspectral images. Afterward, both the hyperspectral images and RGB images were cropped into 361 patches using a sliding window of 200×200  pixels with a step size of 100 pixels. Image patches with little tissue (<50% area of the whole patch) were removed from the dataset. The generated patches from hyperspectral images and coregistered RGB images are called high-resolution hyperspectral patches (HR-HSI) and high-resolution RGB patches (HR-RGB). Then, all HR-HSI were downsampled using a “box” (average) interpolation kernel by 2, 4, 5, 8, and 10 times, respectively, to generate low-resolution hyperspectral patches (LR-HSI) for five different super-resolution reconstruction networks. All image patches from the T and N slides were labeled either as “1” for tumor or “0” for normal according to the slide from which they were taken. Figure 3 shows some image patches with various anatomical structures from cancerous slides and normal slides, respectively. Table 1 shows the number of hyperspectral images and the number of image patches that we finally obtained for each slide.

Fig. 3

Coregistered RGB histological image patches (200×200  pixels) showing anatomical diversity. (a) Patches extracted from images of cancerous slides with various histological features, including nucleus atypia, basaloid SCC, SCC with chronic inflammation, SCC with keratin pearl, and SCC with hemorrhage. (b) Patches generated from images of normal slides with various histological features, including healthy stratified squamous epithelium, salivary glands, stroma, chronic inflammation, and skeletal muscle.

JBO_27_5_056502_f003.png

Table 1

Summary of the hyperspectral histologic dataset.

Patient IDOrganNumber of imagesNumber of patches
T slideN slideTotalCancerousNormalTotal
1Larynx584210016,444975626,200
2Hypopharynx954514021,04017,35038,390
3Larynx386210011,07716,92428,001
4Larynx2228505015681911,834
5Larynx58238114,526550820,034
6Larynx291039690321399042
7Larynx2816447087299210,079
8Buccal mucosa231235488818416729
9Larynx2037573979819912,178
10Larynx131932311146047715
11Larynx53247713,573367117,244
12Larynx231740537526948069
13Larynx58288616,318640122,719
14Larynx46287412,519651819,037
15Larynx3517528750423812,988
16Larynx814012120,00210,07330,075
Total6804481,128170,607109,727280,334

In this study, we first used a small portion of the entire dataset to train, validate, and test our proposed super-resolution reconstruction network. To make sure that the network is able to deal with the staining variation and thickness variation of different slides, which can cause the change of spectral signatures, we selected two images with minimal blank areas for each slide and used them as the “super-resolution dataset” instead of using lots of images from only one patient. In total, 22,053 patches from 64 images of 16 patients were selected, with 17,022 patches from 48 images of 12 patients used for training, 2166 patches from four images of two patients (#3 and #14) used for validation, and 2865 patches from four images of the rest two patients (#5 and #15) used for testing. Then, after the super-resolution reconstruction network was trained, high-resolution hyperspectral image patches were generated for all data using downsampled hyperspectral patches with RGB guidance, and an Inception-based CNN was trained for whole-slide image classification using (1) high-resolution RGB histology image patches, (2) original high-resolution hyperspectral image patches, and (3) generated high-resolution hyperspectral image patches, respectively. The data partition for three types of data was exactly the same. In total 181,711 images patches from 10 patients were used for training, 46,564 image patches from three patients (#3, #4, and #8) were used for validation, and 52,059 image patches from another three patients (#5, #14, and #15) were used for testing. The data that were previously used as the “super-resolution dataset” were not excluded for the image classification network to generate a classification probability map of the whole slide. Nevertheless, those data took only a small portion (<9%) in the entire dataset; thus they did not cause obvious bias for the classification results. The overall workflow and data partition are as shown in Fig. 4. The numbers within the parentheses show how many patients were used for training (TRAIN), validation (VAL), and testing (TEST), and the numbers underneath indicate the specific number of image patches included in each data group.

Fig. 4

Illustration of experiment workflow and data partition. The numbers within the parentheses show how many patients were used for training (TRAIN), validation (VAL), and testing (TEST), with the numbers underneath showing the specific number of image patches used in each data group.

JBO_27_5_056502_f004.png

2.3.

Unsupervised Super-Resolution Reconstruction Network

We developed an unsupervised super-resolution reconstruction network that takes the LR-HSI and HR-RGB as inputs and generates HR-HSI as well as LR-HSI for outputs. To consider different acquisition methods of low-resolution hyperspectral images, we trained five networks with different super-resolution scales. The first scenario is to use the same camera but reduce the magnification of the video adapter (e.g., from 1× to 0.5×), which will result in a smaller image size and lower image resolution. For this case, we trained a 2× super-resolution reconstruction network. The second scenario is to use a camera that has a relatively lower spatial resolution, such as the snapshot hyperspectral camera. For this case, we trained a 4× and a 5× network because 4×4 and 5×5 are the two most common mosaic patterns of snapshot hyperspectral cameras. Two other networks, namely 8× and 10×, were also implemented to evaluate whether the proposed method has the ability for larger-scale spatial enhancement.

Figure 5 shows the architecture of a 4× super-resolution reconstruction network as an example. The three channels (R, G, and B) of the input HR-RGB were duplicated by 35, 35, and 17 times, respectively, and stacked together to form an 87-band “stacked HR-RGB.” Specifically, the blue channel was duplicated less than other two channels because our hyperspectral camera did not cover the wavelength range of 370 to 470 nm, resulting in less blue-color related information in the hyperspectral images. The LR-HSI input (50×50×87) was first upsampled to 200×200×87 using a two-dimensional (2D) deconvolution layer and then concatenated with the stacked HR-RGB along the third dimension, i.e., the spectral dimension, to form a 200×200×174 patch. What followed is a modified U-Net architecture, which took the concatenated 174-band patch as input and output an HR-HSI patch. In addition, a 4×4 average pooling layer was applied to the generated HR-HSI to output an LR-HSI patch with the same size as the input LR-HSI. Note that the dimensions of both the input and generated LR-HSI in Fig. 5 are for a 4× network. For the 2×, 5×, 8×, and 10× networks, the dimensions would be 100×100×87, 40×40×87, 25×25×87, and 20×20×87, respectively. The details of the network are shown in Table 2.

Fig. 5

Architecture of the unsupervised 4× super-resolution reconstruction network, which takes stacked HR-RGB and LR-HSI as inputs and outputs HR-HSI and LR-HSI. Dimensions of the LR-HSI in the figure are for the 4× network only. For the 2×, 5×, 8×, and 10× networks, the dimensions of both the input and generated LR-HSI would be 100×100×87, 40×40×87, 25×25×87, and 20×20×87, respectively.36

JBO_27_5_056502_f005.png

Table 2

Super-resolution reconstruction network architecture.

LayerKernel/strides/paddingOutput shape
Input_1: high-res RGBImage input 1200×200×87
Input_2: low-res HSIImage input 2W×W×87
Conv2D_Transpose(S+1)×(S+1), S, ‘same’200×200×87
ConcatenateConcatenated feature maps200×200×174
Conv2D3×3, 1, ‘same’200×200×196
Conv2D3×3, 1, ‘same’200×200×256
Conv2D3×3, 2, ‘same’100×100×384
Conv2D3×3, 1, ‘same’100×100×512
Conv2D3×3, 2, ‘same’50×50×640
Conv2D3×3, 1, ‘same’50×50×768
Conv2D_Transpose3×3, 2, ‘same’100×100×768
ConcatenateSkip connection100×100×1280
Conv2D3×3, 1, ‘same’100×100×640
Conv2D3×3, 1, ‘same’100×100×512
Conv2D_Transpose3×3, 2, ‘same’200×200×256
ConcatenateSkip connection200×200×512
Conv2D3×3, 1, ‘same’200×200×256
Conv2D3×3, 1, ‘same’200×200×194
Conv2D3×3, 1, ‘same’200×200×160
Conv2D3×3, 1, ‘same’200×200×128
Conv2D3×3, 1, ‘same’200×200×96
Conv2D1×1, 1, ‘same’200×200×87
AveragePoolS×S, S, ‘valid’W×W×87
Note: W is the width of low-resolution image patches, S is the super-resolution scale. For the 2× network, S=2, W=100; for the 4× network, S=4, W=50; for the 5× network, S=5, W=40; for the 8× network, S=8, W=25; and for the 10× network, S=10, W=20.

Our purpose in this work is to generate high-resolution hyperspectral images that have spatial contrast and details as clear as the high-resolution RGB images and spectral signatures close to the original HSI. Thus, the network should be able to extract the spatial information and spectral information from the RGB images and hyperspectral images, respectively, and fuse them in the generated hyperspectral image. This was fulfilled by simultaneously minimizing the spatial loss and spectral loss of the network. For the spatial loss, both the input stacked high-resolution RGB and the generated high-resolution HSI were averaged along the spectral dimension, resulting in two high-resolution panchromatic images; then the mean squared error (MSE) of the two panchromatic images was calculated and used as the spatial loss, as Eq. (2) shows. The spectral loss calculates the MSE between the generated and input LR-HSI, as shown in Eq. (3). The loss weights of the loss functions were 0.5 and 0.5:

Eq. (2)

Lspatial=1M×Ni=1Mj=1N(Gi,jG^i,j)2,

Eq. (3)

Lspectral=1m×n×bi=1mj=1nk=1b(Hi,j,kH^i,j,k)2  ,
where M and N are the spatial dimensions of the high-resolution panchromatic images; m and n are the spatial dimensions of the LR-HSI; b=87 is the number of bands in the LR-HSI; G and G^ are the average panchromatic images of the stacked HR-RGB and the generated HR-HSI, respectively; and H and H^ are the input and generated LR-HSI, respectively.

The unsupervised super-resolution reconstruction network was implemented using Keras on a Titan XP NVIDIA GPU with 12 GB memory. We used the Adam optimizer37 with a learning rate of 104. The network was trained with a batch size of 2. The five networks (2×, 4×, 5×, 8×, and 10×) were trained for 7 to 13 epochs depending on how fast the validation loss stopped decreasing. Each epoch took about 1 h.

2.4.

Whole-Slide Image Classification

To evaluate the usefulness of our proposed super-resolution reconstruction network for image classification, we used the HR-RGB, original HR-HSI, and generated HR-HSI to train, validate, and test an Inception-based 2D CNN and compared the classification results. Specifically, reconstructed HR-HSI were generated for all image patches using the downsampled LR-HSI and the previously trained 4× super-resolution reconstruction network. The Inception-based CNN architecture was modified from the original Inception-v4 network38 to be adapted to our patch size. Each convolutional layer was initialized using the “he_normal” weight initialization39 and was followed by the “ReLU” activation and a 20% dropout, except the activation function of the output layer was sigmoid. The CNN architecture and the input size of each layer/block are shown in Table 3.

Table 3

CNN architecture for whole-slide classification.

Layer/blockInput size
Conv2D, “same”200×200×87
Conv2D, “same”100×100×90
Conv2D, “same”50×50×94
Inception A block ×425×25×96
Reduction A block25×25×384
Inception B block ×711×11×1024
Reduction B block11×11×1024
Inception C block ×35×5×1536
Average pool5×5×1536
Flatten1×1×1536
Dense (two neurons, ‘sigmoid)1536

The CNN was implemented using Keras40 with a Tensorflow backend on the same Titan XP NVIDIA GPU. The optimizer was Adadelta41 with an initial learning rate of 1 and decay rate of rho=0.95. The network was trained with a batch size of 16, and the loss function was binary cross-entropy. All data were split into the training, validation, and testing groups. Data from 10 patients were used for training, three patients (#3, #4, and #8) were used for validation, and three patients (#5, #14, and #15) were used for testing. No data from the same patient were used in two groups at the same time. The same data partition was used for HR-RGB, original HR-HSI, and generated HR-HSI to avoid bias. The image classification network was trained for 10 epochs, with a training time of 4.5 h per epoch using either type of hyperspectral data or 25 min per epoch using the RGB data.

2.5.

Evaluation Metrics

To evaluate the performance of our super-resolution reconstruction network, we calculated different metrics based on the entire hypercube (spatially and spectrally), grayscale image (spatially), and spectral signatures (spectrally). We first used the MATLAB implementations of peak signal-to-noise-ratio (PSNR) and mean absolute error (MAE) to quantify the super-resolution reconstruction quality. PSNR measures the global intensity difference between generated HR-HSI and original HR-HSI and is calculated by averaging the PSNR value in a band-by-band basis across all 87 bands, with one band (B) from the real HR-HSI and the corresponding band (B^) from the generated HR-HSI:

Eq. (4)

PSNRλ(B,B^)=10log10(peak value2/MSE),  

Eq. (5)

PSNR(RealHSI,GeneratedHSI)=  187λ=187PSNRλ(B,B^),

MAE was also calculated a band-by-band basis across all 87 bands:

Eq. (6)

MAEλ(B,B^)=1M×Ni=1Mj=1N|BijB^ij|,

Eq. (7)

MAE(RealHSI,GeneratedHSI)=187λ=187MAEλ(B,B^).

For the spectral evaluation of generated HR-HSI, we used spectral angle mapper (SAM):42,43

Eq. (8)

α=cos1(i=1Mj=1Ntijriji=1Mj=1Ntij2i=1Mj=1Nrij2),
where tij is the spectral signature from the generated HR-HSI, rij is the reference spectral signature from the original HR-HSI, and α is the spectral angle between tij and rij.

Then, we used structural similarity index measure (SSIM)44 and perception-based image quality evaluator (PIQUE)45 to evaluate the spatial character of generated HR-HSI. PIQUE was trained to evaluate the image quality without a reference image; a low score value indicates high perceptual quality. Since it only takes a grayscale or RGB image as input, we generated panchromatic images for generated HR-HSI, original HR-HSI, and stacked HR-RGB by calculating their average across all bands and then calculated PIQUE scores of the three types of panchromatic images. Usually, for super-resolution reconstruction tasks, SSIM is calculated between the entire data cube of the original and generated hyperspectral images to evaluate how “real” the generated images can be. However, due to the obvious improvement of spatial contrast in the generated HR-HSI, which resulted in a disparity between the generated and original hyperspectral images, we only obtained SSIM scores in the range of 0.3 to 0.7. Considering that most spatial information came from stacked HR-RGB, we calculated SSIM between the panchromatic images (G and G^) of stacked HR-RGB and generated HR-HSI:

Eq. (9)

SSIM(G,G^)=(2μGμG^+c1)(2σGG^+c2)(μG2+μG^2+c1)(σG2+σG^2+c2),
where μ and σ are the mean and standard deviation of the panchromatic images, respectively, and c1=(0.01×L)2=0.0001 and c2=(0.03×L)2=0.0009 are two constants chosen as the default values dependent on the dynamic range (L) of the image values.

For the evaluation of whole-slide image classification, we use the area under the receiver operating characteristic (ROC) curve (AUC), as well as accuracy, sensitivity, and specificity as metrics. Accuracy is the ratio of all correctly labeled image patches to the total number of image patches. Sensitivity and specificity are determined by true positive (TP), true negative (TN), false positive (FP), and false negative (FN), where positive is cancerous and negative is normal. Sensitivity measures the percentage of correctly labeled cancerous image patches among all cancerous patches, and specificity measures how well normal image patches are detected:

Eq. (10)

Accuracy=TP+TNTP+FP+TN+FN,

Eq. (11)

Sensitivity=TPTP+FN.

Eq. (12)

Specificity=TNTN+FP.

3.

Results

Our proposed unsupervised super-resolution reconstruction network fuses the spatial information from the high-resolution digital histology images and the spectral information from the low-resolution hyperspectral images and generates high-quality, high-resolution hyperspectral images. The reconstruction time for each 200×200  pixels image patch was about 100 ms. The quantitative evaluation results of five super-resolution reconstruction networks (2×, 4×, 5×, 8×, and 10×) are shown in Table 4. The PSNR, SSIM, and PIQUE measures are more related to the spatial component of the hyperspectral images, whereas the MAE and SAM are more sensitive to the spectral reconstruction errors.46 The PSNR and MAE of five networks indicate a satisfying reconstruction performance both spatially and spectrally. Due to the significant improvement in image quality of the generated HR-HSI, especially the recovered texture details that were missing in the original HR-HSI, we did not get high SSIM values on the entire hyperspectral data cube. However, we calculated SSIM between the panchromatic images of generated HR-HSI and stacked HR-RGB and got an average SSIM of 93.1% (2×), 95.0% (4×), 96.0% (5×), 96.5% (8×), and 97.0% (10×) in the validation data, as well as 89.6% (2×), 93.2% (4×), 94.6% (5×), 95.9% (8×), and 96.7% (10×) in the testing data, which showed high-level similarity of spatial components in generated HR-HSI and HR-RGB. Considering the inherent difference between the two panchromatic images, which would have lowered the SSIM value to a certain extent, the obtained SSIM scores indicate good maintenance of the spatial information. The PIQUE scores for the panchromatic images of generated HR-HSI were comparable to the stacked HR-RGB (28.1 for validation and 30.0 for testing) and were lower than the original HR-HSI (32.5 for validation and 37.5 for testing), also proving a significantly improved image quality of the generated hyperspectral data. Regarding the spectral signatures, the average spectral angles of all networks were lower than 5 deg, showing good maintenance of spectral information. Still, due to the improved image contrast of the generated HR-HSI, the spectral signature of each pixel in generated HR-HSI is not always the same as that of the corresponding pixel in the original HR-HSI. It can be noticed that the network with a larger scale usually results in a higher SSIM score and larger spectral angle because there is less spectral information from the input, and the output is prone to rely more on the spatial information from HR-RGB.

Table 4

Quantitative validation and testing results of the unsupervised super-resolution reconstruction network.

ScalePSNR (dB)MAE (%)SAM (deg)SSIM (%)PIQUE
Validation2×25.14.4±0.62.993.126.6
4×24.34.8±0.93.595.025.1
5×23.25.5±0.93.896.026.6
8×22.75.8±0.84.296.527.4
10×22.36.1±0.94.297.028.6
Testing2×23.85.0±1.12.989.627.6
4×22.85.7±1.53.793.225.4
5×21.66.4±1.73.994.627.5
8×21.17.1±1.74.495.929.8
10×20.77.4±2.04.596.730.9

Figure 6 shows five wavelength bands of the original HR-HSI and generated HR-HSI from different super-resolution reconstruction networks. The intensity of each wavelength band in the generated HR-HSI is close to that in the original HR-HSI. The spatial features were very well reconstructed in all wavelength bands, even if some texture details were missing in the original HR-HSI (e.g., 701 nm). It can be seen that the RGB-guided super-resolution reconstruction network significantly reduced the noise in some wavelength bands (e.g., 470 nm) that was caused by the low sensitivity of the image sensor. Moreover, our network also compensates for the unsatisfying focusing issue on some hyperspectral images. The high-frequency spatial information that was fused from the RGB images also increased the contrast in the generated hyperspectral images, especially when the hyperspectral camera was not perfectly focused and the image quality was compromised, as shown in Fig. 7. Specifically, Fig. 7(a) shows an HSI-synthesized RGB image patch that was out-of-focus due to an accidental impromptu focus adjustment, and Fig. 7(d) is from a partially out-of-focus image caused by uneven slide thickness. Our network was able to extract the spatial information from the digital histology images, as shown Figs. 7(b) and 7(e), and recover the fiber structures as well as clear nuclei edges in the reconstructed images, as shown in Figs. 7(c) and 7(f). Note that the out-of-focus images in this figure are just examples to illustrate the performance of our proposed method, and they do not imply intrinsic low image quality of HSI.

Fig. 6

Single-band images of the original HR-HSI and generated HR-HSI from five different super-resolution reconstruction networks showing high similarity and satisfying spatial reconstruction in various wavelength bands. The major difference is in the first (e.g., 470 nm) and last (e.g., 701 nm) several bands, where noise exists in the original HSI but gets removed from the generated HSI.

JBO_27_5_056502_f006.png

Fig. 7

Improvement of image quality in generated high-resolution hyperspectral images. (a) HSI-synthesized RGB of an original high-resolution hyperspectral image patch, which was out-of-focus and lost the fiber structures due to an accidental impromptu focus adjustment. (b) Coregistered high-resolution RGB histology image patch with fine structures of the tissue. (c) HSI-synthesized RGB of the generated high-resolution hyperspectral image patch, where the fiber structures are recovered. (d) HSI-synthesized RGB patch from a partially out-of-focus image, where the edges of nuclei become fuzzy. (e) Coregistered high-resolution RGB histology image with clear edges of nuclei. (f) HSI-synthesized RGB of the generated HR-HSI, where the shape of nuclei is well recovered.

JBO_27_5_056502_f007.png

Figure 8 shows the spectral signatures of the extracted nucleus, cytoplasm, lymphocyte, and blank area in the slide from the original HR-HSI and the generated HR-HSI, respectively. The shape of spectra from different tissue types was well maintained, which is critical for effective image classification.

Fig. 8

Comparison of spectral signatures of various cellular components from the original HR-HSI and generated HR-HSI. The well-maintained spectral shapes indicate a good spectral reconstruction. (a) A high-resolution digital histology image patch from a head and neck cancer slide, including cancer cells and lymphocytes. (b) Spectral signatures of the same blank area extracted from the original HR-HSI and the generated HR-HSI. (c) Spectral signatures of the same lymphocyte extracted from the original HR-HSI and the generated HR-HSI. (d) Spectral signatures of the same nucelous from a cancer cell, extracted from the original HR-HSI and the generated HR-HSI. (e) Spectral signatures of cytoplasm extracted from the same region in the original HR-HSI and the generated HR-HSI.

JBO_27_5_056502_f008.png

To further validate the effectiveness of our proposed super-resolution reconstruction network, we acquired hyperspectral images of the same region on a histologic slide using four objective lenses with different magnifications, namely 4×, 10×, 20×, and 40×. Then, we reconstructed high-resolution hyperspectral images (equivalent to the optical resolution with the 40× objective lens) using the three relatively low-resolution images, i.e., the images acquired at 20×, 10×, and 4× objective magnification. Figure 9 shows the synthesized RGB images of the hyperspectral images acquired with physically different optical resolutions as well as the reconstructed images. The black square contours in the low-resolution images show the same region of interest as the high-resolution image. It can be seen that the detailed structures were very well recovered in the reconstructed images, and the significant similarity of colors in the synthesized images indicates a satisfying maintenance of the spectral characteristics. Therefore, despite our network being trained using generated low-resolution hyperspectral data, it works well on real low-resolution images.

Fig. 9

Super-resolution reconstruction using images with physically different resolutions. The black contours in the low-resolution images outline the same region as the high-resolution image acquired at 40× magnification.

JBO_27_5_056502_f009.png

In addition to the abovementioned evaluation metrics, we implemented image classification using the original HR-HSI and generated HR-HSI to further prove the usefulness of our proposed super-resolution reconstruction network. The validation and testing results using three types of data are shown in Table 5. RGB histology images, which contained morphological and color information of the slides, had satisfying classification results of 0.9 AUC, 0.85 accuracy, 0.88 sensitivity, and 0.77 specificity in the validation group, as well as 0.86 AUC, 0.80 accuracy, 0.82 sensitivity, and 0.76 specificity in the testing data group. The high image resolution, together with the autofocusing and other image enhancement techniques integrated with the whole-slide scanner, guaranteed sufficient spatial details of the anatomical structures, which were critical for the classificantion. The prediction time using either hyperspectral data or RGB data was <30 ms per patch. For the classification of a whole slide, the prediction time varied from 0.5 to 3 min with the size of the slide.

Table 5

Quantitative results of whole-slide image classification using different data.

DataAUCAccuracySensitivitySpecificity
ValidationHR-RGB0.900.850.880.77
Original HR-HSI0.930.870.890.87
Generated HR-HSI0.940.900.890.90
TestingHR-RGB0.860.800.820.76
Original HR-HSI0.870.820.850.73
Generated HR-HSI0.900.860.870.83

Although the hyperspectral camera was focused by our manual adjustment, which might be subjective and dependent on the operators, it was able to acquire hyperspectral images with a very decent image quality. Regardless of the noise in a few wavelength bands and some partially unfocused images, which would have slightly compromised the spatial characters, the original high-resolution hyperspectral image patches still outperformed RGB because of the extra spectral information, with 0.93 AUC, 0.87 accuracy, 0.89 sensitivity, and 0.87 specificity in validation data, as well as 0.87 AUC, 0.82 accuracy, 0.85 sensitivity, and 0.73 specificity in the testing group. As for the generated HR-HSI, the unsupervised super-resolution reconstruction network fused the spatial components from the RGB images and the spectral information from the hyperspectral images, resulted in an improved image quality in the generated HR-HSI. One significant improvement was the removal of noise, which greatly increased the efficacy of the first and last several bands. In addition, the out-of-focus images, which were caused by the uneven thickness of slides and had fuzzy nuclei edges or blurry fiber structures, also regained sharpness. Due to the improved image quality, using generated HR-HSI made the classification performance even better.

Figure 10 shows the probability maps of whole-slide classification using the HR-RGB, original HR-HSI, and generated HR-HSI. Although using RGB image patches achieved satisfying results, it is prone to generate false positives at deformed epithelium or false negatives where very few cancer nuclei exist, as shown in Figs. 10(a)10(b), 10(e)10(f). Classification using the original hyperspectral image patches obviously reduced the number of false positives and false negatives, which proved the usefulness of the spectral information. However, the probability values of tumor and normal images using the original HR-HSI were not as extremely separated. Instead, most tumor patches had a probability value of around 0.8, and most normal patches had a probability value of 0.2 to 0.3. This does not affect the classification results though, and it is likely to be improved by increasing the dataset. The probability maps of generated HR-HSI show improved classification performance compared with both HR-RGB and original HR-HSI. It seems that the high-contrast spatial components from the high-resolution RGB images might have provided critical cancer-related information, such as the morphology of cancerous nuclei,33 and the spectral components from the hyperspectral images were beneficial especially where RGB images tend to give false results. In addition, at the bottom of the tumor slide in Fig. 10(e), there is a small piece of tissue with dysplasia, which is a type of abnormalty but not tumor. None of the networks misclassified it without being previously trained, although the probability maps of HR-HSI did show abnormalty to some extent.

Fig. 10

Illustration of whole-slide image classification using different types of data. (a) The digital histology image of a normal tissue slide. (b)–(d) Probability map of the normal slide using the RGB image patches, original hyperspectral image patches, and generated hyperspectral image patches, respectively. Note that the green contour in (a) was falsely classified as positive in (b) using RGB image patches but was correctly classified as negative using either hyperspectral image patches (c and d). (e) The digital histology image of a tumor slide with a little dysplasia tissue (regarded as normal) at the bottom. The green rectangle region was falsely classified as negative using RGB image patches but was correctly classified using hyperspectral patches. (f)–(h) Probability maps of the tumor slide using RGB image patches, original hyperspectral image patches, and generated hyperspectral image patches, respectively.

JBO_27_5_056502_f010.png

4.

Discussion and Conclusion

In this work, we developed an unsupervised CNN based on a modified U-Net architecture for hyperspectral super-resolution reconstruction with the guidance of high-quality RGB digital histology images. The network fuses the spatial information from the high-resolution RGB images and the spectral information from the low-resolution hyperspectral images to generate high-resolution hyperspectral images. The generated hyperspectral images from all five super-resolution reconstruction networks with different spatial enhancement scales (2×, 4×, 5×, 8×, and 10×) maintained the shape of the original spectral signatures while being enriched with high-frequency spatial information. In addition, the proposed method improved the image quality for hyperspectral images, including reducing image noise, increasing image contrast, and deblurring the image. Furthermore, the unsupervised method does not require any high-resolution hyperspectral image as ground truth for network training, which minimizes the complexity of the workflow. We implemented image classification using the RGB image patches, original hyperspectral image patches, and hyperspectral patches generated from the 4× network, respectively. The results show the usefulness of the spectral information in the hyperspectral images and prove the ability of our proposed super-resolution reconstruction network for improving image quality. By fusing the high-contrast spatial information from the RGB histology images and the rich spectral information from the hyperspectral images, our proposed method potentially brings benefits to histopathological diagnosis.

One shortcoming of this study is the small dataset volume. To avoid potential data leakage between the super-resolution reconstruction network and the whole-slide image classification network, we only used a very small portion of data (two images per slide) for the super-resolution reconstruction network and left the remainder for the whole-slide image classification network. It is possible that, with more training data, our super-resolution reconstruction network could perform even better for spectral signature reconstructions. As for the whole-slide image classification network, 32 entire slides may not be sufficient due to the anatomical variations of SCC and the significantly increased number of features in the hyperspectral data. In addition, the tissues in all slides were either tumor or normal, which made it slightly easier to get decent image classification results. In the future, we will investigate whole-slide cancer detection in tumor-normal margin slides and see if the generated HR-HSI is able to significantly improve the outcome.

Another shortcoming in this work is that we used the simulated low-resolution hyperspectral data to train the super-resolution reconstruction network, instead of acquiring a large number of hyperspectral images with a physically lower resolution as training data. It is possible that the differences between the real low-resolution hyperspectral images and simulated ones may affect the network outcome. But acquiring images with different optical resolutions and registering the high-resolution RGB images to the low-resolution hyperspectral images potentially introduces misalignment and thus minor errors during the network training. On the other hand, the way that we simulated the LR-HSI for training data (downsampling with “box” kernel) was the same as how the network output the generated LR-HSI, which could have reduced some bias. In addition, we tested the network trained with simulated LR-HSI on real LR-HSI to validate the performance of the proposed method. The results indicate that, even though the network was trained with simulated low-resolution data, it also worked well with images with actual lower resolutions. In the future, we will further investigate this by acquiring more images with physically different resolutions.

With HSI getting more attention in the medical imaging field, the conflict among the acquisition speed, data storage, and resolution of hyperspectral images must be solved to apply this technology in real clinical settings. Our proposed method makes it possible to generate high-quality hyperspectral images for automatic histopathological analysis with a low-resolution HSI camera and low-magnification objective lens, hence greatly reducing the acquisition time and the file size of hyperspectral histologic images. For instance, with a 4× network, the acquisition time of a line-scanning HSI system can be cut to 1/4, and the file size can be cut to 1/16. It also allows for the use of snapshot cameras, which can easily reach video-rate imaging, without the concern of low spatial image resolution. Moreover, our method does not change the routine workflow in pathology. Instead of developing a new complex system, it can be achieved simply by mounting a low-resolution hyperspectral camera onto a commercial whole-slide scanner. The acquired low-resolution hyperspectral images can be stored and used for automatic digital pathological analysis after super-resolution reconstruction, while the RGB digital histology images are still available for pathologists to peruse and confirm the results.

In the future, we plan to develop an automatic whole-slide scanning HSI microscope with a low-resolution hyperspectral camera (e.g., a snapshot camera) and the proposed super-resolution reconstruction network built into the system. By synchronizing the color camera and the hyperspectral camera, the system will be able to acquire both data simultaneously. With the mature autofocusing and image enhancement techniques of color cameras as well as our proposed method, the image quality of both the RGB and hyperspectral images will be secured. We may be able to establish a comprehensive database of whole-slide hyperspectral histologic images, which is extremely beneficial for a thorough investigation of pathological features in hyperspectral images as well as various deep learning algorithms. In addition, the super-resolution reconstruction network and whole-slide image classification network may be combined into a single and straightforward workflow in which high-resolution hyperspectral images can be generated along with the acquisition of low-resolution hyperspectral images and high-resolution RGB images and directly fed into the image classification network, so that histopathological diagnosis can be achieved during the scanning of the slides.

In conclusion, our deep learning-based hyperspectral image super-resolution reconstruction method can bring high-resolution spatial features and rich spectral information to improve pathological diagnosis, while increasing the image acquisition speed and reducing data storage requirement, which will provide wide applications in digital and computational pathology.

Disclosures

The authors have no relevant financial interests in this article and no potential conflicts of interest to disclose. Informed consent was obtained from all patients in accordance with Emory Institutional Review Board policies under the Head and Neck Satellite Tissue Bank (HNSB, IRB00003208) protocol.

Acknowledgments

This research was supported in part by the US National Institutes of Health (NIH) Grants (R01CA156775, R01CA204254, R01HL140325, and R21CA231911) and by the Cancer Prevention and Research Institute of Texas (CPRIT) Grant No. RP190588.

References

1. 

A. Madabhushi and G. Lee, “Image analysis and machine learning in digital pathology: challenges and opportunities,” Medical Image Analysis, 33 170 –175 (2016). https://doi.org/10.1016/j.media.2016.06.037 Google Scholar

2. 

S. Nam et al., “Introduction to digital pathology and computer-aided pathology,” J. Pathol. Transl. Med., 54 (2), 125 –134 (2020). https://doi.org/10.4132/jptm.2019.12.31 Google Scholar

3. 

S. Ortega et al., “Hyperspectral imaging and deep learning for the detection of breast cancer cells in digitized histological images,” Proc. SPIE, 11320 113200V (2020). https://doi.org/10.1117/12.2548609 Google Scholar

4. 

L. Ma et al., “Hyperspectral microscopic imaging for automatic detection of head and neck squamous cell carcinoma using histologic image and machine learning,” Proc. SPIE, 11320 113200W (2020). https://doi.org/10.1117/12.2549369 Google Scholar

5. 

L. Ma et al., “Hyperspectral microscopic imaging for the detection of head and neck squamous cell carcinoma in histologic images,” Proc. SPIE, 11603 116030P (2021). https://doi.org/10.1117/12.2581970 Google Scholar

6. 

X. Zhou et al., “Automatic detection of head and neck squamous cell carcinoma on pathologic slides using polarized hyperspectral imaging and machine learning,” Proc. SPIE, 11603 116030Q (2021). https://doi.org/10.1117/12.2582330 Google Scholar

7. 

S. Ortega et al., “Detecting brain tumor in pathological slides using hyperspectral imaging,” Biomed. Opt. Express, 9 (2), 818 –831 (2018). https://doi.org/10.1364/BOE.9.000818 BOEICL 2156-7085 Google Scholar

8. 

S. Ortega et al., “Hyperspectral imaging for the detection of glioblastoma tumor cells in h&E slides using convolutional neural networks,” Sensors, 20 (7), 1911 (2020). https://doi.org/10.3390/s20071911 SNSRES 0746-9462 Google Scholar

9. 

S. Ortega et al., “Hyperspectral and multispectral imaging in digital and computational pathology: a systematic review,” Biomed. Opt. Express, 11 (6), 3195 –3233 (2020). https://doi.org/10.1364/BOE.386338 BOEICL 2156-7085 Google Scholar

10. 

S. Ortega et al., “Hyperspectral superpixel-wise glioblastoma tumor detection in histological samples,” Appl. Sci., 10 (13), 4448 (2020). https://doi.org/10.3390/app10134448 Google Scholar

11. 

A. N. Pronichev et al., “The use of optical microscope equipped with multispectral detector to distinguish different types of acute lymphoblastic leukemia,” J. Phys.: Conf. Ser., 784 012003 (2017). https://doi.org/10.1088/1742-6596/784/1/012003 JPCSDZ 1742-6588 Google Scholar

12. 

Y. Liu, F. Long, “Acute lymphoblastic leukemia cells image analysis with deep bagging ensemble learning,” ISBI 2019 C-NMC Challenge: Classification in Cancer Cell Imaging, 113 –121 Springer, Singapore (2019). Google Scholar

13. 

A. Panda, R. B. Pachori and N. D. Sinnappah-Kang, “Classification of chronic myeloid leukemia neutrophils by hyperspectral imaging using Euclidean and Mahalanobis distances,” Biomed. Signal Process. Control, 70 103025 (2021). https://doi.org/10.1016/j.bspc.2021.103025 Google Scholar

14. 

F. Fereidouni, A. N. Bader and H. C. Gerritsen, “Spectral phasor analysis allows rapid and reliable unmixing of fluorescence microscopy spectral images,” Opt. Express, 20 (12), 12729 –12741 (2012). https://doi.org/10.1364/OE.20.012729 OPEXFF 1094-4087 Google Scholar

15. 

G. A. Roth et al., “Hyperspectral microscopy as an analytical tool for nanomaterials,” Wiley Interdiscipl. Rev.: Nanomed. Nanobiotechnol., 7 (4), 565 –579 (2015). https://doi.org/10.1002/wnan.1330 Google Scholar

16. 

F. Fereidouni et al., “Dual-mode emission and transmission microscopy for virtual histochemistry using hematoxylin- and eosin-stained tissue sections,” Biomed. Opt. Express, 10 (12), 6516 –6530 (2019). https://doi.org/10.1364/BOE.10.006516 BOEICL 2156-7085 Google Scholar

17. 

D. Nakaya et al., “Digital pathology with hyperspectral imaging for colon and ovarian cancer,” 109560X (2019). https://doi.org/10.1117/12.2512328 Google Scholar

18. 

M. Ishikawa et al., “Detection of pancreatic tumor cell nuclei via a hyperspectral analysis of pathological slides based on stain spectra,” Biomed. Opt. Express, 10 (9), 4568 –4588 (2019). https://doi.org/10.1364/BOE.10.004568 BOEICL 2156-7085 Google Scholar

19. 

L. Ma et al., “Automatic detection of head and neck squamous cell carcinoma on histologic slides using hyperspectral microscopic imaging,” J. Biomed. Opt., 27 (4), 046501 (2022). https://doi.org/10.1117/1.JBO.27.4.046501 Google Scholar

20. 

M. H. Tran et al., “Thyroid carcinoma detection on whole histologic slides using hyperspectral imaging and deep learning,” 120390H (2022). https://doi.org/10.1117/12.2612963 Google Scholar

21. 

T. Akgun, Y. Altunbasak and R. M. Mersereau, “Super-resolution reconstruction of hyperspectral images,” IEEE Trans. Image Process., 14 (11), 1860 –1875 (2005). https://doi.org/10.1109/TIP.2005.854479 IIPRE4 1057-7149 Google Scholar

22. 

N. Sun and H. Li, “Super resolution reconstruction of images based on interpolation and full convolutional neural network and application in medical fields,” IEEE Access, 7 186470 –186479 (2019). https://doi.org/10.1109/ACCESS.2019.2960828 Google Scholar

23. 

D. Qiu et al., “Super-resolution reconstruction of knee magnetic resonance imaging based on deep learning,” Comput. Methods Prog. Biomed., 187 105059 (2020). https://doi.org/10.1016/j.cmpb.2019.105059 Google Scholar

24. 

L. Loncan et al., “Hyperspectral pansharpening: a review,” IEEE Geosci. Remote Sens. Mag., 3 (3), 27 –46 (2015). https://doi.org/10.1109/MGRS.2015.2440094 Google Scholar

25. 

Z. Jiang et al., “Spatial resolution enhancement for pushbroom-based microscopic hyperspectral imaging,” Appl. Opt., 58 (4), 850 –862 (2019). https://doi.org/10.1364/AO.58.000850 APOPAI 0003-6935 Google Scholar

26. 

Y. Zheng et al., “Deep residual learning for boosting the accuracy of hyperspectral pansharpening,” IEEE Geosci. Remote Sens. Lett., 17 (8), 1435 –1439 (2020). https://doi.org/10.1109/LGRS.2019.2945424 Google Scholar

27. 

G. Masi et al., “Pansharpening by convolutional neural networks,” Remote Sens., 8 (7), 594 (2016). https://doi.org/10.3390/rs8070594 Google Scholar

28. 

J. Yang et al., “PanNet: a deep network architecture for pan-sharpening,” in Proc. IEEE Int. Conf. Comput. Vision, 1753 –1761 (2017). https://doi.org/10.1109/ICCV.2017.193 Google Scholar

29. 

W. Yao et al., “Pixel-wise regression using U-Net and its application on pansharpening,” Neurocomputing, 312 364 –371 (2018). https://doi.org/10.1016/j.neucom.2018.05.103 NRCGEO 0925-2312 Google Scholar

30. 

N. Dey et al., “Multi-modal image fusion for multispectral super-resolution in microscopy,” Proc. SPIE, 10949 109490D 2019). https://doi.org/10.1117/12.2512598 Google Scholar

31. 

J. Cai and B. Huang, “Super-resolution-guided progressive pansharpening based on a deep convolutional neural network,” IEEE Trans. Geosci. Remote Sens., 59 (6), 5206 –5220 (2021). https://doi.org/10.1109/TGRS.2020.3015878 IGRSD2 0196-2892 Google Scholar

32. 

W. G. C. Bandara, J. M. J. Valanarasu and V. M. Patel, “Hyperspectral pansharpening based on improved deep image prior and residual reconstruction,” IEEE Trans. Geosci. Remote Sens., 60 1 –16 (2022). https://doi.org/10.1109/TGRS.2021.3139292 Google Scholar

33. 

M. Halicek et al., “Head and neck cancer detection in digitized whole-slide histology using convolutional neural networks,” Sci. Rep., 9 14043 (2019). https://doi.org/10.1038/s41598-019-50313-x SRCEC3 2045-2322 Google Scholar

34. 

M. Halicek et al., “Hyperspectral imaging of head and neck squamous cell carcinoma for cancer margin detection in surgical specimens from 102 patients using deep learning,” Cancers, 11 (9), 1367 (2019). https://doi.org/10.3390/cancers11091367 Google Scholar

35. 

E. Rublee et al., “ORB: an efficient alternative to SIFT or SURF,” in Int. Conf. Comput. Vision, 2564 –2571 (2011). Google Scholar

36. 

L. Ma et al., “Unsupervised super resolution network for hyperspectral histologic imaging,” 120390P (2022). https://doi.org/10.1117/12.2611889 Google Scholar

37. 

D. P. Kingma and J. Ba, “Adam: a method for stochastic optimization,” in Int. Conf. Learn. Represent. (ICLR), (2014). Google Scholar

38. 

C. Szegedy et al., “Inception-v4, inception-resnet and the impact of residual connections on learning,” in Thirty-First AAAI conf. Artif. Intell., (2017). Google Scholar

39. 

K. He et al., “Delving deep into rectifiers: surpassing human-level performance on ImageNet classification,” in Proc. IEEE International Conference on Computer Vision, 1026 –1034 (2015). Google Scholar

40. 

N. Ketkar, Deep Learning with Python: A Hands-on Introduction, Apress, Berkeley, California (2017). Google Scholar

41. 

M. D. Zeiler, “Adadelta: an adaptive learning rate method,” (2012). https://doi.org/10.48550/arXiv.1212.5701 Google Scholar

42. 

J. W. Boardman, “Spectral angle mapping: a rapid measure of spectral similarity,” (1993). Google Scholar

43. 

F. A. Kruse et al., “The spectral image processing system (SIPS)—interactive visualization and analysis of imaging spectrometer data,” Remote Sens. Environ., 44 (2–3), 145 –163 (1993). https://doi.org/10.1016/0034-4257(93)90013-N Google Scholar

44. 

Z. Wang et al., “Image quality assessment: from error visibility to structural similarity,” IEEE Trans. Image Process., 13 (4), 600 –612 (2004). https://doi.org/10.1109/TIP.2003.819861 IIPRE4 1057-7149 Google Scholar

45. 

N. Venkatanath et al., “Blind image quality evaluation using perception based features,” in Twenty First Natl. Conf. Commun. (NCC), 1 –6 (2015). Google Scholar

46. 

M. Halicek et al., “Conditional generative adversarial network for synthesizing hyperspectral images of breast cancer cells from digitized histology,” Proc. SPIE, 11320 113200U (2020). https://doi.org/10.1117/12.2549994 Google Scholar

Biography

Ling Ma is a graduate research assistant in the Quantitative Bioimaging Laboratory (QBIL) in the Department of Bioengineering at the University of Texas at Dallas. She received her BS degree in measuring and controlling technology and instruments from Tianjin University in 2015. Her current research interests include in vivo hyperspectral imaging and hyperspectral microscopic imaging for head and neck cancer detection. She is a student member of SPIE.

Armand Rathgeb is a research assistant in the Quantitative Bioimaging Laboratory (QBIL) in the Department of Bioengineering at the Erik Jonsson School of Engineering and Computer Science at the University of Texas at Dallas.

Hasan Mubarak is a research assistant in the Quantitative Bioimaging Laboratory (QBIL) in the Department of Bioengineering at the University of Texas at Dallas.

Minh Tran is a PhD student in the Quantitative Bioimaging Laboratory (QBIL) in the Department of Bioengineering and the Center for Imaging and Surgical Innovation (CISI) at the University of Texas at Dallas. His research interests are bioimaging and artificial intelligence. He is a student member of SPIE.

Baowei Fei is the Cecil H. and Ida Green Chair in Systems Biology Science, professor of Bioengineering at the University of Texas at Dallas and a professor of Radiology at UT Southwestern Medical Center. He is the director of the Quantitative BioImaging Laboratory ( www.fei-lab.org) and the director of the Center for Imaging and Surgical Innovation. He is a fellow of the International Society for Optics and Photonics (SPIE) and the American Institute for Medical and Biological Engineering (AIMBE).

CC BY: © The Authors. Published by SPIE under a Creative Commons Attribution 4.0 Unported License. Distribution or reproduction of this work in whole or in part requires full attribution of the original publication, including its DOI.
Ling Ma, Armand Rathgeb, Hasan Mubarak, Minh Tran, and Baowei Fei "Unsupervised super-resolution reconstruction of hyperspectral histology images for whole-slide imaging," Journal of Biomedical Optics 27(5), 056502 (16 May 2022). https://doi.org/10.1117/1.JBO.27.5.056502
Received: 18 February 2022; Accepted: 28 April 2022; Published: 16 May 2022
Lens.org Logo
CITATIONS
Cited by 5 scholarly publications.
Advertisement
Advertisement
KEYWORDS
Hyperspectral imaging

Super resolution

RGB color model

Image classification

Cameras

Image quality

Tissues

Back to Top