Latent spectral-spatial diffusion model for single hyperspectral super-resolution

In recent years, significant advances have been achieved in addressing super-resolution (SR) tasks for hyperspectral images, primarily through deep learning-based methodologies. Nevertheless, methods oriented toward optimizing peak signal-to-noise ratio (PSNR) often tend to drive the SR image to an...

Full description

Saved in:
Bibliographic Details
Main Authors: Yingsong Cheng, Yong Ma, Fan Fan, Jiayi Ma, Yuan Yao, Xiaoguang Mei
Format: Article
Language:English
Published: Taylor & Francis Group 2024-12-01
Series:Geo-spatial Information Science
Subjects:
Online Access:https://www.tandfonline.com/doi/10.1080/10095020.2024.2378917
Tags: Add Tag
No Tags, Be the first to tag this record!
_version_ 1832591271528497152
author Yingsong Cheng
Yong Ma
Fan Fan
Jiayi Ma
Yuan Yao
Xiaoguang Mei
author_facet Yingsong Cheng
Yong Ma
Fan Fan
Jiayi Ma
Yuan Yao
Xiaoguang Mei
author_sort Yingsong Cheng
collection DOAJ
description In recent years, significant advances have been achieved in addressing super-resolution (SR) tasks for hyperspectral images, primarily through deep learning-based methodologies. Nevertheless, methods oriented toward optimizing peak signal-to-noise ratio (PSNR) often tend to drive the SR image to an average of several possible SR predictions, resulting in visually over-smoothed outputs. Furthermore, the current landscape of hyperspectral SR techniques exhibits a notable deficiency in accounting for the inherent noise complexities within the realistic data, which hinders their efficacy in real-world scenarios. To address these issues, we propose a novel latent spectral-spatial diffusion model (LSDiff) for single hyperspectral SR. The diffusion model is chosen for its remarkable generative capabilities and noise robustness. However, hyperspectral images are characterized by their exceptionally high spectral dimensions and complex spectral-spatial properties. To address this complexity, we leverage a large-scale pre-trained autoencoder to map the data into a low-dimensional space conducive to diffusion. Additionally, the incorporation of the 3DConvNext module, complemented by meticulously designed loss constraints, empowers the extraction of rich spectral-spatial features that are inherent in hyperspectral data. Subsequently, the diffusion model undergoes efficient optimization through a variant of the variational bound on the data likelihood. During the reverse transformation, LSDiff systematically converts Gaussian noise into SR images, conditioned on the low-resolution input. Comprehensive experiments provide empirical evidence of LSDiff’s clear superiority over existing state-of-the-art SR techniques. It generates images with markedly improved spatial and spectral fidelity while concurrently showcasing robustness to noise.
format Article
id doaj-art-e411962d2e1d4a4c8e0dd1dad4ad0560
institution Kabale University
issn 1009-5020
1993-5153
language English
publishDate 2024-12-01
publisher Taylor & Francis Group
record_format Article
series Geo-spatial Information Science
spelling doaj-art-e411962d2e1d4a4c8e0dd1dad4ad05602025-01-22T14:54:53ZengTaylor & Francis GroupGeo-spatial Information Science1009-50201993-51532024-12-0111610.1080/10095020.2024.2378917Latent spectral-spatial diffusion model for single hyperspectral super-resolutionYingsong Cheng0Yong Ma1Fan Fan2Jiayi Ma3Yuan Yao4Xiaoguang Mei5Department of Electronic Information School, Wuhan University, Wuhan, ChinaDepartment of Electronic Information School, Wuhan University, Wuhan, ChinaDepartment of Electronic Information School, Wuhan University, Wuhan, ChinaDepartment of Electronic Information School, Wuhan University, Wuhan, ChinaDepartment of Electronic Information School, Wuhan University, Wuhan, ChinaDepartment of Electronic Information School, Wuhan University, Wuhan, ChinaIn recent years, significant advances have been achieved in addressing super-resolution (SR) tasks for hyperspectral images, primarily through deep learning-based methodologies. Nevertheless, methods oriented toward optimizing peak signal-to-noise ratio (PSNR) often tend to drive the SR image to an average of several possible SR predictions, resulting in visually over-smoothed outputs. Furthermore, the current landscape of hyperspectral SR techniques exhibits a notable deficiency in accounting for the inherent noise complexities within the realistic data, which hinders their efficacy in real-world scenarios. To address these issues, we propose a novel latent spectral-spatial diffusion model (LSDiff) for single hyperspectral SR. The diffusion model is chosen for its remarkable generative capabilities and noise robustness. However, hyperspectral images are characterized by their exceptionally high spectral dimensions and complex spectral-spatial properties. To address this complexity, we leverage a large-scale pre-trained autoencoder to map the data into a low-dimensional space conducive to diffusion. Additionally, the incorporation of the 3DConvNext module, complemented by meticulously designed loss constraints, empowers the extraction of rich spectral-spatial features that are inherent in hyperspectral data. Subsequently, the diffusion model undergoes efficient optimization through a variant of the variational bound on the data likelihood. During the reverse transformation, LSDiff systematically converts Gaussian noise into SR images, conditioned on the low-resolution input. Comprehensive experiments provide empirical evidence of LSDiff’s clear superiority over existing state-of-the-art SR techniques. It generates images with markedly improved spatial and spectral fidelity while concurrently showcasing robustness to noise.https://www.tandfonline.com/doi/10.1080/10095020.2024.2378917Super-resolutionhyperspectral imageslatent spectral-spatial diffusion modelautoencoder3DConvNext
spellingShingle Yingsong Cheng
Yong Ma
Fan Fan
Jiayi Ma
Yuan Yao
Xiaoguang Mei
Latent spectral-spatial diffusion model for single hyperspectral super-resolution
Geo-spatial Information Science
Super-resolution
hyperspectral images
latent spectral-spatial diffusion model
autoencoder
3DConvNext
title Latent spectral-spatial diffusion model for single hyperspectral super-resolution
title_full Latent spectral-spatial diffusion model for single hyperspectral super-resolution
title_fullStr Latent spectral-spatial diffusion model for single hyperspectral super-resolution
title_full_unstemmed Latent spectral-spatial diffusion model for single hyperspectral super-resolution
title_short Latent spectral-spatial diffusion model for single hyperspectral super-resolution
title_sort latent spectral spatial diffusion model for single hyperspectral super resolution
topic Super-resolution
hyperspectral images
latent spectral-spatial diffusion model
autoencoder
3DConvNext
url https://www.tandfonline.com/doi/10.1080/10095020.2024.2378917
work_keys_str_mv AT yingsongcheng latentspectralspatialdiffusionmodelforsinglehyperspectralsuperresolution
AT yongma latentspectralspatialdiffusionmodelforsinglehyperspectralsuperresolution
AT fanfan latentspectralspatialdiffusionmodelforsinglehyperspectralsuperresolution
AT jiayima latentspectralspatialdiffusionmodelforsinglehyperspectralsuperresolution
AT yuanyao latentspectralspatialdiffusionmodelforsinglehyperspectralsuperresolution
AT xiaoguangmei latentspectralspatialdiffusionmodelforsinglehyperspectralsuperresolution