Generalized Pixel-Aware Deep Function-Mixture Network for Effective Spectral Super-Resolution
33 Pages Posted: 20 Nov 2024
There are 2 versions of this paper
Generalized Pixel-Aware Deep Function-Mixture Network for Effective Spectral Super-Resolution
Generalized Pixel-Aware Deep Function-Mixture Network for Effective Spectral Super-Resolution
Abstract
Recent progress on spectral super-resolution (SR) mainly focuses on directly mapping an RGB image to its HSI counterpart using deep convolutional neural networks, i.e., non-linearly transform the RGB context within a size-fixed receptive field centered at each pixel to its spectrum using a universal deep mapping function. However, in real scenarios, pixels in HSIs inevitably require size-different receptive fields and distinct mapping functions due to their differences in object category or spatial position, and consequently, these existing methods show limited generalization capacity, especially when the imaging scene is complicated. To tackle this issue, we introduce a pixel-aware deep function-mixture network (PADFMN) for SSR, which consists of a novel class of modules called function-mixture (FM) blocks. Each FM block contains several basis functions, represented by parallel subnets with varying receptive field sizes. Additionally, a separate subnet functions as a mixing function, generating pixel-level weights that linearly combine the outputs of the basis functions. This approach allows the network to dynamically adjust the receptive field size and mapping function for each pixel based on its specific characteristics.Through stacking several such FM blocks together and fusing their intermediate feature representations, we can obtain an effective SSR network with flexibility in learning pixel-wise deep mapping functions as well as better generalization capacity. Moreover, with the aim of employing the proposed PADFMN to cope with two more challenging SSR tasks, including cross-sensor SSR (i.e., test on RGB image shot by a new sensor with unseen spectral response function) and scale-arbitrary SSR (i.e., the spectral resolution of HSI to reconstruct can be arbitrarily determined), we extend the core FM blocks to two more generalized versions, namely sensor-guided FM block and scale-guided FM block. The former is able to cast the sensor related information (e.g., spectral response function) into guidance via dynamic filters to assist the spectral reconstruction using the basic FM block. This is beneficial to reduce the distribution shift between the training and test images incurred by unseen RGB sensors in terms of establishing the deep mapping function, thus leading to pleasing performance in cross-sensor SSR tasks. On the other hand, the latter encodes the user-determined spectral resolution to precisely control the channel dimension of the feature output by the last basic FM block via dynamically generating corresponding convolution filters, so that the network can reconstruct HSI with an arbitrarily determined scale meanwhile keeping the spectrum accuracy. We test the proposed method on three benchmark datasets and it achieves state-of-the-art performance in SSR, cross-sensor SSR, and scale-arbitrary SSR tasks.
Keywords: Spectral super-resolution, function-mixture, dynamic network.
Suggested Citation: Suggested Citation