Goto

Collaborating Authors

 esrgan



Deep Self-Dissimilarities as Powerful Visual Fingerprints Supplementary Material 1 Experimental Setting

Neural Information Processing Systems

Tables 4 and 5 provide descriptions of the network architectures we use in each of the experiments. In Tab. 4 both networks consist of We provide additional visualization results for DSD as in Sec. Figure 11 shows an additional visualization of the effect of the DSD loss. Figure 13 shows a visual comparison between different losses that utilize feature distributions. Figure 16 shows additional motion-debluring comparisons. We do this for two pairs of scales (between full res.


Diffusion-Based, Data-Assimilation-Enabled Super-Resolution of Hub-height Winds

Ma, Xiaolong, Dong, Xu, Tarrant, Ashley, Yang, Lei, Kotamarthi, Rao, Wang, Jiali, Yan, Feng, Kettimuthu, Rajkumar

arXiv.org Artificial Intelligence

High-quality observations of hub-height winds are valuable but sparse in space and time. Simulations are widely available on regular grids but are generally biased and too coarse to inform wind-farm siting or to assess extreme-weather-related risks (e.g., gusts) at infrastructure scales. To fully utilize both data types for generating high-quality, high-resolution hub-height wind speeds (tens to ~100m above ground), this study introduces WindSR, a diffusion model with data assimilation for super-resolution downscaling of hub-height winds. WindSR integrates sparse observational data with simulation fields during downscaling using state-of-the-art diffusion models. A dynamic-radius blending method is introduced to merge observations with simulations, providing conditioning for the diffusion process. Terrain information is incorporated during both training and inference to account for its role as a key driver of winds. Evaluated against convolutional-neural-network and generative-adversarial-network baselines, WindSR outperforms them in both downscaling efficiency and accuracy. Our data assimilation reduces WindSR's model bias by approximately 20% relative to independent observations.


Assessing the Impact of Image Super Resolution on White Blood Cell Classification Accuracy

Nagarhalli, Tatwadarshi P., Pawar, Shruti S., Dahanukar, Soham A., Aswalekar, Uday, Save, Ashwini M., Patil, Sanket D.

arXiv.org Artificial Intelligence

Accurately classifying white blood cells from microscopic images is essential to identify several illnesses and conditions in medical diagnostics. Many deep learning technologies are being employed to quickly and automatically classify images. However, most of the time, the resolution of these microscopic pictures is quite low, which might make it difficult to classify them correctly. Some picture improvement techniques, such as image super-resolution, are being utilized to improve the resolution of the photos to get around this issue. The suggested study uses large image dimension upscaling to investigate how picture-enhancing approaches affect classification performance. The study specifically looks at how deep learning models may be able to understand more complex visual information by capturing subtler morphological changes when image resolution is increased using cutting-edge techniques. The model may learn from standard and augmented data since the improved images are incorporated into the training process. This dual method seeks to comprehend the impact of image resolution on model performance and enhance classification accuracy. A well-known model for picture categorization is used to conduct extensive testing and thoroughly evaluate the effectiveness of this approach. This research intends to create more efficient image identification algorithms customized to a particular dataset of white blood cells by understanding the trade-offs between ordinary and enhanced images.


SU-ESRGAN: Semantic and Uncertainty-Aware ESRGAN for Super-Resolution of Satellite and Drone Imagery with Fine-Tuning for Cross Domain Evaluation

Ramkumar, Prerana

arXiv.org Artificial Intelligence

Generative Adversarial Networks (GANs) have achieved realistic super-resolution (SR) of images however, they lack semantic consistency and per-pixel confidence, limiting their credibility in critical remote sensing applications such as disaster response, urban planning and agriculture. This paper introduces Semantic and Uncertainty-Aware ESRGAN (SU-ESRGAN), the first SR framework designed for satellite imagery to integrate the ESRGAN, segmentation loss via DeepLabv3 for class detail preservation and Monte Carlo dropout to produce pixel-wise uncertainty maps. The SU-ESRGAN produces results (PSNR, SSIM, LPIPS) comparable to the Baseline ESRGAN on aerial imagery. This novel model is valuable in satellite systems or UAVs that use wide field-of-view (FoV) cameras, trading off spatial resolution for coverage. The modular design allows integration in UAV data pipelines for on-board or post-processing SR to enhance imagery resulting due to motion blur, compression and sensor limitations. Further, the model is fine-tuned to evaluate its performance on cross domain applications. The tests are conducted on two drone based datasets which differ in altitude and imaging perspective. Performance evaluation of the fine-tuned models show a stronger adaptation to the Aerial Maritime Drone Dataset, whose imaging characteristics align with the training data, highlighting the importance of domain-aware training in SR-applications.


Learning Enhanced Structural Representations with Block-Based Uncertainties for Ocean Floor Mapping

Minoza, Jose Marie Antonio

arXiv.org Artificial Intelligence

Published as a workshop paper at "Tackling Climate Change with Machine Learning", ICLR 2025 Accurate ocean modeling and coastal hazard prediction depend on high-resolution bathymetric data; yet, current worldwide datasets are too coarse for exact numerical simulations. While recent deep learning advances have improved earth observation data resolution, existing methods struggle with the unique challenges of producing detailed ocean floor maps, especially in maintaining physical structure consistency and quantifying uncertainties. This work presents a novel uncertainty-aware mechanism using spatial blocks to efficiently capture local bathymetric complexity based on block-based conformal prediction. Compared to conventional techniques, experimental results over several ocean regions show notable increases in both reconstruction quality and uncertainty estimation reliability. This framework increases the reliability of bathymetric reconstructions by preserving structural integrity while offering spatially adaptive uncertainty estimates, so opening the path for more solid climate modeling and coastal hazard assessment.Figure 1: Learning Enhanced Structural Representations with Block-Based Uncertainties 1 Simple diffusion equations to complex Navier-Stokes equations used in computational fluid dynamics (CFD) span these physical models, all of which depend on thorough bathymetric data to properly forecast tsunami propagation, storm surges, and the effects of sea level rise on coastal communities. The GEBCO project (General Bathymetric Chart of the Oceans), fuses multibeam sonar, satellite altimetry, and shipborne soundings, yet filling in sub-kilometer details globally would take on the order of two centuries at current survey rates Mayer et al. (2018). Enhancement is further complicated by three interrelated factors: (1) heterogeneous data sources with distinct error characteristics and regional resolution gaps; (2) the need to preserve sharp morphological boundaries, such as ridges, canyons, and trenches, that are critical for physical simulations; and (3) spatially varying data quality arising from different acquisition techniques (direct soundings vs. altimetry) that induce nonuniform uncertainty patterns.


Super resolution of histopathological frozen sections via deep learning preserving tissue structure

Yoshai, Elad, Goldinger, Gil, Haifler, Miki, Shaked, Natan T.

arXiv.org Artificial Intelligence

Histopathology plays a pivotal role in medical diagnostics. In contrast to preparing permanent sections for histopathology, a time-consuming process, preparing frozen sections is significantly faster and can be performed during surgery, where the sample scanning time should be optimized. Super-resolution techniques allow imaging the sample in lower magnification and sparing scanning time. In this paper, we present a new approach to super resolution for histopathological frozen sections, with focus on achieving better distortion measures, rather than pursuing photorealistic images that may compromise critical diagnostic information. Our deep-learning architecture focuses on learning the error between interpolated images and real images, thereby it generates high-resolution images while preserving critical image details, reducing the risk of diagnostic misinterpretation. This is done by leveraging the loss functions in the frequency domain, assigning higher weights to the reconstruction of complex, high-frequency components. In comparison to existing methods, we obtained significant improvements in terms of Structural Similarity Index (SSIM) and Peak Signal-to-Noise Ratio (PSNR), as well as indicated details that lost in the low-resolution frozen-section images, affecting the pathologist's clinical decisions. Our approach has a great potential in providing more-rapid frozen-section imaging, with less scanning, while preserving the high resolution in the imaged sample.


Pixel Perfect: ESRGAN-powered High-Resolution Image Upscaling Platform

#artificialintelligence

The MLOps-based application is designed to upscale images by a factor of 4 using ESRGAN, a deep learning-based technique for image super-resolution. The application is hosted on Render's web services and is implemented as a Flask-based API. Users can upload their low-resolution images to the API, and the application will use ESRGAN to upscale them to four times their original resolution. The API is optimised for scalability and can simultaneously handle large volumes of image requests. With this application, users can easily enhance the quality of their images and produce high-resolution versions for their various use cases.


MLP-SRGAN: A Single-Dimension Super Resolution GAN using MLP-Mixer

Mitha, Samir, Choe, Seungho, Maralani, Pejman Jahbedar, Moody, Alan R., Khademi, April

arXiv.org Artificial Intelligence

We propose a novel architecture called MLP-SRGAN, which is a single-dimension Super Resolution Generative Adversarial Network (SRGAN) that utilizes Multi-Layer Perceptron Mixers (MLP-Mixers) along with convolutional layers to upsample in the slice direction. MLP-SRGAN is trained and validated using high resolution (HR) FLAIR MRI from the MSSEG2 challenge dataset. The method was applied to three multicentre FLAIR datasets (CAIN, ADNI, CCNA) of images with low spatial resolution in the slice dimension to examine performance on held-out (unseen) clinical data. Upsampled results are compared to several state-of-the-art SR networks. For images with high resolution (HR) ground truths, peak-signal-to-noise-ratio (PSNR) and structural similarity index (SSIM) are used to measure upsampling performance. Several new structural, no-reference image quality metrics were proposed to quantify sharpness (edge strength), noise (entropy), and blurriness (low frequency information) in the absence of ground truths. Results show MLP-SRGAN results in sharper edges, less blurring, preserves more texture and fine-anatomical detail, with fewer parameters, faster training/evaluation time, and smaller model size than existing methods. Code for MLP-SRGAN training and inference, data generators, models and no-reference image quality metrics will be available at https://github.com/IAMLAB-Ryerson/MLP-SRGAN.


Generative models in a nutshell. Have you ever wished you could create…

#artificialintelligence

Have you ever wished you could create your own realistic images, translate text into multiple languages, or even compose original music? If so, you'll be interested to learn about generative models -- a type of machine learning algorithm that has the power to bring these capabilities to life. Generative models are trained on a dataset and learn the underlying distribution of the data. This allows them to generate new, synthetic examples that are similar to the original dataset. But what makes generative models so special?