Now showing 1 - 7 of 7
  • Placeholder Image
    Publication
    Single noisy image super resolution by minimizing nuclear norm in virtual sparse domain
    (01-01-2018)
    Mandal, Srimanta
    ;
    Super-resolving a noisy image is a challenging problem, and needs special care as compared to the conventional super resolution approaches, when the power of noise is unknown. In this scenario, we propose an approach to super-resolve single noisy image by minimizing nuclear norm in a virtual sparse domain that tunes with the power of noise via parameter learning. The approach minimizes nuclear norm to explore the inherent low-rank structure of visual data, and is further augmented with coarse-to-fine information by adaptively re-aligning the data along the principal components of a dictionary in virtual sparse domain. The experimental results demonstrate the robustness of our approach across different powers of noise.
  • Placeholder Image
    Publication
    Matte super-resolution for compositing
    (22-11-2010)
    Prabhu, Sahana M.
    ;
    Super-resolution of the alpha matte and the foreground object from a video are jointly attempted in this paper. Instead of super-resolving them independently, we treat super-resolution of the matte and foreground in a combined framework, incorporating the matting equation in the image degradation model. We take multiple adjacent frames from a low-resolution video with non-global motion for increasing the spatial resolution. This ill-posed problem is regularized by employing a Bayesian restoration approach, wherein the high-resolution image is modeled as a Markov Random Field. In matte super-resolution, it is particularly important to preserve fine details at the boundary pixels between the foreground and background. For this purpose, we use a discontinuity-adaptive smoothness prior to include observed data in the solution. This framework is useful in video editing applications for compositing low-resolution objects into high-resolution videos. © 2010 Springer-Verlag.
  • Placeholder Image
    Publication
    Unsupervised class-specific deblurring
    (01-01-2018)
    Madam, Nimisha Thekke
    ;
    Kumar, Sunil
    ;
    In this paper, we present an end-to-end deblurring network designed specifically for a class of data. Unlike the prior supervised deep-learning works that extensively rely on large sets of paired data, which is highly demanding and challenging to obtain, we propose an unsupervised training scheme with unpaired data to achieve the same. Our model consists of a Generative Adversarial Network (GAN) that learns a strong prior on the clean image domain using adversarial loss and maps the blurred image to its clean equivalent. To improve the stability of GAN and to preserve the image correspondence, we introduce an additional CNN module that reblurs the generated GAN output to match with the blurred input. Along with these two modules, we also make use of the blurred image itself to self-guide the network to constrain the solution space of generated clean images. This self-guidance is achieved by imposing a scale-space gradient error with an additional gradient module. We train our model on different classes and observe that adding the reblur and gradient modules helps in better convergence. Extensive experiments demonstrate that our method performs favorably against the state-of-the-art supervised methods on both synthetic and real-world images even in the absence of any supervision.
  • Placeholder Image
    Publication
    PIRM challenge on perceptual image enhancement on smartphones: Report
    (01-01-2019)
    Ignatov, Andrey
    ;
    Timofte, Radu
    ;
    Van Vu, Thang
    ;
    Luu, Tung Minh
    ;
    Pham, Trung X.
    ;
    Van Nguyen, Cao
    ;
    Kim, Yongwoo
    ;
    Choi, Jae Seok
    ;
    Kim, Munchurl
    ;
    Huang, Jie
    ;
    Ran, Jiewen
    ;
    Xing, Chen
    ;
    Zhou, Xingguang
    ;
    Zhu, Pengfei
    ;
    Geng, Mingrui
    ;
    Li, Yawei
    ;
    Agustsson, Eirikur
    ;
    Gu, Shuhang
    ;
    Van Gool, Luc
    ;
    de Stoutz, Etienne
    ;
    Kobyshev, Nikolay
    ;
    Nie, Kehui
    ;
    Zhao, Yan
    ;
    Li, Gen
    ;
    Tong, Tong
    ;
    Gao, Qinquan
    ;
    Hanwen, Liu
    ;
    Michelini, Pablo Navarrete
    ;
    Dan, Zhu
    ;
    Fengshuo, Hu
    ;
    Hui, Zheng
    ;
    Wang, Xiumei
    ;
    Deng, Lirui
    ;
    Meng, Rang
    ;
    Qin, Jinghui
    ;
    Shi, Yukai
    ;
    Wen, Wushao
    ;
    Lin, Liang
    ;
    Feng, Ruicheng
    ;
    Wu, Shixiang
    ;
    Dong, Chao
    ;
    Qiao, Yu
    ;
    Vasu, Subeesh
    ;
    Thekke Madam, Nimisha
    ;
    Kandula, Praveen
    ;
    ;
    Liu, Jie
    ;
    Jung, Cheolkon
    This paper reviews the first challenge on efficient perceptual image enhancement with the focus on deploying deep learning models on smartphones. The challenge consisted of two tracks. In the first one, participants were solving the classical image super-resolution problem with a bicubic downscaling factor of 4. The second track was aimed at real-world photo enhancement, and the goal was to map low-quality photos from the iPhone 3GS device to the same photos captured with a DSLR camera. The target metric used in this challenge combined the runtime, PSNR scores and solutions’ perceptual results measured in the user study. To ensure the efficiency of the submitted models, we additionally measured their runtime and memory requirements on Android smartphones. The proposed solutions significantly improved baseline results defining the state-of-the-art for image enhancement on smartphones.
  • Placeholder Image
    Publication
    Inpainting in multi-image stereo
    (22-11-2010)
    Bhavsar, Arnav V.
    ;
    In spite of numerous works on inpainting, there has been little work addressing both image and structure inpainting. In this work, we propose a new method for inpainting both image and depth of a scene using multiple stereo images. The observations contain unwanted artifacts, which can be possibly caused due to sensor/lens damage or occluders. In such a case, all the observations contain missing regions which are stationary with respect to the image coordinate system. We exploit the fact that the information missing in some images may be present in other images due to the motion cue. This includes the correspondence information for depth estimation/inpainting as well as the color information for image inpainting. We establish our approaches in the belief propagation (BP) framework which also uses the segmentation cue for estimation/inpainting of depth maps. © 2010 Springer-Verlag.
  • Placeholder Image
    Publication
    Scale-recurrent multi-residual dense network for image super-resolution
    (01-01-2019)
    Purohit, Kuldeep
    ;
    Mandal, Srimanta
    ;
    Recent advances in the design of convolutional neural network (CNN) have yielded significant improvements in the performance of image super-resolution (SR). The boost in performance can be attributed to the presence of residual or dense connections within the intermediate layers of these networks. The efficient combination of such connections can reduce the number of parameters drastically while maintaining the restoration quality. In this paper, we propose a scale recurrent SR architecture built upon units containing series of dense connections within a residual block (Residual Dense Blocks (RDBs)) that allow extraction of abundant local features from the image. Our scale recurrent design delivers competitive performance for higher scale factors while being parametrically more efficient as compared to current state-of-the-art approaches. To further improve the performance of our network, we employ multiple residual connections in intermediate layers (referred to as Multi-Residual Dense Blocks), which improves gradient propagation in existing layers. Recent works have discovered that conventional loss functions can guide a network to produce results which have high PSNRs but are perceptually inferior. We mitigate this issue by utilizing a Generative Adversarial Network (GAN) based framework and deep feature (VGG) losses to train our network. We experimentally demonstrate that different weighted combinations of the VGG loss and the adversarial loss enable our network outputs to traverse along the perception-distortion curve. The proposed networks perform favorably against existing methods, both perceptually and objectively (PSNR-based) with fewer parameters.
  • Placeholder Image
    Publication
    Deep decoupling of defocus and motion blur for dynamic segmentation
    (01-01-2016)
    Punnappurath, Abhijith
    ;
    Balaji, Yogesh
    ;
    Mohan, Mahesh
    ;
    We address the challenging problem of segmenting dynamic objects given a single space-variantly blurred image of a 3D scene captured using a hand-held camera. The blur induced at a particular pixel on a moving object is due to the combined effects of camera motion, the object’s own independent motion during exposure, its relative depth in the scene, and defocusing due to lens settings. We develop a deep convolutional neural network (CNN) to predict the probabilistic distribution of the composite kernel which is the convolution of motion blur and defocus kernels at each pixel. Based on the defocus component, we segment the image into different depth layers. We then judiciously exploit the motion component present in the composite kernels to automatically segment dynamic objects at each depth layer. Jointly handling defocus and motion blur enables us to resolve depth-motion ambiguity which has been a major limitation of the existing segmentation algorithms. Experimental evaluations on synthetic and real data reveal that our method significantly outperforms contemporary techniques.