Have you heard of the VDSR upscaling algorithm?

  • Thread starter Thread starter bhobba
  • Start date Start date
  • Tags Tags
    Algorithm
Click For Summary
SUMMARY

The VDSR (Very Deep Super Resolution) algorithm is a notable method for upscaling images, utilizing 20 convolutional layers with 3x3x64 filters. The key innovation lies in the 'residual learning' approach, which allows the model to generate a high-resolution image by learning the difference between the low-resolution input and the high-resolution output. This technique significantly reduces memory requirements by producing a single residual image rather than a vast number of intermediate images. Understanding the residual step is crucial for grasping the algorithm's efficiency and effectiveness.

PREREQUISITES
  • Understanding of convolutional neural networks (CNNs)
  • Familiarity with image processing concepts
  • Knowledge of residual learning techniques
  • Basic understanding of sparse matrix techniques
NEXT STEPS
  • Research the VDSR algorithm in detail, focusing on the 'Residual-Learning' section of the original paper
  • Explore convolutional neural network architectures and their applications in image processing
  • Study sparse matrix techniques and their relevance in deep learning
  • Learn about other image upscaling algorithms, such as SRCNN and EDSR
USEFUL FOR

Machine learning practitioners, image processing engineers, and researchers interested in advanced image upscaling techniques and convolutional neural networks.

Messages
11,011
Reaction score
3,862
Hi All

Here is an interesting algorithm for up-scaling images:
https://arxiv.org/pdf/1511.04587.pdf

It has 20 3x3x64 (some I have read use 3x3x16) convolutions. Now my understanding of convolutions means, the first produces 64 images from the 64 filers, then in the second convolution each of the 64 images has 64 filters applied etc so the last convolution would have 20^64 images. That's simply impossible memory wise. Either my understanding of convolutions is wrong or something else is going on. Maybe because its producing a difference output between the up-scaled image and bi-cubic up-scaling nearly all elements are zero and you can use sparse matrix techniques? I have read pruning generally is not used, which to me also suggests sparse array techniques

Thanks
Bill
 
Technology news on Phys.org
I've only skimmed it but I think you are missing the 'residual' step between each convolution. So each step is a convolution of 64 filters over the LR image which are then reduced down to 1 residual HR image. The details of this residual step are I think the essence of the novelty of the method described in the paper; see the sub-section titled 'Residual-Learning' in section 3.2.

But this is not my area of expertise (although it looks fascinating and the results appear remarkable) so I may be completely wrong :smile:
 
  • Like
Likes   Reactions: bhobba

Similar threads

  • · Replies 30 ·
2
Replies
30
Views
7K
  • · Replies 13 ·
Replies
13
Views
5K
  • · Replies 1 ·
Replies
1
Views
2K
  • · Replies 6 ·
Replies
6
Views
4K
  • · Replies 7 ·
Replies
7
Views
2K
Replies
4
Views
2K
  • · Replies 1 ·
Replies
1
Views
2K
  • · Replies 3 ·
Replies
3
Views
3K
  • · Replies 3 ·
Replies
3
Views
3K
Replies
29
Views
6K