SPS Webinar: 24 August 2022, presented by Dr. Wenming Yang, Dr. Yucheng Hang, and Dr. Bin Xia

You are here

Inside Signal Processing Newsletter Home Page

Top Reasons to Join SPS Today!

1. IEEE Signal Processing Magazine
2. Signal Processing Digital Library*
3. Inside Signal Processing Newsletter
4. SPS Resource Center
5. Career advancement & recognition
6. Discounts on conferences and publications
7. Professional networking
8. Communities for students, young professionals, and women
9. Volunteer opportunities
10. Coming soon! PDH/CEU credits
Click here to learn more.

News and Resources for Members of the IEEE Signal Processing Society

SPS Webinar: 24 August 2022, presented by Dr. Wenming Yang, Dr. Yucheng Hang, and Dr. Bin Xia

Upcoming SPS Webinar!

Title: Attention Mechanisms for Image Super-Resolution: A Brief Review and Beyond
Date: 24 August 2022
Time: 7:30 AM Eastern (New York time)
Duration: Approximately 1 Hour
Presenters: Dr. Wenming Yang, Dr. Yucheng Hang, and Dr. Bin Xia

Based on the IEEE Xplore® article: Deep Learning for Single Image Super-Resolution: A Brief Review
Published: IEEE Transactions on Multimedia, December 2029, available in IEEE Xplore®

Download: The riginal article is available for download.


Register for the Webinar



Recently, deep convolutional neural network (CNNs) have been widely used in Single Image Super-Resolution (SISR) and have obtained great success.

However, most of the existing methods are limited to local receptive field and equal treatment of different types of information; existing methods cannot effectively aggregate hierarchical feature information. To address these issues, we propose an attention cube network (A-CubeNet). Specifically, the adaptive spatial attention branch (ASAB) and the adaptive channel attention branch (ACAB) constitute the adaptive dual attention module (ADAM), which can capture the long-range spatial and channel-wise contextual information to expand the receptive field and distinguish different types of information. Furthermore, the adaptive hierarchical attention module (AHAM) can capture the long-range hierarchical contextual information to flexibly aggregate different feature maps depending on the global context.

Moreover, Non-Local Attention (NLA) brings significant improvement for SISR by leveraging intrinsic feature correlation in natural images. However, NLA gives noisy information large weights and consumes quadratic computation resources with respect to the input size, limiting its performance and application. Therefore, we propose a novel Efficient Non-Local Contrastive Attention (ENLCA), specifically, ENLCA consists of two parts: Efficient Non-Local Attention (ENLA) and Sparse Aggregation. ENLA adopts the kernel method to approximate exponential function and obtains linear computation complexity. For Sparse Aggregation, it makes the network focus on informative features and applies contrastive learning to further separate relevant and irrelevant features.

Experiments demonstrate the superiority of the above two methods over state-of-the-art approaches in both quantitative comparison and visual analysis.


Wenming Yang

Wenming Yang received the Ph.D. degree in information and communication engineering from Zhejiang University in 2006.

He is an Associate Professor in Shenzhen International Graduate School / Department of Electronic Engineering, Tsinghua University. His research interests include image processing, pattern recognition, computer vision and AI in medicine.



Yucheng Hang

Yhanghua Gao received the B.S. degree in telecommunication engineering from Northeastern University, China, in 2006. He is currently pursuing the M.S. degree in communication engineering at Tsinghua University, China.

His research interests include low-level vision and computational photography.




Bin Xia

Bin Xia received the B.S degree in space communication science and technology from Xidian University, Xian, China, in 2020. He is currently pursuing the M.S. degree in Communication Engineering at Tsinghua University, Shenzhen, China.

His research interests include low-level vision and model compression.

SPS on Twitter

  • DEADLINE EXTENDED: The 2023 IEEE International Workshop on Machine Learning for Signal Processing is now accepting… https://t.co/NLH2u19a3y
  • ONE MONTH OUT! We are celebrating the inaugural SPS Day on 2 June, honoring the date the Society was established in… https://t.co/V6Z3wKGK1O
  • The new SPS Scholarship Program welcomes applications from students interested in pursuing signal processing educat… https://t.co/0aYPMDSWDj
  • CALL FOR PAPERS: The IEEE Journal of Selected Topics in Signal Processing is now seeking submissions for a Special… https://t.co/NPCGrSjQbh
  • Test your knowledge of signal processing history with our April trivia! Our 75th anniversary celebration continues:… https://t.co/4xal7voFER

IEEE SPS Educational Resources

IEEE SPS Resource Center

IEEE SPS YouTube Channel