ICCV 2025
Visual quality assessment plays a crucial role in computer vision, serving as a fundamental step in tasks such as image quality assessment (IQA), image super-resolution, document image enhancement, and video restoration. Traditional visual quality assessment techniques often rely on scalar metrics like Peak Signal-to-Noise Ratio (PSNR) and Structural Similarity Index Measure (SSIM), which, while effective in certain contexts, fall short in capturing the perceptual quality experienced by human observers. This gap emphasizes the need for more perceptually aligned and comprehensive evaluation methods that can adapt to the growing demands of applications such as medical imaging, satellite remote sensing, immersive media, and document processing. In recent years, advancements in deep learning, generative models, and multimodal large language models (MLLMs) have opened up new avenues for visual quality assessment. These models offer capabilities that extend beyond traditional scalar metrics, enabling more nuanced assessments through natural language explanations, open-ended visual comparisons, and enhanced context awareness. With these innovations, VQA is evolving to better reflect human perceptual judgments, making it a critical enabler for next-generation computer vision applications.
The VQualA Workshop aims to bring together researchers and practitioners from academia and industry to discuss and explore the latest trends, challenges, and innovations in visual quality assessment. We welcome original research contributions addressing, but not limited to, the following topics:
Submission Details
Papers will be peer-reviewed and comply with the ICCV 2025 proceedings style, format and length. The camera-ready deadline aligns with the main conference. Accepted papers must be registered and presented to ensure their inclusion in the IEEE Xplore Library.
For details, refer to the ICCV 2025 Author Guidelines.
Prof. Alan Bovik (HonFRPS) holds the Cockrell Family Endowed Regents Chair in Engineering in the Chandra Family Department of Electrical and Computer Engineering in the Cockrell School of Engineering at The University of Texas at Austin, where he is Director of the Laboratory for Image and Video Engineering (LIVE). He is a faculty member in the Department of Electrical and Computer Engineering, the Wireless Networking and Communication Group, and the Institute for Neuroscience. His research interests include digital television, digital photography, visual perception, social media, and image and video processing. His work broadly focuses on creating new theories and algorithms that allow for the perceptually optimized streaming and sharing of visual media. The outcomes of his work have the benefits of ensuring the visual satisfaction of billions of viewers worldwide, while substantially reducing global bandwidth consumption. He has published over 1,000 technical articles in these areas. His publications have been cited more than 175,000 times in the literature, his H-index is above 135, and he is listed as a Highly-Cited Researcher by The Web of Science Group. His several books include the Handbook of Image and Video Processing (Academic Press, 2000, 2005), Modern Image Quality Assessment (2006), and the companion volumes The Essential Guides to Image and Video Processing (Academic Press, 2009).
Dr. Balu Adsumilli (IEEE Fellow) is the Head of Media Algorithms group at YouTube/Google, where he and his team research and develop algorithms to transform the uploaded videos to formats played across all your devices. Over the past years, he was instrumental in building and scaling technologies in the areas of video processing, computer vision, video compression, and video quality, which garnered Two Technology and Engineering Emmy awards for Google. Prior to YouTube, he was the Director of Advanced Technology at GoPro, where he led the Camera Architecture, and the Advanced Software teams, and developed their ProTune mode in collaboration with ACES and Technicolor. This paved the way for GoPro cameras capturing Industry neutral formats, and enabled their widespread applicability in the movie and television industry. Dr. Adsumilli serves on the board of the Television Academy, on the Visual Effects Society board, on the NATAS technical committee, on the IEEE Multimedia Signal Processing (MMSP) Technical Committee, the IEEE Image, Video, Multidimensional Signal Processing (IVMSP) Technical Committee, and on ACM Mile High Video Steering Committee. He has co-authored 125+ technical publications and holds 200+ US patents. He is on TPCs and organizing committees for various conferences and organized numerous workshops. He is a Fellow of IEEE, and an active member of ACM, SMPTE, VES, SPIE, and the Internet Society. He received his PhD from the University of California Santa Barbara, and masters from the University of Wisconsin Madison.
Dr. Ying Chen (IEEE M’05 - SM’11) received a B.S. in Applied Mathematics and an M.S. in Electrical Engineering & Computer Science, both from Peking University, in 2001 and 2004, respectively. He received his PhD in Computing and Electrical Engineering from Tampere University of Technology (TUT), Finland, in 2010. Dr. Chen joined Alibaba Group in 2018 as a Senior Director. Before joining Alibaba, his earlier working experiences included Principal Engineer/Manager at Qualcomm Incorporated, San Diego, CA, USA, from 2009 to 2018; Researcher at TUT and Nokia Research Center, Finland, from 2006 to 2009; and Research Engineer at Thomson Corporate Research, Beijing, from 2004 to 2006. Dr. Chen is currently leading the Audiovisual Technology Group in Taobao, Alibaba, supporting end-to-end multimedia features and applications within Taobao. Dr. Chen has been focusing on multimedia algorithms. His group has won various winner awards at CVPR NTIRE, including the 2023 challenge on Quality Assessment of Video Enhancement, the 2022 challenge on Super-Resolution and Enhancement of Compressed Video, as well as the MSU Video Codecs Comparisons (2020, 2021, and 2023-2024). Dr. Chen contributed to three generations of video coding standards, including H.264/AVC, H.265/HEVC, and H.266/VVC, as well as video file format and transport standards. Dr. Chen has served as an editor and a software coordinator for H.264/AVC and H.265/HEVC (both for Multiview and 3D Video extensions). Dr. Chen has been serving as an Associate Editor for IEEE Transactions on Circuits and Systems for Video Technology (T-CSVT). Dr. Chen’s research areas include video coding and transmission, image/video restoration and enhancement, image/video quality assessment, and generative AI for image/video. He has authored or co-authored about 90 academic papers and over 250 granted US patents in the above areas. His publications have been cited more than 20,000 times.
The participants’ submissions will be evaluated on the test set based on the metrics presented in the related paper for each respective task. The top 3 teams for each task will be invited to present their posters during the workshop.