PapersFlow Research Brief

Physical Sciences · Computer Science

Image and Video Quality Assessment
Research Guide

What is Image and Video Quality Assessment?

Image and Video Quality Assessment is the development of objective methods to evaluate the perceptual quality of images and videos by modeling human visual perception, encompassing full-reference, reduced-reference, and no-reference approaches across distortions like blur and compression.

The field includes 30,978 works focused on perceptual quality, no-reference assessment, deep learning applications, structural similarity index, HTTP adaptive streaming, blur assessment, and quality of experience in multimedia. Wang et al. (2004) introduced the structural similarity index in 'Image quality assessment: from error visibility to structural similarity,' which has garnered 53,503 citations by shifting from error visibility to structural distortions. Mittal et al. (2012) advanced no-reference methods in 'Making a “Completely Blind” Image Quality Analyzer' and 'No-Reference Image Quality Assessment in the Spatial Domain,' enabling quality prediction without reference images.

Topic Hierarchy

100%
graph TD D["Physical Sciences"] F["Computer Science"] S["Computer Vision and Pattern Recognition"] T["Image and Video Quality Assessment"] D --> F F --> S S --> T style T fill:#DC5238,stroke:#c4452e,stroke-width:2px
Scroll to zoom • Drag to pan
31.0K
Papers
N/A
5yr Growth
453.1K
Total Citations

Research Sub-Topics

Why It Matters

Image and Video Quality Assessment underpins video coding standards like H.264/AVC, where Wiegand et al. (2003) detailed enhanced compression and network-friendly representations in 'Overview of the H.264/AVC video coding standard,' achieving 8,001 citations for applications in streaming and broadcasting. In HTTP adaptive streaming, these metrics optimize quality of experience by balancing bitrate and perceptual fidelity. For instance, BRISQUE from Mittal et al. (2012) in 'No-Reference Image Quality Assessment in the Spatial Domain' (5,447 citations) predicts quality in the spatial domain without distortion-specific features, deployed in real-time monitoring for platforms like Netflix and YouTube to detect artifacts in user-generated content.

Reading Guide

Where to Start

'Image quality assessment: from error visibility to structural similarity' by Wang et al. (2004), as it provides the foundational shift to structural metrics with 53,503 citations and clear explanations of human visual system modeling.

Key Papers Explained

Wang et al. (2004) in 'Image quality assessment: from error visibility to structural similarity' establishes SSIM basics, extended by Wang et al. (2004) in 'Multiscale structural similarity for image quality assessment' for scale invariance and Wang and Bovik (2002) in 'A universal image quality index' for loss of correlation, contrast, and luminance. Mittal et al. (2012) in 'No-Reference Image Quality Assessment in the Spatial Domain' builds NR-IQA using spatial natural scene statistics, while Zhang et al. (2011) in 'FSIM: A Feature Similarity Index for Image Quality Assessment' refines with phase congruency. Horé and Ziou (2010) in 'Image Quality Metrics: PSNR vs. SSIM' quantifies their relationship.

Paper Timeline

100%
graph LR P0["A universal image quality index
2002 · 5.6K cites"] P1["Overview of the H.264/AVC video ...
2003 · 8.0K cites"] P2["Image quality assessment: from e...
2004 · 53.5K cites"] P3["Multiscale structural similarity...
2004 · 5.7K cites"] P4["Making a “Completely Blind” Imag...
2012 · 6.0K cites"] P5["No-Reference Image Quality Asses...
2012 · 5.4K cites"] P6["Neural Collaborative Filtering
2017 · 6.3K cites"] P0 --> P1 P1 --> P2 P2 --> P3 P3 --> P4 P4 --> P5 P5 --> P6 style P2 fill:#DC5238,stroke:#c4452e,stroke-width:2px
Scroll to zoom • Drag to pan

Most-cited paper highlighted in red. Papers ordered chronologically.

Advanced Directions

Recent works build on NR-IQA like BRISQUE for video streaming, but no preprints in the last 6 months indicate focus on integrating these with deep learning for generalization. Frontiers include extending spatial domain models to stereoscopic images and HTTP adaptive streaming quality of experience.

Papers at a Glance

# Paper Year Venue Citations Open Access
1 Image quality assessment: from error visibility to structural ... 2004 IEEE Transactions on I... 53.5K
2 Overview of the H.264/AVC video coding standard 2003 IEEE Transactions on C... 8.0K
3 Neural Collaborative Filtering 2017 6.3K
4 Making a “Completely Blind” Image Quality Analyzer 2012 IEEE Signal Processing... 6.0K
5 Multiscale structural similarity for image quality assessment 2004 5.7K
6 A universal image quality index 2002 IEEE Signal Processing... 5.6K
7 No-Reference Image Quality Assessment in the Spatial Domain 2012 IEEE Transactions on I... 5.4K
8 FSIM: A Feature Similarity Index for Image Quality Assessment 2011 IEEE Transactions on I... 5.0K
9 Image Quality Metrics: PSNR vs. SSIM 2010 4.2K
10 Image information and visual quality 2006 IEEE Transactions on I... 3.9K

Frequently Asked Questions

What is the structural similarity index?

The structural similarity index, introduced by Wang et al. (2004) in 'Image quality assessment: from error visibility to structural similarity,' models human visual perception by comparing luminance, contrast, and structure between images. It outperforms traditional error metrics like MSE by aligning with subjective quality ratings. The index has 53,503 citations, reflecting its foundational role in full-reference assessment.

How does no-reference image quality assessment work?

No-reference methods like BRISQUE from Mittal et al. (2012) in 'No-Reference Image Quality Assessment in the Spatial Domain' use natural scene statistics in the spatial domain to predict quality without a reference image. The model extracts features from naturalness deviations caused by distortions, trained on human opinion scores. It achieves distortion-generic predictions, cited 5,447 times.

What distinguishes SSIM from PSNR?

PSNR measures pixel-wise error via mean squared error, while SSIM from Wang et al. (2004) assesses structural similarity through luminance, contrast, and structure. Horé and Ziou (2010) derived a mathematical relationship in 'Image Quality Metrics: PSNR vs. SSIM,' showing SSIM's superiority for perceptual quality across degradations like Gaussian noise. SSIM correlates better with human judgments.

What is FSIM in image quality assessment?

FSIM, proposed by Zhang et al. (2011) in 'FSIM: A Feature Similarity Index for Image Quality Assessment,' extends SSIM by incorporating phase congruency and gradient magnitude for feature similarity. It improves accuracy on phase-based structures, earning 5,004 citations. The metric aligns closely with subjective scores on standard databases.

Why use multiscale structural similarity?

Multiscale SSIM from Wang et al. (2004) in 'Multiscale structural similarity for image quality assessment' applies SSIM across image scales to capture perception at varying resolutions. It provides a more robust quality measure than single-scale versions, with 5,688 citations. This approach handles multi-resolution distortions effectively.

What role does H.264/AVC play in video quality?

H.264/AVC, overviewed by Wiegand et al. (2003), enhances video compression for better quality at lower bitrates in adaptive streaming. It supports network-friendly representations, cited 8,001 times. Quality assessment metrics evaluate its perceptual performance in multimedia delivery.

Open Research Questions

  • ? How can no-reference models generalize across unseen distortion types without prior knowledge, as limited in current NR-IQA like BRISQUE?
  • ? What features beyond phase congruency and gradients can further improve feature similarity indices like FSIM for diverse image contents?
  • ? How to integrate natural scene statistics with deep learning for blind video quality assessment in streaming scenarios?
  • ? Which multiscale approaches best capture human perception of quality in stereoscopic images and HTTP adaptive streaming?
  • ? How do information-theoretic measures like those in Sheikh and Bovik (2006) extend to dynamic video sequences?

Research Image and Video Quality Assessment with AI

PapersFlow provides specialized AI tools for Computer Science researchers. Here are the most relevant for this topic:

See how researchers in Computer Science & AI use PapersFlow

Field-specific workflows, example queries, and use cases.

Computer Science & AI Guide

Start Researching Image and Video Quality Assessment with AI

Search 474M+ papers, run AI-powered literature reviews, and write with integrated citations — all in one workspace.

See how PapersFlow works for Computer Science researchers