Image quality perception, I've often found, varies massively from person to person. Some can't tell the difference between a game running with DLSS set to Performance and one running at Native, while others can easily ignore the blurriness of a poor TAA implementation while their peers are busy climbing the walls. Intel's new tool, however, attempts to drill down on image quality and provide a quantifiable end result to give game developers a helping hand.
The (CVGM) tool aims to detect and rate distortions introduced by modern rendering techniques and aids, like neural supersampling, path tracing, and variable rate shading, in order to provide a useful evaluation result.
I'm sure you were waiting for this part: A 3D CNN model (ie, the sort of AI model used in many traditional AI-image enhancement techniques) was then to predict image quality by comparing the reference and distorted videos. The tool then uses the model to detect and rate visual errors, and provides a global quality score
along with per-pixel error maps, which highlight artifacts—and even attempts to identify how they may have occurred.
What you end up with after all those words, according to Intel, is a tool that outperforms all the other current metrics when it comes to predicting how humans will judge visual distortions. Not only does it predict how distracting a human player will find an error, but it also provides easily-interpretable maps to show exactly where it's [[link]] occurring in a scene. Intel hopes it will be used to optimise quality and performance trade-offs when implementing upscalers, and provide smarter reference generation for training denoising algorithms.
"Whether you’re training neural renderers, evaluating engine updates, or testing new upscaling techniques, having a perceptual metric that aligns with human judgment is a huge advantage", says Intel.
"While [CGVQM's] current reliance on reference videos limits some applications, ongoing work aims to expand CGVQM’s reach [[link]] by incorporating saliency, motion coherence, and semantic awareness, making it even more robust for real-world scenarios."
Cool. You don't have to look far on the interwebs to find people complaining about visual artifacts introduced by some of these modern image-quality-improving and frame rate-enhancing techniques (). So, anything that allows devs to get a better bead on how distracting they might be seems like progress to me. The tool is as a PyTorch implementation, so [[link]] have at it, devs.

👉👈
1. Best overall:
2. Best value:
3. Best budget:
4. Best mid-range:
5. Best high-end: