Material fictions

Comparing physically based renderings and generative AI images through material perception

Journal Article (2026)
Author(s)

Yuguang Zhao (TU Delft - Perceptual Intelligence)

Jeroen Stumpel (Universiteit Utrecht)

Huib de Ridder (TU Delft - Human Technology Relations)

Jan Jaap R. van Assen (TU Delft - Perceptual Intelligence)

Maarten W.A. Wijntjes (TU Delft - Perceptual Intelligence)

Research Group
Perceptual Intelligence
DOI related publication
https://doi.org/10.1167/jov.26.3.7 Final published version
More Info
expand_more
Publication Year
2026
Language
English
Research Group
Perceptual Intelligence
Journal title
Journal of vision
Issue number
3
Volume number
26
Downloads counter
26
Reuse Rights

Other than for strictly personal use, it is not permitted to download, forward or distribute the text or part of it, without the consent of the author(s) and/or copyright holder(s), unless the work is under an open content license such as Creative Commons.

Abstract

Generative artificial intelligence (AI) models unlock new ways to create images, emerging as a new medium alongside paintings, photographs, physically based renderings (PBR), etc. Generative AI images can be perceptually convincing without being physically plausible, allowing to investigate the boundaries of visual perception. This study examines whether generative AI images adhere to a medium-independent perceptual space converged from previous studies. We compared the perceptual similarity of images from three generative AI models against a bidirectional reflectance distribution functions (BRDFs) PBR image dataset, using human similarity judgments. In experiment 1, we used the text descriptions of 32 materials (e.g., blue acrylic) from the Mitsubishi Electric Research Laboratories (MERL) BRDF dataset, prompting two text-to-image models, DALL-E 2 and Midjourney v2, to generate 32 sphere-shaped stimuli per model. Perceptual spaces derived from similarity judgments revealed that both AI models resulted in two-dimensional spaces whereas the MERL space was confined to one dimension, probably owing to a lack of surface texture. These unrelated perceptual spaces suggest the AI models generated unique and different images from identical text prompts. In experiment 2 we used the text-to-image model Stable Diffusion v1.5 with ControlNet for additional depth-map constraints. Using the same 32 descriptions, we generated 3 sets using 3 different depth maps. The three resulting perceptual spaces are all two-dimensional, exhibiting high similarity, indicating a robust and non-random structure. They also show a similar structure to the MERL space and perceptual spaces from other material studies using photographs, PBR, and depictions, suggesting AI-generated imagery may indeed be used as a new medium to explore material perception.