Abstract
Digital representation of materials is crucial in fields such as virtual reality, industrial design and quality control. However, predicting human perception of materials from image data is challenging due to the complexity of material appearances and the intricacies of human vision. This study introduces a perceptual representation termed the 'visual fingerprint', linking image-based measurements of materials to intuitive, human-understandable attributes. We conducted psychophysical studies using standardized video sequences of 347 diverse real-world materials, including fabrics and wood, selected to encompass a broad spectrum of textures, colours and reflective properties. Sixteen key appearance attributes were identified, and over 110 000 human ratings were collected to map perceptual attributes across material categories. By integrating CLIP-derived image features with a multi-layer perceptron model, we developed a predictive framework for material perception. Our results demonstrate that human judgements of appearance and similarity can be accurately predicted using only two images of a material. This work offers a practical and interpretable approach to material representation, enabling intuitive comparisons and retrievals in applications where material appearance is crucial. The proposed material fingerprint and its prediction directly from image data represent a significant step towards simplifying the understanding and interoperability of material properties in diverse digital environments.