联合分析
生成语法
计算机科学
人工智能
生成模型
自然语言处理
认知心理学
心理学
数学
统计
偏爱
作者
Ankit Sisodia,Alex Burnap,Vineet Kumar
标识
DOI:10.1177/00222437241276736
摘要
This article develops a method to automatically discover and quantify human-interpretable visual characteristics directly from product image data. The method is generative and can create new visual designs spanning the space of visual characteristics. It builds on disentanglement methods in deep learning using variational autoencoders, which aim to discover underlying statistically independent and interpretable visual characteristics of an object. The impossibility theorem in the deep learning literature indicates that supervision with ground truth characteristics would be required to obtain unique disentangled representations. However, these are typically unknown in real-world applications, and are in fact exactly the characteristics that need to be discovered. Extant machine learning methods are unsuitable since they require ground truth labels for each visual characteristic. In contrast, this method postulates the use of readily available product characteristics (such as brand and price) as proxy supervisory signals to enable disentanglement. This method discovers and quantifies human-interpretable and statistically independent characteristics without any specific domain knowledge on the product category. It is applied to a dataset of watches to automatically discover interpretable visual product characteristics, obtain consumer preferences over visual designs, and generate new ideal point designs targeted to specific consumer segments.
科研通智能强力驱动
Strongly Powered by AbleSci AI