核(代数)
嵌入
代表(政治)
核方法
计算机科学
组分(热力学)
特征向量
特征(语言学)
字符串内核
模式识别(心理学)
人工智能
支持向量机
数学
算法
径向基函数核
纯数学
物理
语言学
哲学
政治
政治学
法学
热力学
作者
Minoru Kusaba,Yoshihiro Hayashi,Chang Liu,Araki Wakiuchi,Ryo Yoshida
出处
期刊:Physical review
[American Physical Society]
日期:2023-10-16
卷期号:108 (13)
被引量:5
标识
DOI:10.1103/physrevb.108.134107
摘要
For using machine learning to predict material properties, the feature representation of the materials given to the model plays a fundamental role. A model describes material properties as a function of any given material system expressed as a fixed-length numeric vector, often called a descriptor. However, in most cases, the variables of interest are nontrivial for encoding their compositional or structural features, such as molecules, crystal systems, chemical compositions, and composite materials, into a fixed-length vector. Conventionally, to translate such a multicomponent system into a fixed-length vector, the distribution of predefined component features is summarized into a few summary statistics. The disadvantage of this reduction operation is that some distributional information, such as multimodality, is lost in the vectorization process. Here, we present a general class of material descriptors motivated by the machine-learning theory of kernel mean embedding. Unlike conventional descriptors, kernel mean embedding can retain all information regarding the distribution of component features in the vectorization process. Furthermore, the kernel mean descriptor uniquely determines the inverse map to the original material space. We demonstrate the expressive power and versatility of the kernel mean descriptor in various applications, including the prediction of the formation energy of inorganic compounds, prediction of the chemical composition to form quasicrystalline materials, and the use of force-field parameters to characterize polymeric materials.
科研通智能强力驱动
Strongly Powered by AbleSci AI