Ensemble probabilistic quantization encoding for information preservation of numerical variables in convolutional neural networks

卷积神经网络中数值变量信息保持的集成概率量化编码

阅读:1

Abstract

One-hot encoding is a prevalent method used to convert numeric variables into categorical variables. But one-hot encoding omits crucial quantitative data, which compromises the performance of convolutional neural networks (CNNs). This study introduces the ensemble probabilistic quantization encoding, where each class is treated as a quantum with distinct values and the classes collaborate in an ensemble manner to preserve numerical information. This method uses the cross-entropy loss function, enhancing its robustness to outliers. Moreover, classes collaborate in an ensemble fashion to yield more diverse and enriched outcomes. We compared three encoding techniques-ensemble probabilistic quantization, one-hot encoding, label smoothing, and mean square error-using the same dataset and model structure. Our investigations into the impact of quantitative information loss on CNN performance revealed that omitting this information significantly undermines CNN functionality. Ensemble probabilistic quantization proved less dependent on the number of classes than the other methods, thus maintaining effectiveness even with fewer classes. In conclusion, the efficient transmission of quantitative information from numerical to categorical variables is essential for optimal CNN performance. Ensemble probabilistic quantization effectively conveys diverse quantitative information with fewer classes, outperforming one-hot encoding and label smoothing when class numbers are limited.

特别声明

1、本页面内容包含部分的内容是基于公开信息的合理引用;引用内容仅为补充信息,不代表本站立场。

2、若认为本页面引用内容涉及侵权,请及时与本站联系,我们将第一时间处理。

3、其他媒体/个人如需使用本页面原创内容,需注明“来源:[生知库]”并获得授权;使用引用内容的,需自行联系原作者获得许可。

4、投稿及合作请联系:info@biocloudy.com。