论文标题
修复模型问题的实际见解
Practical Insights of Repairing Model Problems on Image Classification
论文作者
论文摘要
对深度学习模型的额外培训可能会对结果造成负面影响,从而将最初的阳性样本变成负面样本(降解)。由于样本特征的多样性,在现实世界中的用例中可能会降解。也就是说,一组样本是关键的样本的混合物,不应错过且不重要的样本。因此,我们不能仅靠准确性理解性能。尽管现有的研究旨在防止模型退化,但需要对相关方法的见解来掌握其收益和局限性。在本次演讲中,我们将从减少降解的方法的比较中提出含义。特别是,我们就数据集的安排为工业环境制定了用例。结果表明,由于准确性和防止降解之间的权衡,从业者应该关心更好的方法,考虑AI系统的数据集可用性和生命周期。
Additional training of a deep learning model can cause negative effects on the results, turning an initially positive sample into a negative one (degradation). Such degradation is possible in real-world use cases due to the diversity of sample characteristics. That is, a set of samples is a mixture of critical ones which should not be missed and less important ones. Therefore, we cannot understand the performance by accuracy alone. While existing research aims to prevent a model degradation, insights into the related methods are needed to grasp their benefits and limitations. In this talk, we will present implications derived from a comparison of methods for reducing degradation. Especially, we formulated use cases for industrial settings in terms of arrangements of a data set. The results imply that a practitioner should care about better method continuously considering dataset availability and life cycle of an AI system because of a trade-off between accuracy and preventing degradation.