论文标题
通过CNN的最佳特征,用于食品检测的新型多粉核极限学习机
Novel Multicolumn Kernel Extreme Learning Machine for Food Detection via Optimal Features from CNN
论文作者
论文摘要
自动食品检测是一个有趣的主题,因为它的广泛应用程序从在社交媒体平台上检测食品图像到饮食评估应用程序中用户的非食品照片的广泛应用。最近,在Covid-19大流行期间,它通过自动从公共场所的相机自动检测饮食活动来促进禁令。因此,为了应对以高精度识别食物图像的挑战,我们提出了一个混合框架的想法,用于从高效的神经网络中提取和选择最佳特征。在此,采用非线性分类器来区分以非常精确的线性不可分割的特征向量。根据这个想法,我们的方法从MobilenetV3中提取特征,通过使用Shapley添加说明(SHAP)值选择最佳的属性子集,并由于其非线性决策边界和良好的概括能力而利用内核极限学习机(KELM)。但是,由于具有大量隐藏节点的内核矩阵的复杂计算,KELM遭受了大型数据集的“维数问题的诅咒”。我们通过提出一种新颖的多柱核极限学习机(McKelm)来解决这个问题,该机器(McKelm)利用K-D树算法将数据分为N子集并在每个数据子集上训练单独的Kelm。然后,该方法将KELM分类器结合到并行结构中,并通过使用K-D树搜索进行分类输入而不是整个网络,从而在测试过程中选择顶部K最近的子集。为了评估提出的框架,大型食品/非食品数据集是使用九个公开可用数据集准备的。实验结果表明,我们方法在一组集成措施上的优越性,同时解决了大型数据集中KELM中维数的诅咒问题。
Automatic food detection is an emerging topic of interest due to its wide array of applications ranging from detecting food images on social media platforms to filtering non-food photos from the users in dietary assessment apps. Recently, during the COVID-19 pandemic, it has facilitated enforcing an eating ban by automatically detecting eating activities from cameras in public places. Therefore, to tackle the challenge of recognizing food images with high accuracy, we proposed the idea of a hybrid framework for extracting and selecting optimal features from an efficient neural network. There on, a nonlinear classifier is employed to discriminate between linearly inseparable feature vectors with great precision. In line with this idea, our method extracts features from MobileNetV3, selects an optimal subset of attributes by using Shapley Additive exPlanations (SHAP) values, and exploits kernel extreme learning machine (KELM) due to its nonlinear decision boundary and good generalization ability. However, KELM suffers from the 'curse of dimensionality problem' for large datasets due to the complex computation of kernel matrix with large numbers of hidden nodes. We solved this problem by proposing a novel multicolumn kernel extreme learning machine (MCKELM) which exploited the k-d tree algorithm to divide data into N subsets and trains separate KELM on each subset of data. Then, the method incorporates KELM classifiers into parallel structures and selects the top k nearest subsets during testing by using the k-d tree search for classifying input instead of the whole network. For evaluating a proposed framework large food/non-food dataset is prepared using nine publically available datasets. Experimental results showed the superiority of our method on an integrated set of measures while solving the problem of 'curse of dimensionality in KELM for large datasets.