论文标题
通过文本增强域的适应性进行半监督协作过滤
Semi-supervised Collaborative Filtering by Text-enhanced Domain Adaptation
论文作者
论文摘要
数据稀疏性是推荐系统中固有的挑战,其中大多数数据都是从用户的隐式反馈中收集的。这在设计有效算法时造成了两个困难:首先,大多数用户与系统仅有一些交互,并且没有足够的数据进行学习。其次,隐式反馈中没有负样本,进行负面抽样以产生负样本是一种常见的做法。但是,这导致了许多潜在的积极样本被错误标记为负面样本,并且数据稀疏会加剧错误标签的问题。为了解决这些困难,我们将有关稀疏隐式反馈的推荐问题视为半监督的学习任务,并探索域的适应性来解决它。我们将学习的知识从密集的数据转移到稀疏数据,我们专注于最具挑战性的情况 - 没有用户或项目重叠。在这种极端情况下,直接对齐两个数据集的嵌入是相当最佳的,因为两个潜在空间编码非常不同的信息。因此,我们采用域不变的文本特征,因为锚指向了潜在空间。要对齐嵌入,我们为每个用户和项目提取文本功能,然后将它们与用户和项目的嵌入为域分类器。对嵌入式训练,以使分类器和文本特征作为锚点固定。根据域的适应,源域中的分布模式转移到目标域。由于可以通过域的适应来监督目标部分,因此我们放弃目标数据集中的负采样,以避免标签噪声。我们采用三对现实数据集来验证转移策略的有效性。结果表明,我们的模型表现明显优于现有模型。
Data sparsity is an inherent challenge in the recommender systems, where most of the data is collected from the implicit feedbacks of users. This causes two difficulties in designing effective algorithms: first, the majority of users only have a few interactions with the system and there is no enough data for learning; second, there are no negative samples in the implicit feedbacks and it is a common practice to perform negative sampling to generate negative samples. However, this leads to a consequence that many potential positive samples are mislabeled as negative ones and data sparsity would exacerbate the mislabeling problem. To solve these difficulties, we regard the problem of recommendation on sparse implicit feedbacks as a semi-supervised learning task, and explore domain adaption to solve it. We transfer the knowledge learned from dense data to sparse data and we focus on the most challenging case -- there is no user or item overlap. In this extreme case, aligning embeddings of two datasets directly is rather sub-optimal since the two latent spaces encode very different information. As such, we adopt domain-invariant textual features as the anchor points to align the latent spaces. To align the embeddings, we extract the textual features for each user and item and feed them into a domain classifier with the embeddings of users and items. The embeddings are trained to puzzle the classifier and textual features are fixed as anchor points. By domain adaptation, the distribution pattern in the source domain is transferred to the target domain. As the target part can be supervised by domain adaptation, we abandon negative sampling in target dataset to avoid label noise. We adopt three pairs of real-world datasets to validate the effectiveness of our transfer strategy. Results show that our models outperform existing models significantly.