## ABSTRACT

Recently, convolutional neural networks (CNNs) have achieved excellent performance for the recommendation system by extracting deep features and building collaborative filtering models. However, CNNs have been verified susceptible to adversarial examples. This is because adversarial samples are subtle non-random disturbances, which indicates that machine learning models produce incorrect outputs. Therefore, we propose a novel model of Adversarial Neural Collaborative Filtering with Embedding Dimension Correlations, named ANCF in short, to address the adversarial problem of CNN-based recommendation system. In particular, the proposed ANCF model adopts the matrix factorization to train the adversarial personalized ranking in the prediction layer. This is because matrix factorization supposes that the linear interaction of the latent factors, which are captured between the user and the item, can describe the observable feedback, thus the proposed ANCF model can learn more complicated representation of their latent factors to improve the performance of recommendation. In addition, the ANCF model utilizes the outer product instead of the inner product or concatenation to learn explicitly pairwise embedding dimensional correlations and obtain the interaction map from which CNNs can utilize its strengths to learn high-order correlations. As a result, the proposed ANCF model can improve the robustness performance by the adversarial personalized ranking, and obtain more information by encoding correlations between different embedding layers. Experimental results carried out on three public datasets demonstrate that the ANCF model outperforms other existing recommendation models.

## 1 INTRODUCTION

Since recommendation systems (RS) can alleviate information overload and provide an effective solution for users’ information search, they are widely adopted in web applications such as E-business, social software, and so on. Generally, the collaborative filtering (CF) approaches are one of crucial methods among various recommendation technologies because of their capabilities of both higher efficiency and accuracy.

In particular, matrix factorization (MF) method is one of the most popular CF approaches since the vectors in the MF can represent latent features of each user and item. Moreover, the inner products of latent features vectors can approximate the user-item interaction well, and are powerful for catching the low-rank structure of sparse data of the interaction between user and item, however, its concision and linearity limit the representation of the predictive function [1, 2]. Recently, a growing number of attempts have been made to address the issues, including two main groups: One improves the model itself to learn user and item representations via deep neural networks (DNNs); the other enhances the learning strategy, e.g. Bayesian Personalized Ranking (BPR) [3], learned MF in pairwise ranking perspective [4], etc.

However, DNNs-based approaches have been verified susceptible to adversarial examples recently. This is because adversarial samples are subtle non-random disturbances, which indicates that machine learning (ML) models produce incorrect outputs. A large number of studies have reported the failure of ML-based RS models against adversarial attacks. To improve the robustness, Goodfellow et al. [5] and Moosavi-Dezfooli et al. [6] developed adversarial training approaches that can correctly classify the dynamically generated adversarial examples. Inspired by adversarial learning, He et al. [7] designed the Adversarial Personalized Ranking (APR) to replace the traditional BPR [3], but the effect is neglect, especially in top recommendation with a small *k* value.

To highlight the importance of modeling dimensional correlations and improve on the performance of the robustness for the RS, we present a novel CF-based model with the adversarial training, named Adversarial Neural Collaborative Filtering with Embedding Dimension Correlations, ANCF in short. In particular, the proposed ANCF model adopts the matrix factorization to train the adversarial personalized ranking in the prediction layer. This is because matrix factorization supposes that the linear interaction of the latent factors, which are captured between the user and the item, can describe the observable feedback, so the ANCF can learn a much more complicated representation of latent factors to improve the performance of recommendation. In addition, ANCF utilizes the outer product instead of the inner product or concatenation to learn pairwise embedding dimensional correlations explicitly, and obtain the interaction map from which CNNs can utilize its strengths to learn high-order correlations. Therefore, the proposed ANCF model can improve the robustness performance by the adversarial personalized ranking, and obtain more information by encoding correlations between different embedding layers. Experimental results from three public datasets demonstrate that the ANCF model outperforms other existing recommendation models.

The contribution of the proposed model is described as follows:

The proposed model learns high-order correlations from feature map

*E*via CNN.The proposed model can solve the adversarial problems via an adversarial matrix factorization.

## 2. RELATED RESEARCH WORK

The paper focuses on the CNN-based CF and adversarial training. Therefore, we introduce their latest developments and applications in the RS in this section.

### 2.1 CNN-based Collaborative Filtering

With the development of DNNs in the area of RS, neural collaborative filtering (NCF) has recently become the most popular framework among the DNN-based CF approaches [8]. This is because NCF utilizes DNNs to improve either the user and item representation learning or the predictive function much better [9, 10, 11, 12, 13, 14]. However, there is still a problem to be addressed in these NCF models recently. That is the correlations of the embedding dimensions resulted from the predictive function. Generally, traditional NCF models often utilize a multi-layer perceptron (MLP) based on the concatenation or the element-wise product of embedding between the user and the item [8, 14]. Afterward, Du et al. presented a model named ConvNCF to learn the high-order correlations of the embedding dimensions by utilizing CNNs-based model via the outer product [15].

Inspired by the ConvNCF model [15], this paper adopts matrix factorization trained with APR (i.e., Adversarial Matrix Factorization, AMF) to solve the adversarial problem via a different way.

### 2.2 Adversarial Recommendation Systems

Adversarial machine learning (AML) focuses on the learning algorithms resisting adversarial attacks and studying benefits and drawbacks of attackers to support appropriate solutions [16, 17]. In recent years, many works have pointed out the failure of machine learning recommendation models. Therefore, He et al. [7] proposed an adversarial learning framework for recommendation at first. The proposed adversarial personalized ranking (APR) model checked both the robustness to adversarial perturbations of users and embedded items of BPR-MF [3]. Afterward, Anelli et al. [16] researched iterative perturbation technologies and proved the ineffectiveness of the APR in protecting the RS from attacks.

Generally, adversarial training involves appending adversarial samples, generated by particular attack models such as FGSM [5] or BIM [17], into the training process. According to reports, both in RS [18, 19] and ML [20], this kind of training process results in the robustness against adversary samples, and achieves better performance of generalization against clean samples.

Afterward, AML has been utilized to create fresh generative models, known as generative adversarial networks (GANs). According to different applications, GAN-based models could improve the negative sampling for the learning sequencing objective function [21, 22], predict missing scores [23, 24] by using time [24, 25], and auxiliary information fitting synthesizers, or enhance training datasets [26, 27]. However, we here focus on the Adversarial Matrix Factorization (AMF) instead of GAN due to its computation consumption [28].

## 3 PROPOSED MODEL

We propose a novel neural network approach named Adversarial Convolutional Neural CF with Embedding Correlations (ANCF), inspired by the work of [15].

This paper selects CNN as the fundamental neural structure due to three advantages as follows.

CNN can deal with the feature map well due to its presence as a 2D matrix;

The sub-region of the feature map has a dimensional relationship represented by CNN;

CNN can capture the correlations of features both locally and globally.

As shown in Figure 1, the ANCF framework consists of four components as follows:

The first layer is the embedding and input layer, which contains two embedding functions:

*f*^{U}(*u*) and*f*^{l}(*i*). It produces two vectors (of size 64) which represents user*u*and item*i*respectively.The second layer is interaction map layer, which computes the pairwise correlations of the vector after the embedding and input layer by the Interaction Map

*E*fed to the ConvNCF Layers.The third layer is ConvNCF Layers including six convolutional layers, following a tower structure with 32 feature maps in each CNN Layer and outputting a tensor in the last CNN layer.

The last prediction layer obtains prediction

*ŷ*trained with the APR to output the final result._{ui}

### 3.1 Layer of Input and Embedding

Given a user *u* and an item *i* and their features, we first encode their features by one-hot encoding and get their embedding *f*^{U}(*u*) and *f*^{l}(*i*) via the equation 1:

where,

$v$

_{u}^{U}: the feature vector of user*u*;$v$

_{j}^{l}: the feature vector for item*i*;ℙ ɛ ℝ

^{M × K}: the embedding matrix for user features;ℚ ɛ ℝ

^{N × K}: the embedding matrix for item features;M: the number of user features;

K: the embedding size;

N: the number of item features.

### 3.2 Layer of the Interaction Map

Although recent works have shown the superiority of inner-product over complex neural networks (CNNs, MLPs), in terms of efficiency and effectiveness, and the applying outer product with CNNs has more time complexity, we replace the inner product with the outer product, to construct interaction map of the user and the item embedding. This is because the advantages of outer product are reflected in the following four aspects:

It does not have the disadvantage of element product only considering the diagonal elements of the interaction map;

It can obtain more information by encoding correlations among various embedding vectors;

It is more effective than the connection operation merely preserving the original information of the embedding vector and does not model any other correlation.

The interaction map layer allows the two embedding vectors (*f*^{U}(*u*), *f*^{l}(*i*)) to do outer product to get the interaction map E, shown in the following equation 2:

where the (*k*_{1}, *k*_{2}) - *th* element in *E* is: $ek1,k2=fU(u)k1\u010bfI(i)k2T$ Obviously, all correlations of the pairwise embedding dimension are encoded in *E*.

### 3.3 ConvNCF Layer

#### 3.3.1 Neural Collaborative Filtering

Neural collaborative filtering is a set of CF models based on the DNNs, in which side information is defined as *S _{u}*

^{user}and

*S*

_{i}^{item}, the scoring function is shown as the equation 3:

where,

function f(·) is the multi-layer perceptron;

θ is the parameters of the network.

Recently, multi-layer perceptron (MLP) has been extensively investigated in the NCF tasks. This is because many existing RS models are linear methods in essence. However, MLP can improve recommendation performance via adding nonlinear transformations and interpreting them into neural extensions [8].

Despite MLP's success, there are still some shortcomings. MLP is easy to overfit and needs more computing resources due to many parameters. For explicit feedback, the whole network can be trained with weighted square loss. And for the implicit feedback, the whole network can be trained with weighted binary cross-entropy loss. Equation 4 is the definition of the cross-entropy loss.

#### 3.3.2 ConvNCF

Based on the NCF, we designed a ConvNCF layer which sets up 32 kernel for each convolution layer and generates a feature map c. A *2D* matrix *E*^{lc} represents a feature map *c* in the convolutional layer l, and its size is the half of its previous layer *l* - 1 since the stride is 2. For layer l, a 3D tensor ε^{l} represented all feature maps together. There are 2 x 2 sizes with no padding of convolutional kernels.

Given the interaction map *E* of the input, we can obtain the feature maps from each layer in the equation 5 as follows:

*e*, the entry in the interaction map;_{x, y}*E*, a product of*f*and^{U}(u)_{x}*f*;^{I}(i)_{y}[xs:xe], a row range;

*[ys:ye]*, a column range;*E*, the entries in the adjacent sub-region;_{xs:xe},_{ys:ye}sub-region, all the basic correlations between

*f*and^{U}(u)_{xs:xe}*f*;^{l}(i)_{ys:ye}*b*_{l}_{+ u}the bias term for layer $l+1,T1=[ta,b,c1]2\xd72\xd732$ where*l*= 0 is a*3D*tensor;$Tl+1=[ta,b,c,dl+1]2\xd72\xd732\xd732$,

*where*1 ≤*l*≤ 5.

According to the equation 5, this feature e_{1}^{x,y} is the compound correlation of the four items in the interaction graph E, presented as [e_{2x,2y;} e_{2x,2y+1}; e_{2x+1,2y}; e_{2x+1,2y+1}]. Therefore, e_{x,y}^{1} is a feature of combined correlation of E_{2x:2x+1,2y:2y+1}, namely second-order correlation. As a result, E^{1} consists of second-order correlation. The rest can be done in the same manner. *E*^{2} consists of 4-order correlation.

We can conclude that only all the entries of the lower feature map can be covered by the entries of the higher feature map. Thus, correlations among all dimensions can be encoded by an entry of the last hidden layer. Based on the 2D interaction map E, high-order correlations of the embedding dimensions can be learned by the ConvNCF Layers both locally and globally according to stacking multiple convolutional layers.

### 3.4 Prediction Layer

#### 3.4.1 Adversarial Personal Ranking

Bayesian Personalized Ranking (BPR) overcomes the challenge that pairwise approaches cannot explicitly model the ranking information among items with stochastic gradient descent (SGD) [3]. Normally, BPR objective function is denoted in equation 6:

where,

the σ(·) is sigmoid function;

λ

_{Θ}is the regularization parameter of the model;$D$ is the set of pairwise training instances;

$D:={(u,i,j)|i\u2208Iu+\u2227j\u2208I\Iu+}$

*l*_{u}^{+}, the set of items that user*u*has interacted before;*l*, the whole item set.

However, BPR model is weak and vulnerable to certain perturbations, when added small perturbations on its parameters. Thus, an adversarial personalized ranking (APR) has been presented to address the adversarial interference via the objective function optimization [7]. Formally, the objective function of the adversarial personalized ranking defined as equation 7:

where,

Δ

_{adv}, the adversarial perturbations aiming to maximize the BPR object function;Δ, the disturbance on model parameters;

ɛ ≥ 0 decides the strength of the disturbance;

$\Theta \u0302$, the present parameters of model;

Θ, aims to minimize the objective function.

The adversarial term $LBPR(D|\Theta +\Delta adv)$ controlled by λ is denoted as a regularization for stabilizing the function in the BPR. -ɛ and λ- are two hyper-parameters in BPR. A training instance (u, *i, j*) is minimized by the local objective function as equation 8, and θ is updated by the SGD rule in the equation 9:

where η refers to the learning rate.

While models trained with APR are robust to adversarial perturbations, they might not be appropriate approaches for personalized ranking due to their weak effectiveness.

#### 3.4.2 Adversarial Matrix Factorization

Give a pair (u, i), the predictive function of AMF is defined in equation 10:

where,

ω, a trainable weight vector in the prediction layer;

Δ

_{u}ɛ ℝ^{K}, the perturbation vector for user*u*;Δ

_{i}ɛ ℝ^{K}, the perturbation vector for item*i*.

We utilize the mini-batch training to get updating rules for parameters in AMF. Firstly, given the mini batch (of size S) extracts training instances *S* as $D$’. Based on the mini batch $D$’, the parameters are trained. The APR objective function for AMF is defined in equation 11:

where *l _{APR}*((

*u, i, j*)|Θ

*)*has been defined in the equation 8. Likewise, the updating rule for Θ is defined in equation 12:

Iterate over the above two steps until the AMF converges or performance begins to degrade. Formally, the objective function for ANCF can be defined in equation 13:

where λ_{*} are the hyper-parameters of the regularization, Θ* _{u}* is the parameters in f

^{U}(·), Θ

_{l}is the parameters in f

^{T}(·),

^{Θ}

_{ConvNCF}is the parameters in ConvNCF and ω for the prediction layer.

## 4 EXPERIMENTAL RESULTS AND ANALYSIS

### 4.1 Datasets and Evaluation Protocols

This paper conducts experiments on three datasets including Yelp, Pinterest and Ml-1M.

Yelp: a data set of user ratings provided by the Yelp Challenge including 25,677 items, 25,815 users, and 730,791 ratings.

Pinterest: an implicit feedback dataset constructed by He et al. [8] for content-based image recommendation, including 55,187 users, 9,916 items, and 1,500,809 ratings.

Ml-1M: A data set on movie ratings including 3,706 movies, 6,040 MovieLens users, and 1,000,209 anonymous ratings.

In the dataset, the latest user interaction is set up as the test set, the training set is set up as the remaining user interactions. After the model is trained, the next phrase is to obtain a personalized ranking list for the user via sorting the items in the training set that have no interaction with the user.

To study the performance of Top-*k* recommendation, this paper truncates the sorted list at position *k* ɛ {5, 10, 20}. Evaluation ranking lists in the paper consists of Hit Rate (HR@*k*), Normalized Discounted Cumulative Gain (NDCG@*k*) and Mean Reicprocal Rank (MRR@*k*). HR@*k* is a metric based on recalls measuring whether or not the test item is in the Top-*k* list. NDCG@*k* presents the ranking order, the higher the ranking item, the higher the calculated NDCG value. MRR@*k* is a statistic measure by producing a list of possible items to a sample of queries. For these three indicators, the larger the value, the better the personalized ranking list generated, and the better the recommendation effect. To eliminate the influence of stochastic oscillations, this paper reports the average score of last 10 epochs on convergence.

### 4.2 Parameter Settings

Parameters in ConvNCF: (1) the learning rate for embedding parameters is 0.01; (2) the learning rate for CNN parameters is 0.05; (3) λ_{1}, λ_{2}, λ_{3}, λ_{4} (in equation 13) hyper parameters for regularization are [0.01, 0.01, 10, 1].

Parameters in AMF: (1) the learning rate for AMF is 0.05; (2) λ (in equation 7) hyper parameter for regularization is 1; (3) e (in equation 7) that controls adversarial perturbation is 0.5.

### 4.3 Baselines and Effectiveness Evaluation

All experiments are conducted under tensorflow-1.12 and python-2.7. To justify the proposed approach effectiveness, this paper compares the proposed approach with other approaches as follows:

MF-BPR [3]: This approach optimizes MF with BPR, which is a competitive CF-based approach

AMF [7]: Adversarial training is added to MF-BPR, which is also a part of the proposed approach.

FISM [29]: Compared with MF which only embeds the user ID, this model integrates the history of interaction with the user to represent the user embedding.

SVD++ [30]: CF model based on the MF and FISM for the user embedding.

MLP [8]: an NCF model that concatenates the user embedding and the item embedding without encoding the embedding dimensional correlations.

JRL [14]: It is an NCF model that improves the performance of GMF [8] by adding hidden layers.

NeuMF [8]: It is an advanced recommendation model that integrates GMF and MLP to learn user-item interaction information.

ConvNCF-MF, ConvNCF-FISM, and ConvNCF-SVD++ [15]: the dimensional correlation is obtained through the outer product, based on MF, FISM, and SVD++ respectively.

Dataset . | Model . | HR@k
. | NDCG@k
. | ||||
---|---|---|---|---|---|---|---|

k=5
. | k=10
. | k=20
. | k=5
. | k=10
. | k=20
. | ||

Yelp | MLP | 0.1766 | 0.2831 | 0.4203 | 0.1103 | 0.1446 | 0.1792 |

JRL | 0.1858 | 0.2922 | 0.4343 | 0.1177 | 0.1519 | 0.1877 | |

NeuMF | 0.1881 | 0.2958 | 0.4385 | 0.1189 | 0.1536 | 0.1895 | |

ConvNCF-FISM | 0.1925 | 0.3028 | 0.4423 | 0.1243 | 0.1598 | 0.1949 | |

ConvNCF-SVD++ | 0.1991 | 0.3092 | 0.4457 | 0.1275 | 0.1629 | 0.1973 |

Dataset . | Model . | HR@k
. | NDCG@k
. | ||||
---|---|---|---|---|---|---|---|

k=5
. | k=10
. | k=20
. | k=5
. | k=10
. | k=20
. | ||

Yelp | MLP | 0.1766 | 0.2831 | 0.4203 | 0.1103 | 0.1446 | 0.1792 |

JRL | 0.1858 | 0.2922 | 0.4343 | 0.1177 | 0.1519 | 0.1877 | |

NeuMF | 0.1881 | 0.2958 | 0.4385 | 0.1189 | 0.1536 | 0.1895 | |

ConvNCF-FISM | 0.1925 | 0.3028 | 0.4423 | 0.1243 | 0.1598 | 0.1949 | |

ConvNCF-SVD++ | 0.1991 | 0.3092 | 0.4457 | 0.1275 | 0.1629 | 0.1973 |

Dataset . | Model . | HR@k
. | NDCG@k
. | MRR@k
. | ||||||
---|---|---|---|---|---|---|---|---|---|---|

k=5
. | k=10
. | k=20
. | k=5
. | k=10
. | k=20
. | k=5
. | k=10
. | k=20
. | ||

Yelp | ConvNCF-MF | 0.1978 | 0.3086 | 0.4430 | 0.1243 | 0.1600 | 0.1939 | 0.2264 | 0.2258 | 0.2261* |

ANCF | 0.5308* | 0.6649* | 0.7859* | 0.3821* | 0.4246* | 0.4535* | 0.2321* | 0.2275* | 0.2250 | |

Pinterest-20 | ConvNCF-MF | 0.5953 | 0.7594 | 0.8800 | 0.4211 | 0.4738 | 0.5032 | 0.2634* | 0.2631 | 0.2621 |

ANCF | 0.5978* | 0.7604* | 0.8859* | 0.4241* | 0.4746* | 0.5071* | 0.2624 | 0.2639* | 0.2631* | |

Ml-1M | ConvNCF-MF | 0.4688 | 0.6500 | 0.8085 | 0.3272 | 0.3827 | 0.4233 | 0.2229 | 0.2307* | 0.2297* |

ANCF | 0.4885* | 0.6596* | 0.8165* | 0.3371* | 0.3878* | 0.4329* | 0.2332* | 0.2279 | 0.2287 |

Dataset . | Model . | HR@k
. | NDCG@k
. | MRR@k
. | ||||||
---|---|---|---|---|---|---|---|---|---|---|

k=5
. | k=10
. | k=20
. | k=5
. | k=10
. | k=20
. | k=5
. | k=10
. | k=20
. | ||

Yelp | ConvNCF-MF | 0.1978 | 0.3086 | 0.4430 | 0.1243 | 0.1600 | 0.1939 | 0.2264 | 0.2258 | 0.2261* |

ANCF | 0.5308* | 0.6649* | 0.7859* | 0.3821* | 0.4246* | 0.4535* | 0.2321* | 0.2275* | 0.2250 | |

Pinterest-20 | ConvNCF-MF | 0.5953 | 0.7594 | 0.8800 | 0.4211 | 0.4738 | 0.5032 | 0.2634* | 0.2631 | 0.2621 |

ANCF | 0.5978* | 0.7604* | 0.8859* | 0.4241* | 0.4746* | 0.5071* | 0.2624 | 0.2639* | 0.2631* | |

Ml-1M | ConvNCF-MF | 0.4688 | 0.6500 | 0.8085 | 0.3272 | 0.3827 | 0.4233 | 0.2229 | 0.2307* | 0.2297* |

ANCF | 0.4885* | 0.6596* | 0.8165* | 0.3371* | 0.3878* | 0.4329* | 0.2332* | 0.2279 | 0.2287 |

### 4.4 The Effectiveness of Adversarial Learning

To ensure the good performance during the adversarial training, this paper pre-trained MF-BPR for 500 epochs (close to complete convergence), and then trained MF-APR (AMF); for comparison, this paper continues to complete the training of MF-BPR, so that the training epoch of the two is the same.

Under the condition of Top-k@10, all the diagrams in Figure 3 reflect that training MF with APR has achieved good results after 500 training epochs, while using BPR the outcome is not pleasing. It even declined slightly (in Pinterest and Ml-1M).

### 4.5 The Effectiveness of CNN

It can be seen from Figure 4 that under the condition of Top-*k*, *k* ɛ {1, 2, …, 100}, both HR@*k* [31] and NDCG@*k* [32] have been improved, but they are still at a low level, especially the metric NDCG@*k*; This is because AMF cannot learn enough information.

To address this problem, this paper utilized ANCF for training. The outer product layer in ANCF can explicitly encode the dimensional relationship between embeddings, and CNN can also handle feature maps well.

Under the condition of Top-*k*@10, AMF is utilized to be pre-trained 1500 epochs, and then ConvNCF is utilized to be trained 1500 epochs to learn high-dimensional information. As shown in Figure 5, the ANCF proposed in this paper has achieved remarkable results on all datasets. In the Yelp, using ANCF, HR@10 and NDCG@10 almost increased to 0.6524 and 0.4187 respectively; in the Pinterest, HR@10 and NDCG@10 are as high as 0.7600 and 0.4764, respectively; in the Ml-1M, HR@10 and NDCG@10 reach to around 0.6596 and 0.3591, respectively.

## 5 CONCLUSIONS

We present a novel ANCF model, which can obtain both the potential dimensional information among embeddings via the outer product and the preference information via multiple convolutional layers. Particularly, through the proposed adversarial training, ANCF can improve the overall robustness performance.

Experimental results demonstrated the flexibility and necessity of proposed schemes as follows:

it is significant to utilize both adversarial training and calculation of potential dimensional information in the CF model.

the ANCF performance is much better than the existing advanced models in the context of Top-

*k*item recommendation.

Our future work will focus on the attention mechanisms via the graph neural networks for the RS. In addition, we will look at the negative sampling mechanism in BPR and APR; The existing content-based recommendation model may also be utilized in the design of embedding vectors.

## ACKNOWLEDGMENTS

This work is supported by National Natural Science Foundation of China (61902116).

## AUTHOR CONTRIBUTION STATEMENT

Yi Gao (E-mail: [email protected], ORCID: 0000-0003-2645-2227): has participated sufficiently in the work to take public responsibility for the content, including participation in the coding, the experiment and analysis, writing the manuscript.

Jianxia Chen (E-mail: [email protected], ORCID: 0000-0001-6662-1895): has participated sufficiently in the work to take public responsibility for the content, including participation in the model design, problem analysis, writing and revision of the manuscript.

Liang Xiao (E-mail: [email protected], ORCID: 0000-0002-1564-2466): has participated sufficiently in the work to take public responsibility for the content, including participation in the model design and revision of the manuscript.

Hongyang Wang (E-mail: [email protected], ORCID: 0000-0002-8202-6655): has participated sufficiently in the work to take public responsibility for the content, including participation in the part of the experiment of recommend system.

Liwei Pan (E-mail: [email protected], ORCID: 0000-0003-2645-2227): has participated sufficiently in the work to take public responsibility for the content, including participation in the part of the experiment of deep learning.

Xuan Wen (E-mail: [email protected], ORCID: 0000-0001-9278-2377): has participated sufficiently in the work to take public responsibility for the content, including participation in the revision of the experiment in the manuscript.

Zhiwei Ye (E-mail: [email protected], ORCID: 0000-0001-6668-4634): has participated sufficiently in the work to take public responsibility for the content, including participation in the revision of the manuscript.

Xinyun Wu (E-mail: [email protected], ORCID: 0000-0002-7525-0114): has participated sufficiently in the work to take public responsibility for the content, including participation in the revision of the manuscript.