 Research
 Open Access
 Published:
Multicriteria collaborative filtering recommender by fusing deep neural network and matrix factorization
Journal of Big Data volume 7, Article number: 34 (2020)
 The Correction to this article has been published in Journal of Big Data 2020 7:45
Abstract
Recommender systems have been an efficient strategy to deal with information overload by producing personalized predictions. Recommendation systems based on deep learning have accomplished magnificent results, but most of these systems are traditional recommender systems that use a single rating. In this work, we introduce a multicriteria collaborative filtering recommender by combining deep neural network and matrix factorization. Our model consists of two parts: the first part uses a fused model of deep neural network and matrix factorization to predict the criteria ratings and the second one employs a deep neural network to predict the overall rating. The experimental results on two datasets, including a realworld dataset, show that the proposed model outperformed several stateoftheart methods across different datasets and performance evaluation metrics.
Introduction
In the era of big data, while the World Wide Web keeps growing exponentially, the size and complexity of many websites (Google, YouTube, Netflix, Amazon, and others) grow along with it, making it increasingly difficult and timeconsuming for the users of these websites to find the item (movie, music, restaurant, book, and product in general) they are looking for. Recommender systems (RSs) provide personalized suggestion for items that the user might like [1]. Exploiting the information from users’ ratings can be useful to solve one of the problems recommender systems suffer from, predicting users’ preferences about an item using a single rating. This is a clear limitation, since the user who makes a choice might take into consideration more than one aspect of the item. For an example, in a movie recommender system, some users may like a movie based on its plot, direction, or conflict, while others may like the same movie but for its acting, characters, or any other attribute of that movie. So, MultiCriteria Recommender System (MCRS) that utilizes multicriteria ratings to evaluate different attributes of the item can improve the accuracy of the recommendations [2]. RSs use many techniques, Collaborative Filtering (CF) is the most commonly used, and it makes recommendations based only on interactions between users and items such as ratings [3]. Matrix Factorization (MF), is the most popular CF techniques, which maps users and items into a joint latent space, using a vector of latent features to represent a user or an item [4]. Then the interaction of a user on an item is obtained from the inner product of their latent vectors. Although MF is effective, the simple choice of the interaction function is insufficient to model the complex relation between the users and items.
Deep Learning (DL) has achieved immense results in many research fields, such as speech recognition, natural language processing, computer vision, and recently in recommender systems, where Deep Neural Networks (DNNs) have proved their capability of modeling nonlinear users and items relationships and approximating any continuous function [5].
Therefore, it is convenient to fuse the DNN with MF to formulate a more general model that makes use of both the nonlinearity of DNN and the linearity of MF to enhance recommendation accuracy.
The objective of this paper is just to propose a multicriteria collaborative filtering recommender by fusing DNN and MF, to improve collaborative filtering performance. Our model consists of two parts: in the first part, we get the users and items’ features and feed them as an input to a fused model of a DNN and MF to predict the criteria ratings, and we use a deep neural network to predict the overall rating in the second part.
By doing experiments on two datasets including a realworld dataset, it can be observed that our proposed model achieves significant improvement compared to the other stateoftheart methods.
The main contributions of this work are as follows:

1.
We present a multicriteria collaborative filtering recommender by fusing DNN and MF, to combine the nonlinearity of DNN and linearity of MF in a multicriteria RS for modeling useritem latent structures.

2.
We do comprehensive experiments on two datasets including a realworld dataset to show the efficiency of our model and the importance of using multicriteria and deep learning in collaborative filtering.
The rest of this paper is organized as the following: In Sect. 2, we survey the related work. We will give a detailed overview of the system in Sect. 3. Section 4 presents the experimental evaluations and discussion and in Sect. 5, we provide a brief conclusion and potential future work.
Related work
Multicriteria recommendation techniques can be divided into two general classes: memorybased and modelbased methods. In memorybased methods, the similarity can be computed in two ways: the first approach calculates the similarity on each criteria rating separately using the traditional way and then aggregates the values into a single similarity using an aggregation method such as average [6], worstcase [6], and weighted sum of individual similarities [7]. The second approach uses the multidimensional distance metrics (Euclidean, Manhattan, and Chebyshev distance metrics) to calculate the distance between multicriteria ratings [6]. Modelbased methods use the useritem ratings to learn a predictive model and later use this model to predict unknown ratings, like Probabilistic Modeling [8, 9], Multilinear Singular Value Decomposition (MSVD) [10], Support Vector Regression (SVR) [11], and aggregationfunctionbased [6], this approach assumes that there is a relation between the item’s overall rating and the multicriteria ratings and it is not independent of other ratings.
Many efforts have been made to improve MF, Koren [12] merged MF and neighborhood models. Wang et al. [13] combined MF with topic models of item content. Rendle [14] introduced Factorization Machines that combine Support Vector Machines with factorization models. He et al. [15] proposed Neural Matrix Factorization (NeuMF) model that changed the linearity nature of MF by combining it with MultiLayer Perceptron (MLP).
There are very few researches on applying deep learning to Collaborative Filtering model. Salakhutdinov et al. [16] proposed restricted boltzmann machines model to learn the item ratings correlation, thereafter Georgiev et al. [17] expanded the former work by adding both user–user and item–item correlations. Ouyang et al. [18] used autoencoder in autoencoderbased collaborative filtering to model users ratings on items. He et al. [15] introduced neural collaborative filtering model that uses MLP to learn the interaction function. Nassar et al. [19] presented deep multicriteria collaborative filtering (DMCCF) model which is the only attempt in applying deep learning and multicriteria to collaborative filtering. The model follows the aggregationfunctionbased approach, where they used a deep neural network to predict the criteria ratings and another DNN to learn the relationship between the criteria ratings and the overall rating in order to predict the overall rating.
Method
The proposed model is based on the model that Nassar et al. [19] presented. It contains three steps:

a.
Predict criteria ratings \( r_{1} , r_{2} , \ldots , r_{k} \) using a DNN.

b.
Learn aggregation function \( f \), which represents the relationship between the criteria ratings and the overall rating using a DNN.

c.
Predict overall ratings using the predicted criteria ratings and the aggregation function \( f \).
In our model, we used a fused model of DNN and MF to predict the criteria ratings \( r_{1} , r_{2} , \ldots , r_{k} \) in the first step, while in the second step, we kept using a DNN to learn aggregation function \( f \), as illustrated in Fig. 1.
Criteria ratings model
This model is used to predict the criteria ratings for a user on an item. In this model, we will fuse MF and DNN as He et al. [15] proposed in their NeuMF framework. MF that applies a linear kernel to model the latent feature interactions, and DNN that uses a nonlinear kernel to learn the interaction function from data.

1.
MF
We can use the item ID or user ID as a feature since it is unique, but the ID is a categorical feature.
Therefore, we converted the IDs into embedding vectors initialized with random values that adjusted to minimize the loss function during the model training.
In the input layer, the input vector \( x \) is given by:
$$ x = v_{u} \odot v_{i} $$(1)where \( v_{u} \) and \( v_{i} \) are user and item embedding vectors and \( \odot \) is the elementwise product of vectors. The formula of the output layer is as follows:
$$ y_{ui} = a_{out} \left( {w^{T} \left( {v_{u} \odot v_{i} } \right)} \right) $$(2)where \( a_{out} \) and \( w^{T} \) are the activation function and weights of the output layer.

2.
DNN
The input vector \( x \) is given by:
$$ x = Concatenate\left( {v_{u} , v_{i} } \right) = \left[ {\begin{array}{*{20}c} {v_{u} } \\ { v_{i} } \\ \end{array} } \right] $$(3)where \( v_{u} \) and \( v_{i} \) are user and item embedding vectors.
Then followed by a number of dense Rectified Linear Units (ReLU) layers, we choose ReLU activation function because it is the most efficient [20]. The formulation of which is given as follows:
$$ ReLU\left( z \right) = max\left( {z,0} \right) $$(4)The output of a hidden layer \( l \) is formulated as:
$$ h_{l} = ReLU\left( {W_{l} h_{l  1} + b_{l} } \right) $$(5)where \( W_{l} , \) and \( b_{l} \) are the weight matrix and bias vector for the layer \( l \) and \( h_{1} = x \).
In the output layer, we have:
$$ y_{ui} = ReLU\left( {W_{L} h_{L  1} + b_{L} } \right) $$(6)where \( L \) is the number of layers.

3.
Fusion of MF and DNN
For more flexibility to the fused model, we let DNN and MF learn separate embeddings, as shown in Fig. 2, and then we combine them by concatenating the last hidden layer. The outputs of the DNN and MF denoted as \( o^{D} , o^{M} \) respectively, are given as follows:
$$ o^{D} = ReLU\left( {W_{L} \left( {ReLU\left( {W_{L  1} \left( { \ldots ReLU\left( {W_{2} \left[ {\begin{array}{*{20}c} {v_{u}^{D} } \\ { v_{i}^{D} } \\ \end{array} } \right] + b_{2} } \right) \ldots } \right) + b_{L  1} } \right)} \right) + b_{L} } \right) $$(7)$$ o^{M} = v_{u}^{M} \odot v_{i}^{M} $$(8)where \( v_{u}^{D} \) and \( v_{u}^{M} \) are user embedding vectors for DNN and MF part, and \( v_{i}^{D} \) and \( v_{i}^{M} \) are item embedding vectors.
In the model output layer, we predict the user criteria ratings \( r_{1} , r_{2} , \ldots , r_{k} \) using Eq. (9) and (10):
$$ y_{ui} = ReLU\left( {W\left[ {\begin{array}{*{20}c} {o^{D} } \\ { o^{M} } \\ \end{array} } \right] + b} \right) $$(9)$$ \left[ {r_{1} , r_{2} , \ldots , r_{k} } \right]^{T} = y_{ui} $$(10)
Overall rating deep neural network
The overall rating DNN is used to learn the aggregation function \( f \), which represents the relationship between the overall rating \( r_{0} \) and the criteria ratings \( r_{1} , r_{2} , \ldots , r_{k} \), in order to predict the overall rating:
In the input layer, the input vector is the criteria ratings \( r_{1} , r_{2} , \ldots , r_{k} \) for user \( u \) and item \( i \), as shown in Fig. 2. We normalize the continuous features \( r_{1} , r_{2} , \ldots , r_{k} \) because the DNNs are sensitive to the inputs scaling and distribution [21]. The normalization of a sample \( r_{i} \) is calculated as:
with \( m_{i} \) the mean of the training samples for rating \( r_{i} \) and \( s_{i} \) the standard deviation of the training samples for rating \( r_{i} \). The input vector becomes:
Then followed by a number of hidden layers, the output of a hidden layer is given again by Eq. (5).
In the output layer, we predict the overall rating \( r_{0} \) in Eq. (6), where:
We used Adam optimizer [22] and MAE loss function in both parts.
Recommendation
After the finishing of the model training, where each part was trained individually without knowing each other. We can use the model to predict the user’s overall rating on the new items.
The recommendation process happens as shown in Fig. 1, for each user \( u \) and item \( i \) pair we:

a)
Get the user ID and the item ID, and use them as an input for the Criteria Ratings model, to compute the criteria ratings \( r^{\prime}_{1} , r^{\prime}_{2} , \ldots , r^{\prime}_{k} \).

b)
Normalize the previously computed criteria ratings \( r^{\prime}_{1} , r^{\prime}_{2} , \ldots , r^{\prime}_{k} \) and use them as an input for the Overall Rating DNN, to compute the overall rating \( r^{\prime}_{0} \).

c)
Recommend the user the items as in traditional recommender systems using the overall rating \( r^{\prime}_{0} \).
Results and discussion
Dataset
We evaluated our model on two multicriteria rating datasets, a realworld TripAdvisor^{Footnote 1} dataset and a Movies^{Footnote 2} dataset.

1.
TripAdvisor dataset
Multicriteria rating dataset for hotels. It includes an overall rating and seven criteria ratings Value, Rooms, Location, Cleanliness, Check in/front desk, Service, and Business service, the rating range between 1 and 5. Table 1 demonstrates the statistics of the dataset and Table 2 the distribution of values for the different criteria ratings and the overall rating.

2.
Movies dataset
Multicriteria rating dataset for movies, available on GitHub. It contains four criteria ratings and an overall rating, the rating range between 1 and 13. Tables 3 and 4 demonstrate the dataset statistics and the distribution of the different criteria ratings and the overall rating.
Evaluation
To evaluate the performance of our model we used the same metrics used in [19].

a.
Mean Absolute Error (MAE) [23]
$$ MAE = \frac{1}{M}\mathop \sum \limits_{u,i} \left {r_{ui}  \tilde{r}_{ui} } \right $$(15)where \( r_{ui} \) the true rating of user \( u \) for item \( i \), \( \tilde{r}_{ui} \) the predicted rating, and \( M \) is the size of the test set.

b.
Fmeasure (\( {\text{F}}_{1} \) and \( {\text{F}}_{2} \)) [23]
$$ F_{1} = \frac{2PR}{P + R} $$(16)$$ F_{2} = \frac{5PR}{4P + R} $$(17)where \( P \) the precision and \( R \) the recall.

c.
Fraction of Concordant Pairs (FCP) [24]
$$ n_{c}^{u} = \left {\left\{ {\left( {i,j} \right) \left { \tilde{r}_{ui} } \right\rangle \tilde{r}_{uj} and r_{ui} > r_{uj} } \right\}} \right $$(18)where \( n_{c}^{u} \) the number of concordant pairs for user \( u \), and we calculate \( n_{d}^{u} \) the number of discordant pairs for user \( u \) in a similar way.
$$ n_{c} = \mathop \sum \limits_{u} n_{c}^{u} $$(19)$$ n_{d} = \mathop \sum \limits_{u} n_{d}^{u} $$(20)$$ FCP = \frac{{n_{c} }}{{n_{c} + n_{d} }} $$(21) 
d.
Mean Average Precision (MAP) [23] is the average precision for all users.

e.
Mean Reciprocal Rank (MRR) [25] is the average of the first relevant item rank for all users.
Settings
We used Keras^{Footnote 3} with TensorFlow^{Footnote 4} as a backend to implement our model.

1.
Criteria Ratings Model Settings
We conducted several experiments to find the optimal parameters for DNN and MF.

For DNN, we randomly initialized the DNN parameters using a normal distribution with mean of 0 and standard deviation of 0.05. We used Adam optimizer, with 0.001 learning rate and parameters values as provided in [22].
For TripAdvisor dataset, we used batch size of 512 and set epochs to 2. We set each of user and item embedding vector size to 64, and we selected the [128 → 64 → 32 → 16 → 8] hidden layers. For Movies dataset, we used batch size of 64 and set epochs to 4. We set each of user and item embedding vector size to 32, and we selected the [64 → 32 → 16 → 8] hidden layers.

For MF, we tried to find the optimal embedding vector size. As shown in Fig. 3, for TripAdvisor dataset, we set user and item embedding vector size to 16, and for Movies dataset, we set them to 8.

In the model output layer, for TripAdvisor dataset, there are 7 neurons, equal to the number of the criteria ratings, and 4 neurons for Movies dataset.

Overall Rating DNN Settings
We initialized randomly the DNN parameters like the previous DNN, using a normal distribution with mean of 0 and standard deviation of 0.05. We also used Adam optimizer with 0.001 learning rate and the same parameters values. For TripAdvisor dataset, we set the epochs to 50, and for Movies dataset, we set the epochs to 100. While for both datasets, we set the batch size to 512. We used [64 → 32 → 16 → 8] hidden layers. Finally, in the output layer, there is 1 neuron, for the overall rating.
Results
We used fivefold crossvalidation method, to split the data randomly into 20% test set and 80% training set including a 1% validation set. We repeated this process 5 times and calculated the mean value of each metric. We compared the performance of our model to the DMCCF model [19], then to a single DNN that predicts the overall rating directly where the finest results of this DNN were acquired from settings shown in Table 5. In addition, we compared our model to a number of famous single rating recommendation methods such as SVD [26], SVD++ [12], Baseline Estimates [27], and SlopeOne [28].
This comparison was done on the overall rating. The results are illustrated in Table 6. We can see that our model achieves the best performance on both datasets, significantly outperforming each of DMCCF model, single DNN, and the other stateoftheart methods on all the evaluation metrics. This indicates the high expressiveness of our model by fusing the nonlinear DNN and linear MF models to capture the methods. F1 and F2 of our model are better than the other methods. Our model surpasses the other models in FCP and it exceeds useritem interactions. According to the results, in MAE, our model excelled the other compared methods at MAP. In MRR, our model is the best.
Conclusion and future work
In this paper, we proposed multicriteria collaborative filtering recommender by fusing DNN and MF. The model consists of two parts: in the first part, we get the users and items’ features and feed them as an input to a fused model of a DNN and MF to predict the criteria ratings, and in the second part, we use a DNN to predict the overall rating. Then we demonstrated the efficiency of our proposed model, where the experimental results showed that our model significantly outperformed the other methods on both datasets for all the evaluation metrics, which proved that the application of deep learning and multicriteria in collaborative filtering is a successful attempt and it can be enhanced using different deep learning techniques or by building more complex models.
In future work, we will study the use of different deep learning techniques, such as Autoencoder, Convolutional Neural Network (CNN), and Recurrent Neural Network (RNN) in recommendation systems and attempt to further improve the performance of our model, we will also try other feature representation methods precisely to solve the cold start problem by using user and item content features.
Availability of data and materials
The datasets used during the current study are available in the Datasets repository, https://github.com/nournassar/Datasets.
Change history
07 July 2020
An amendment to this paper has been published and can be accessed via the original article.
Notes
 1.
The dataset is crawled from the TripAdvisor website: https://www.tripadvisor.com/.
 2.
 3.
 4.
https://github.com/tensorflow/tensorflow.
Abbreviations
 CF:

Collaborative filtering
 DL:

Deep learning
 DMCCF:

Deep multicriteria collaborative filtering
 DNN:

Deep neural network
 FCP:

Fraction of Concordant Pairs
 MAE:

Mean Absolute Error
 MAP:

Mean Average Precision
 MCRS:

MultiCriteria Recommender System
 MF:

Matrix factorization
 MLP:

Multilayer perceptron
 MRR:

Mean Reciprocal Rank
 NeuMF:

Neural Matrix Factorization
 ReLU:

Rectified Linear Unit
 RS:

Recommender system
References
 1.
Zhao J, Geng X, Zhou J, Sun Q, Xiao Y, Zhang Z, Fu Z. Attribute mapping and autoencoder neural network based matrix factorization initialization for recommendation systems. Knowl Based Syst. 2019;166:132–9.
 2.
Turk AM, Bilge A. Robustness analysis of multicriteria collaborative filtering algorithms against shilling attacks. Expert Syst Appl. 2019;115:386–402.
 3.
Zhang S, Yao L, Sun A, Tay Y. Deep learning based recommender system: a survey and new perspectives. ACM Comput Surv. 2019;52:5.
 4.
Fu M, Qu H, Yi Z, Lu L, Liu Y. A novel deep learningbased collaborative filtering model for recommendation system. IEEE Trans. Cybern. 2018. https://doi.org/10.1109/tcyb.2018.2795041.
 5.
Hornik K, Stinchcombe M, White H. Multilayer feedforward networks are universal approximators. Neural Networks. 1989;2:359–66.
 6.
Adomavicius G, Kwon Y. New recommendation techniques for multicriteria rating systems. IEEE Intell Syst. 2007;22:48–55. https://doi.org/10.1109/MIS.2007.58.
 7.
Tang TY, Mccalla G. The pedagogical value of papers: a collaborativefiltering based paper recommender. J. Digit. Inf. 2009. https://doi.org/10.1109/wkdd.2009.132.
 8.
L. Si, R. Jin, Flexible Mixture Model for Collaborative Filtering. In: Proc. Twent. Int. Conf. Int. Conf. Mach. Learn., AAAI Press, 2003: pp. 704–711. http://dl.acm.org/citation.cfm?id=3041838.3041927.
 9.
Sahoo N, Krishnan R, Duncan G, Callan J. The halo effect in multicomponent ratings and its implications for recommender systems: the case of yahoo! movies. Inf. Syst. Res. 2012;23:231–46. https://doi.org/10.1287/isre.1100.0336.
 10.
Q. Li, C. Wang, G. Geng, Improving personalized services in mobile commerce by a novel multicriteria rating approach. In: Proceeding 17th Int. Conf. World Wide Web  WWW’08. (2008) 1235. https://doi.org/10.1145/1367497.1367743.
 11.
Drucker H, Burges CJC, Kaufman L, Smola A, Vapnik VN. Support vector regression machines. Adv Neural Inf Process Syst. 1997;9(1):155–61.
 12.
Koren Y. Factorization meets the neighborhood: a multifaceted collaborative filtering model. In: Proc. 14th ACM SIGKDD Int. Conf. Knowl. Discov. Data Min., ACM, New York, NY, USA, 2008: pp. 426–34. https://doi.org/10.1145/1401890.1401944.
 13.
Wang H, Wang N, Yeung DY. Collaborative deep learning for recommender systems. In: Proc. 21th ACM SIGKDD Int. Conf. Knowl. Discov. Data Min., 2015. pp. 1235–44.
 14.
Rendle S. Factorization machines. In: Data Min. (ICDM), 2010 IEEE 10th Int. Conf., 2010: pp. 995–1000.
 15.
He X, Liao L, Zhang H, Nie L, Hu X, Chua TS. Neural Collab Filter. 2017. https://doi.org/10.1145/3038912.3052569.
 16.
Salakhutdinov R, Mnih A, Hinton G. Restricted boltzmann machines for collaborative filtering. In: Proc. 24th Int. Conf. Mach. Learn., ACM, New York, NY, USA, 2007: pp. 791–98. https://doi.org/10.1145/1273496.1273596.
 17.
Georgiev K, Nakov P. A nonIID framework for collaborative filtering with restricted boltzmann machines. In: S. Dasgupta, D. McAllester (Eds.), Proc. 30th Int. Conf. Mach. Learn., PMLR, Atlanta, Georgia, USA, 2013. pp. 1148–56. http://proceedings.mlr.press/v28/georgiev13.html.
 18.
Ouyang Y, Liu W, Rong W, Xiong Z. Autoencoderbased collaborative filtering. In: Loo CK, Yap KS, Wong KW, Beng AT, Huang K, editors. Neural Inf. Cham: Process. Springer International Publishing; 2014. p. 284–91.
 19.
Nassar N, Jafar A, Rahhal Y. A novel deep multicriteria collaborative filtering model for recommendation system. KnowledgeBased Syst. 2019. https://doi.org/10.1016/j.knosys.2019.06.019.
 20.
Fuentes O, Parra J, Anthony EY, Kreinovich V. Why rectified linear neurons are efficient: symmetrybased, complexitybased, and fuzzybased explanations. Dep. Tech. Reports (CS). (2017). https://digitalcommons.utep.edu/cs_techrep/1171.
 21.
Ioffe S, Szegedy C. Batch normalization: accelerating deep network training by reducing internal covariate shift. arXiv. 2015. https://doi.org/10.1007/s1339801401737.2.
 22.
D.P. Kingma, J. Ba. Adam: A Method for Stochastic Optimization. In ICLR: International Conference on Learning Representations. 2014. .
 23.
Herlocker JL, Konstan JA, Terveen LG, Riedl JT. Evaluating collaborative filtering recommender systems. ACM Trans Inf Syst. 2004;22:5–53.
 24.
Koren Y, Sill J. Collaborative filtering on ordinal user feedback. In: IJCAI Int. Jt. Conf. Artif. Intell., 2013: pp. 3022–26. https://doi.org/10.1145/1985793.1986027.
 25.
Abel F, Araújo S, Gao Q, Houben GJ. Analyzing crosssystem user modeling on the social web. In: Int. Conf. Web Eng., 2011: pp. 28–43.
 26.
Koren Y, Bell R, Volinsky C. Matrix factorization techniques for recommender systems. 2009, p. 42–49. https://doi.org/10.1109/mc.2009.263.
 27.
Koren Y. Factor in the neighbors. ACM Trans Knowl Discov Data. 2010;4:1–24. https://doi.org/10.1145/1644873.1644874.
 28.
Lemire D, Maclachlan A. Slope one predictors for online ratingbased collaborative filtering. 2007. doi: 10.1137/1.9781611972757.43
Acknowledgements
The authors thank N. Mohanna for careful reading of the manuscript. Special thanks are due to S. Tarbouche for advice and support.
Funding
Not applicable.
Author information
Affiliations
Contributions
All the corresponding authors contributed equally to the conduct of the present study. All authors read and approved the final manuscript.
Corresponding author
Ethics declarations
Competing interests
The authors declare that they have no competing interests.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
The original version of this article was revised: the swapped figures 1 and 2 are corrected.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Nassar, N., Jafar, A. & Rahhal, Y. Multicriteria collaborative filtering recommender by fusing deep neural network and matrix factorization. J Big Data 7, 34 (2020). https://doi.org/10.1186/s40537020003096
Received:
Accepted:
Published:
Keywords
 Collaborative filtering
 Deep learning
 Deep neural network
 Matrix factorization
 Multicriteria
 Recommender system