Design and Comparative Analysis of New Personalized Recommender Algorithms With Specific Features For Large Scale Datasets
Design and Comparative Analysis of New Personalized Recommender Algorithms With Specific Features For Large Scale Datasets
Article
Design and Comparative Analysis of New
Personalized Recommender Algorithms with Specific
Features for Large Scale Datasets
S. Bhaskaran, Raja Marappan * and B. Santhi
School of Computing, SASTRA Deemed University, Thanjavur 613401, India; [email protected] (S.B.);
[email protected] (B.S.)
* Correspondence: [email protected]
Received: 27 May 2020; Accepted: 30 June 2020; Published: 6 July 2020
Abstract: Nowadays, because of the tremendous amount of information that humans and machines
produce every day, it has become increasingly hard to choose the more relevant content across a
broad range of choices. This research focuses on the design of two different intelligent optimization
methods using Artificial Intelligence and Machine Learning for real-life applications that are used
to improve the process of generation of recommenders. In the first method, the modified cluster
based intelligent collaborative filtering is applied with the sequential clustering that operates on the
values of dataset, user0 s neighborhood set, and the size of the recommendation list. This strategy
splits the given data set into different subsets or clusters and the recommendation list is extracted
from each group for constructing the better recommendation list. In the second method, the specific
features-based customized recommender that works in the training and recommendation steps by
applying the split and conquer strategy on the problem datasets, which are clustered into a minimum
number of clusters and the better recommendation list, is created among all the clusters. This strategy
automatically tunes the tuning parameter λ that serves the role of supervised learning in generating
the better recommendation list for the large datasets. The quality of the proposed recommenders for
some of the large scale datasets is improved compared to some of the well-known existing methods.
The proposed methods work well when λ = 0.5 with the size of the recommendation list, |L| = 30 and
the size of the neighborhood, |S| < 30. For a large value of |S|, the significant difference of the root mean
square error becomes smaller in the proposed methods. For large scale datasets, simulation of the
proposed methods when varying the user sizes and when the user size exceeds 500, the experimental
results show that better values of the metrics are obtained and the proposed method 2 performs
better than proposed method 1. The significant differences are obtained in these methods because the
structure of computation of the methods depends on the number of user attributes, λ, the number
of bipartite graph edges, and |L|. The better values of the (Precision, Recall) metrics obtained with
size as 3000 for the large scale Book-Crossing dataset in the proposed methods are (0.0004, 0.0042) and
(0.0004, 0.0046) respectively. The average computational time of the proposed methods takes <10
seconds for the large scale datasets and yields better performance compared to the well-known
existing methods.
1. Introduction
Recommender systems are broadly utilized to assist users in handling the great amount of
information available on the web, particularly in searching for the most appropriate content tailored to
the specific preferences of the user0 s. Because of the large esteem of this type of organization and required
to guarantee that such interfaces provide the user with huge relevance and quality, the mechanisms
for making these recommendations should be updated continuously [1]. The recommendations can
also be personalized and non-personalized in approach as per the user’s characteristics. A list of
best-recommended products on the site is provided for a personalized recommendation. The product
may be recommended based on an analysis of users0 past behavior or the statistical advice provided by
the other user, whereas non-personalized are simple to make, as they are independent of user’s actions.
Conventional recommender systems are necessarily a content-based and Collaborative Filtering
(CF) system [2]. The traditional method utilized for recommendations is CF. Recommender systems
based on CF calculate user preferences for products or services by learning past user–item relationships
from a group of users sharing the same interests and tastes. An additional standard scheme when
designing referral systems is content-based filtering. Content-based filtering schemes are according to
the description of the item and profile of the user0 s preferences. These schemes are well-matched for
situations where the known data of an object (name, location, description, etc.) are present, but not
in the user list. Despite the achievement of these two filtration techniques, several drawbacks have
been recognized. Most of the issues are connected with content-based filtering methods that include
limited content analysis, over-specialization, and spacing of data [3]. Moreover, joint approaches reveal
cold-start, spacing, and scaling problems. These issues generally decrease the quality of referrals.
To alleviate some of the issues identified, hybrid filtration has been proposed combining two or
more filtration methods in various ways to enlarge the accuracy and efficiency of the recommender
systems [4–18].
The important research in recommender applications is the development of a good
recommendation system that is expected to create a better recommendation list, based on the specific
needs of the users [13,19–30]. To resolve the problems in the existing methods, this research focuses
on the design of two different intelligent optimization methods using Artificial Intelligence (AI) and
Machine Learning (ML) for real-life applications that are used to improve the process of generation
of recommenders. In the first method, the modified cluster based intelligent CF is applied with the
sequential clustering that operates on the values of dataset, user0 s neighborhood set, and |L|. In the
second method, the specific features-based customized recommender that works in the training and
recommendation steps by applying the split and conquer strategy on the problem datasets, which are
clustered into a minimum number of clusters and the better recommendation list, is created among all
the clusters.
Some of the research gaps in the existing recommender methods are: having more deviations in
the performance measurements and taking a lot of computational complexity, which results in less
accuracy in the generation of the recommendation list [13,23,24,31–39]. Hence, it is necessary to design
the new recommender strategies to offset the issues in the well-known methods of solving real-life
applications [6,33,38,40–47]. Section 2 focuses on the survey on recent recommender algorithms in
the generation of a better recommendation list. The requirement of the design of new recommender
methods for real-world applications is also discussed in Section 2. The notations and the definitions
applied in the proposed methods are explained in Section 3. The proposed recommender algorithms
are presented in Section 4. The simulation of the proposed strategies, along with the analysis of the
experimental results, is focused on in Section 5. The conclusions of this research with the future
research areas are discussed in Section 6.
2. Literature Survey on Recent Recommenders & the Need for New Recommender Strategies
with the familiarity of products that are functioning in two different phases such as training and
recommendation. The training phase of UPOD performs operations such as data prediction, defining
the values of users0 features, feature-based clustering of users, creating a bilateral map of interactions,
and searching the profile of every cluster. The training phase trains a classifier, which provides features
of the target user. The referral phase of UPOD requires an input map, a target user, a set of user
attributes, a classifier trained in the training operation, and the size of the referral list. In this strategy,
the tuning parameter is automatically adjusted to evaluate the amount of mixing in mass diffusion
activity for different sparse datasets. This strategy generates recommendations based on certain
features of the user’s profile. Most importantly, this strategy includes the user’s profile information
for refining and personally recommending the content to the users. This strategy is simulated with
the parameters: λ = 0.5, the size of the recommendation list |L| = 30, size of the neighborhood is 30,
k-minimum to 100, k-maximum to 200.
without strongly affecting their time efficiency [67]. The parallel hardware implementation based
algorithm is developed for embedded CF applications with large datasets [68]. The online
recommendation algorithm is designed, which combines clustering and CF techniques to improve the
accuracy of online recommendation systems for group-buying applications [69]. The recommender
system development is discussed that uses several algorithms to obtain groupings [70].
Nowadays, the new recommender algorithms are required for real-world applications, because of
the following reasons [1–18,21–24,29–32,43–47,66–70]:
One of the main reasons why we need a recommender system in modern society is that there are
many ways for people to use the Internet. For example, Netflix has an enormous collection of
movies. Despite the increase in the amount of information available, a new problem arose due to
the difficulty of selecting the items that people wanted to see;
A recommender system attempts to assess and predict user content preferences related to games,
stories, or videos. The system draws from data usage history aimed at making recommendations
based on the user0 s (current) interests;
In the e-commerce system, recommender systems improve revenue because they are the best way
to sell more products;
A company with a list of thousands and thousands of products will be hard-pressed for all its
products with hardcore product recommendations, and such standard recommendations will
soon be outdated or inappropriate for many customers by using a variety of methods for filtering;
you can find business hours to recommend new products that you can buy (whether on their site,
via email, or otherwise);
The recommender system should provide more precise and personalized recommendations than
the existing systems. The outcome of the recommender should prove the correctness of the
recommendation based on the specific needs of the users.
Cov(X, Y)
r(x, y) = (1)
σx σ y
Pn
i=1 (xi − x)( yi − y)
r(x, y) = q q (2)
Pn 2 Pn 2
i = 1 ( xi − x ) i=1 ( yi − y)
In general, 0 ≤ r(x, y) ≤ 1. The values of r(x, y) = −1 and + 1 signify the perfect negative and
perfect positive correlation, respectively. Cov(X, Y), the covariance between variables X and Y becomes
zero when X and Y are independent. Cov(X, Y) is defined as follows:
The expectation of a random variable X is given by E(X) = x. Then, the equation r(x, y) = 0
represents the random variables X and Y, which are independent.
Mathematics 2020, 8, 1106 6 of 27
v
n
u
t
1 X 2
σy = yj − y (6)
n−1
j=1
where Ixy defines the set of items rated by variables of two users x and y [32]. The rating values of jth
item are defined as rx,j and r y,j for users x and y, respectively. The average ratings of all items that
interacted with users x and y are defined as rx and r y , respectively. PearsonCorrCoe f f (x, y) ∈ R and
−1 ≤ PearsonCorrCoe f f (x, y) ≤ 1.
where S is the set of users similar to user x concerning the values of the neighborhood [31].
PearsonCorrCoe f f (x, a) defines the similarity between two variables x and a. The rating value of
jth item is defined as ra, j for the user a. The mean rating of all items interacted with users a is defined
as ra . Typical values of the cardinality of the set S lie between 20 and 40 to maximize the item prediction
for a better recommendation.
where δ( yi , xi ) = 0 when xi = yi and δ( yi , xi ) = 1 when xi , yi . For example, consider the two vectors
X and Y with sample attribute values such as a1 = profession, a2 = age, and a3 = sex and assume
the sample values of vector X and Y are x1 = engineer, x2 = 20–25 and x3 = female and y1 = doctor,
y2 = 25–35 and y3 = female respectively. The dissimilarity measurement between X and Y is obtained
as follows: Dissimilarity (Y, X) = 1 + 1 + 0 = 2, resulting in only one attribute; a3 = sex and is common
to both vectors X and Y. This measurement finds the number of identical and non-identical attributes
in both vectors X and Y. This function does not define a distance metric. This measurement is applied
in the k-modes clustering algorithm in assigning the classifiers for each object in each cluster. The mode
of clusters is evaluated using Dissimilarity (Y, X) & Dissimilarity (Q, X) measurements. Then the
effective data partition is evaluated for each cluster.
k
1 X
E Ck = nk H (Ck ) (18)
n
k =1
1
Optimize Ck = H (X) − E Ck (19)
c
g(L)
Precision (List, User) = (21)
|L|
Applying the split and conquer strategy on a large scale datasets into different clusters and
generating better recommenders from each cluster.
Updating the similar knowledge needs of other users in all clusters in a database and storing the
better recommendation lists in a database for all the clusters.
Mathematics 2020, 8, 1106 10 of 27
Applying the Machine Learning in the identification of knowledge requirements for every
partition by extracting the better previously-stored knowledge information from the database to
reduce the computational complexity.
The proposed strategy works well for large scale datasets compared to the feature-based
recommender developed in solving ERP System and E-Agribusiness datasets, which require the
computation of some configuration functions initially and its evaluation process requires some offline
evaluation of parametric optimization [48]. Compared to this strategy, the proposed strategy obtains
the better recommendation list in each of the clusters and it is updated in the database for future
recommendation purposes. The favorite items are combined in the recommendation list based on the
user profile of a target user.
The proposed methods are compared with the CF developed for both user-based and object-based
cases in bipartite networks where the filtering is based on the degree of nodes in the bipartite
network [49]. However, in the proposed CF, the split and conquer strategy predicts the overall ratings
for all unrated items and recommends the best list for each cluster, and then the better recommendation
from the ratings of the entire list is chosen artificially from the database. The proposed methods work
well in solving the MovieLens dataset compared to the other methods.
The group recommendation model is proposed based on factors such as sparsity, dynamics,
and timeliness [50]. However, the proposed collaborative recommender is designed to update the
similar knowledge needs of other users in all clusters of the same group, and the information is
updated in a database for providing the overall better recommendation. Even when more additional
features are considered, this intelligent optimization strategy further minimizes the RMSE for large
scale datasets to provide a better recommendation.
For the datasets that require a higher number of categories, latent class methods are computationally
slow and provide an infeasible solution using k-modes clustering developed in [51]. Hence for large
datasets that involve more categorical variables to improve the performance of cluster analysis,
the k-modes clustering with specific features-based personalized recommender algorithm is required.
Then k-modes method is applied to construct the different clusters by preprocessing the updated
training dataset. The proposed k-modes strategy is a frequency-based method which evaluates the
mode of clusters using Dissimilarity (Y, X) & Dissimilarity (Q, X) measurements. In the proposed
k-modes strategy, for each object from the classifier: (training-dataset, attributes-set), the object is assigned
to the cluster whose mode is the nearest to it according to Dissimilarity (Y, X) & Dissimilarity (Q, X)
measurements. This strategy is used to evaluate the effective data partition for each cluster in obtaining
a better recommendation. The training and recommendation steps are applied with some preprocessing
and automatically tune the parameter λ to serve the role of supervised learning.
Figure1.1. The
Figure The Flowchart
Flowchart of
of the
the Proposed
ProposedMethod
Method1.
1.
Figure 2. The Flowchart of the Split and Conquer Strategy of the Proposed Method 2.
Figure 2. The Flowchart of the Split and Conquer Strategy of the Proposed Method 2.
The proposed Profile-based Customized Recommender is modeled as a bi-partition graph G of a
The proposed
with three Profile-based
sets: (I, U, E)Customized Recommender
set is {u1 , is
u2 ,modeled
u3 . . . uNas a bi-partition
item set is {igraph G of a
user-item where the user }, the 1 , i2 , i3 . . . iN }
user-item with three sets: (I, U, E)
and the edge set is {e1 , e2 , e3 . . . ek }. where the user set is {u 1, u2, u3 … uN}, the item set is {i1, i2, i3 …iN}
and the
Thisedge set is {eoperates
algorithm 1, e2, e3 … ek}.
based on user profile information for each cluster. It applies a tuning
This algorithm operates
parameter λ, which lies between based
0 and on1.user
Theprofile information
two important stagesforineach
this cluster.
algorithmIt applies a tuning
are training and
parameter λ, which lies between 0 and 1. The two important stages in this
recommendation. The stage 1 operations are dataset preprocessing, defining user features, deleting algorithm are training and
recommendation. The stage 1 operations are dataset preprocessing, defining
information for invalid users and empty data, performing basic feature-based operations, constructing user features, deleting
information through
interactions for invalid
graphs, users and empty
preprocessed datadata, performing
features into clusteringbasic usingfeature-based
a k-modesoperations,
clustering
constructing interactions through graphs, preprocessed data features into
algorithm, mapping the clusters to λ values, defining the pair (λ, cluster) values, and transformingclustering using a k-modes
attributes into categories. The required parameters are n, the data size; k, the cluster size; and t, and
clustering algorithm, mapping the clusters to λ values, defining the pair (λ, cluster) values, the
transforming
number attributes
of iterations. Theinto categories.
proposed Thefeatures
specific required parameters
based recommenderare n, the data size;
is depicted k, the cluster
in Algorithm 2.
size; and t, the number of iterations. The proposed specific features based recommender is depicted
in Algorithm 2.
Mathematics 2020, 8, 1106 13 of 27
Algorithm3:3:Training
Algorithm Training Steps
Steps
Inputs:training-dataset,
Inputs: training-dataset, attributes-set,
attributes-set, k-minimum,
k-minimum, k-maximum
k-maximum
1: Partition the required knowledge
1: Partition the required knowledge needs and apply split and needs and apply splitstrategy
conquer and conquer strategy on
on the training-dataset andthe
training-dataset
extract the better and
stored extract
similarthe better stored
knowledge needssimilar
from theknowledge
database needs from the database
2: Apply
2: Applythe thepreprocessing
preprocessing operations
operations on theontraining-dataset
the training-dataset and update
and update it it
3: Initialize
3: Initializeevaluation
evaluation to to
zerozero
4:
4: For
Forthe
thecluster
clustersize k =k k-minimum
size = k-minimum to k-maximum:
to k-maximum:
5: Constructclusters
5: Construct clusters(k),(k),
that is, the
that k - clusters
is, the by applying
k - clusters Algorithm
by applying 4.
Algorithm 4.
6: Evaluate clusters (k) using entropy-based clustering criterion (Eqn. 19) andevaluation
6: Evaluate clusters (k) using entropy-based clustering criterion (Eqn. 19) and update (k)
update evaluation (k)
7: Compare evaluation (k) with evaluation and update evaluation, best-clusters and best (k)
7: Compare evaluation (k) with evaluation and update evaluation, best-clusters and best (k)
8: End For
8: End For
9: Construct a bipartite graph G for the training-dataset
9: Construct a bipartite graph G for the training-dataset
10: Initialize training-set to null
10: Initialize
11: For j = 1 totraining-set
best (k): to null
11: For j = 1 to best (k):
12: Compute λ-best (j) by applying Algorithm 5
12: For
13: Compute
each user λ-best (j) by applying
in best-clusters (j): Algorithm 5
13: For each user in best-clusters (j):
14: Update user-features by extracting the features from user attributes-set
14: Include
15: Updatethe user-features by extracting
classifier: (user-features, the(j))
λ-best features from user attributes-set
in the training-set
15: End
16: Include
For the classifier: (user-features, λ-best (j)) in the training-set
17:
16: End
EndFor For
17: End Forthe classifier by training to the training-set
18: Update
19:
18: Return
Updatethe thebipartite graph
classifier byand the classifier
training to the training-set
19: Return the bipartite graph and the classifier
Mathematics 2020, 8, 1106 15 of 27
This algorithm computes the λ-best, the best value of the profile of the user’s λ for each cluster. It determines
λ-best, by evaluating each recommendation list, which is generated with each value of λ by applying the
metrics recall, precision and ranking score. The set Λ consists of the typical λ values that lie between 0.0 and 1.0.
The output of the trained classifier from Algorithm 3 determines the best λ for a target user in its
recommender phase. This step requires the following inputs: a bipartite graph G, attributes set, target
user, a trained classifier that is the result obtained from Algorithm 3, and |L|. The complete operations
of the recommendation steps are described in Algorithm 3. The algorithm starts with extracting the
target user0 s feature values and is assigned to user features. Then, λUser , the proper λ value for the
user, will be predicted by giving the user-features values into the trained classifier. The target user
profile is reflected in the variable λUser with Mass Diffusion Heat Spreading algorithm that selects the
novelty-based popular items. The recommendation list is updated by applying the Mass Diffusion Heat
Spreading algorithm with the tuning parameter λUser and |L|. Finally, the lesser-known and favorite
items are combined in the recommendation list based on the user profile of a target user. The flowchart
of the recommendation step of the proposed method 2 is depicted in Figure 4. The specific features
based recommendation process is described in Algorithm 6.
Mathematics 2020, 8, 1106 16 of 27
Mathematics 2020, 8, x FOR PEER REVIEW 16 of 29
Algorithm 6:6:Specific
Algorithm Specific Features
Featuresbased Recommendation
based Recommendation
Inputs:
Inputs: G,G,attributes-set,
attributes-set, target-user,
target-user, |L| |L|
classifier,
classifier,
1: Update user-features
Update user-features byby extracting
extracting the features
the features from from user attributes-set
user attributes-set
2: Predict λ
Predict λUser , , the proper λ value for the user, by giving
the proper λ value for the user, by giving the user-features values into
the user-features the trained
values classifier
into the trained
3: Apply
classifier the Mass Diffusion Heat Spreading algorithm and update the Recommendation-List
3: Return Recommendation-List
4: Apply the Mass Diffusion Heat Spreading algorithm and update the Recommendation-List
4: Return Recommendation-List
In general, the worst-case complexity of the proposed recommenders is O (ni) for n users with the
numberIn general, the worst-case
of data items, complexity
i. The complexity of theconstruction
of model proposed recommenders is O based
of proposed cluster (ni) for n users with
recommenders
the
withnumber of data
the partition of c items,
clustersi. isThe complexity
O (cni), a linear of model construction
complexity for one ratingof prediction
proposed with clusterthe based
space
recommenders with the partition of c clusters is O (cni), a linear complexity
requirement of O (ci + n). The asymptotic complexity of the proposed k-modes clustering strategy is for one rating prediction
with
Θ(nki), the spacecomplexity
a linear requirement sinceoftheO k(ci + n). are
clusters Theevaluated
asymptotic complexity
for each value of of the proposed
k ranging k-modes
from k-minimum
clustering
to k-maximum. strategy is Θ(nki),
However, a linear complexity
in practical applications, since
it isthe k clusters
expected are evaluated
to have for each value
O (n + i) complexity sinceoffor
k
ranging from k-minimum to k-maximum. However, in practical applications,
each user only a finite number of items are considered. Since one loop is used on n users to compute it is expected to have
O
the(nsimilarity
+ i) complexity
and one since for ieach
on the itemsuser only a finite
to compute number of items are considered. Since one loop
the prediction.
is used on n users to compute the similarity
The implementation cost in realistic recommender and one on thesystems
i items to compute
depends onthetheprediction.
number of user
attributes, size of the recommendation list, the tuning parameter, and the numbernumber
The implementation cost in realistic recommender systems depends on the of graphofedges
user
attributes,
produced in size
theofbipartite
the recommendation
graph. There list, the tuning
are some solutionsparameter, and the the
for addressing number of graph edges
implementation cost
produced in the bipartite graph. There are some solutions for addressing
in realistic recommender systems: discarding the users with minimal required popular items and the implementation cost in
realistic recommender systems: discarding the users with minimal required
discarding very popular items, since the items are partitioned into different datasets and clustering the popular items and
discarding
data is applied.very popular items, since the items are partitioned into different datasets and clustering
the data is applied.
5. Simulations & Results
5. Simulations & Results
This section focuses on the simulation of the proposed algorithms on the sample datasets with
This
its outcomes section
andfocuses on the
analysis. Thesimulation
algorithms of are
the implemented
proposed algorithms on the
in the Java sample datasets
language. with
It has been
its outcomes and
experimentally analysis.
found that theThe algorithms
computation timearedepends
implemented in the Java
on the structure of thelanguage.
computation,It has been
number
experimentally
of user attributes, found
size ofthat
thethe
graph computation
edges, and time depends on the structure of the computation,
the dataset.
number of user attributes, size of the graph edges, and the dataset.
5.1. Datasets
5.1. Datasets
The proposed algorithms are implemented and executed on the following data sets [19]:
MovieLens
The The dataset,
proposed which are
algorithms consists of the users,
implemented items, and
and executed on interactions
the followingvalues as 910,
data sets [19]:1672,
The
MovieLens dataset, which consists of the users, items, and interactions values as 910, 1672, 95,579
Mathematics 2020, 8, 1106 17 of 27
95,579 respectively, with user’s features age, location, and gender. Last.FM dataset consists of the users,
items, and interaction values of 2846, 4995, 14,583 respectively, with users, features country, gender,
and age. The dataset Book-Crossing consists of the users, items, and interactions values of 3421, 26811,
35,572 respectively with users, features age and location.
Strategies Ranking Score (User) Recall (List, User) Precision (List, User)
CF [11,23] 0.497 0.014 0.003
MDHS [36,46] 0.080 0.299 0.086
UPOD [6] 0.074 0.308 0.091
Proposed Method 1 0.070 0.312 0.091
Proposed Method 2 0.069 0.312 0.092
Strategies Ranking Score (User) Recall (List, User) Precision (List, User)
CF [11,23] 0.002 0.002 0.001
MDHS [36,46] 0.001 0.008 0.001
UPOD [6] 0.001 0.007 0.002
Proposed Method 1 0.001 0.008 0.002
Proposed Method 2 0.001 0.008 0.003
Mathematics 2020, 8, 1106 18 of 27
Strategies Ranking Score (User) Recall (List, User) Precision (List, User)
CF [11,23] 0.444 0.018 0.001
MDHS [36,46] 0.252 0.119 0.006
UPOD [6] 0.251 0.174 0.009
Proposed Method 1 0.251 0.175 0.010
Proposed Method 2 0.250 0.176 0.011
Strategies Ranking Score (User) Recall (List, User) Precision (List, User)
CF [11,23] 0.004 0.004 0.000
MDHS [36,46] 0.007 0.006 0.000
UPOD [6] 0.006 0.014 0.001
Proposed Method 1 0.006 0.015 0.001
Proposed Method 2 0.004 0.019 0.001
Strategies Ranking Score (User) Recall (List, User) Precision (List, User)
CF [11,23] 0.406 0.000 0.000
MDHS [36,46] 0.395 0.002 0.000
UPOD [6] 0.394 0.003 0.000
Proposed Method 1 0.395 0.004 0.000
Proposed Method 2 0.398 0.004 0.000
Strategies Ranking Score (User) Recall (List, User) Precision (List, User)
CF [11,23] 0.007 0.000 0.000
MDHS [36,46] 0.007 0.001 0.000
UPOD [6] 0.008 0.001 0.000
Proposed Method 1 0.008 0.001 0.000
Proposed Method 2 0.007 0.005 0.000
The proposed methods are evaluated and compared with the other well-known strategies, such as
CF, MDHS, and UPOD. The parameter λ has been set to 0.5 when comparing with MDHS and UPOD
methods. The proposed methods are evaluated using the n group cross-validation in which the data
set is divided into n groups of equal sizes. ValidUsers , that is, the users are those who are available both
in testing and the training set, are considered for the evaluation.
The performance of the proposed methods with the existing methods is evaluated using the three
metrics: Recall (List, User), Precision (List, User), and Ranking Score (User), and the recommendation list
generated by the system is validated. The simulation outcomes are tested using the statistical t-test with
a level of significance α = 0.05 to check if there is a significant difference between the proposed methods,
with the existing methods being statistically significant. The measures kµ and kσ are computed for
each dataset, after applying the clustering with 10 executions. The inferences are analyzed from the
experimental results, which are tabulated from tables Tables 1–6. The figures which are indicated in
bold conclude the better performance of the proposed methods over the existing methods.
The comparison of the mean µ and standard deviation σ to the Movielens dataset is shown in
Tables 1 and 2 respectively. For this dataset, the measures are calculated as kµ = 198 and kσ = 4.21.
For this dataset, it has been found that there is no significant difference obtained in terms of parameters
µ and σ concerning the Ranking Score (User) and Recall (List, User) measures in the proposed methods.
Mathematics 2020, 8, 1106
x FOR PEER REVIEW 19 of 27
29
proposed methods. However, the proposed method 2 behaves well in terms of Precision (List, User)
However,
measurementsthe proposed
compared method 2 behaves
to other methods well in terms
with of Precision
the size (List, User) measurements
of the recommendation compared
list |L| = 30.
to other methods with the size of the recommendation list |L| = 30.
The accuracy of the proposed methods is also evaluated for the MovieLens dataset with the same
The accuracy
parameters of the proposed
considered in [49]. To methods is also
test the evaluated for
performance of thetheMovieLens
proposeddataset with the same
recommenders, the
parameters considered in [49]. To test the performance of the proposed
MovieLens dataset is divided into 90% of the training set and 10% of probe data. The dataset recommenders, the MovieLens
can also
dataset is divided
be divided into 20%),
into (80%, 90% of(70%,
the training
30%), and set and 10%
so on. Theof only
probeknown data. The dataset can
information also be divided
is available in the
into (80%,set
training 20%),
and (70%, 30%), andisso
no prediction on. The
made only
in the known
probe setinformation
of data. Foristhe available
jth userin uj,the
thetraining
positionsetofand
an
no predictionobject
uncollected is madeoj isinmeasured
the probein setthe
of ordered the jthThen
data. Forqueue. user uthej , the position
position of of
oj isanobtained
uncollected object oj
by dividing
is
themeasured
particularin the ordered
location fromqueue.
the Then
top by thethe
position
total of oj is obtained
number by dividing
of uncollected the particular
movies. Hence, location
a good
from the top by the total number of uncollected movies. Hence,
recommender is expected to produce a small Ranking Score (User), which shows the better a good recommender is expected
accuracy
to
of produce a small Ranking
the recommender. Score (User),comparison
The performance which shows theproposed
of the better accuracy
methods of with
the recommender.
other methodsThe for
performance
the MovieLenscomparison
dataset over of the
the three
proposed methods
metrics within
is shown other
Figuremethods
5 [49].for
The MovieLensisdataset
thesimulation over
conducted
the
withthree
10%metrics
of probe is shown
data, Lin= Figure
50. The5 values
[49]. The simulation is to
corresponding conducted
the proposedwith 10% of probe
methods aredata, = 50.
betterL ones
The values corresponding
concerning all three metrics. to the proposed methods are better ones concerning all three metrics.
Figure
Figure 5. The Performance
5. The Performance Comparison
Comparison of
of the
the Proposed
Proposed Methods
Methods with
with Other
Other Methods
Methods [49]:
[49]:
MovieLens dataset.
MovieLens dataset.
The
The comparison
comparisonofofthe themean
mean µ and
𝜇 andstandard
standard deviation σ to the
deviation Last.FM
𝜎 to datasetdataset
the Last.FM is shown isin Tablesin
shown 3
and 4 respectively. For this dataset, the measures are
Tables 3 and 4 respectively. For this dataset, the measures are calculatedcalculated as kµ = 195 and
as 𝑘 = 195 k = 2.75. For
σ and 𝑘 =2.75. the
Last.FM dataset, the proposed method 2 performs better than the
For the Last.FM dataset, the proposed method 2 performs better than the existing methods existing methods in terms of all
in terms
performance metrics considered. For this dataset, there is no significant difference
of all performance metrics considered. For this dataset, there is no significant difference obtained obtained even when
the
evensize of thethe
when recommendation list |L| becomes 30,
size of the recommendation listwhile keeping the
|L| becomes 30, size of the
while neighborhood
keeping the size asof <30
the
in all executions.
neighborhood as < 30 in all executions.
For n,
For n,the
thenumber
number of of
ratings present
ratings in the
present in test
the set,
testthe quality
set, of the predicted
the quality rating is obtained
of the predicted rating is
using RMSE [50]. This metric compares the prediction ratings with the probe
obtained using RMSE [50]. This metric compares the prediction ratings with the probe test set. The test set. The accuracy of
the proposed methods is also evaluated for the Last.FM dataset and the
accuracy of the proposed methods is also evaluated for the Last.FM dataset and the results areresults are compared with the
methods
comparedpresented
with the in [50]. The
methods minimum
presented inRMSE is considered
[50]. The minimum for someisof
RMSE the methods
considered for presented
some of the in
the DGR. The
methods RMSE in
presented comparison
the DGR.ofThe the proposed methods with
RMSE comparison DGR
of the presented
proposed in [50] iswith
methods shown DGRin
Figure 6. In the Figure, the size of the groups is plotted on the X-axis and the
presented in [50] is shown in Figure 6. In the Figure, the size of the groups is plotted on the X-axisRMSE is plotted on the
Y-axis.
and theThe RMSEpredictions
is plottedgenerated by the proposed
on the Y-axis. recommenders
The predictions generatedarebybetter
the than the existing
proposed methods.
recommenders
The experimental results further conclude that the proposed recommenders
are better than the existing methods. The experimental results further conclude that the proposed consider the features of
the individual preferences of the group members and the specific features.
recommenders consider the features of the individual preferences of the group members and the It has also been found that
specific features. It has also been found that the accuracy is slightly decreasing while the group size
Mathematics 2020, 8, 1106 20 of 27
Mathematics 2020, 8, x FOR PEER REVIEW 20 of 29
Mathematics 2020, 8, x FOR PEER REVIEW 20 of 29
the accuracy is slightly
is increasing decreasing
since the while the groupdepends
group recommendation size is increasing sinceset
on the diverse theofgroup
users recommendation
and specific
is increasing
depends on the since the
diverse setgroup
of recommendation
users and specific depends features.
personal on the diverse set of users and specific
personal features.
personal features.
Figure 6. The
Figure RMSE
6. The RMSEComparison
Comparison of the Proposed
of the ProposedMethods
Methods with
with DGR
DGR [50].[50].
Figure 6. The RMSE Comparison of the Proposed Methods with DGR [50].
The The
performance
performance of of
specific
specificfeature
featureselection
selection isis also
alsoanalyzed
analyzed based
based on on varying
varying the values
the values of of
The performance
neighborhood |S|. The
sizessizes of specific feature selection is alsoduring
analyzedthebased on varying
whenthe values of
neighborhood |S|.interesting result
The interesting is obtained
result is obtained during simulation
the simulation the experiments
when the
neighborhood sizes |S|. The interesting result is obtained during the simulation when the
are conducted
experiments forare
small, medium,
conducted for and
small,large neighborhood
medium, and largesizes |S| and aresizes
neighborhood shown
|S|in Figure
and 7. TheinRMSE
are shown
experiments are conducted for small, medium, and large neighborhood sizes |S| and are shown in
values of 41% and 37% are obtained in the proposed methods when |S| < 20. When the value
Figure 7. The RMSE values of 41% and 37% are obtained in the proposed methods when |S| < 20. of |S|
Figure 7. The RMSE values of 41% and 37% are obtained in the proposed methods when |S| < 20.
When
increases, the value of |S| increases, RMSE also increases. For
of |S|,For large values of |S|, there is a significant
When RMSE
the value also
of increases.
|S| increases,ForRMSE
largealso
values
increases. there is avalues
large significant
of |S|, difference that becomes
there is a significant
difference
smaller in the that becomes
proposed smaller in the proposed methods.
methods.
difference that becomes smaller in the proposed methods.
Figure 7. The RMSE Comparison of the Proposed Methods for Different Neighborhood Sizes.
Figure 7. The
Figure RMSE
7. The Comparison
RMSE Comparisonof
ofthe
the Proposed Methodsforfor
Proposed Methods Different
Different Neighborhood
Neighborhood Sizes.Sizes.
The average (kµ ) & standard deviation (kσ ) of observed ratings for the proposed datasets are
shown in Table 7.
Mathematics 2020, 8, x FOR PEER REVIEW 21 of 29
The average (𝑘 ) & standard deviation (𝑘 ) of observed ratings for the proposed datasets are
shown in 2020,
Mathematics Table 7.
8, 1106 21 of 27
The
The comparison
comparisonofofthe themean
mean µ and
𝜇 and standard
standard deviation σ to the
deviation 𝜎 to Book-Crossing dataset dataset
the Book-Crossing is shown is
in Tablesin5Tables
shown and 6 respectively. For this dataset,
5 and 6 respectively. For this the measures
dataset, are calculated
the measures as kµ = 194
are calculated 𝑘 =kσ194
asand = 7.83.
and
𝑘 =this
For dataset,
7.83. For thisit has beenitfound
dataset, that found
has been there isthat
a significant
there is a difference
significantobtained
differencein obtained
terms of parameters
in terms of
µparameters
and σ concerning
𝜇 and the 𝜎 concerning
Precision (List,
theUser) and Recall
Precision User) User)
(List, (List, and measures
Recall (List,in User)
the proposed
measures methods.
in the
However,
proposed the proposed
methods. methods
However, theare competitive
proposed methods withare
thecompetitive
existing methods.
with theThe proposed
existing method
methods. The2
performs well compared to proposed method 1. It has also been found that,
proposed method 2 performs well compared to proposed method 1. It has also been found that, for for this dataset, there is no
significant
this dataset, difference
there is noobtained evendifference
significant when the size of theeven
obtained recommendation
when the sizelist of |L|
thebecomes 50 while
recommendation
keeping the size of
list |L| becomes 50the neighborhood
while keeping theassize < 30
of in
theallneighborhood
executions. as < 30 in all executions.
The
The proposed
proposed methods are simulated simulated for forthe
thelarge
largescale
scaledata
dataset Book-Crossing
setBook-Crossing which
which consists
consists of
of 3421
3421 users
users andand 26,811
26,811 items
items on on varying
varying thethe tuning
tuning parameter
parameter λ for
λ for the the different
different values
values as 0.3,
as 0.3, 0.4, 0.4,
0.5,
0.5,
andand0.6. 0.6. The significant
The significant resultsresults are obtained
are obtained and are and are shown
shown in Figure
in Figure 8. been
8. It has It has beenthat
found found when thatλ
when λ < Ranking
< 0.5, the 0.5, the Ranking Score
Score (User) (User) increases
increases while thewhile Precision
the (List,
Precision User) andUser)
(List, Recalland Recall
(List, User)(List, User)
measures
measures
decrease. decrease.
When λ > When > 0.5, the
0.5, theλ Ranking Ranking
Score (User)Score
also (User) alsogradually
increases increases andgradually
no such and no such
significant
significant
differencesdifferences
are obtained areinobtained
the Precision Precision
in the(List, andUser)
User)(List, Recalland Recall
(List, (List,
User) User) measures.
measures. In this
In this case, the
case, the experimental
experimental results
results show show
that the that the proposed
proposed methodmethod 2 can provide
2 can provide a bettera recommendation
better recommendation based
based
on theon the defined
defined metrics metrics
when whenλ = 0.5.λ = 0.5.
For theBook-Crossing
For the Book-Crossingdataset,
dataset,thethemetric
metric Precision
Precision(List, User)
(List, is evaluated
User) for for
is evaluated different sizessizes
different of the
of
recommendation list, for example, for the values of |L| = 10, 20, 30, 40, 50, and
the recommendation list, for example, for the values of |L| = 10, 20, 30, 40, 50, and 100. The 100. The corresponding
metric is plotted
corresponding as shown
metric in Figure
is plotted as shown9. When this metric
in Figure 9. When becomes higher,
this metric the items
becomes thatthe
higher, in L
areitems
corresponding to the items list
that are in L corresponding that
to the is corresponding
items to the users to
list that is corresponding present in the
the users testing
present in set
the values.
testing
By the tuning parameter as λ =
set values. By keeping the tuning parameter as λ = 0.5, the proposed method 2 provides while
keeping 0.5, the proposed method 2 provides better performance better
|L|
performance |L| > 30,
≤ 30. Whenwhile |L|it ≤seems that the
30. When |L|proposed methodthat
> 30, it seems 1 provides better recommendation
the proposed method 1 provides based on
better
this metric.
recommendation based on this metric.
Mathematics 2020, 8, 1106 22 of 27
Mathematics 2020, 8, x FOR PEER REVIEW 22 of 29
Figure 9. Size of the Recommendation List versus Precision (List, User) for Book-Crossing.
Figure 9. Size of the Recommendation List versus Precision (List, User) for Book-Crossing.
The experiments are also conducted for the Book-Crossing dataset when varying the user sizes,
The experiments are also conducted for the Book-Crossing dataset when varying the user sizes,
for example, for the values of 500, 1000, 1500, 2000, 2500, and 3000, and the corresponding Precision
for example, for the values of 500, 1000, 1500, 2000, 2500, and 3000, and the corresponding Precision
User)User)
(List, (List, & Recall (List, User) metrics are plotted as shown in Figure 10. The experimental results
& Recall (List, User) metrics are plotted as shown in Figure 10. The experimental results
analyze that for small
analyze that for small values of user
values sizes
of user (≤500),
sizes higher
(≤500), highervalues
valuesofofthese
thesemetrics
metrics are
are obtained. When the
obtained. When
user sizes
the user sizes exceed 500, these metrics decrease gradually and the proposed method 2 performsbetter
exceed 500, these metrics decrease gradually and the proposed method 2 performs
than better
the proposed
than the method
proposed1.method
The significant differences
1. The significant are obtained
differences in these
are obtained methods
in these because
methods
because the
the structure of structure
computation of computation of the methods
of the methods depends depends
on theon the number
number of user
of user attributes,
attributes, sizeof the
size
of the recommendation
recommendation list, the
list, the tuning tuning parameter,
parameter, and the and the number
number of graphof graph
edges edges produced
produced in theinbipartite
the
bipartite graph. The maximum values of the (Precision, Recall) metrics obtained for
graph. The maximum values of the (Precision, Recall) metrics obtained for the Book-Crossing dataset in the Book-Crossing
dataset in 2020,
the proposed methods are (0.0004, 0.0042) and (0.0004, 0.0046) respectively.
the proposed
Mathematics methods are
8, x FOR (0.0004,
PEER REVIEW 0.0042) and (0.0004, 0.0046) respectively. 23 of 29
Figure 10. Sizes of the Users versus Precision & Recall Metrics for Book-Crossing.
Figure 10. Sizes of the Users versus Precision & Recall Metrics for Book-Crossing.
results conclude that the tuning parameter λ serves the role of supervised learning and obtains a better
recommendation list for the considered sparse datasets. For the MovieLens dataset, it has been found
that there is no significant difference obtained in terms of parameters µ and σ concerning the Ranking
Score (User) and Recall (List, User) measures in the proposed methods. For this dataset, the simulation
is conducted with 10% of probe data, L = 50. The values corresponding to the proposed methods are
better ones concerning all three metrics. For the Last.FM dataset, the proposed method 2 performs
better than the existing methods in terms of all performance metrics considered. The predictions
generated by the proposed recommenders are better than the existing methods concerning the RMSE
metric. The experimental results further conclude that the proposed recommenders consider the
features of the individual preferences of the group members and the specific features. It has also
been found that the accuracy is slightly decreasing while the group size is increasing since the group
recommendation depends on the diverse set of users and specific personal features. The performance
of specific feature selection is also analyzed based on varying the values of neighborhood sizes |S|.
The interesting result is obtained during the simulation when the experiments are conducted for small,
medium, and large neighborhood sizes |S|. The RMSE values of 41% and 37% are obtained in the
proposed methods when |S| < 20. When the value of |S| increases, RMSE also increases. For large
values of |S|, there is a significant difference that becomes smaller in the proposed methods. For
the Book-Crossing dataset, it has been found that there is a significant difference obtained in terms
of parameters µ and σ concerning the Precision (List, User) and Recall (List, User) measures in the
proposed methods. However, the proposed methods are competitive with the existing methods. The
proposed method 2 performs well compared to proposed method 1. For the considered datasets, there
is no significant difference obtained even when the size of the recommendation list |L| becomes 30,
while keeping the size of the neighborhood as < 30 in all executions; in addition, also in the proposed
method 2, better recommendations are provided based on the defined metrics when λ is equal to 0.5.
By keeping the tuning parameter as λ = 0.5, the proposed method 2 provides better performance while
|L| ≤ 30. When |L| > 30, it seems that the proposed method 1 provides better recommendation based on
this metric. The experimental results analyze that for small values of user sizes (≤ 500), higher values
of these metrics are obtained. When the user sizes exceed 500, these metrics decrease gradually and the
proposed method 2 performs better than proposed method 1. The significant differences are obtained
in these methods because the structure of computation of the methods depends on the number of
user attributes, size of the recommendation list, the tuning parameter, and the number of graph edges
produced in the bipartite graph. The maximum values of the (Precision, Recall) metrics obtained for the
Book-Crossing dataset in the proposed methods are (0.0004, 0.0042) and (0.0004, 0.0046) respectively.
The average computational time of the proposed methods takes <10 seconds.
The following are the future research directions:
Enhancing the resource values computation and increasing the item rating values,
Designing the strategy to filter the better-rated items and ignoring the poorly rated items,
Applying the local search strategies to improve resource propagation,
Generating a recommender based on multiple attributes selection.
Features based trust or reputation in recommendations can be integrated into the proposed
algorithms interestingly with some evolutionary operators for e-learning and other real-world
applications [59–61,71–73].
Author Contributions: Methodology, software, resources, writing—original draft and formal analysis: S.B.;
conceptualization, formal analysis, validation, writing—review and editing, proofreading: R.M.; revise the paper:
B.S. All authors have read and agreed to the published version of the manuscript.
Funding: This research received no external funding.
Acknowledgments: The authors would like to acknowledge the support rendered by the Management of SASTRA
Deemed University for providing financial support.
Conflicts of Interest: The authors declare no conflict of interest.
Mathematics 2020, 8, 1106 25 of 27
References
1. Bobadilla, J.; Ortega, F.; Hernando, A.; Gutiérrez, A. Recommender systems survey. Knowl. Based Syst. 2013,
46, 109–132. [CrossRef]
2. Zhang, F.; Gong, T.; Lee, V.E.; Zhao, G.; Rong, C.; Qu, G. Fast algorithms to evaluate collaborative filtering
recommender systems. Knowl. Based Syst. 2016, 96, 96–103. [CrossRef]
3. Adomavicius, G.; Tuzhilin, A. Toward the next generation of recommender systems: A survey of the
state-of-the-art and possible extensions. IEEE Trans. Knowl. Data Eng. 2005, 17, 734–749. [CrossRef]
4. Göksedef, M.; Gündüz-Öğüdücü, Ş. Combination of Web page recommender systems. Expert Syst. Appl.
2010, 37, 2911–2922. [CrossRef]
5. Mobasher, B. Recommender Systems. In Kunstliche Intelligenz, Special Issue on Web Mining; BottcherIT Verlag:
Bremen, Germany, 2007; pp. 41–43.
6. Bertani, R.M.; Bianchi, R.A.; Costa, A.H.R. Combining novelty and popularity on personalised
recommendations via user profile learning. Expert Syst. Appl. 2020, 146, 113–149. [CrossRef]
7. Shu, J.; Shen, X.; Liu, H.; Yi, B.; Zhang, Z. A content-based recommendation algorithm for learning resources.
Multimedia Syst. 2018, 24, 163–173. [CrossRef]
8. Syed, M.A.; Rakesh, K.L.; Gopal, K.N.; Rabindra, K.B. Movie recommendation system using genome tags
and content-based filtering. In Advances in Data and Information Sciences; Springer: Singapore, 2018; pp. 85–94.
[CrossRef]
9. Klašnja-Milićević, A.; Vesin, B.; Ivanović, M.; Budimac, Z. E-Learning personalization based on hybrid
recommendation strategy and learning style identification. Comput. Educ. 2011, 56, 885–899. [CrossRef]
10. Cui, G.; Luo, J.; Wang, X. Personalized travel route recommendation using collaborative filtering based on
GPS trajectories. Int. J. Digit. Earth 2018, 11, 284–307. [CrossRef]
11. Beel, J.; Gipp, B.; Langer, S.; Breitinger, C. Research-paper recommender systems: A literature survey. Int. J.
Digit. Libr. 2016, 17, 305–338. [CrossRef]
12. Betru, B.T.; Onana, C.A. Deep learning methods on recommender system: A survey of state-of-the-art. Int. J.
Comput. Appl. 2017, 162, 975–8887. [CrossRef]
13. Ai, Q.; Azizi, V.; Chen, X.; Zhang, Y. Learning heterogeneous knowledge base embeddings for explainable
recommendation. Algorithms 2018, 11, 137. [CrossRef]
14. Celma, O. Music Recommendation and Discovery; Springer: Berlin, Germany, 2010; ISBN 978-3-642-13287-2.
15. Chen, B.; Zeng, A.; Chen, L. The effect of heterogeneous dynamics of online users on information filtering.
Phys. Lett. A 2015, 379, 2839–2844. [CrossRef]
16. Chen, K.; Liu, L. Best K: Critical clustering structures in categorical datasets. Knowl. Inf. Syst. 2009, 20, 1–33.
[CrossRef]
17. Deng, X.; Zhong, Y.; Lü, L.; Xiong, N.; Yeung, C. A general and effective diffusion-based recommendation
scheme on coupled social networks. Inf. Sci. 2017, 417, 420–434. [CrossRef]
18. Fu, M.; Qu, H.; Moges, D.; Lu, L. Attention based collaborative filtering. Neurocomputing 2018, 311, 88–98.
[CrossRef]
19. Harper, F.M.; Konstan, J.A. The MovieLens datasets: History and context. ACM Trans. Interact. Intell. Syst.
2015, 5, 19:1–19:19. [CrossRef]
20. Huang, Z. Extensions to the k-means algorithm for clustering large data sets with categorical values. Data Min.
Knowl. Disc. 1998, 2, 283–304. [CrossRef]
21. Javari, A.; Jalili, M. A probabilistic model to resolve diversity-accuracy challenge of recommendation systems.
Knowl. Inf. Syst. 2014, 44, 609–627. [CrossRef]
22. Kaminskas, M.; Bridge, D. Diversity, serendipity, novelty, and coverage. ACM Trans. Interact. Intell. Syst.
2016, 7, 1–42. [CrossRef]
23. Katarya, R.; Verma, O.P. Recent developments in affective recommender systems. Physica A 2016, 461,
182–190. [CrossRef]
24. Kotkov, D.; Wang, S.; Veijalainen, J. A survey of serendipity in recommender systems. Knowl. Based Syst.
2016, 111, 180–192. [CrossRef]
25. Lacerda, A. Multi-Objective ranked bandits for recommender systems. Neurocomputing 2017, 246, 12–24. [CrossRef]
26. Liu, H.; Hu, Z.; Mian, A.; Tian, H.; Zhu, X. A new user similarity model to improve the accuracy of
collaborative filtering. Knowl. Based Syst. 2014, 56, 156–166. [CrossRef]
Mathematics 2020, 8, 1106 26 of 27
27. Lu, J.; Shambour, Q.; Xu, Y.; Lin, Q.; Zhang, G. A web-based personalized business partner recommendation
system using fuzzy semantic techniques. Comput. Intell. 2013, 29, 37–69. [CrossRef]
28. Lu, J.; Wu, D.; Mao, M.; Wang, W.; Zhang, G. Recommender system application developments: A survey.
Decis. Support Syst. 2015, 74, 12–32. [CrossRef]
29. Ma, T.; Zhou, J.; Tang, M.; Tian, Y.; Al-Dhelaan, A.; Al-Rodhaan, M.; Lee, S. Social network and tag sources
based augmenting collaborative recommender system. IEICE Trans. Inf. Syst. 2015, 98, 902–910. [CrossRef]
30. Ma, W.; Ren, C.; Wu, Y.; Wang, S.; Feng, X. Personalized recommendation via unbalance full-connectivity
inference. Physica A 2017, 483, 273–279. [CrossRef]
31. Patra, B.K.; Launonen, R.; Ollikainen, V.; Nandi, S. A new similarity measure using Bhattacharyya coefficient
for collaborative filtering in sparse data. Knowl. Based Syst. 2015, 82, 163–177. [CrossRef]
32. Pearson, K. Notes on the history of correlation. Biometrika 1920, 13, 25–45. [CrossRef]
33. Ricci, F.; Rokach, L.; Shapira, B.; Kantor, P.B. Recommender Systems Handbook, 1st ed.; Springer:
New York, NY, USA, 2010. [CrossRef]
34. Shambour, Q.; Lu, J. An effective recommender system by unifying user and item trust information for B2B
applications. J. Comput. Syst. Sci. 2015, 81, 1110–1126. [CrossRef]
35. Sánchez-Moreno, D.; Gil González, A.B.; Muñoz Vicente, M.D.; López Batista, V.F.; Moreno García, M.N.
A collaborative filtering method for music recommendation using playing coefficients for artists and users.
Expert Syst. Appl. 2016, 66, 1339–1351. [CrossRef]
36. Wang, X.; Liu, Y.; Zhang, G.; Zhang, Y.; Chen, H.; Lu, J. Mixed similarity diffusion for recommendation on
bipartite networks. IEEE Access 2017, 5, 21029–21038. [CrossRef]
37. Witten, I.H.; Frank, E.; Trigg, L.; Hall, M.; Holmes, G.; Cunningham, S.J. Weka: Practical machine learning tools
and techniques with Java implementations. In Proceedings of the ICONIP/ANZIIS/ANNES’99 Workshop on
Emerging Knowledge Engineering and Connectionist-Based Information Systems, Dunedin, New Zealand,
22–23 November 1999; Kasabov, N., Ko, K., Eds.; 1999; pp. 192–196.
38. Yang, Z.; Wu, B.; Zheng, K.; Wang, X.; Lei, L. A survey of collaborative filtering-based recommender systems
for mobile internet applications. IEEE Access 2016, 4, 3273–3287. [CrossRef]
39. Yu, F.; Zeng, A.; Gillard, S.; Medo, M. Network-based recommendation algorithms: A review. Physica A 2015,
452, 192–208. [CrossRef]
40. Zeng, W.; An, Z.; Liu, H.; Shang, M.-S.; Zhou, T. Uncovering the information core in recommender systems.
Sci. Rep. 2014, 4, 6140. [CrossRef] [PubMed]
41. Zeng, W.; Zeng, A.; Shang, M.S.; Zhang, Y.C. Information filtering in sparse online systems: Recommendation
via semi-local diffusion. PLoS ONE 2013, 8, e79354. [CrossRef]
42. Zhang, F.-G.; Zeng, A. Information filtering via heterogeneous diffusion in online bipartite networks.
PLoS ONE 2015, 10, e0129459. [CrossRef]
43. Zhang, S.; Yao, L.; Sun, A. Deep learning based recommender system: A Survey and new perspectives.
arXiv 2017, arXiv:1707.07435. [CrossRef]
44. Zhang, Y.-C.; Blattner, M.; Yu, Y.-K. Publisher’s note: Heat conduction process on community networks as a
recommendation model. Phys. Rev. Lett. 2007, 99, 169902. [CrossRef]
45. Zhou, T.; Kuscsik, Z.; Liu, J.-G.; Medo, M.; Wakeling, J.R.; Zhang, Y.-C. Solving the apparent diversity-accuracy
dilemma of recommender systems. Proc. Natl. Acad. Sci. USA 2010, 107, 4511–4515. [CrossRef]
46. Zhou, T.; Ren, J.; Medo, M.; Zhang, Y.C. Bipartite network projection and personal recommendation. Phys. Rev.
E 2007, 76, 1–7. [CrossRef] [PubMed]
47. Ziegler, C.-N.C.; McNee, S.M.S.; Konstan, J.a.J.; Lausen, G. Improving recommendation lists through topic
diversification. In Proceedings of the 14th International Conference on World Wide Web, Chiba, Japan,
10–14 May 2005; Volume 5, pp. 22–32. [CrossRef]
48. Juliana, A.P.; Pawel, M.; Sebastian, K.; Myra, S.; Gunter, S. A Feature-based personalized recommender
system for product-line configuration. In Proceedings of the 2016 ACM SIGPLAN International Conference
on Generative Programming: Concepts and Experiences, Amsterdam, The Netherlands, 20 October 2016;
pp. 120–131. [CrossRef]
49. Runran, L.; Chunxiao, J.; Tao, Z.; Duo, S.; Binghong, W. Personal recommendation via modified collaborative
filtering. Physica A 2009, 388, 462–468. [CrossRef]
50. Jinpeng, C.; Yu, L.; Deyi, L. Dynamic group recommendation with modified collaborative filtering and
temporal factor. Int. Arab J. Inf. Technol. 2016, 13, 294–301.
Mathematics 2020, 8, 1106 27 of 27
51. Chaturvedi, A.; Green, P.; Caroll, J. K-modes clustering. J. Classif. 2001, 18, 35–55. [CrossRef]
52. Kourosh, M. Recommendation system based on complete personalization. Procedia Comput. Sci. 2016, 80,
2190–2204. [CrossRef]
53. Yonghong, T.; Bing, Z.; Yanfang, W.; Yue, Z.; Qi, W. College library personalized recommendation system
based on hybrid recommendation algorithm. Procedia CIRP 2017, 83, 490–494. [CrossRef]
54. Marchela, A.; Christos, C. Personalized micro-service recommendation system for online news. Procedia Comput.
Sci. 2019, 160, 610–615. [CrossRef]
55. Julián, M.-P.; Jose, A.; Edwin, M.; Camilo, S. Autonomous recommender system architecture for virtual
learning environments. Appl. Comput. Inf. 2020. [CrossRef]
56. Anand, S.T. Generating items recommendations by fusing content and user-item based collaborative filtering.
Procedia Comput. Sci. 2020, 167, 1934–1940. [CrossRef]
57. Pradeep, K.R.; Sarabjeet, S.C.; Rocky, B. A machine learning approach for automation of resume
recommendation system. Procedia Comput. Sci. 2020, 167, 2318–2327. [CrossRef]
58. Hanane, Z.; Souham, M.; Chaker, M. New contextual collaborative filtering system with application to
personalized healthy nutrition education. J. King Saud Univ.–Comput. Inf. Sci. 2020. [CrossRef]
59. Christian, R.; Michael, W.; Günther, P. State of the art of reputation-enhanced recommender systems.
Web Intell. 2018, 16, 273–286.
60. Pasquale, D.M.; Lidia, F.; Fabrizio, M.; Domenico, R.; Giuseppe, M.L.S. Providing recommendations in social
networks by integrating local and global reputation. Infor. Syst. 2018, 78, 58–67. [CrossRef]
61. Barry, S.; Maurice, C.; Peter, B.; Kevin, M.; Michael, P.O. Collaboration, Reputation and Recommender Systems
in Social Web Search. In Recommender Systems Handbook; Springer: Boston, MA, USA, 2015; pp. 569–608.
[CrossRef]
62. Stai, E.; Kafetzoglou, S.; Tsiropoulou, E.E.; Papavassiliou, S. A holistic approach for personalization, relevance
feedback & recommendation in enriched multimedia content. Multimed. Tools Appl. 2018, 77, 283–326.
[CrossRef]
63. Vasiliki, P.; Stella, K.; Eirini, E.T.; Aggeliki, D.; Symeon, P. Personalized multimedia content retrieval through
relevance feedback techniques for enhanced user experience. In Proceedings of the 13th International
Conference on Telecommunications (ConTEL), Graz, Austria, 13–15 July 2015. [CrossRef]
64. Simon, D. Dynamic generation of personalized hybrid recommender systems. In Proceedings of the RecSys’13:
Proceedings of the 7th ACM Conference on Recommender Systems, Hong Kong, China, 12–16 October 2013;
pp. 443–446. [CrossRef]
65. Thai, M.T.; Wu, W.; Xiong, H. Big Data in Complex and Social Networks, 1st ed.; Chapman & Hall/CRC Big Data
Series; CRC Press: Boca Raton, FL, USA, 2016; ISBN-10: 1498726844, ISBN-13: 978-1498726849.
66. Abbas, S.M.; Alam, K.A.; Shamshirband, S. A soft-rough set based approach for handling contextual sparsity
in context-aware video recommender systems. Mathematics 2019, 7, 740. [CrossRef]
67. Sardianos, C.; Ballas Papadatos, G.; Varlamis, I. Optimizing parallel collaborative filtering approaches for
improving recommendation systems performance. Information 2019, 10, 155. [CrossRef]
68. Pajuelo-Holguera, F.; Gómez-Pulido, J.A.; Ortega, F. Performance of two approaches of embedded
recommender systems. Electronics 2020, 9, 546. [CrossRef]
69. Bai, L.; Hu, M.; Ma, Y.; Liu, M. A hybrid two-phase recommendation for group-buying e-commerce
applications. Appl. Sci. 2019, 9, 3141. [CrossRef]
70. Cintia Ganesha Putri, D.; Leu, J.-S.; Seda, P. Design of an unsupervised machine learning-based movie
recommender system. Symmetry 2020, 12, 185. [CrossRef]
71. Bhaskaran, S.; Santhi, B. An efficient personalized trust based hybrid recommendation (TBHR) strategy for
e-learning system in cloud computing. Cluster Comput. 2019, 22, 1137–1149. [CrossRef]
72. Marappan, R.; Sethumadhavan, G. Solution to graph coloring using genetic and tabu search procedures.
Arab. J. Sci. Eng. 2018, 43, 525–542. [CrossRef]
73. Marappan, R.; Sethumadhavan, G. Complexity analysis and stochastic convergence of some well-known
evolutionary operators for solving graph coloring problem. Mathematics 2020, 8, 303. [CrossRef]
© 2020 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access
article distributed under the terms and conditions of the Creative Commons Attribution
(CC BY) license (http://creativecommons.org/licenses/by/4.0/).