Skip to main content
Top
Published in: International Journal of Data Science and Analytics 4/2022

Open Access 22-01-2022 | Regular Paper

Personalized multi-faceted trust modeling to determine trust links in social media and its potential for misinformation management

Authors: Alexandre Parmentier, Robin Cohen, Xueguang Ma, Gaurav Sahu, Queenie Chen

Published in: International Journal of Data Science and Analytics | Issue 4/2022

Activate our intelligent search to find suitable subject content or patents.

search-config
loading …

Abstract

In this paper, we present an approach for predicting trust links between peers in social media, one that is grounded in the artificial intelligence area of multiagent trust modeling. In particular, we propose a data-driven multi-faceted trust modeling which incorporates many distinct features for a comprehensive analysis. We focus on demonstrating how clustering of similar users enables a critical new functionality: supporting more personalized, and thus more accurate predictions for users. Illustrated in a trust-aware item recommendation task, we evaluate the proposed framework in the context of a large Yelp data set. We then discuss how improving the detection of trusted relationships in social media can assist in supporting online users in their battle against the spread of misinformation and rumors, within a social networking environment which has recently exploded in popularity. We conclude with a reflection on a particularly vulnerable user base, older adults, in order to illustrate the value of reasoning about groups of users, looking to some future directions for integrating known preferences with insights gained through data analysis.
Notes

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

1 Introduction

Online sources of information are increasingly relied upon by many. According to yearly studies by the Pew Research Center, the percentage of American adults using the internet has jumped from 52% in 2000 to 90% in 2019 [3]. In addition to the established institutions that have made the jump from paper and TV to the web, many new blogs, content aggregators, and social networks have become a vital source in the information diet: up to 62% of American adults rely on information shared through social media for their news [35]. A study conducted in the wake of the 2016 American election found that, among American voters, Facebook ranked as the third most relied upon source for news about the election (after Fox News and CNN), far outranking local TV and newspapers, and many national stations [14]. It is clear that the power to influence and inform has shifted drastically away from traditional institutions and into the hands of individuals.
While this democratization of information and influence may strike one as appealing, there are reasons to be concerned about this new paradigm. According to Facebook, throughout the 2016 American election thousands of ads designed to incite panic over gun rights and LGBTQ 1 rights were purchased by accounts believed to be funded by the Russian government, some of them specifically targeting voters in swing districts [42]. Also in 2016, a heavily armed man broke into a neighborhood pizza parlor during business hours and fired shots after having become convinced by an online conspiracy popular on Twitter that the basement of the restaurant was used by the Clintons and other Washington elite to murder and rape children [11].
With an increasing number of individuals garnering attention from provocative posts, it becomes critical to assess whether the content that is shown to users can in fact be trusted. One way to address the existence of untrustworthy information online is to deploy message recommender systems. Rather than showing users a random sampling or chronologically ordered list of the content that has been added to the network since their last visit, artificial intelligence (AI) systems can be designed to reason about which messages should be shown to which users. The subfield of multiagent trust modeling is especially relevant: the future trustworthiness of an agent can be predicted based on reported past experiences of peers with this agent. While this research has traditionally been applied in contexts such as selecting trusted sellers in e-marketplace environments [48], a few efforts in recent times have focused on using the methods for reasoning about reputable content in online social networks [38]. One promising new direction has been to recognize that multiple features of the data may be relevant, and thus that a proper weighting of these different contributing factors, when reasoning about trustworthiness, is important. This is the basic premise of the very novel pursuit known as multi-faceted trust modeling [9, 29].
In this paper, we first of all expand the horizons of multi-faceted trust modeling in order to offer a more comprehensive treatment of the different features under consideration. We then introduce a very important new focus on supporting personalized solutions of trust modeling for users.
We do this by adding an unsupervised clustering step before trust formulation models are fit, and learning a distinct model for each cluster of users. This approach allows groups of similar users, who potentially express trust in similar ways, to have a model fit for their community, rather than receiving trust predictions that have been smoothed out to apply well to the entire population of the network. As will be explained, we consider trust to be subjective and thus believe it critical to move beyond the standard view of current multiagent systems trust modeling which adopts, for all agents in the network, a “one size fits all” approach to trust prediction; we reveal how recommendations that support personalization can lead to improved predictions.
In order to demonstrate the value of our approach, we apply our methods in the context of recommending items to users, making use of a Yelp data set of reviews which indicates user preferences. We demonstrate the value of integrating multi-faceted trust modeling which explicitly reasons about how to weight the different trust indicators, and of supporting personalized predictions of trust links when recommending content to users. Following our results, we reflect further on how to extend our current algorithms and implementation, and then explicitly discuss how the methods proposed in this paper can be used toward helping to moderate online social network content for users. We also illustrate the value of reasoning about certain classes of users such as older adults, to consider solutions which cater to the general needs and preferences of this user base.

2 Background

Trust modeling is a subfield of study within the artificial intelligence research area of multiagent systems [44]. We consider trustworthiness as defined in [6], namely the quality of being worthy of trust: in essence, the truster shows willingness to take risk based on a belief that the trustee is expected to exhibit reliable behavior, drawing from an assessment of past experience.

2.1 Multiagent trust modeling related work

Multiagent trust modeling algorithms seek to predict the trustworthiness of another agent based upon first hand experience and reports provided by other agents in the environment, sometimes referred to as advisors [48].
The beta reputation system (BRS) proposed by Jøsang and Ismail [23] estimates reputation of selling agents using a probabilistic model. The beta distributions are a family of statistical distribution functions that are characterized by two parameters \(\alpha \) and \(\beta \). The beta probability density function is defined as follows:
$$\begin{aligned} beta(p|\alpha , \beta ) = \frac{\Gamma (\alpha + \beta )}{\Gamma (\alpha ) \Gamma (\beta )} p^{\alpha - 1}(1 - p)^{\beta - 1} \end{aligned}$$
(1)
where \(\Gamma \) is the gamma function, \(p \in [0, 1]\) is a probability variable, and \(\alpha , \beta > 0\). This function shows the relative likelihood of the values for the parameter p, given the fixed parameters \(\alpha \) and \(\beta \).
Ratings from peers (used to estimate the reputation of a seller) are binary in this model (1 or 0, to represent that the advisor considers the seller to be satisfactory or dissatisfactory in a transaction). Individual ratings received are combined by simply accumulating the number of ratings supporting the conclusion that the seller has good reputation and the number of ratings supporting the conclusion that the seller has bad reputation.
The prior distribution of the parameter p is assumed to be the uniform beta probability density function with \(\alpha = 1\) and \(\beta = 1\). The posteriori distribution of p is the beta probability density function after observing \(\alpha - 1\) ratings of 1 and \(\beta - 1\) ratings of 0.
The reputation of the seller s can then be represented by the probability expectation value of the beta distribution, which is the most likely frequency value, used to predict whether the seller will act honestly in the future. The formalization of this is given as follows:
$$\begin{aligned} Tr(s) = E(p) = \frac{\alpha }{\alpha + \beta }. \end{aligned}$$
(2)
Zhang and Cohen [48] suggest a personalized trust model (PTM) to determine whom to listen to among a network of buyers and sellers in thee-marketplace domain. In particular, they address whether a buyer, b, should purchase a product from a seller, s, based on a combination of global advice from other buyers (i.e., advisors, a), and b’s own local past experiences with s.
The PTM global metric is further broken down to combine private and public trust estimates of advisors. The intuition is that b may have radically different expectations or preferences regarding s’s product than a, and so b should have some notion of how much to trust a. To the extent that b relies on past common experiences to evaluate a’s trustworthiness, b uses a private trust metric to incorporate a’s recommendation. To the extent that b relies on a’s similarity to the global rating of various sellers (i.e., how fair are a’s ratings), b uses a public trust metric to incorporate a’s recommendation.
In particular, b’s private reputation according to a, \(R_{pri}(a,b)\), is modeled by the expectation of a beta distribution where \(\alpha \) is the number of times a and b have agreed in the past about the reputation of other agents, and \(\beta \) corresponds to how many times they have disagreed. The public reputation of b, \(R_{pub}(b)\), is again modeled by the expectation of a beta distribution, where \(\alpha \) corresponds to the number of times b’s advice has agreed with majority opinion, and \(\beta \) the number of times it has not. The final reputation of b fora is then a linear combination of the private and public reputation of b, weighted by a factor w which reflects how much comparable experience a has had with b (i.e., the number of agents commonly rated agents).
$$\begin{aligned} T(a, b) = w R_{pri}(a, b) + (1 - w) R_{pub}(b). \end{aligned}$$
(3)

2.2 Multi-faceted trust modeling

Multi-faceted trust modeling (MFTM) is a flexible and data-driven approach to trust modeling. Inspired by work in the social sciences which have outlined the numerous variables which influence the formation of trust relationships [30], MFTM incorporates arbitrarily many indicators of trustworthiness into a single (optionally context-dependent) trustworthiness score. Operationalizing this core idea for trust and social tie prediction has been proposed by multiple researchers (e.g., [9, 13, 22, 26, 29]). As is evident in these works, there is little agreement over whether this technique should be called multi-dimensional, multi-faceted or composite trust modeling, and this confusion has likely led to some difficulty in coordinating efforts in this research direction. We use the term “multi-faceted,” in keeping with the most recent works.
The defining feature of an MFTM is a customizable vector of trust indicators, where each indicator is a real number based on two agents:
$$\begin{aligned} \varvec{\Psi }(a_1,a_2) = \langle \psi _1(a_1,a_2), \psi _2(a_1,a_2), ..., \psi _n(a_1, a_2)\rangle . \end{aligned}$$
(4)
A “trust indicator” can be thought of as a piece of evidence for or against trusting an agent under a particular context. For example, \(\psi _1(a_1,a_2) = friendCount(a_2)\) may be relevant to assessing the reputation of \(a_2\) in a domain where only popular and reputable agents can accrue large numbers of friends. The indicators \(\psi _i\) must be computable given A and E (the set of agents and their attributes and the history of events). One important feature of MFTM is its flexibility to tune its parameters to different domains of use. The customizability of MFTM is highly attractive for application to social networks, as it is rare to find explicit statements of trust encoded into the feature set of online environments2. Instead, an arbitrary number of “imperfect” indicators of trustworthiness, such as popularity, friendship, reputation, interaction history, preference similarity, and institutional credibility can be considered as each contributing to a final tally of trustworthiness. Clearly, the underlying assumption of this model is that the existence of trustworthiness between two agents can be predicted based on a comparison of the attributes and behaviors of those agents.
The consideration of multiple indicators of trust can be viewed as an emulation of the way in which humans consider multiple sources of evidence when deciding to trust or not [30]. For example, consider the problem of choosing an auto mechanic shortly after having moved to a new town. In this case, one has no interaction history with any nearby mechanics and must weigh available evidence in order to choose which mechanic to trust. In a simple case, one might only consider two pieces of evidence toward or against a mechanic: has any colleague recommended them (\(\psi _1\)), and have their prices been posted clearly online (\(\psi _2\)). In this case both indicators are binary, and it seems likely that the mechanic \(a_j\) who satisfies both indicators, \(\Psi (\overset{\rightharpoonup }{a_i}, \overset{\leftharpoonup }{a_j}) = \langle 1, 1 \rangle \), will be a good candidate to trust. (Note that we use \(\overset{\rightharpoonup }{}\) for the trustor and \(\overset{\leftharpoonup }{}\) for the trustee at times in our discussion below, for additional clarity).
In order to predict trustworthiness, the relevance of each indicator can be learned using an off-the-shelf machine learning technique given A and E to train with. To do this, a trust link is chosen as a target of prediction : y (e.g., explicit statements of trust/friendship, high degrees of preference alignment). Then, given the set of existing implicit/explicit trust links, a machine learning model fits a classifier \(\hat{f}\) to the function that determines how trust indicators are related to trust links \(f: \Psi (a_1,a_2) \rightarrow y\). For example, in the case where logistic regression is used, y will be binary and we have:
$$\begin{aligned} T_c(\overset{\rightharpoonup }{a_1}, \overset{\leftharpoonup }{a_2}) = P_{A,E}(\overset{\rightharpoonup }{a_1}, \overset{\leftharpoonup }{a_2}, c) = \frac{1}{1+\exp ^{-({\theta \cdot \Psi (a,b))}}} \end{aligned}$$
(5)
where \(\theta \) is the vector of weights learned through the logistic regression process and \(T_c\) is trustworthiness under context c. (We believe that whether someone is trusted may truly vary according to the context; for the remainder of the paper, we drop the context variable c in the equations). We wish to emphasize that while logistic regression is an elegant and natural choice with some popularity in the literature (e.g., [9]), it is by no means the only choice.
The ability to define custom indicators appropriate to whichever application domain one is pursuing offers a tremendous amount of flexibility. As we will show in Sect. 3, both highly generic as well as application-specific trust indicators can be defined.
Finally, we wish to emphasize how MFTM can be seen as a generalization of a number of existing trust modeling techniques. Primarily this is because many trust modeling techniques do in fact consider multiple sources of evidence, but they weigh or combine this evidence in a non-data-driven manner. For example, the beta reputation system can be configured so that old advice is considered less important than new advice. However, a method for specifying how much more important newer advice should be treated compared to older advice is not specified. A similar situation occurs in the personalized trust model [48], where private and public reputation are weighed against each other. The weighting function chosen has a good statistical justification3, but ultimately does not specify how error bounds should be chosen, and thus how exactly to weigh personal and private reputation. MFTM can consider arbitrarily many sources of information, and learns the weights for them directly from data. For example, PTM could be roughly replicated by treating private and public reputation as trust indicators, and learning an appropriate function for combining them.
Another example of how MFTM is data driven is that it does not specify which distributions should be used to model beliefs. For example, both the foundational trust modeling Beta Reputation System (BRS) [23] and PTM rely heavily on the beta distribution. By allowing arbitrary machine learning methods to combine many forms of trust evidence into prediction, MFTM loses Bayesian rigor, but gains a large degree of flexibility and generalizability.

2.3 Trust-aware recommendation systems

Trust modeling in the context of recommender systems has been examined by several researchers, dating back to the seminal paper of O’Donovan and Smyth [31]. More recent work has examined such issues as addressing cold start recommendation using trust modeling [17] or examining how to speed up trust-aware recommendation through improvements from matrix factorization [15]. In this paper, trust-aware recommendation arises as a central element of the validation of our proposed framework.
To explain: one of the recurring challenges in the development of trust models is finding grounds for the validation of the accuracy of the models [6]. Trust models aim to predict new trust links but independent agents may choose to follow or ignore these predictions. It is therefore difficult to truly evaluate the effectiveness of models without deploying a system on an active service and measuring the real effects of trust link prediction. As this is expensive and requires the cooperation of an active social network service, many models validate their effectiveness on data generated by an agent simulation instead (e.g., [4, 23, 36, 48]). While this is a useful approach for contrasting the effectiveness of various models and gives the researcher a large amount of control for simulating specific types of agent behavior, it clearly adds a layer of ambiguity between the reported effectiveness of the model and its potential for real-world application. In some cases, merely changing simulation parameters can defeat systems that had performed well on the simulations their creators had designed [24].
A rising trend in this field is to validate models by applying their predictions to a recommendation task (e.g., [9, 29]): that is, using the trust model to predict novel trust links, \(\hat{\Gamma }\), in a multiagent system (MAS), then feeding those predicted links into a trust-aware item recommendation system. These trust-aware recommender systems incorporate both user–item rating behavior and user–user social/trust connections to better recommend items by leveraging the fact that social/trust connections exert influence on the preferences of agents (e.g., you are more likely to watch/enjoy a film a trusted friend recommends). The logic of this two-part process is that when a trust model is able to accurately predict trust links in the context of peer to peer item recommendation, then the resulting accuracy of the recommender system trained with those links will improve.
For the validation of the model, we present in Sect. 3, we implement a task-aware item recommendation task. As will be explained in more detail in Sect. 3.2.3 we introduce two distinct trust-aware recommendation systems, TrustMF [46] and MTR [29]. MTR belongs to a class of recommenders based on k-nearest neighbors [19]. This approach requires a good selection of the value of k and an appropriate distance metric to determine closeness. In Sect. 3.2.3, we provide more insights into how these were chosen for our experimentation with MTR. TrustMF belongs to a class of systems known as latent factor models. To be clearer about how these systems operate, we provide below additional explanation. As will be seen, this method works well with data-driven trust recommendation, in seeking to leverage the most relevant factors of the users.

2.3.1 Latent factor models for recommendation

Latent factor models for recommendation are a popular approach to collaborative filtering-based recommendation derived from matrix factorization technique called Singular Value Decomposition (SVD) [39]. Specifically, by applying an SVD technique, a \(m \times n\) matrix R of rank \(\ell \) can be decomposed into three matrices of rank \(k \le \ell \): \(R = Q \cdot S \cdot V\), where Q is \(m \times k\), S is \(k \times k\) and V is \(k \times n\). While S has a number of interesting mathematical properties, in recommender system literature it is frequently ignored by substituting \(U = Q \cdot S\).
SVD can be applied to recommender systems when R is the user–item matrix of review scores such that \(r_{ij}\) is the rating user i gave to item j. Naturally, this matrix is sparse—in practice, the vast majority of the entries in R are unknown, as most users have only given feedback on a small number of items. While SVD cannot be applied directly to a sparse matrix like R, we can imagine that the defined entries in R comprise a subset of the entries in the (unknown) dense matrix \(R'\) where every user has expressed an opinion on every item. By SVD, \(R'\), is guaranteed to have a minimal rank-k decomposition. This line of reasoning serves as inspiration for the following loss function [25]:
$$\begin{aligned} {{\min _{u_*, v_*}}} \sum _{(i,j) \in \kappa } (r_{ij} - u_i^T v_j)^2 + \lambda (||u_i||^2 + ||v_j||^2) \end{aligned}$$
(6)
where \(u_i\) is a length k vector corresponding to user i and \(v_j\) is a length k vector corresponding to item j, and \(\kappa \) is the set of indices (ij) such that \(r_{ij}\) is defined in R. \(\lambda \) simply controls the strength of the regularization penalty. By optimizing Equation 6, one constructs matrices \(\hat{U}\) and \(\hat{V}\), where the i’th row of \(\hat{U}\) is \(u_i^T\) and the j’th column of \(\hat{V}\) is \(v_j\). Then, \( \hat{R} = \hat{U} \cdot \hat{V}\) is a matrix where the distance between defined members of R and their corresponding entries in \(\hat{R}\) has been minimized. At the same time, estimates for every undefined entry in R are present in \(\hat{R}\). A user i can then be recommended items where \(r_{ij}\) is undefined (the user has not yet rated the item) but \(\hat{r}_{ij}\) is high (the user is predicted to rate the item highly).
This approach is particularly amenable to the recommendation task, as it makes the optimization far more tractable. In particular, rather than grappling with the O(mn) user–item ratings directly, the \(O(k(m+n))\) values in \(\hat{U}\) and \(\hat{V}\) are all that need to be optimized. This offers considerable performance improvements when \(k<< min(m, n)\) (in many applications there may be millions of users and items, but \(10 \le k \le 100\) factors are sufficient for good modeling of the system [25]).
Koren et al. [25] describe the intuition behind this procedure in illuminating way. For the task of recommending movies, we can imagine that each movie can be measured on k dimensions. Each user will have some level of preference for various dimensions of a movie. Rather than explicitly defining these k dimensions and laboriously categorizing each movie in this way, the SVD recommendation procedure infers factors directly from rating patterns. These so-called latent factors are essentially learned via error minimization over available data regarding users and movies.
TrustMF is used to test the accuracy of our trust link prediction algorithm in Sect. 3. This system can be roughly characterized as combining the optimization described above with an optimization over a matrix of user–user trust links that shares a latent space with the user–item matrix. Conceptually, a user’s preference for items shares a space with that user’s preferences for trusting other users. Thus, the presence of trust links exerts an influence over the latent factors that are discovered, incorporating social trust into the recommendation process. The ability to incorporate social trust data makes this recommender system “trust-aware.”

3 Personalized multi-faceted trust modeling

In this section, we describe an experiment which explores the influence of personalization and context on a multi-faceted trust model. We aim to demonstrate the benefit of learning trust formulation behaviors at the level of clusters of users, rather than on the entire population of agents. We argue that this increase in resolution constitutes a form of personalization (albeit, performed at a group level rather than at an individual level). In addition, we explore the impact of considering differing contexts of trust by testing the effect of predicting two types of trust links.
At the heart of our solution is an effort to predict novel trust links in a social network by using machine learning methods to determine how to weight feature importance, and to approximate trust formulation procedures among groups of similar agents. Our approach makes use of the flexibility of MFTM, which we demonstrate by combining features drawn from two existing proposals with our own novel features. Evaluation is performed by measuring the error rates on a recommendation task that incorporates trust information. This is performed on a data set collected from Yelp4, a content rating site with social network features.
On Yelp, users can indicate binary social trust toward other users (friends) and submit ratings for products, businesses, or websites (taken together, and following the trend in recommender systems literature, these entities are called “items”) that they have experienced, indicating their satisfaction with that item. These ratings are integers in the range [1, 5], illustrated as stars, where higher numbers indicate a stronger recommendation. An example 5-star review for Schwartz’s Deli is presented in Fig. 1. We used this data set particularly because it is amenable to validation of trust model effectiveness via a downstream item recommendation task. Yelp was in fact used by [29], one of the central multi-faceted trust modeling papers which motivated our work.
As will be seen in our experimentation below, trust links for the Yelp environment will be predicted both on the basis of friendship relations and through the discovery of similar rating behavior.

3.1 Personalization

The rationale for testing the effect of personalization is simple: we expect trust formulation procedures to vary from person to person, therefore learning approximations of trust formulation procedures may be more accurate on a more personalized scale.
The inherent subjectivity of trust has important implication from a machine learning perspective. In particular, it implies that trust predictors trained on large data sets representing the behavior of many individuals are not necessarily more correct than those trained on smaller groups. While the predictor trained in the former case will likely have a higher accuracy across the broad population, it is essentially learning the “average” trust formulation procedure, potentially disadvantaging agent’s whose preferences are not aligned with the population at large.
We note that learning distinct predictors based on the data associated with clusters of users was suggested by Fang et al. [9], but was not pursued. While our approach is useful for capturing the variance of trust formulation in smaller groups, it does not attempt to learn the preferences of individual agents. We will discuss possible avenues for truly individual personalization of trust modeling and other approaches in Sect. 5.2.1.
Table 1
Experiment descriptions
Experiment name
Experiment description
RealFriends
Perform no prediction of trust links whatsoever. Use the real, explicit trust/friend links in the data set.
FriendPrediction
Predict trust/friendship links with no personalization step (i.e., learn one trust predictor for the entire population of agents).
PrefpRedict
Predict positive review score correlation with no personalization.
PrefCluster-PrefPredict
Determine clusters of agents with similar preferences for items (i.e., positive item review score correlation) and predict positive review score correlation links for each cluster.
PrefCluster-Friendpredict
Determine clusters of agents with similar preferences for items (i.e., positive item review score correlation) and predict trust/friendship links for each cluster.
SocialCluster-PrefPredict
Determine clusters of agents with high overlaps in their social circles and predict positive review score correlation links for each cluster.
SocialCluster-FriendPredict
Determine clusters of agents with high overlaps in their social circles and predict trust/friendship links for each cluster.

3.2 Methodology

In this work, we test the effect of personalization or trust link prediction on an item recommendation task. This is the case where agent \(a_i\) encourages agent \(a_j\) to invest resources into accessing or consuming item k based on their own experience with it. In particular, we test whether clustering agents and learning trust link predictors on the basis of clusters of similar agents, as opposed to learning a single trust link predictor for the entire population of agents, can increase the accuracy of trust-aware item recommendation systems. We also test the effect of altering the type of trust link prediction by either (a) attempting to predict the presence of an explicit friend/trust link or (b) predicting positive correlation in item review scores (i.e., two agents having scores that are the same).5
Our final analysis will report the recommendation accuracy of 7 configurations, where each configuration uses an identical set of agents and recommendation procedures, but a different procedure for predicting the trust links between agents. Each configuration is given a name reflecting which (if any) type of clustering was performed, and which type of trust link was predicted. The complete list is presented in Table 1.
The entire procedure can be described sequentially, as follows. Each step will be briefly explained, noting its inputs and outputs, then will be more carefully considered in subsections below.
When the first step is skipped, no personalization is performed. When the second step is skipped, no trust link prediction is performed. The rationale of skipping certain steps is to compare and contrast the effect applying these steps has on the final accuracy of the recommendation task.
The overview presented below assists readers in clearly understanding the interplay between the central processes of our approach: clustering, trust link prediction, and recommendation evaluation, before the details of our methodology are provided.
  • Clustering
    • Input All agents A and an agent–agent similarity matrix, S.
    • Output An assignment of every agent to a cluster, C
    • Description Partition the agents into groups of highly similar agents. We used social circle overlap (Jaccard Similarity) and review score correlation (Pearson Correlation Coefficient) as similarity measures. We developed two clustering methods for this step.
  • Trust Link Prediction
    • Input Clusters of agents, C, and trust indicator function \(\Psi (a,b)\).
    • Output A matrix of trust link predictions, \(\hat{\Gamma }\)
    • Description For each cluster \(c_l\) of agents a logistic regression learns a distinct MFTM trust prediction function for that cluster. We experimented with predicting friendship links and positive review score correlation. Output a \(|A| \times |A|\) matrix, \(\hat{\Gamma }\), where \(\hat{\Gamma }_{ij} = 1\) if the classifier for the i’th agent’s cluster predicts a trust link between agents i and j and 0 otherwise.
  • Recommendation Evaluation
    • Input Agent-item rating matrix R, trust link prediction matrix, \(\hat{\Gamma }\).
    • Output A agent-item matrix of predicted review scores, \(\hat{R}\).
    • Description Given reviews present in the original data set and the predictions from the previous step train a trust-aware recommender system to predict review scores. After training, we evaluate the correctness of the recommender on a reserved testing set using mean absolute error (MAE) and root mean squared error (RMSE) metrics.

3.2.1 Clustering

We note that truly individual personalization is difficult to test. This is because explicit elicitation of factors which influence trust on a personal level is rare on most services, and most agents have not participated in enough activity in order to accurately measure the patterns of their preferences implicit in their behavior. Therefore, we focus on clusters of similar agents rather than considering each agent distinctly. We posit that if personalization at this level of granularity is sufficient to increase the accuracy of our trust models, then we will have found evidence that some level of personalization is indeed useful for the trust modeling task, and will have motivated further research in the area.
Clustering procedures generally rely on the definition of a distance or closeness (alternatively, similarity) metric between elements to be clustered [37]. In this work, we tested two separate similarity functions. Specifically, we tested clustering agents on the basis of preference similarity and social circle similarity, as defined in Equations 7 and 8, respectively:
$$\begin{aligned}&prefSim'(a_i,a_j) \nonumber \\&\quad =\frac{\sum _{k \in R_{i,j}} (r_{ik} - \bar{r}_{k}) (r_{jk} - \bar{r}_{k}) }{\sqrt{\sum _{k \in R_{i,j}} (r_{ik} - \bar{r}_{k})^2}\sqrt{\sum _{k \in R_{i,j}} (r_{jk} - \bar{r}_{k})^2}} \end{aligned}$$
(7)
$$\begin{aligned}&socialSim'(a_i, a_j) = \frac{|friends(a_i) \cap friends(a_j)|}{|friends(a_i) \cup friends(a_j)|} \end{aligned}$$
(8)
where \(R_{i,j}\) is the set of items which both agents \(a_i\) and \(a_j\) have reviewed, \(r_{ik}\) is the rating given by agent \(a_i\) to item k, \(\bar{r}_{k}\) is the average rating for item k, and \(friends(a_i)\) is the set of agents \(a_i\) has entered into mutual friendship with. Put otherwise, we clustered agents on the basis of the Pearson Correlation Coefficient of scores they had given in reviews to items and on the basis of the Jaccard Similarity of their friend groups.
The choice of both metrics was motivated by a desire to extract metrics from our data which:
1.
Are relatively generic (i.e., could likely be applied to similar data sets).
 
2.
Could plausibly be argued to constitute a basis for determining which agents are similar enough that we might expect their trust formulation procedures to also be similar.
 
Since our context of trust is based on recommending items, we argue that both criteria are met. For 1., we argue it is reasonable to assume that on any online service with an item review and recommendation component, it will be possible to calculate Equation 7. Similarly, it is reasonable to assume that Equation 8 will often be computable on these services, as it is widely believed that friend relationships are a useful tool for expressing preference alignment among agents in such domains [18]. For 2., we argue that \(socialSim'\) directly satisfies this criteria by its definition, as \(socialSim'\) measures the observed similarity in the output of a trust-like relationship formation procedure (friendship). For \(prefSim'\), we argue that if two agents a and b have demonstrated a strong preference for similar items, then it is reasonable to conclude that their procedures for choosing who to trust for new recommendations should be similar. Thus, it is reasonable to cluster them under this context.
While we have argued that these similarity metrics are relevant for our goals, they do present challenges as metrics for clustering algorithms. Specifically:
  • Both metrics violate the triangle inequality.
  • Both metrics can sometimes be undefined (when the denominator is 0).
As many clustering algorithms are defined over Euclidean spaces, these caveats represent significant restrictions of possible approaches. It is somewhat helpful that the second caveat can be addressed by simply substituting default values in the case where division by zero would occur. Accordingly, we used the following metrics in our final procedure:
$$\begin{aligned} prefSim(a_i,a_j)= & {} \ 1 \ \text {if} \nonumber \\&\ |R_{i,j}| < 4 \nonumber \\&\ \text {or} \sum _{k \in R_{i,j}} (r_{i, k} - \bar{r}_{k})^2 = 0 \nonumber \\&\ \text {or} \sum _{k \in R_{i,j}} (r_{j, k} - \bar{r}_{k})^2 = 0; \nonumber \\&1 + prefSim'(a_i,a_j) \ \text {otherwise} \nonumber \\ \end{aligned}$$
(9)
$$\begin{aligned} socialSim(a_i,a_j)= & {} \ 0 \ \text {if} \ |friends(a_i) \cup friends(a_j)| = 0; \nonumber \\&socialSim'(a_i,a_j) \ \text {otherwise} \end{aligned}$$
(10)
Note that \(0 \le prefSim(a_i, a_j) \le 2\), where values below 1 indicate a negative correlation. Therefore, the most appropriate default value is 1. Similarly, \(0 \le socialSim(a_i, a_j) \le 1\), where values near 0 indicate very few common friends between \(a_i\) and \(a_j\), thus, the most appropriate default value when neither agent has any friends is 0. In addition, in Equation 9 we have also substituted a default value when \(|R_{i,j}| < 4\). This is because correlation tests produce noisy results with small data sets, making it prudent to choose a cutoff point under which no correlation metrics are considered. Meanwhile, if this cutoff is too high, then potentially useful data is ignored to avoid error. We chose the cutoff at 4 arbitrarily.
While this at least leaves the similarity functions well defined, it also creates a situation where the vast number of pairs of agents have a default distance between them, as any two randomly picked agents in a large enough environment will be unlikely to have any interaction history. This is a potential issue as it may causes clusters to appear significantly less cohesive than they actually are, e.g., in the case where agents a and b have “default” distance between them, but are both close to agent c. In this case, a and b should likely be in the same cluster as c, even if they don’t themselves appear to share any relationship.
In addition to the challenges described above, our clustering task had the additional goal of finding relatively large clusters. This is because our “downstream” goal was to learn personalized classifiers for each cluster of agents. If clusters are too small, then the accuracy of classifiers will suffer.
Given these goals and constraints, our first attempt at a clustering was a simple, non-iterative greedy algorithm, shown in Algorithm 1. This algorithm takes as input the set of agents to be clustered, A, the similarity matrix between agents S (where \(S_{i,j} = sim(a_i, a_j)\) for some similarity function) and the desired size of clusters \(\eta \).
In the above, freeAgents(AC) returns the set of agents not yet assigned to a cluster in C (unassigned agents), pickCentroid(ACS) returns the unassigned agent with the greatest mean similarity to all other agents, and pickNext(cACS) returns the unassigned agent with the greatest mean similarity to the agents in c.
Roughly, Algorithm 1 partitions the set of agents into at least \(\lfloor |A| / \eta \rfloor \) clusters of size \(\eta \). It does this by picking the most central unassigned agent as the core of a new cluster \(c_i\), then adding agents to that cluster in order of greatest mean similarity to agents already in cluster \(c_i\) until \(|c_i| = \eta \). The process repeats for \(c_{i+1}\), except only agents not already assigned to a cluster are considered. This continues until less than \(\eta \) agents remain unassigned, at which point all unassigned agents are added to a final cluster of unspecified size.
Clearly this algorithm is quite simple, but it is appropriate for the constraints outlined above. Firstly, all clusters of agents except for one will have a guaranteed minimum size \(\eta \), allowing control over the minimum training data size for the downstream prediction task. More importantly, it handles the non-Euclidean nature of the data by using the mean distance of all points in a cluster as a similarity metric, rather than a geometric center6.
We improved this algorithm by transforming it into an iterative version listed below (Algorithm 2).
In the above, greedilyPartition(ACS) assigns each agent to a cluster using the procedure outlined in Algorithm 1. computeClusterSims(ACS) computes a new similarity matrix, \(S'\), between agents and clusters, where \(S'_{i,j}\) is the average similarity between agent i and all agents in the j’th cluster (other than themselves):
$$\begin{aligned} S'_{i,j} = \frac{1}{|c_j| - \mathbbm {1}(a_i \in c_j)}\sum _{a_k \in c_j} \mathbbm {1}(i \ne k)S_{k, i} \end{aligned}$$
where \(\mathbbm {1}(cond)\) is the function which is equal to 1 when cond is true and 0 otherwise. \(assignToNearestCluster(a_i, S', C)\) computes a modification of the current set of clusters C by moving agent \(a_i\) to the cluster \(c_j\) that maximizes \(S'_{i, j}\), that is, the cluster for which they have the highest average similarity with other cluster members, (with ties broken randomly). This process repeats for a predetermined maximum number of iterations m.
This process is much closer to classic k-means clustering, again with the modification that distances between clusters and points must be calculated on the basis of mean distances rather than distances to the cluster’s geometric center. In addition, rather than picking random points to serve as initial cluster centers, the initial clusters are determined by a greedy partitioning method. These modifications result in an algorithm that, in our experiments, tended to produce relatively large and cohesive clusters.
Both Algorithms 1 and 2 require a parameter used to control the number of clusters (\(\eta \) and k, respectively). When performing our experiments, we determined values for these parameters by running the clustering step multiple times over a range of parameters with a relatively low maximum iteration setting, then choosing the best performing parameter to proceed with.
Data set selection and filtering: We pause to provide further details on the data set used within our experimentation, so that our continued description of the clustering process can be deepened by discussing it within this context.
Yelp is an ideal candidate to explore relationships between agents against a validation of recommended content. It is a product review site and social network of crowd-sourced reviews targeting brick-and-mortar businesses. In addition to writing reviews of services, users of the site can form mutual friendships and follow other users in order to receive the recommendations of these trusted users first. Data describing users, reviews, and businesses are made public by Yelp on a regular basis7. The full data set from 2019 contained descriptions of 1,637,138 users, 192,609 businesses, and 6,685,900 reviews.
We filtered the data set both to reduce the massive amount of data and to narrow down the context of trust in focus. Specifically, we only considered users who had reviewed at least 20 businesses that were tagged as restaurants. This narrows the context of trust from “recommending businesses or services” to “recommending restaurants” and reduces the data set down to 30,721 users and 4,432,064 reviews concerning 74,560 businesses. This filtering procedure was inspired by [29].
Table 2
Yelp filtered data statistics
 
Mean
Median
Mode
Min
Max
Friends per user
153.23
45
1
1
9564
Review per user
49.66
33
20
20
3159
Average user rating
3.74
3.77
4
1.33
5
Reviews per item
59.33
17
3
3
8349
Global review scores
3.72
4
4
1
5
Statistics from the filtered data set are presented in Table 2 and in the histograms in Figs. 2 and 3. As can be seen, there is a relatively well spread out distribution of scores given to items8, centered around 4 stars. This leads to a relatively difficult prediction task, as predicting the median review score is only correct in 35% of cases. Counts of friends and reviews are plotted on a logarithmic scale, and show a “long tail” distribution that is common in online phenomenon [40].
Clustering methods used with the data set Against the backdrop of the Yelp data set, in Figs. 4 and 5 we illustrate the performance of clustering techniques as the number of clusters (k) is altered. We measure cluster cohesiveness using two metrics, mean intra-cluster distance, and silhouette score. In the below, dist(ij) is the appropriate distance measure for the similarity function chosen, i.e., if sim(ij) is high when i and j are similar, then dist(ij) is low when i and j are similar. Mean intra-cluster distance is defined as follows:
$$\begin{aligned} meanintra(C) = \frac{1}{|C|}\sum _{c_i \in C}\sum _{j \in c_i}\sum _{k \in c_i} \frac{dist(i, j)}{|c_i|}. \end{aligned}$$
(11)
That is, the average distance between all elements in a cluster and the other elements in that cluster, averaged over all clusters.
Silhouette score s(j) for a single clustered point j which has been assigned to cluster \(c_i\) is defined as follows:
$$\begin{aligned} a(j)&= \frac{1}{|c_i| - 1} \sum _{k \in c_i, j \ne k} dist(j, k) \end{aligned}$$
(12)
$$\begin{aligned} b(j)&= \min _{\ell \ne i} \frac{1}{c_\ell } \sum _{k \in c_\ell } dist(j, k) \end{aligned}$$
(13)
$$\begin{aligned} s(j)&= {\left\{ \begin{array}{ll} \frac{b(j) - a(j)}{max(a(j), b(j))} &{} if |c_i| > 1 \\ 0 &{} otherwise \end{array}\right. } \end{aligned}$$
(14)
That is, a(j) is the average distance point j has to other points in its cluster. b(j) is the minimum distance from j to any other point not in the same cluster as j. s(j) is the silhouette score for the point j. When a point exists which is not in the same cluster as j but is closer to j than the average point in j’s cluster, then the score is negative. When the closest point to j outside of its cluster is farther away than the average distance of point in j’s cluster, then the score is positive. The silhouette score for a set of clusters C is calculated by taking the average of a random sample of points from different clusters.
Both metrics capture a sense of the cohesiveness of a set of clusters, and can be used to judge the relative merits of different clustering schemes and parameter settings. While both metrics are interesting, there are a few caveats to consider. First, for this data set and clustering algorithm, it is expected that intra-cluster distance will decrease as cluster count increases. This is shown to be true in the results reported. Thus, this metric is better suited for showing the difference in performance between clustering methods than it is for choosing a value of k. The silhouette metric does a better job of outlining the tradeoff for k values, as it will punish a method for assigning two close points to separate clusters. Therefore, a value of k that maximizes the silhouette score over a range is a more appropriate guide to choosing a value of k.
Noticeably, performance on both scores is low in an absolute sense. Silhouette score ranges from \([-1, 1]\), where positive scores are good and negative bad. Our clustering algorithms achieve scores in the range of \([-0.06,\) 0.02]—a tiny portion of the possible range near 0. Similarly, the scores for intra-cluster dist should range from [0, 2], where small scores are good, and our algorithms find scores in the [0.95, 0.99] range. Why is this the case? The answer is related to the sparsity of defined links between agents when all \(|A|^2\) possible pairs of agents are considered. As most agents do not know most other agents, and there is no basis for determining the similarity (distance) between them, in the vast majority of cases \(sim(a_i, a_j)\) is equal to a default value for randomly picked i and j. Therefore, as both metrics take some kind of average of the distances between pairs of agents, the metrics will always be close to the default distance.
Should these low absolute scores deter us from this method? We argue they should not. First, as we have briefly argued above, the nature of this data implies that average measures of cluster cohesiveness will always be close to a default. Secondly, the trend lines show that appropriate choice of k and cluster methodology can affect the sign and magnitude of silhouette scores in consistent ways—for example, when k goes above 60 in Fig. 4b. We take this as an indication that positive results are not simply a coincidence.
Table 3
Trust indicators used for Yelp data
Name
Description
Source
Benevolence
Equation 9, the similarity in rating behavior between truster and trustee.
Fang
Integrity
How similar the trustee’s ratings are to the global average.
Fang
Competence
How often the trustee’s ratings are within an acceptable range of other agents’ ratings
Fang
Predictability
How consistently the trustee’s ratings are more/less positive than the truster’s
Fang
SocialJacc
\(rel_{ab}\), Equation 10, the Jaccard similarity in the truster and trustees friend sets
Mauro
EliteYears
\(elite_a\), the number of elite years the trustee has
Mauro
ProfileUp
\(lup_a\), the number of compliments on the trustee’s profile
Mauro
Fans
\(opLeader_a\), the number of fans the trustee has
Mauro
Visibility
\(vis_a\), the ratio of compliments received to amount of content produced by the trustee
Mauro
GlobalFeedback
\(fb_a\), the number of compliments the trustee’s content has received
Mauro
EliteNorm
EliteYears divided by trustee account age in years
 
ProfileNorm
ProfileUp divided by trustee account age in years
 
FansNorm
Fans divided by trustee account age in years
 
FeedbackNorm
GlobalFeedback divided by trustee account age in years
 
ItemJacc
Jaccard similarity relative to items reviewed.
 
CategoryJacc
Jaccard similarity relative to categories of items reviewed
 
AreFriends
Are truster and trustee friends
 
AreFoF
Are truster and trustee friends of friends
 
Our trust link prediction procedure was intended to combine what we perceived to be the best traits of the work of Fang et al. [9] and Mauro et al. [29]. Both works tested the effects of predicting trust links using multi-faceted trust modeling (MFTM) on an item recommendation task.
In [29], a relatively large number of domain-specific trust indicators are proposed for the Yelp data set; however, the importance of each trust indicator is not learned in a data-driven way. Instead, they selectively enable and disable indicators for each performance test, combining their values by simply taking the average of the enabled indicators. In [9], a relatively small number of generic trust indicators are proposed for an Epinions data set. The importance of each indicator is learned via logistic regression and performance is tested under a number of different sparsity conditions9.
We will compare our work more closely to the works of Fang et al. and Mauro et al. in Sect. 4. In our work, we combined indicators proposed in both work and adopted a data-driven indicator importance weighting procedure.
Trust indicator list To quickly restate the goals of MFTM10, we wish to define a vector of trust indicators over every pair of agents, \(\Psi (\overset{\rightharpoonup }{a_i}, \overset{\leftharpoonup }{a_j})\) then use machine learning to approximate the function \(f: \Psi (\overset{\rightharpoonup }{a_i},\overset{\leftharpoonup }{a_j}) \rightarrow y\), where y is some type of trust link.
In Table 3, we have listed all trust indicators that we calculated for Yelp data. When an indicator was proposed in the works of either Fang et al. [9] or Mauro et al. [29], we have indicated this in the last column (although some were also adjusted by us, as clarified below). Some indicators are defined specifically for pairs of agents (e.g., the similarity of rating behavior for two agents), while others are defined on the basis of a single agent. In Yelp data, the only explicit trust link present is mutual friendship.
Here, we describe some of these indicators in full detail, in order to illustrate some less obvious indicators. Complete descriptions of the indicators not described here are available in [9] and [29].
Benevolence: Already described in Equation 9. In [9], \(\bar{r}_j\), the average rating given to item j was replaced with \(\bar{r_i}\), the average score agent i gave to items. We made this replacement as a common rating behavior in the Yelp data set was for an agent to only submit 5-star reviews, causing frequent divisions by zero. By comparing to the global average rating of an item, this behavior is no longer an issue. Intuitively, when \(benevolence(a_i,a_j)\) is high, agents \(a_i\) and \(a_j\) may be inclined to trust each other’s reviews, as they have reviewed items similarly in the past.
Competence: A threshold value \(\epsilon \) is used to determine how often the trustee’s ratings where “close enough” to the ratings of other agents who had also rated those items to be considered “correct.”
$$\begin{aligned} competence(a_i) = \frac{\sum _{j \in R_{i}}\sum _{k \in I_j}\mathbbm {1}(|r_{ij} - r_{kj}| < \epsilon )}{\sum _{j \in R_{i}} |I_j|} \end{aligned}$$
(15)
where \(R_i\) is all the items the i’th agent has rated and \(I_j\) is the set of all agents who have rated item j and \(r_{ij}\) is the rating agent i gave to item j. Competence is high when an agent’s rating behavior is similar to the plurality of agents. Since ratings on Yelp use a 5-star scale, we used the threshold value 0.5. Intuitively, when \(competence(a_j)\) is high, \(a_j\) may be trustworthy for agents who consider agreement with popular consensus to be important.
Predictability: A threshold value \(\theta \) is used to determine how often a trustee’s preferences are consistently higher, lower, or similar to the truster’s :
$$\begin{aligned}&n_u = \sum _{k \in R_{i, j}} \mathbbm {1}(|r_{ik} - r_{jk}| \le \theta ) \end{aligned}$$
(16)
$$\begin{aligned}&n_n = \sum _{k \in R_{i, j}} \mathbbm {1}(r_{ik} - r_{jk} < \theta ) \end{aligned}$$
(17)
$$\begin{aligned}&n_p = \sum _{k \in R_{i, j}} \mathbbm {1}(r_{ik} - r_{jk} < -\theta ) \end{aligned}$$
(18)
$$\begin{aligned}&predictability(a_i, a_j) = \frac{max(n_u, n_n, n_p)-min(n_u,n_p,n_n)}{|R_{i,j}|} \end{aligned}$$
(19)
where \(n_u\), \(n_n\), and \(n_p\) count how many times the trustee rated an item about the same as the truster, lower than the truster, and higher than the truster, respectively. Accordingly, predictability is lowest when \(n_u = n_n = n_p\), meaning the trustee rates items better, worse, and equivalent to the truster in equal amounts. This would mean there isn’t a justification to expect that the trustee has a bias in any particular direction, relative to the truster. Similar to Competence, we used a threshold value of 0.5. Intuitively, \(predictability(a_i, a_j)\) may be important to \(a_i\) deciding whether or not to trust \(a_j\), as it is useful to know whether \(a_j\)’s ratings have a consistent bias compared to \(a_i\).
Visibility: The relative popularity of agent, taking into consideration how much content the agent has produced and the popularity of the most popular agent.
$$\begin{aligned} visibility(a_i) = \frac{appr(i)}{max_{a_j \in A}(appr(j) \times contr(i))} \end{aligned}$$
(20)
where appr(i) is the number of public “appreciations” an agent has received from other agents (e.g., likes) and contr(i) is the number of contributions an agent has made (e.g., posts, reviews). Intuitively, when \(visibility(a_j)\) is high, \(a_j\) may be trustworthy to agents who consider consistent popularity important.
Some of the indicators listed in Table 3 were developed by us. For example, we normalized a number of the indicators proposed by [29] by dividing by how many years the target user had been on the site. This is useful for giving newer users a chance to compete with older users on certain attributes (e.g., how many “fans” they’ve accrued). We also computed the Jaccard similarity between users with respect to the sets of items they had reviewed and the categories of items they had reviewed, reasoning that these indicators would help to outline the case when users have similar areas of interest. Finally, we checked to see if pairs of users were friends of friends, a potentially useful feature for integrating trust transitivity into reasoning.
When computing these trust indicators, it appears necessary to consider all ordered pairs of agents in the environment, as trust is directed and can occur between any two agents. This presents a significant computational bound on the number of agents that can be considered. Discussion of this issue, and our approach to minimizing this impact is presented in Appendix A. In brief, only pairs of agents where there is significant evidence that the pair have an overlap in interests / social circle are actually considered as candidates for novel trust link prediction.
Classification process The trust indicators listed in Table 3 were used to predict two types of trust links (1) whether the truster had explicitly expressed trust in the trustee (friendship), and (2) whether the truster and trustee had a positive correlation in review scores. Note, in the case where expressed trust was the target of prediction, review score correlation was considered as evidence (e.g., included in \(\Psi (a_i, a_j)\)) and vice versa, although the target of prediction was obviously not considered as evidence.
Following the example set by Fang et al [9], we use logistic regression to learn functions that predict the presence of statistically likely trust links based on the vector of trust indicators computed between pairs of agents. We refer to these functions as “trust link classifiers,” as once learned, they classify each ordered pair of agents as either being linked by trust (the former should trust the latter) or not. We used the SAGA solver logistic regression classifier included in the sklearn Python package [34] to learn these functions.
In the case where no clustering was performed, a single classifier was learned for all agents. When clustering was performed, a classifier was learned for each cluster of agents. Thus, each cluster-specific classifier learns how the agents in the cluster form trust links in their role as trusters. This makes obvious a substantial tradeoff to this approach to personalization: the more clusters are found (increasing cluster cohesiveness up to a point), the less data available to train machine learning classifiers (decreasing prediction accuracy). We will discuss other potential approaches to personalization in Sect. 4. For our purposes, we only learned cluster-specific classifiers for clusters that had at least 100 agents and at least 1000 positive outgoing trust links. When a cluster failed to meet these standards, it was assigned a generic classifier, trained on examples from a random sample of users across all clusters. We implemented this strategy in order to avoid training wildly inaccurate classifiers. When training the cluster-specific classifiers, all available data relevant to each cluster was used for training. This is because we are not directly interested in how well each classifier is able to fit each cluster, only on whether this personalization process increases the accuracy of the downstream recommendation task. Therefore, it is not necessary to reserve a test/validation set for any trust link classifier. Note, at this point, test sets of ratings data are already reserved for the recommendation task outlined in the next section.
A common problem in link prediction generally is the large class imbalance between positive and negative examples. Put simply, the number of negative examples of trust links in a community of agents grows with \(O(|A|^2)\), while positive examples have much more conservative linear growth. This can be either because humans have an upper limit on how many others they will trust, or, like on Yelp, technical limitations are imposed on the number of allowed friends. Compounding the problem is that there are two kinds of negative examples, which are often difficult to distinguish between. On the one hand, agents \(a_i\) and \(a_j\) may not be friends simply because they have never met. On the other hand, they may have interacted and prefer not to do so again in the future.
Therefore, it is necessary to devise a strategy for training classifiers to deal with this imbalance and ambiguity. One popular method, which we have used here, is to construct balanced training sets by including a random negative link for every positive link. This method has the advantage of requiring no further tinkering to classifiers in order to accommodate a class imbalance. The ambiguity in negative links is ignored as best as possible by simply sampling negative links randomly.
After training classifiers for each cluster, trust link prediction is performed by feeding the trust indicator vector \(\Psi (a_i, a_j)\) to the appropriate classifier for the cluster of agent \(a_i\). Ultimately a matrix of trust link predictions \(\hat{\Gamma }\) is produced, where \(\hat{\Gamma }_{ij} = 1\) if the classifier for the i’th agent’s cluster predicts a trust link from \(a_i\) to \(a_j\) with probability greater than 0.5. It is noted once again that predictions were only made for pairs of agents that were considered to be in the same neighborhood, as described in Appendix A.

3.2.3 Recommender evaluation

Two trust-aware recommender systems were used to measure the accuracy of the trust links predicted in the previous step: TrustMF and MTR. Two systems were evaluated in order to reduce the risk (e.g., of using a flawed implementation that skews results). TrustMF leverages matrix factorization and gradient descent to optimize predictions of user–item ratings. We used the Librec implementation of TrustMF for our experiments [16]. MTR is a trust-aware modification of a similarity-based KNN recommendation model proposed by Mauro et al. [29]. Under this system, the predicted rating for an agent i for an item j is:
$$\begin{aligned} \hat{r}_{ij} = \bar{r}_i + \frac{\sum _{k \in N^\kappa _j(i)} inf_{ki}(r_{kj} - \bar{r}_k)}{\sum _{k \in N^\kappa _j(i)} |inf_{ki}|} \end{aligned}$$
(21)
where \(\bar{r}_i\) is the mean score agent i has given in ratings, \(N^\kappa _j(i)\) is the set of the top \(\kappa \) most influential agents on i who have also rated item j, and \(inf_{ki}\) is the influence agent k’s recommendation exerts on agent i: a linear combination of the similarity between k and i’s past rating behavior and a trust metric. In our case, \(inf_{ki}\) is the probability that k is trustworthy for i according to the predictions of the trust model in the previous step
$$\begin{aligned} inf_{ki} = \beta \cdot \sigma (ik) + (1-\beta ) \cdot \hat{\Gamma }_{ik} \end{aligned}$$
(22)
where \(\beta \) is simply a parameter for controlling the weight of trust modeling on the recommendation process. When \(\hat{\Gamma }_{ik}\) was undefined (e.g., in the case where i and k are not in the same neighborhood, see Appendix A), a value of 0 was substituted. We modified an implementation of a KNN-based recommender system distributed in the Surprise library [20] to test this method.
Accuracy of recommendation was measured by mean absolute error (MAE) and root mean squared error (RMSE):
$$\begin{aligned} MAE(\hat{R})&= \frac{1}{|\hat{R}|} \sum _{r_{ij} \in \hat{R}} |\hat{r}_{ij} - r_{ij}| \end{aligned}$$
(23)
$$\begin{aligned} RMSE(\hat{R})&= \sqrt{\frac{1}{|\hat{R}|} \sum _{r_{ij} \in \hat{R}} (\hat{r}_{ij} - r_{ij})^2} \end{aligned}$$
(24)
where R and \(\hat{R}\) are a set of real agent-item ratings and predicted agent-item ratings, respectively, and \(r_{ij}\) is the rating given by user i to item j. MAE simply captures the average unsigned error in predictions across all ratings, while RMSE is more heavily penalized for gratuitously erroneous ratings and less penalized for very nearly correct ratings. These measures can be analogized to the mean and variance of a distribution over prediction error. As measures of error, we prefer recommendations that minimize these measures. Thus for all of the following graphs lower values are better. The sensitivity of RMSE to outliers is a useful property for this application, as grossly inaccurate predictions can erode user trust in future recommendations.
The latent factors model adopted by TrustMF was outlined already in Section 2.3.1. TrustMF has the following significant hyperparameters: the regulation penalty \(\lambda \), the weight given to fitting the user–user trust matrix (as opposed to the user–item rating matrix), \(\lambda _t\), and the number of dimensions of the latent space d. We kept the number of dimensions at the default of 10 and the regulation penalty at 0.01. In order to determine an appropriate setting for \(\lambda _t\), we sampled 10000 users from the filtered Yelp data set and plotted MAE and RMSE over the change in \(\lambda _t\). Results of this tuning are presented in Fig. 6. For readability, we have only plotted the best performing experiments from each of the main groups11 (RealFriends as a baseline, FriendPrediction as a non-personalized (MFTM) example, and PrefCluster-PrefPredict as a personalized (PMFTM) example). Each data point is the average of three runs with different random seeds, and an iteration limit of 200 epochs. For these preliminary tests, we set the number of clusters at 10.
Recommendation accuracy changes little for the actual trust links in the data set (RealFriends) as the importance of trust for recommendation increases, but the MFTM and PMFTM lines form roughly convex curves, reaching a range of minimal values around \(0.8< \lambda _t < 0.125\). For future experiments, we set \(\lambda _t = 0.11\).
Figure 6 also serves as some encouraging early results, showing both that the impact of predicting trust links reduces recommendation error and a personalized approach can reduce this error further, given the correct weighting of trust importance.
MTR has two significant hyperparameters: the maximum neighborhood size for a user (e.g., the maximum number of peer recommendations that will be taken into consideration), \(\kappa \), and a social weighting parameter, the value of \(\beta \) in Equation 22. In their original work, Mauro et al. [29] set \(\beta \) at 0.1, but did not report on how modifying this variable affects recommendation accuracy12. We used all Yelp users from the filtered data set and computed the recommendation accuracy as \(\beta \) changed using a set of predictions based on a single social classifier (i.e., the FriendPrediction setup). Results are illustrated in Fig. 7, showing a clear tradeoff between only considering user–user similarity and incorporating trust. Similar results were seen for the PrefPredict experiments. Accordingly, future experiments were run with a value of \(\beta = 0.3\).
We ran experiments with \(\kappa \) set to 50, but also experimented later with modifying the value of \(\kappa \) (to simulate sparsity). As a reminder, this value is the maximum number of ratings that are considered when recommending to a user.
Test sets were created by reserving 20% of each user’s reviews. For all figures in this section, except in the results shown in Fig. 6, these reviews were excluded from every step of the process13, that is, the clustering and link prediction steps did not have access to these reviews. Due to the computation time required to generate and evaluate many of these experiments, only the results reported in Table 4 were cross-validated. In this case, fivefold cross-validation was used with respect to users, so each user had a distinct 20% of their reviews reserved as a validation set for each of the folds. This validation set was hidden from every step in the pipeline. This validation approach is similar to the ones used in [29] and [9], where results were reported based on the average across folds of a tenfold cross-validation and the average across a complete leave-one-out cross-validation, respectively.
Table 4
Recommendation error results
 
MTR
TrustMF
 
MAE
RMSE
MAE
RMSE
RealFriends
0.8610
1.1196
0.8879
1.1205
FriendPrediction
0.8453
1.0960
0.9063
1.1179
SocialCluster-FriendPredict
0.8434
1.0932
0.9120
1.1179
PrefCluster-FriendPredict
0.8436
1.0936
0.9077
1.1189
PrefPredict
0.8551
1.1105
0.8984
1.1109
SocialCluster-PrefPredict
0.8551
1.1103
0.8987
1.1111
PrefCluster-PrefPredict
0.8551
1.1103
0.8987
1.1111

3.3 Results

In our original tests, we attempted to set the number of clusters, k, by evaluating the silhouette scores of clusters in a large range for each data set, then simply choosing k based on whichever cluster count had the highest silhouette score. Unfortunately, this method was fickle, as the silhouette evaluation is based on a random sample of the clusters, and a single outlier could achieve a minimal score even if other nearby values of k were not optimal. Further, it is basically a heuristic to use cluster cohesion to choose the number of clusters, when ultimately we are interested in improving the personalized trust links. Therefore, we iterated over a range of cluster values and repeated the entire experiment with each choice of cluster score, using the MTR recommender system. Results are illustrated in Figs. 8, 9, 10, and 11. In general, results show that as the number of clusters searched for (k) increases, the error in the task follows a consistent trend of reduction. Note that when \(k=1\), the situation is equivalent to a non-personalized approach (as only searching for a single cluster is equivalent to doing no clustering), and as k increases the granularity of personalization increases. When predicting whether two users should be friends or not, MAE can be lowered by 0.003 points by adding personalization, while when predicting aligned preferences it is only lowered by 0.0005 points. These results are less impactful than the early results seen using the TrustMF classifier in Fig. 6. That said, the results indicate a consistent trend of improvement as clustering-based personalization is applied: a fairly consistent line of decrease in error can be observed in all lines.
We wished to verify that the results illustrated in these figures are indeed improving because our clustering technique was finding groups of similar users, which allowed the prediction techniques to learn more personalized classifiers for these groups. For instance, it is conceivable that splitting users into groups and learning multiple classifiers is helpful regardless of the groups picked, as this procedure would be similar to bootstrap aggregating [1], which allows simple classifiers to model multiple weak correlations in data. To test this, we repeated the FriendPredict experiment, but clustered agents into k clusters randomly14. The results illustrated in Fig. 12 compare this random clustering to clustering by social circle overlap. This figure clearly shows that the reduction in error is largely due to the non-random clustering approach. The solid and dashed lines start off identically at \(k=1\) on the x-axis (no clustering) but as the numbers of clusters increases, the error decreases, for the case where social clusters are used. We take this as evidence that the clustering technique is improving accuracy because clustering genuinely enables more personalized predictions, not simply because the number of models being learned has increased.
We also experimented with modifying the \(\kappa \) variable on MTR, effectively simulating sparsity, as this variable controls how many peer advisers can be considered for a recommendation. Results illustrated in Fig. 13 compare an accuracy on the SocialCluster-FriendPredict task, comparing the error rates for a single cluster (unclustered) and for 55 clusters (clustered). Overall, the gap in error rate is most dramatic when a larger \(\kappa \) value is used, but the advantage for a clustered approach applies over the range of values.
Finally, in Table 4 we present the results of a fivefold cross-validation over user ratings for these tasks. Best results are bolded. There are a number of interesting results. First, the conceptually simple MTR system outperforms TrustMF across the board, despite the fact that the TrustMF system was allowed to run for a much greater period of time in order to reach convergence. This gap in performance is often dramatic, for example, in the best cases for each system, MTR has a MAE that is 5% lower than TrustMF, and a RMSE that is 1.6% lower.
On the better performing MTR recommender system, the best results are achieved when predicting friendship links rather than predicting preference correlation. This makes, sense, as the MTR system already considers observable user preference (see Equation 22), thus predicting new instances of aligned preferences is not likely to add much new information. The best performing task, by a small margin, is the SocialCluster-FriendPredict task, which basically reconfirms the findings presented earlier in this section (with the added certainty of being averaged across folds). Clustering did not have an appreciable effect (at least not at the scale of \(10^{-4}\)) for the preference alignment prediction task.
Note that, in the case where an improvement was seen, although the scale of the effect appears small it is clear from the previous graphs that this is not merely due to statistical variance. For example, Fig. 8 clearly shows that the decrease in error between FriendPrediction and SocialCluster-FriendPredict is due to the increasing the number of clusters.
On TrustMF, results are more tightly grouped and there is quite little appreciable difference between experiments. As this system is more conceptually complex than MTR, it is difficult to interpret exactly why this might be the case. Interestingly, the best performing task for MTR is the worst performing task on TrustMF. Clustering does not have a positive effect in these experiments, and in the SocialCluster-FriendPredict task actually seems to harm the performance. Our early experiments with this recommender system (presented in Fig. 6) suggested that there might be more interesting differences between approaches, but this was not the case in these final results. We speculate that because these earlier results were computed using different techniques to select Yelp users (randomly selecting 10000 users versus our final strategy of selecting all 30000 users with more than 20 reviews submitted), the underlying distributions of ratings may have been different enough to cause this change.

3.4 Conclusion

In this work, we evaluated the effect that personalization via clustering had on the accuracy of a trust link prediction task. We accomplished this by predicting novel trust links on a data set of Yelp users and measuring accuracy of these predicted trust links via an item recommendation task.
Our results show that the option of predicting novel trust links results in better performance than using the explicitly stated trust links for the recommendation task. Further, our results show a small but consistent improvement in recommendation accuracy when clustering is used to determine groups of similar agents and distinct trust prediction models are learned for each group of agents with the MTR recommender system (e.g., Figs. 8, 9, 10, and 11). We showed that this improvement was not simply the result of the fact that more classifiers were trained, as randomly splitting users into groups does not improve accuracy nearly as much as the clustering technique does (Fig. 12). While our early results with TrustMF inspired confidence, and we hoped to see more dramatic improvement in recommender accuracy from these experiments, the final results show that, while consistent, the improvements in accuracy from the procedures outlined here are small. We will comment on ways these techniques could be improved, and potential avenues for future work, in Sect. 5.2.1.
In addition to the experiments with personalization (which explored multiple approaches for clustering users), we produced a comprehensive MFTM solution, combining techniques from the literature with novel features. We also make clear the applicability of MFTM to social networks. We experimented with predicting two types of trust links: explicit friendship (the FriendPredict experiments) and implicitly stated preference alignment (the PrefPredict experiments) and evaluated the utility of the predicted links derived by our methods, using two distinct trust-aware recommender systems. We found that the preferred target of trust link prediction can vary with the desired use-case: it was not clearly preferable to predict friendship links or preference alignment links. On the MTR system, which already strongly considers user preference alignment, our experiments performed better when predicting friend links, while on the TrustMF system predicting preference alignment between users produces (slightly) better recommendation accuracy.

4 Discussion

In this section, we first reflect on how our work compares to those of other researchers, with respect to both multi-faceted trust modeling and to personalizing solutions for trust modeling. We then discuss how our methods for predicting trust links using our particular personalized multi-faceted trust modeling can serve as a useful starting point for handling misinformation in social networking environments.

4.1.1 Multi-faceted trust modeling

Our work in Sect. 3 was heavily inspired by the works of Mauro et al. [29] and Fang et al. [9]. All works have a similar structure: they propose a multi-faceted trust model and test it on a recommendation task on a data set harnessed from a site with item rating component. We sought to extend these works by combining the best features from each of them while testing the effects of a personalization step to increase the accuracy of trust prediction. In particular, Mauro’s work developed a large set of trust indicators on the Yelp data set, while Fang’s work proposed a smaller set of relatively generic indicators that could be used on the Epinions data set. In our work, we combined these indicators when testing on the Yelp data set, with the goal of achieving a more comprehensive model of user to user trust formulation. While Mauro’s work proposes a large number of trust indicators, it does not seek to weight the importance of those indicators in a data-driven manner: they instead experiment by taking a non-weighted average of a subset of the indicators. Like in Fang’s work, we have used a logistic regression to find weights for these indicators that fit the data set, believing this method to be a more principled approach to the problem. In addition, we did some preliminary investigation of Epinions data in order to expand the environments examined under our approach. Conceptually, the approaches taken to personalize recommendations we undertook on the Yelp data would be easily transferable to this data set. See Sect. 5.2.1 for more details.
Another work which has relevance is that of Gilbert and Karahalios [13]. While not an artificial intelligence paper, the authors present a multi-faceted statistical analysis of the factors which affect tie strength between pairs of users in social media. They found that a set of 74 variables collected from a the Facebook accounts of participants could be used to predict, with high accuracy, the answers these participants gave to survey questions designed to model social tie strength with their friends on Facebook (e.g., “How comfortable would you feel asking this person for a loan?”). This work presents strong evidence for the notion that trust (i.e., as an aspect of a strong social tie) can be predicted between agents based on relatively simple data extracted from interaction history on social media.

4.1.2 Personalization approaches

Given the subjective nature of trust, it is clear that accurate trust models need to incorporate some level of agent-specific personalization. It is feasible to model reputation or popularity without such personalization. However, given the sparsity of data in most networks, the cold start problem15, and computational limitations, it is not typically feasible to give each agent a completely distinct model. Our approach to personalization in Sect. 3 was to determine clusters of similar users and learn trust link classifiers on the basis of these clusters.
The Personalized Trust Model developed by Zhang and Cohen [48] (described in Sect. 2.1) can be seen as an extension of the Beta Reputation System [23] that computes both a private and public trust factors for integrating the advice of some other agent. The weighting of these factors is based on the overlap in the number of common items the truster and trustee have advised on (rated), thus giving higher weight to private trust when there is more basis for comparing the two agents directly with respect to past behavior. Effectively, this system implements personalization by using generic predictions under uncertainty about individual preferences. However, their formula for assigning weight to personal and private trust factors is basically a heuristic, as the settings for appropriate error and confidence bounds are not derived in a data-driven manner. Our work attempts to implement personalization in a data-driven way, by identifying clusters of similar users and learning their trust formulation procedures at a cluster level16.
The usefulness of stereotypes toward improved trust modeling is an approach examined by other researchers who may also derive benefit from examining our data-driven methods. The StereoTrust Model developed by Liu et al. [28] implements personalization by allowing each agent to define its own grouping function for partitioning the set of other agents via stereotypes. For example, an agent may decide to stereotype based on stated interest, location, seniority, etc. This is intended to model the subjective assumptions humans apply in every day life. The agent then uses a trust estimation function (inspired by the Beta Reputation System [23]) to reason about their trust with respect to groups defined by stereotypes, rather than with respect to individuals. The trust an agent \(\overset{\rightharpoonup }{a}\) has in another agent \(\overset{\leftharpoonup }{b}\) is then computed as a weighed average of the trust \(\overset{\rightharpoonup }{a}\) has in all the groups that \(\overset{\leftharpoonup }{b}\) is a part of.
This system implements personalization by allowing each agent to specify its own stereotypes, although in practice it’s not clear how this information would be elicited from real users. This approach relies on the notion that members of a group will act similarly, but by allowing individual agents to define groups arbitrarily, the usefulness of this notion is under a certain strain. Without the ability to statistically analyze large amounts of data from the environment, it is unclear how individual agents could be expected to create stereotypes that define groups which actually have some cohesiveness of behavior. In their actual implementation, stereotypes were implemented based on rating similarity, so no agents actually had an opportunity to specify their stereotypes, and in practice this solution turns out to be a complex approach to reach the same end goal as, for instance, clustering users based on rating behavior (as we have done).

4.2 Toward improved handling of misinformation

In order to improve the lives of users of social media, we have presented an approach for predicting trust links between peers within these networks. Our framework makes it possible to assess whether the content created online is a good candidate to display to a user or not (where options may include flagging messages coming from sources that are not established to be well trusted).
Our work aims to improve online experiences by supporting distinct presentation of content to differing users, achieved by reasoning about relationships with peers and the concept of trust. Our concern with trustworthiness of content relates well to companion efforts devoted to detect digital misinformation [7, 21, 47]. There is a spectrum of possible outcomes when messages which are of questionable quality are shown to users, including special attention in contexts such as healthcare where the consequences may be more troubling [32]. Note that there will still be various options for actions to take, once trust modeling has provided some insights into messages of concern.
The methods we present here are designed to be self-contained algorithms which can be provided to any party which has the data at hand, to reason about trustworthiness. It would be possible, for instance, to have platform owners flag less trustworthy posts and individual users have agency to choose what kinds of information should be filtered for them. Our algorithms would be able to indicate, for a particular user, whether the other users in the environment are predicted to be trusted (based on whether a trust link is predicted to exist between them). There is then an interesting dilemma about whether top-down control of the social network (e.g., dictated by government) or bottom-up management of the content (e.g., under the control of individual users) should be launched in order to take actions with respect to the messages of agents with questionable trustworthiness. While our model promotes a solution that is attuned to an individual’s preferences, in cases where these may be in conflict with interests of the public (for instance, promoting hate) a tension may exist in deciding where the control should lie. If a decision is made to consult reputable outside sources to determine acceptability, this could potentially be integrated into the trust models to discourage inappropriate behavior. We do not propose an answer to this challenge of determining appropriate control but merely acknowledge this as a concern for anyone trying to address content recommendation in social media.
As our work has drawn out the value of personalized solutions, the models that we have presented should be flexible enough to support a variety of overall preferences with respect to final outcomes. With respect to the array of concerns for this Special Issue, our work is best viewed as focused on computational approaches grounded in artificial intelligence methods which assist in the detection of misinformation and disinformation. We introduce novel perspectives on this particular agenda for improving online social networks, through techniques for personalizing the analysis and with highlighting of the potential provided by performing trust modeling.

5 Conclusions and future work

5.1 Summary

In this paper, we considered the problem of improving the experience of users on social networks, particularly with respect to content overload and the propagation of untrustworthy information. We argued that a trust modeling approach could be appropriate for social networks and could be used to enhance message recommendation systems. We then outlined some of the issues involved in applying these models as they currently exist. The types of trust models that can be applied need to be highly flexible, capable of capturing many different kinds of data, and personalizable.
We argued that a multi-faceted trust model was ideal for application to social networks. This is because the multi-faceted model can incorporate arbitrarily many signals from the agents and their environment into a data-driven model of how trust is apportioned by agents in an environment. We argued that this flexibility was a key feature, as it allows the model to adapt to many different kinds of social networks. In Sect. 3, we designed a comprehensive MFTM and applied it to a large data set, including multiple new features and features proposed in previous works. We experimented with personalizing the predictions generated by a multi-faceted model, by clustering similar users and learning distinct models for each cluster of users. We argued that although this approach is not “truly individualized” personalization, a data-driven model like MFTM imposes a tradeoff between the number of users a model is learned for, as smaller numbers of users will have less data available to train classifiers with. We showed that this approach can lower error rates in a downstream trust-aware recommendation task.
In brief, our primary contributions with this work are to:
  • identify a critical challenge in applying trust models to social networks, namely to personalize trust prediction
  • develop a clustering-based approach to personalization on a large data set, applying predictions to a downstream recommendation task and showing consistent improvements in error rates
  • assemble a comprehensive array of multi-faceted trust indicators to incorporate into data-driven reasoning about trustworthiness as an advance to this method for multiagent trust modeling
  • outline the potential for our approach to trust link prediction (reasoning either about rating behavior of users or social circles) to assist in efforts to address misinformation in social networks, clarifying as well challenges which remain to be addressed
When misinformation abounds in social media, being able to judge which sources are trusted is a critical step in assisting the users in these networks to navigate the waters. In the sections below, we elaborate on ways to extend the models we have developed, and how to assist users in handling misinformation once our trust link prediction process has been run. We follow this with some suggested steps forward to assist some of the most vulnerable online users, older adults, illustrating how reasoning with clusters of users, the centerpiece of our proposed model, can be quite valuable in allowing unique experiences for this user base.

5.2 Future work

5.2.1 Expanding upon personalized multi-faceted trust modeling

There are a number of ways the project of personalizing multi-faceted trust predictions can be extended.
Clustering For example, we spent considerable time in this paper explaining the difficulties involved in clustering points that represent agents in a social network. While the approach we took was ultimately geometrically inspired, graph clustering algorithms could potentially offer a better fit to this type of data. This is an especially attractive option, as the sparsity of defined similarities between agents when considered geometrically is a major issue for applying and accurately measuring performance of geometric clustering approaches. There is also merit in examining hierarchical clustering methods, as certain types of data may fit this model well; however, parameters will need to be tuned to produce groupings with sufficient moderate sized clusters.
Two other challenges related to the clustering aspect of this work are finding new methods of determining the optimal number of clusters (k), and considering other distance functions. In this work, we ran the entire experiment from beginning to end (cluster, predict, recommend) many times in order to measure the effect of cluster count changes. Searching for new methods of determining k which are more computationally tractable than an exhaustive search, or finding heuristics that can guide this search, would be a useful and interesting research project. Second, we clustered agents in this work on the basis of social circle overlap (Jaccard similarity of trusted users) and preference similarity (Pearson Correlation Coefficient observed in train set ratings). While we’ve argued that each of these is fairly natural metrics, it would be interesting to explore new metrics, including those based on implicit preferences (e.g., browsing behavior), categories of interest (e.g., types of items enjoyed), and other biographical factors of the agents (e.g., geographic location, age). Each of these can plausibly be argued to be indicative of some facet of agent similarity, which in turn may be correlated to similarities in trust formulation procedures.
Other clustering options that we could explore include seeing whether meaningful clusters exist first using a statistic such as the Hopkins Statistic and delving further to determine the number of clusters in the data set using a metric such as the Bayesian Information Criterion [19]. This then may help to direct the choices for clustering that are used.
Our machine learning methods could also be broadened. We settled on logistic regression as the central method used but exploring further the use of other choices such as SVM [19] may help to determine whether additional robustness with performance could be achieved. Logistic regression is valuable as it admits a simple probabilistic interpretation, is quick to optimize and the weight vector learned is highly interpretable; however, there may be limitations of dealing with linearly separable data, making it more challenging to compute interesting feature combinations.
Experimental setup Our work does not consider dynamic changes in the network or agent preferences over time. For example, our method did not consider agents who had no preference data associated with them, that is, new agents joining the network. In practice, this could be handled by simply assigning generic predictions for agents who lacked sufficient preference data on which to cluster them. A periodic re-training of the models would also allow the system to account for changing preferences over time. This dynamic process of agents entering the network could be simulated for our experiments by leaving out a sample of users from the initial processing, then adding them after clusters have been created already.
Another area for possible expansion is in our use of the personalized cluster classifiers. We did not learn a classifier for a cluster when that cluster had less than 1000 positive examples of outgoing trust links and 100 agents in it. This step was taken to avoid learning very inaccurate classifiers, but some of the classifiers learned still fit the data related to the cluster significantly worse than a classifier trained on larger sample of random agents. Therefore, it is worth exploring better ways of combining the “local” (cluster specific) predictions with the “global” predictions, similar to the procedure taken in the Personalized Trust Model [48]. Perhaps the weight given to a local trust model could be based on the difference in accuracy between the fit of that model to the agents it represents and the accuracy a generic classifier would achieve for those agents. This way, local irregularities could still be learned, but in cases where data is sparse, a little help from a generic classifier can nudge predictions toward a more accurate final outcome. This approach could also be taken to enable more “truly individual” personalization. For users with a large amount of activity (thousands of friends and other users to compare preferences with), a single-user classifier could be trained, and the results of this classifier combined linearly with a cluster or global classifier, allowing truly individualized personalization, and a gradual ramp-up from generic to individual solutions as more data becomes available.
In our work, we excluded users from experimentation who had fewer than 20 reviews. This filtering procedure was inspired by Mauro et al. [29], but it imposes certain biases on the following evaluations. Under this procedure, only the opinions and activities of the most active users are taken into account (only about 2% of Yelp users have submitted at least 20 reviews). In our earlier experiments, we sampled users randomly, and the results from this time tended to show a more dramatic difference between personalized and non-personalized approaches using TrustMF (e.g., in Fig. 6). It would be valuable to experiment with different procedures for sampling users from this data set.
There is also merit in examining how our model operates in other social networking contexts. Epinions is a reasonable second case for us to explore, as it was also examined by [9]. We conducted a preliminary study of Epinions data sets and noticed that the chance of a randomly picked review score being 5 (the highest) is over 70%, while on Yelp the distribution is much more spread out, with the highest probability being only 35% on a score of 4. With this kind of bias in the data, we would expect even better score accuracy on the score prediction task when applying our methods. It is also interesting to note that Epinions users typically have fewer friends (trusted users) and that with Epinions users submitting ratings to written text (rating others’ reviews), there is vastly more feedback to examine. All of these differences may provide greater insights into the conditions under which our model has the most value. Expanding our study to other more elaborate data sets will also help to shed light on the scalability of our particular approach.
There may be additional challenges when examining other social networks. While many recent projects in trust modeling focus on data from social networks with a significant item rating component (as it is convenient to measure trust-aware recommendation accuracy on a set of reserved ratings as a proxy measure for the quality of novel predicted trust links), we acknowledge that many popular networks such as Twitter and Facebook lack a significant item rating component. In cases like these, it would likely be necessary to engage in a user study (like the one in [13]) and survey actual users whether the predicted trust links appeal to them or not. This work would be useful, especially if a data set can be publicly released, as more data where preferences are explicitly indicated by users (rather than inferred) will be a boon to future trust modeling research.

5.2.2 Integrating with efforts to address digital misinformation

Integrating our proposed approaches directly into the larger effort aimed at combating digital misinformation would be a rich area for future work. Some subtopics which would be especially valuable to explore include connecting to efforts on detecting content which has been generated by bots [10]. Our methods may be able to provide more insights into bot detection algorithms or our algorithms may be able to adjust their predictions based on information revealed to us about suspected bot nodes within the network. It would also be useful to adjust our predictions of trust links and the use of these outcomes toward addressing misinformation, in view of the networking behavior in the social media environment. Work such as that of Tong et al. [43] conducts an analysis of how rumors spread among the network’s peers. They suggest where to seed factual information in order to increase the odds of halting false information. Shao et al. [41] also proposes a way to limit attention to those nodes which are most critical for the flow of information. Cho et al. [5] reflect both on stemming false informers and promoting true informers by examining more closely how beliefs of users are updated over time, considering various types of network centrality. What we are able to learn about trusted links, together with a study of the accompanying network relationships, may provide important insights for where to focus effort aimed not just at identifying misinformation but also at stemming its tide.

5.2.3 Exploring new avenues with the use of data sets

A persistent difficulty in applying trust models to social networks has been in finding appropriate data set and evaluation procedures. Most previous attempts to apply trust models to real social networks have relied on networks that included a significant content rating component, such as Yelp, Epinions, and FilmTrust. These networks are attractive primarily because of the ease of harvesting objective test sets from the data extracted from them. How to tell if two agents should really trust each other? Simply check the correlation between the ratings they have given to content—if it’s positive, they should trust each other.
One concern is the fact that some of these data sets, such as that of Epinions, represent dated information (with the site now defunct). The most popular networks (Facebook, Reddit, and Twitter) may be overlooked because they lack a significant content rating component. The issue of securing publically available data sets for some of these platforms arises at times as well.
There may be value to making more of an effort in the future to interrogate actual users of systems. We note the work of Gilbert and Karahalios [13], where 35 participants were recruited for the experiment. The authors had access to the data that the participants agreed to share with them—it was not necessary to convince Facebook to produce a data set for the researchers. After the statistical analysis, a qualitative analysis was performed to help contextualize the errors in the system by interviewing the participants. Cooperating more closely with the users of online social networks in this way will likely be impactful for future trust modeling research.

5.2.4 Considering vulnerable users: the case of older adults

Another direction for future research that would be especially valuable to explore is making use of information about the needs and preferences of a cluster of users that are known independently before performing the data analysis proposed here to determine trust relationships (of value in assisting users in coping with misinformation).
Below, we sketch our current thoughts for how to integrate this prior information about the user base into the overall solution. In this approach, we could in fact have some preconceived notions of the user at hand due to what the user modeling community refers to as stereotypes [2] (what that entire class of users is likely to generally prefer).
We have thought of this direction forward as part of our particular interest in offering support for certain groups of users who are especially vulnerable online. One such community is that of older adults. It would be valuable to be able to carefully advise this demographic about misleading content, and personalized trust link prediction via clustering may thus be of use. We have begun to examine the special considerations of this user base when it comes to misinformation; we note that other research has already identified notable differences for older adults in social media [8, 27, 45]. The framework presented in this paper could expand to integrate prior knowledge of its users in the following way.
Suppose we had a group of older adult users. Per the algorithm of Sect.3, the unsupervised learning method could assign them to the same cluster, suggesting the same weight of their trust indicators (if they have similar trust profiles).
For Trust Link prediction, besides defining trust indicators as we discuss in Sect. 3.3, we could also take the general preference of the older adults into consideration, allowing some finer granularity to the reasoning. In other words, we could consider factors which generally influence the preferences of this particular user base. Ideally, predicting a trust link from this set of users to a certain peer should be predicated both on what the algorithm of Sect. 3 suggests from its data-driven analysis and also by the known prior preferences of the user base.
A more detailed view of how to expand the overall process (our preliminary ideas for doing so) is as follows:
  • We have some priors about the needs of users who fit certain stereotypes (e.g., older adults). We’d like to help those particular users.
  • We do the clustering based on the methods in Sect. 3.
  • We examine the clusters and observe that a large portion of users in some cluster embody one of our known stereotypes, e.g., in some cluster over half the users are older adults.
  • We combine data driven with stereotype-based predictions for those clusters where large proportions of the users embody a particular stereotype, therefore applying our prior about the needs of a stereotype group to a cluster of users that seem to largely embody that stereotype.
One way this process could be operationalized once clustering is performed as in Sect. 3, with clusters of older adults located:
  • Trust Link Prediction
    • Input Trust indicators of older adults \(I_1(a_i,a_j),...,\) \(I_m(a_i,a_j)\) and preference effective function: \(f: I -> \{0,1\}\).
    • Output \(T(a_i, a_j)\), the score of \(a_i\) trusting \(a_j\)’s recommendation.
    • Process Predictor1 takes all indicators and makes a prediction. Predictor2 makes an independent prediction on indicators with \(f(I)=1\). Get \(T(a_i, a_j)\) by combining the two prediction results.
  • Information Recommendation
    • Input Trust-link Prediction \(T(a_i, a_j)\) and information scores from \(a_j\)’s feedback, \(s_j\).
    • Output Recommendation scores for \(a_i\), \(R(a_i, s_j)\)
It would be interesting to delve further into these options for combining prior knowledge and trust link prediction, in order to provide richer recommendations to users. While an approach such as this could be used for any subgroup with known preferences, we feel it especially worthwhile to continue to learn more about the specific needs of older adults, and would explore our new direction with respect to this user base, as a first step.

Acknowledgements

We are grateful to the reviewers for their valuable feedback on an earlier version of this paper.

Open Access

This article is distributed under the terms of the Creative Commons Attribution 4.0 International License (http://​creativecommons.​org/​licenses/​by/​4.​0/​), which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made.

Declarations

Conflicts of interest/Competing interests

The authors declare that they have no conflict of interest.

Availability of data and material

The data set used in this work can be downloaded from https://​www.​yelp.​com/​dataset (the 2019 full data set).

Code availability

Algorithms as explained in the paper.
Open AccessThis article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://​creativecommons.​org/​licenses/​by/​4.​0/​.

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Appendix

A computation of trust indicators

In this appendix, we discuss the challenge of the computation of trust indicators between pairs of agents and adjustments that we made. The trust indicator function \(\Psi (a_i, a_j)\) is expected to be computed for all ordered pairs of agents. Of course, as there are \(O(n^2)\) possible pairs of agents, this rapidly becomes a computational issue as the number of agents considered grows. In our experiments we worked with groups of agents where \(|A| \approx 30000\), implying approximately 900,000,000 pairs—a large but tractable computation on modern consumer hardware. However, the unfiltered Yelp data set contains descriptions of 1,637,138 agents, and we can be assured that other large online environments contain many millions of users. At this scale, the \(O(n^2)\) computation time becomes a serious barrier, and storing the trillions of resulting vectors for further processing would likely be extremely costly.
However, it is not necessary to consider every possible pair of agents. For example, if \(a_i\) and \(a_j\) have never interacted in any meaningful way and share no known interests—in sum, we have no evidence of any way they might know or be interested in each other—then it is likely safe to conclude, without any complex trust modeling, that they need not trust each other. Further, we can conclude that the lack of a trust link between them is most likely the result of ignorance rather than opinion. To analogize, the potential trust relationship between a university professor in China and a wheat farmer in Canada need not be explicitly modeled and computed if no evidence can be found that the two may in fact share a communication channel or desire to interact in the future.
Thus, a solution to the computation barrier presents itself: simply defining a neighborhood function, N(a), on individual agents and only computing trust indicators and trust predictions between pairs of agents in the same neighborhood. So long as computing N(a) is efficient, the execution time of computing all relevant trust indicator pairs then becomes linear with a constant bounded by the maximum neighborhood size.
The definition of N(a) can be very liberal and still result in a substantial speed up. For example, when computing trust indicators for the Yelp and Epinions data set, we used:
$$\begin{aligned} \begin{aligned}&a_j \in N(a_i) \; \iff \\&|R_{ij}| > 0 \vee friends(a_i, a_j)\; \vee friendOfFriend(a_i, a_j). \end{aligned} \end{aligned}$$
(25)
That is, \(a_j\) is in the neighborhood of \(a_i\) if they have both reviewed at least one item in common, if they are friends, if they are friends of friends, or if they are friends of friends.
Applying this neighborhood function drastically reduces the number of pairs of agents that need to be considered in the following stages.
Footnotes
1
The acronym stands for Lesbian, Gay, Bisexual, Transgender, and Queer.
 
2
For example, social network designers could elicit explicit statements of trust from users, but such a feature is not currently popular online.
 
3
Based on the Chernoff bound theorem.
 
5
Predicting positive review score correlation between agents directly may be helpful, since some research suggests that friendship only correlates weakly with similarity in reviewing behavior [16]. This may be viewed as an implicit link between agents.
 
6
This is inspired by the average linkage criterion used in hierarchical clustering algorithms [37]. We tested clustering this data hierarchically, but had little success producing clusters of reasonable size.
 
8
Compared to the popular Epinions data set, where nearly 80% of reviews are 5 stars.
 
9
We use logistic regression because of its simplicity and interpretability. Note that while this model can only learn a linear boundary between classes, using a nonlinear model is also possible. However, simply using a nonlinear model without clustering would not by itself lead to more personalized recommendations.
 
10
A full treatment is given in Sect. 2.
 
11
See Table 1 for the complete list of experiments.
 
12
Note, \(0 \le \beta \le 1\).
 
13
In earlier versions of this work [33], we only split reviews into test and train sets at the last step (recommender evaluation). This would allow, for example, the clustering step to use data to form clusters which was later being tested on.
 
14
This random clustering essentially partitions the data set into k random samples (a close emulation of bootstrap aggregation).
 
15
That is, the problem of giving personalized recommendations to a user who has just joined the network and has not expressed any beliefs, opinions or preferences.
 
16
Fleming [12] also proposes a progression in user modeling from assumptions about general users to ones about individuals but they also suggest an intermediate phase of learning more about groups, via stereotypes. Our consideration of clusters fits well within this vision.
 
Literature
1.
go back to reference Breiman, L.: Bagging predictors. Mach. Learn. 24(2), 123–140 (1996)MATH Breiman, L.: Bagging predictors. Mach. Learn. 24(2), 123–140 (1996)MATH
2.
go back to reference Burnett, C., Norman, T., Sycara, K.: Bootstrapping trust evaluations through stereotypes. In: Proceedings of the 9th International Conference on Autonomous Agents and Multiagent Systems, pp. 241–248 (2010) Burnett, C., Norman, T., Sycara, K.: Bootstrapping trust evaluations through stereotypes. In: Proceedings of the 9th International Conference on Autonomous Agents and Multiagent Systems, pp. 241–248 (2010)
4.
go back to reference Champaign, J., Zhang, J., Cohen, R.: Coping with poor advice from peers in peer-based intelligent tutoring: The case of avoiding bad annotations of learning objects. In: Proceedings of the Nineteenth International Conference on User , Adaptation, and Personalization, UMAP’11, pp. 38–49. Springer (2011) Champaign, J., Zhang, J., Cohen, R.: Coping with poor advice from peers in peer-based intelligent tutoring: The case of avoiding bad annotations of learning objects. In: Proceedings of the Nineteenth International Conference on User , Adaptation, and Personalization, UMAP’11, pp. 38–49. Springer (2011)
7.
go back to reference Ciampaglia, G.L., Mantzarlis, A., Maus, G., Menczer, F.: Research challenges of digital misinformation: toward a trustworthy web. AI Mag. 39(1), 65–74 (2018) Ciampaglia, G.L., Mantzarlis, A., Maus, G., Menczer, F.: Research challenges of digital misinformation: toward a trustworthy web. AI Mag. 39(1), 65–74 (2018)
12.
go back to reference Fleming, M.: The use of increasingly specific user models in the design of mixed-initiative systems. In: A.Y. Tawfik, S.D. Goodwin (eds.) Advance Artificial Intelligence, pp. 434–438 (2004) Fleming, M.: The use of increasingly specific user models in the design of mixed-initiative systems. In: A.Y. Tawfik, S.D. Goodwin (eds.) Advance Artificial Intelligence, pp. 434–438 (2004)
13.
go back to reference Gilbert, E., Karahalios, K.: Predicting tie strength with social media. In: Proceedings of the 27th international conference on Human factors in computing systems (2009) Gilbert, E., Karahalios, K.: Predicting tie strength with social media. In: Proceedings of the 27th international conference on Human factors in computing systems (2009)
14.
go back to reference Gottfried, J., Barthel, M., Mitchell, A.: Internet use. Pew Research Center (2017 (Accessed March 23, 2020)) Gottfried, J., Barthel, M., Mitchell, A.: Internet use. Pew Research Center (2017 (Accessed March 23, 2020))
15.
go back to reference Guo, G., Yang, E., Shen, L., Yang, X., He, X.: Discrete trust-aware matrix factorization for fast recommendation. In: Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence, IJCAI-19, pp. 1380–1386. International Joint Conferences on Artificial Intelligence Organization (2019). https://doi.org/10.24963/ijcai.2019/191 Guo, G., Yang, E., Shen, L., Yang, X., He, X.: Discrete trust-aware matrix factorization for fast recommendation. In: Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence, IJCAI-19, pp. 1380–1386. International Joint Conferences on Artificial Intelligence Organization (2019). https://​doi.​org/​10.​24963/​ijcai.​2019/​191
16.
go back to reference Guo, G., Zhang, J., Sun, Z., Yorke-Smith, Librec, N.: A java library for recommender systems. In: UMAP Workshops 4, (2015) Guo, G., Zhang, J., Sun, Z., Yorke-Smith, Librec, N.: A java library for recommender systems. In: UMAP Workshops 4, (2015)
17.
go back to reference Guo, G., Zhang, J., Yorke-Smith, N.: Trustsvd: Collaborative filtering with both the explicit and implicit influence of user trust and of item ratings. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 29 (2015) Guo, G., Zhang, J., Yorke-Smith, N.: Trustsvd: Collaborative filtering with both the explicit and implicit influence of user trust and of item ratings. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 29 (2015)
21.
go back to reference Hui, P.M., Shao, C., Flammini, A., Menczer, F., Ciampaglia, G.L.: The hoaxy misinformation and fact-checking diffusion network. In: Proceedings of the 12th International AAAI Conference on Web and Social Media (2018) Hui, P.M., Shao, C., Flammini, A., Menczer, F., Ciampaglia, G.L.: The hoaxy misinformation and fact-checking diffusion network. In: Proceedings of the 12th International AAAI Conference on Web and Social Media (2018)
22.
go back to reference Jia, D., Zhang, F., Liu, S.: A robust collaborative filtering recommendation algorithm based on multidimensional trust model. J. Softw. 8(1), 11–18 (2013)CrossRef Jia, D., Zhang, F., Liu, S.: A robust collaborative filtering recommendation algorithm based on multidimensional trust model. J. Softw. 8(1), 11–18 (2013)CrossRef
23.
go back to reference Jøsang, A., Ismail, R.: The beta reputation system. In: Proceedings of the 15th Bled Conference on Electronic Commerce (2002) Jøsang, A., Ismail, R.: The beta reputation system. In: Proceedings of the 15th Bled Conference on Electronic Commerce (2002)
24.
go back to reference Kerr, R., Cohen, R.: Treet: the trust and reputation experimentation and evaluation testbed. Electron. Commer. Res. 10(3–4), 271–290 (2010)CrossRef Kerr, R., Cohen, R.: Treet: the trust and reputation experimentation and evaluation testbed. Electron. Commer. Res. 10(3–4), 271–290 (2010)CrossRef
25.
go back to reference Koren, Y., Bell, R., Volinsky, C.: Matrix factorization techniques for recommender systems. Computer 42(8), 30–37 (2009)CrossRef Koren, Y., Bell, R., Volinsky, C.: Matrix factorization techniques for recommender systems. Computer 42(8), 30–37 (2009)CrossRef
27.
go back to reference Lehtinen, V., Näsänen, J., Sarvas, R.: “a little silly and empty-headed” - older adults’ understandings of social networking sites 10.14236/ewic/HCI2009.6,(2009) Lehtinen, V., Näsänen, J., Sarvas, R.: “a little silly and empty-headed” - older adults’ understandings of social networking sites 10.14236/ewic/HCI2009.6,(2009)
28.
go back to reference Liu, X., Datta, A., Rzadca, K., Lim, E.P.: StereoTrust: a group based personalized trust model. In: Proceedings of the 18th ACM conference on Information and knowledge management, CIKM ’09, pp. 7–16 (2009) Liu, X., Datta, A., Rzadca, K., Lim, E.P.: StereoTrust: a group based personalized trust model. In: Proceedings of the 18th ACM conference on Information and knowledge management, CIKM ’09, pp. 7–16 (2009)
29.
go back to reference Mauro, N., Ardissono, L., Hu, Z.F.: Multi-faceted trust-based collaborative filtering. In: Proceedings of the 27th ACM Conference on User Modeling, Adaptation and Personalization, UMAP ’19, pp. 216–224 (2019) Mauro, N., Ardissono, L., Hu, Z.F.: Multi-faceted trust-based collaborative filtering. In: Proceedings of the 27th ACM Conference on User Modeling, Adaptation and Personalization, UMAP ’19, pp. 216–224 (2019)
31.
go back to reference O’Donovan, J., Smyth, B.: Trust in recommender systems. In: Proceedings of the 10th international conference on Intelligent user interfaces, pp. 167–174 (2005) O’Donovan, J., Smyth, B.: Trust in recommender systems. In: Proceedings of the 10th international conference on Intelligent user interfaces, pp. 167–174 (2005)
32.
go back to reference Ohashi, D., Cohen, R., Fu, X.: The current state of online social networking for the health community: Where trust modeling research may be of value. In: Proceedings of the 2017 International Conference on Digital Health, DH ’17, p. 23–32. Association for Computing Machinery, New York, NY, USA (2017) Ohashi, D., Cohen, R., Fu, X.: The current state of online social networking for the health community: Where trust modeling research may be of value. In: Proceedings of the 2017 International Conference on Digital Health, DH ’17, p. 23–32. Association for Computing Machinery, New York, NY, USA (2017)
33.
go back to reference Parmentier, A., Cohen, R.: Personalized multi-faceted trust modeling in social networks. In: C. Goutte, X. Zhu (eds.) Advances in Artificial Intelligence - 33rd Canadian Conference on Artificial Intelligence, Canadian AI 2020, Ottawa, ON, Canada, May 13-15, 2020, Proceedings, Lecture Notes in Computer Science, vol. 12109, pp. 445–450. Springer (2020). https://doi.org/10.1007/978-3-030-47358-7_46 Parmentier, A., Cohen, R.: Personalized multi-faceted trust modeling in social networks. In: C. Goutte, X. Zhu (eds.) Advances in Artificial Intelligence - 33rd Canadian Conference on Artificial Intelligence, Canadian AI 2020, Ottawa, ON, Canada, May 13-15, 2020, Proceedings, Lecture Notes in Computer Science, vol. 12109, pp. 445–450. Springer (2020). https://​doi.​org/​10.​1007/​978-3-030-47358-7_​46
34.
go back to reference Pedregosa, F., Varoquaux, G., Gramfort, A., Thirion, V.M.B., Grisel, O., Blondel, M., Prettenhofer, P., Weiss, R., Dubourg, V., Vanderplas, J., Passos, A., Cournapeau, D., Brucher, M., Perrot, M.: Édouard Duchesnay: scikit-learn: machine learning in python. J. Mach. Learn. Res. 12, 2825–2830 (2011)MathSciNetMATH Pedregosa, F., Varoquaux, G., Gramfort, A., Thirion, V.M.B., Grisel, O., Blondel, M., Prettenhofer, P., Weiss, R., Dubourg, V., Vanderplas, J., Passos, A., Cournapeau, D., Brucher, M., Perrot, M.: Édouard Duchesnay: scikit-learn: machine learning in python. J. Mach. Learn. Res. 12, 2825–2830 (2011)MathSciNetMATH
36.
go back to reference Regan, K., Poupart, P., Cohen, R.: Bayesian reputation modeling in e-marketplaces sensitive to subjectivity, deception and change. In: Proceedings of the National Conference on Artificial Intelligence, vol. 2 (2006) Regan, K., Poupart, P., Cohen, R.: Bayesian reputation modeling in e-marketplaces sensitive to subjectivity, deception and change. In: Proceedings of the National Conference on Artificial Intelligence, vol. 2 (2006)
37.
go back to reference Rokach, L., Maimon, O.: Clustering Methods, pp. 321–352. Springer, New York (2005) Rokach, L., Maimon, O.: Clustering Methods, pp. 321–352. Springer, New York (2005)
39.
go back to reference Sarwar, B., Karypis, G., Konstan, J., Riedl, J.: Application of dimensionality reduction in recommender system-a case study. Minnesota Univ Minneapolis Dept of Computer Science, Tech. rep. (2000) Sarwar, B., Karypis, G., Konstan, J., Riedl, J.: Application of dimensionality reduction in recommender system-a case study. Minnesota Univ Minneapolis Dept of Computer Science, Tech. rep. (2000)
40.
go back to reference Seely Brown, J., Adler, R.: Open education, the long tail, and learning 2.0. Educause Rev. 43(1), 16–20 (2008) Seely Brown, J., Adler, R.: Open education, the long tail, and learning 2.0. Educause Rev. 43(1), 16–20 (2008)
41.
go back to reference Shao, C., Hui, P.M., Wang, L., Jiang, X., Flammini, A., Menczer, F., Ciampaglia, G.L.: Anatomy of an online misinformation network. PloS one 13(4), 1 (2018)CrossRef Shao, C., Hui, P.M., Wang, L., Jiang, X., Flammini, A., Menczer, F., Ciampaglia, G.L.: Anatomy of an online misinformation network. PloS one 13(4), 1 (2018)CrossRef
43.
go back to reference Tong, A., Du, D.Z., Wu, W.: On misinformation containment in online social networks. In: Advances in Neural Information Processing Systems 31, pp. 341–351 (2018) Tong, A., Du, D.Z., Wu, W.: On misinformation containment in online social networks. In: Advances in Neural Information Processing Systems 31, pp. 341–351 (2018)
44.
go back to reference Weiss, G.: Multiagent systems. MIT press, Cambridge (2013) Weiss, G.: Multiagent systems. MIT press, Cambridge (2013)
45.
go back to reference Wylie, L., Patihis, L., Mcculler, L., Davis, D., Brank, E., Loftus, E., Bornstein, B.: Misinformation effect in older versus younger adults a meta-analysis and review. In: Toglia, M.P., Ross, D.F., Pozzulo, J., Pica, E. (eds.) The Elderly Eyewitness in Court, pp. 38–66. Psychology Press (2014) Wylie, L., Patihis, L., Mcculler, L., Davis, D., Brank, E., Loftus, E., Bornstein, B.: Misinformation effect in older versus younger adults a meta-analysis and review. In: Toglia, M.P., Ross, D.F., Pozzulo, J., Pica, E. (eds.) The Elderly Eyewitness in Court, pp. 38–66. Psychology Press (2014)
46.
go back to reference Yang, B., Lei, Y., Liu, D., Liu, J.: Social collaborative filtering by trust. In: Proceedings of the Twenty-Third IJCAI International Joint Conference on Artificial Intelligence, pp. 2747–2753 (2013) Yang, B., Lei, Y., Liu, D., Liu, J.: Social collaborative filtering by trust. In: Proceedings of the Twenty-Third IJCAI International Joint Conference on Artificial Intelligence, pp. 2747–2753 (2013)
47.
go back to reference Yang, S., Shu, K., Wang, S., Gu, R., Wu, F., Liu, H.: Unsupervised fake news detection on social media: A generative approach. In: Proc. AAAI (2019) Yang, S., Shu, K., Wang, S., Gu, R., Wu, F., Liu, H.: Unsupervised fake news detection on social media: A generative approach. In: Proc. AAAI (2019)
48.
go back to reference Zhang, J., Cohen, R.: Evaluating the trustworthiness of advice about seller agents in e-marketplaces: a personalized approach. Electron. Commer. Res. Appl. 7, 330–340 (2008)CrossRef Zhang, J., Cohen, R.: Evaluating the trustworthiness of advice about seller agents in e-marketplaces: a personalized approach. Electron. Commer. Res. Appl. 7, 330–340 (2008)CrossRef
Metadata
Title
Personalized multi-faceted trust modeling to determine trust links in social media and its potential for misinformation management
Authors
Alexandre Parmentier
Robin Cohen
Xueguang Ma
Gaurav Sahu
Queenie Chen
Publication date
22-01-2022
Publisher
Springer International Publishing
Published in
International Journal of Data Science and Analytics / Issue 4/2022
Print ISSN: 2364-415X
Electronic ISSN: 2364-4168
DOI
https://doi.org/10.1007/s41060-021-00294-w

Other articles of this Issue 4/2022

International Journal of Data Science and Analytics 4/2022 Go to the issue

Premium Partner