TY - JOUR AU - Zhang, Yangwen AB - Introduction Recommender Systems (RS), in recent years, have become essential in online applications, tackling information overload by learning personalized user preferences [1,2]. They achieve this by leveraging user-item interaction histories [3]. However, RS that depend exclusively on these histories encounter challenges like sparse interaction data (i.e. most users have only acted on a few items, which can affect the accuracy of the RS) and the cold-start problem (i.e. it difficult to provide effective recommendations for new users or items without sufficient user behavior data). Recommending cold items in RS is a long-standing challenge, as hot items are recommended based on user behavior, while cold items are recommended based on content characteristics [4]. Collaborative Filtering (CF) is a classic machine learning technique in RS (e.g., [5–7]). While traditional CF has been successful, it often struggles with the cold-start problem and sparse interaction data, and it is limited in handling graph-structured data. With the advancement of deep learning, research has shifted toward Graph Neural Networks (GNNs), which excel at capturing complex relationships in graph-structured data and provide a robust framework for learning node embeddings. This has led to the development of GNN-based recommendation models [3,8]. These methods effectively model user-item interaction graphs using graph neural encoders, enhancing user representation learning. However, sparse data and cold-start problem still pose challenges to the performance of prediction models [9]. Previous research [10] has shown that social networks, such as trust links on shopping websites [11], significantly influence users’ online behavior. Combining social links with user-item interactions (Fig 1) enhances recommendation performance [12,13], known as the social recommendation problem. This approach relies on Heterogeneous Information Networks (HINs), which integrate various entities and relationships, strengthening connections between users and items, as well as explore potential collaborative correlations between various entities [14–16]. Researchers hypothesize that users with shared social connections exhibit similar interests. By integrating social information, both social connections and interaction data can be captured simultaneously, improving recommendation accuracy and diversity and mitigating sparse data and cold-start problem. This hypothesis encourages exploring more heterogeneous collaborative relationships beyond the user domain to enhance recommendation accuracy. If validated, it could provide new perspectives for addressing sparse data and cold-start problem in RS, especially when considering the potential performance degradation with privacy-preserving technologies like federated learning and sensitive data collection techniques [17–21]. Download: PPT PowerPoint slide PNG larger image TIFF original image Fig 1. Social network, user-item interaction and item-category relationship. https://doi.org/10.1371/journal.pone.0313491.g001 In real-world recommendation scenarios, significant dependencies exist not only among users but also among items. Users’ preferences for items can be inferred through their social relationships and fine-grained relational knowledge of items, such as item-category relationships (Fig 1) or temporal knowledge (like ODETKE [22]). For instance, user interactions like purchasing, liking, or sharing may be influenced by item features such as brand, color, or director [23,24]. The characteristics of items are composed of many factors such as brand, type, color, and how the item interacts with other behaviors (such as sharing or purchasing between users). In short, there is often a potential connection and mutual influence between the characteristics of users and items. Thus, exploring the rich semantic relationships between users and items is crucial for understanding users’ complex interests [25,26]. Incorporating relational heterogeneity and cross-node dependencies into social recommendation and GNN models is challenging due to the diversity of relationships. One solution is dependency-based heterogeneous graph learning methods [15,27], which integrate structural connections between users and items into graph models. However, these often rely on manually designed meta-paths, requiring domain-specific knowledge and lacking adaptability to different scenarios. Additionally, existing methods typically adjust transformation weights in the feature representation process of node or edge types, limiting their ability to capture the complete heterogeneous relationship context from both user and item perspectives. Most current social recommendation solutions either do not explore or only superficially explore the rich collaborative semantics among items (e.g., ConsistRec [28]). Moreover, with the continuous development of algorithms, data and computing power, the increase in size and complexity of these models has led to an increase in training costs [29]. To address these challenges, we propose a novel recommendation model, CoHet4Rec, based on Collaborative Heterogeneous Information Networks (CHIN). This model integrates various information sources and learns latent features without relying on specific nodes and edges. First, we incorporate user-user social relationships and item-category relationships into RS to extend user-item interactions and construct semantically richer collaborative heterogeneous networks. Then we use GNNs to learn node embeddings. To handle the heterogeneity and collaboration of relationships in the data, we incorporate external collaborative heterogeneous relation units and differentiable embedding propagation operators into the GNN architecture. We introduce an adaptive first-order collaborative heterogeneous neighbor attention mechanism, which combines multiple collaborative semantics from both user and item domains, enhancing user-item interactions. By assigning different attention weights to heterogeneous neighbors, we capture their varying importance. This model does not require manually designed meta-paths, offers high scalability, and can accommodate personalized information dissemination needs by either designing meta-paths or considering broader contextual collaborative semantics. In summary, the key contributions of our work are: We provided a new insight into alleviating sparse data and cold-start problem by enhancing user-item interaction and adopting extended heterogeneous relational contexts in both user and item domains. We proposed a GNN-based and CHIN-based recommendation framework that no longer relies on manually designed meta-paths to coordinate the context of different information sources. Extensive experiments conducted on three real-world datasets have demonstrated that CoHet4Rec outperforms existing methods. For instance, HR@5 and NDCG@5 are increased by 16.00% and 18.18% on average on Ciao dataset. Related work Social RS Early approaches in social RS focused on CF techniques and integrating social network information to address the cold-start problem and enhance recommendation diversity. For instance, SoRec [5] decomposed the user rating matrix and social link matrix using coefficients, while HGMF [6] introduced hierarchical group matrix factorization to learn user group features in social networks for recommendations. However, traditional machine learning methods do not perform outstandingly on graph structured data. Recent advancements in deep learning have propelled social RS, with GNN playing a key role in capturing complex dependencies in social networks. GraphRec [8] and SAMN [30] employed Graph Attention Networks to learn user and item features effectively, while MHCN [31] utilized self-supervised learning for data augmentation. SENGR [32] introduced an emotion-enhanced neural graph recommender that integrates text comments and bipartite graphs, and GNNRec [33] is a conversation-based social recommendation model that simulates users’ interests using their conversation and social data. While these studies highlight the importance of incorporating social information into RS, they often overlook the collaborative aspects of heterogeneous relationships on the item side, limiting their ability to fully exploit diverse user preferences for accurate recommendations. To fill this gap, this work aims to explore a new social RS that explores potential collaborative heterogeneous semantics under GNN architecture. HIN-based RS Heterogeneous graph-based RS is an emerging paradigm that harnesses diverse, multi-relational data across various domains [34,35]. For example, DisenHAN [36] focuses on learning different user/item aspects in a HIN by utilizing meta relationships to decompose high-order connectivity between nodes. DGNN [37] disentangles heterogeneous relationships by propagating information separately and aggregating edges within the same relationship during node aggregation. MML [16] maps users and items to a unified latent vector space, modeling not only the observed relationships between users and items (user-user, item-item, user-item), but also the latent relationships. However, there is still a gap in their exploration of potential collaboration between various relationships. Most works in this field have used meta-path-based techniques to explore intricate relationships within heterogeneous networks. For instance, a dual attention mechanism was been proposed in HAN [27] to effectively learn the significance of different meta-paths. SMIN [38] designed a meta-path guided heterogeneous GNN that incorporates the relationship structure between social and knowledge awareness into user preference representation. MAHGE [39] proposed a meta-path aggregation method to model time as a relationship on the graph edge, capturing complex dependencies between users and time. MNRec [34] derives user and item embeddings through meta-paths based on domain knowledge and a commuting matrix-based approach for heterogeneous network embedding. Additionally, MAERec [35] employs automatic meta-path extraction from HIN and a hierarchical attention network to learn explicit representations of context-based meta-paths. While meta-path-based methods mentioned above are effective for recommendation in HIN, they often rely on domain experts to create precise meta-paths or complex methods to generate them. Social RS Early approaches in social RS focused on CF techniques and integrating social network information to address the cold-start problem and enhance recommendation diversity. For instance, SoRec [5] decomposed the user rating matrix and social link matrix using coefficients, while HGMF [6] introduced hierarchical group matrix factorization to learn user group features in social networks for recommendations. However, traditional machine learning methods do not perform outstandingly on graph structured data. Recent advancements in deep learning have propelled social RS, with GNN playing a key role in capturing complex dependencies in social networks. GraphRec [8] and SAMN [30] employed Graph Attention Networks to learn user and item features effectively, while MHCN [31] utilized self-supervised learning for data augmentation. SENGR [32] introduced an emotion-enhanced neural graph recommender that integrates text comments and bipartite graphs, and GNNRec [33] is a conversation-based social recommendation model that simulates users’ interests using their conversation and social data. While these studies highlight the importance of incorporating social information into RS, they often overlook the collaborative aspects of heterogeneous relationships on the item side, limiting their ability to fully exploit diverse user preferences for accurate recommendations. To fill this gap, this work aims to explore a new social RS that explores potential collaborative heterogeneous semantics under GNN architecture. HIN-based RS Heterogeneous graph-based RS is an emerging paradigm that harnesses diverse, multi-relational data across various domains [34,35]. For example, DisenHAN [36] focuses on learning different user/item aspects in a HIN by utilizing meta relationships to decompose high-order connectivity between nodes. DGNN [37] disentangles heterogeneous relationships by propagating information separately and aggregating edges within the same relationship during node aggregation. MML [16] maps users and items to a unified latent vector space, modeling not only the observed relationships between users and items (user-user, item-item, user-item), but also the latent relationships. However, there is still a gap in their exploration of potential collaboration between various relationships. Most works in this field have used meta-path-based techniques to explore intricate relationships within heterogeneous networks. For instance, a dual attention mechanism was been proposed in HAN [27] to effectively learn the significance of different meta-paths. SMIN [38] designed a meta-path guided heterogeneous GNN that incorporates the relationship structure between social and knowledge awareness into user preference representation. MAHGE [39] proposed a meta-path aggregation method to model time as a relationship on the graph edge, capturing complex dependencies between users and time. MNRec [34] derives user and item embeddings through meta-paths based on domain knowledge and a commuting matrix-based approach for heterogeneous network embedding. Additionally, MAERec [35] employs automatic meta-path extraction from HIN and a hierarchical attention network to learn explicit representations of context-based meta-paths. While meta-path-based methods mentioned above are effective for recommendation in HIN, they often rely on domain experts to create precise meta-paths or complex methods to generate them. Preliminary In this paper, we have three collections: users, items, and categories, represented as , , and , respectively, where N, M, and K are the numbers of users, items, and categories. Social relationships are represented by a social matrix . Interaction behavior between users and items is captured by an interaction matrix . Item-category relationships are depicted by a relation matrix . In matrix X, if there is an interaction (e.g., rating) between user n and an item m, it is denoted as ; otherwise, . In matrix S, a social link between users n and is denoted as ; otherwise, . It is important to note that a user can have social connections with multiple other users. In matrix T, if an item m belongs to category k, it is denoted as ; otherwise, . An item can belong to multiple categories. CoHet4Rec aims to train a recommendation model using X, S, and T to accurately predict user preferences for unseen items. Method We introduce the CoHet4Rec framework (Fig 2), which comprises three essential modules: the embedding layer, the collaborative heterogeneous neighbor attention layer, and the prediction and optimization stage. The embedding layer embeds the features of nodes, the collaborative heterogeneous neighbor attention layer mainly updates the features of nodes, and the prediction and optimization stage mainly trains and optimizes the model. Download: PPT PowerPoint slide PNG larger image TIFF original image Fig 2. Overall framework of CoHet4Rec. Expanding on the conventional user-item interaction network, we incorporate the social network and item-category relation network to extend information in both user and item domains. This enables the joint construction of a CHIN, which is input into the embedding layer to obtain user, item, and category node embeddings. These embeddings are then fed into CoHet4Rec for collaborative heterogeneous information propagation, enhanced by meta-path adaptive heterogeneous relation attention and collaborative heterogeneous neighbor information aggregation. https://doi.org/10.1371/journal.pone.0313491.g002 Embedding layer To preserve social connections, user-item interaction, item-category relationships and their rich semantic information, we combine the interaction, social and relation sub-graph into a unified CHIN G. (1) where N = U ∪ V ∪ C is the set of user/item/category nodes, and E = X ∪ S ∪ T denotes the set of edges connecting these nodes. We maintain an embedding layer , wherein each row corresponds to a learnable node representation mapping parameter. By indexing the layer, we can obtain the embedding for a given node v ∈ N. Collaborative heterogeneous neighbor attention layer The CoHet4Rec framework utilizes a CHIN G as the input computational graph, which encompasses three types of heterogeneous relations (or can be extended to include multiple relations) for collaborative heterogeneous information propagation (as depicted in the left portion of Fig 2). The attention layer primarily encompasses two steps: collaborative heterogeneous relation information propagation with attention enhancement and multi-heterogeneous collaborative relation information aggregation. We can formally represent the processes of information propagation and aggregation at two distinct levels (from layer to layer ): Edge-wise: (2) where ϕ ( ⋅ ) is the message function used to propagate messages between node u and v along the connecting edge ( u , v ) according to the attention weights of the heterogeneous relations. Node-wise: (3) where ψ ( ⋅ ) is the message update function used to update the own message of u. represents u’s embedding at the -th layer. N ( u ) represents the set of 1st-order neighbors of u. The function Agg ( ⋅ ) serves as a message aggregation mechanism, responsible for combining the messages received from the 1st-order neighboring nodes. Specifically: (1) To capture the enhanced collaborative heterogeneous features, which are influenced by the knowledge of user domain and item domain during the process of collaborative heterogeneous information propagation with attention enhancement, we employ external heterogeneous relation attention units to parameterize the heterogeneity and collaboration of relations as embedding projections. As a result, the graph encoder is capable of acquiring meaningful non-linear characteristics from collaborative heterogeneous relation representations. In order to maintain the collaborative semantics among different nodes and edges, a shared hyper-parameter space is utilized to encode the collaborative heterogeneous relations with heterogeneous relation attention. This approach enables us to learn collaborative semantic features that capture various potential collaborative influencing factors within the graph G. Specifically, the heterogeneous relation attention encoder is given by Eqs (4) and (5): (4)(5) where, since the representations learned by nodes from different relationships are heterogeneous, we employ a Multi-Layer Perceptron (MLP) based attention calculation network to learn the heterogeneous relationships between nodes and project these representations into the same semantic space. The trainable transformation matrix in Linear1 ( ⋅ ) is represented as and is represented as in Linear2 ( ⋅ ) . The trainable bias terms in Linear2 ( ⋅ ) is represented as . The negative slope is set to 0.2 for better gradient back-propagation. u is the source node and v is the target node. (2) Regarding the aggregation of multi-heterogeneous collaborative relation information: after conducting collaborative heterogeneous information propagation with attention enhancement, the subsequent step involves aggregating the heterogeneous relation semantics from the 1st-order collaborative heterogeneous neighbors of the nodes. We need to aggregate messages from various sources of information, such as users, items, and categories (as depicted in Fig 3). For instance, in the case of node (which can represent a user node, item node, or category node), we combine the messages from the 1st-order heterogeneous relation context in the message aggregator, as represented by Eq (6). After multiple information propagation and aggregation processes, the model can obtain high-order information of nodes in the same domain and different domains (as shown by the unidirectional dashed line and bidirectional solid line in Fig 3). (6) where | N ( u ) | represents the number of 1st-order collaborative heterogeneous neighbors of a given node u. Download: PPT PowerPoint slide PNG larger image TIFF original image Fig 3. High-order information propagation of CoHet4Rec. https://doi.org/10.1371/journal.pone.0313491.g003 To ensure stable network training, following [37,40], we also utilize a combination of self-propagation and layer normalization techniques. This helps to enhance the overall stability of the training process. In terms of self-loop, we treats the self node as a neighbor and applies attention encoding. (7) where, we introduce learnable scaling factors α and bias terms β. μ and σ are the mean and standard deviation of , respectively. eps is a small constant. We also conducted cross-layer residual feature aggregation (as depicted in Eq (8) and the left portion of Fig 2) to fully leverage multi-layer node embedding. (8) where represents the final node embedding. ‘ ∣ ’ represents embedding concatenation. Prediction and optimization During the prediction phase, we calculate user preferences for items by calculating the inner product of user and item features. Our optimization objective is defined by Eq (9), which includes a composite loss: the pairwise Bayesian Personalized Ranking (BPR) loss and a regularization term for weight decay. (9)(10) where represents the difference in preference between the model’s predictions for positive sample item i and negative sample item j for user u. λ represents the weight of the regularization term to avoid over-fitting, θ represents the training parameters. and represent the embeddings of the positive and negative item i and j, respectively. To maximize , which represents the probability that the user’s preference for the positive sample is higher than the negative one, is the optimization objective. During the optimization phase, we iteratively update θ by gradient descent (Eq 11). (11) where η is the learning rate. The process of CoHet4Rec is detailed in Algorithm 1. Algorithm 1. CoHet4Rec Embedding layer To preserve social connections, user-item interaction, item-category relationships and their rich semantic information, we combine the interaction, social and relation sub-graph into a unified CHIN G. (1) where N = U ∪ V ∪ C is the set of user/item/category nodes, and E = X ∪ S ∪ T denotes the set of edges connecting these nodes. We maintain an embedding layer , wherein each row corresponds to a learnable node representation mapping parameter. By indexing the layer, we can obtain the embedding for a given node v ∈ N. Collaborative heterogeneous neighbor attention layer The CoHet4Rec framework utilizes a CHIN G as the input computational graph, which encompasses three types of heterogeneous relations (or can be extended to include multiple relations) for collaborative heterogeneous information propagation (as depicted in the left portion of Fig 2). The attention layer primarily encompasses two steps: collaborative heterogeneous relation information propagation with attention enhancement and multi-heterogeneous collaborative relation information aggregation. We can formally represent the processes of information propagation and aggregation at two distinct levels (from layer to layer ): Edge-wise: (2) where ϕ ( ⋅ ) is the message function used to propagate messages between node u and v along the connecting edge ( u , v ) according to the attention weights of the heterogeneous relations. Node-wise: (3) where ψ ( ⋅ ) is the message update function used to update the own message of u. represents u’s embedding at the -th layer. N ( u ) represents the set of 1st-order neighbors of u. The function Agg ( ⋅ ) serves as a message aggregation mechanism, responsible for combining the messages received from the 1st-order neighboring nodes. Specifically: (1) To capture the enhanced collaborative heterogeneous features, which are influenced by the knowledge of user domain and item domain during the process of collaborative heterogeneous information propagation with attention enhancement, we employ external heterogeneous relation attention units to parameterize the heterogeneity and collaboration of relations as embedding projections. As a result, the graph encoder is capable of acquiring meaningful non-linear characteristics from collaborative heterogeneous relation representations. In order to maintain the collaborative semantics among different nodes and edges, a shared hyper-parameter space is utilized to encode the collaborative heterogeneous relations with heterogeneous relation attention. This approach enables us to learn collaborative semantic features that capture various potential collaborative influencing factors within the graph G. Specifically, the heterogeneous relation attention encoder is given by Eqs (4) and (5): (4)(5) where, since the representations learned by nodes from different relationships are heterogeneous, we employ a Multi-Layer Perceptron (MLP) based attention calculation network to learn the heterogeneous relationships between nodes and project these representations into the same semantic space. The trainable transformation matrix in Linear1 ( ⋅ ) is represented as and is represented as in Linear2 ( ⋅ ) . The trainable bias terms in Linear2 ( ⋅ ) is represented as . The negative slope is set to 0.2 for better gradient back-propagation. u is the source node and v is the target node. (2) Regarding the aggregation of multi-heterogeneous collaborative relation information: after conducting collaborative heterogeneous information propagation with attention enhancement, the subsequent step involves aggregating the heterogeneous relation semantics from the 1st-order collaborative heterogeneous neighbors of the nodes. We need to aggregate messages from various sources of information, such as users, items, and categories (as depicted in Fig 3). For instance, in the case of node (which can represent a user node, item node, or category node), we combine the messages from the 1st-order heterogeneous relation context in the message aggregator, as represented by Eq (6). After multiple information propagation and aggregation processes, the model can obtain high-order information of nodes in the same domain and different domains (as shown by the unidirectional dashed line and bidirectional solid line in Fig 3). (6) where | N ( u ) | represents the number of 1st-order collaborative heterogeneous neighbors of a given node u. Download: PPT PowerPoint slide PNG larger image TIFF original image Fig 3. High-order information propagation of CoHet4Rec. https://doi.org/10.1371/journal.pone.0313491.g003 To ensure stable network training, following [37,40], we also utilize a combination of self-propagation and layer normalization techniques. This helps to enhance the overall stability of the training process. In terms of self-loop, we treats the self node as a neighbor and applies attention encoding. (7) where, we introduce learnable scaling factors α and bias terms β. μ and σ are the mean and standard deviation of , respectively. eps is a small constant. We also conducted cross-layer residual feature aggregation (as depicted in Eq (8) and the left portion of Fig 2) to fully leverage multi-layer node embedding. (8) where represents the final node embedding. ‘ ∣ ’ represents embedding concatenation. Prediction and optimization During the prediction phase, we calculate user preferences for items by calculating the inner product of user and item features. Our optimization objective is defined by Eq (9), which includes a composite loss: the pairwise Bayesian Personalized Ranking (BPR) loss and a regularization term for weight decay. (9)(10) where represents the difference in preference between the model’s predictions for positive sample item i and negative sample item j for user u. λ represents the weight of the regularization term to avoid over-fitting, θ represents the training parameters. and represent the embeddings of the positive and negative item i and j, respectively. To maximize , which represents the probability that the user’s preference for the positive sample is higher than the negative one, is the optimization objective. During the optimization phase, we iteratively update θ by gradient descent (Eq 11). (11) where η is the learning rate. The process of CoHet4Rec is detailed in Algorithm 1. Algorithm 1. CoHet4Rec Results We provide the outcomes of extensive experiments performed to evaluate the efficacy of CoHet4Rec, employing three openly accessible datasets. We proposed several research questions (RQs) to evaluate the performance of our method: RQ1: Is CoHet4Rec better than existing methods? RQ2: Are these components necessary in CoHet4Rec? RQ3: Are these CHINs important? RQ4: How does data sparsity affect model performance? RQ5: How efficient is the model? RQ6: What is the impact of hyper-parameters? RQ1 section conducted a comparative study on the performance of the model and roughly analyzed several important contributions to the success of the model; Then, detailed experimental verification and analysis were conducted on these contributions in RQ2 and RQ3 section, respectively. In addition, RQ4 and RQ5 section studied the robustness of the model to data sparsity and time cost. Finally, RQ6 section explored the influence of some important hyper-parameters that affect model performance. Experimental setup Datasets: In this paper, we performed experiments and analysis using three widely-used social RS datasets (Ciao, Epinions and Yelp) which were collected from various online review systems in real-world. Ciao and Epinions are available at https://www.cse.msu.edu/ tangjili/ datasetcode/truststudy.htm, Yelp is available at https://www.yelp.com/dataset. Table 1 provides comprehensive statistical information about these datasets. Download: PPT PowerPoint slide PNG larger image TIFF original image Table 1. Statistics of datasets. https://doi.org/10.1371/journal.pone.0313772.t001 Baselines: Following [37], we comprehensively evaluate the performance of CoHet4Rec with 15 baselines from 7 different types: Graph-CF RS GCCF [41]: A linear residual GCN-based CF recommendation model. It aims to solve the problems of GCN CF models in terms of training difficulties, nonlinear activation functions, and deep stacking. NGCF [42]: This CF model utilizes graph convolution to leverage the user-item graph structure. It effectively represents and captures high-order connectivity relationships by propagating features on the graph. GNN-based RS GraphRec [8]: A GNN-based social RS algorithm that can simultaneously consider the influence of interaction and related viewpoints. DiffNet [43]: Utilizes graph information propagation to establish a hierarchical diffusion architecture for modeling users’ social relationships. It simulates dynamic social diffusion to capture recursive social influence. MHCN [31]: A hypergraph convolutional networks-based social RS that utilizes high-order user relationships to improve recommendation quality. Disentangled graph RS DGNN [37]: It explores social recommendation by learning disentangled heterogeneous factors and improves social recommendation by disentangling the heterogeneous relationships between users and item domains. DGCF [24]: It focuses particularly on user-item relationships with a finer granularity of user intent and models the intention distribution of each user-item interaction and iteratively refines the intention aware interaction graph and representation. DisenHAN [36]: A disentangled heterogeneous GAT for Top-K recommendation. It learns user/item representations for disentanglement in HINs from different perspectives. Time-aware social RS DGRec [44]: It models dynamic user behavior using RNN and contextual social impact using GAT. Knowledge-aware RS KGAT [25]: This model employs attention mechanisms to aggregate information from both interactions and the item knowledge. Attention-based social RS SAMN [30]: In this model, a two-stage attention mechanism is employed to capture user influence and identify relevant friends for preference modeling. It introduces a friend-style attention to explore the social influence between users. EATNN [45]: It utilizes attention mechanisms to integrate interactive and social information. It combines an optimization scheme to implement a multitasking learning framework. Heterogeneous graph embedding-based RS HERec [15]: A HIN embedding-based method. It learns node embeddings by generating meaningful node sequences through a random walk strategy based on meta-path. HGT [14]: In this method, specific parameters are designed to capture the heterogeneous attention on each edge, taking into account the node and edge types. This enables the model to maintain distinct embeddings for various types of nodes and edges. HAN [27]: A attention mechanism-based heterogeneous GNN method. It considers the diversity of nodes and edges in heterogeneous networks, adopts attention mechanisms at both node and semantic levels, and generates node embedding representations through hierarchical aggregation of node features. Evaluation Metrics: In our study, CoHet4Rec focuses on providing recommendations of the Top-K items for users. To evaluate and compare the models, we use two widely adopted metrics in RS: Hits Ratio (HR@K) and Normalized Discounted Cumulative Gain (NDCG@K) [13]. HR@K and NDCG@K provide performance evaluations of RS in the face of data sparsity and cold-start problem, helping to determine whether the model can still provide effective recommendations under these challenges. HR@K measures the proportion of correctly recommended items in the Top-K list that match the user’s actual interactions (positive items). It emphasizes accuracy, ranging from 0 to 1, with higher values indicating better performance. NDCG@K evaluates the balance between item relevance and ranking within the Top-K recommendation list, focusing on the orderliness of the predicted results. Its numerical range and significance are consistent with HR@K. To compute HR@K and NDCG@K, we choose 100 items as negative samples (i.e., items that the user has not interacted with) and combine them with positive instances. By evaluating these combinations, we can evaluate the model’s performance based on the metrics. The calculations and interpretations of these metrics are as follows: (12) where, N denotes the number of test users. The variable hits ( i , j ) represents whether the item j predicted by the model for the user i is positive. Specifically, hits ( i , j ) takes the value 1 if the item is positive and 0 otherwise. (13)(14) where the Discounted Cumulative Gain (DCG@K) quantifies the relevance of the items and their positions within the recommendation list. It evaluates the balance between item relevance and ranking order among the Top-K recommended items. IDCG@K refers to the Ideal situation (the possible maximum value) of DCG@K. Parameter Settings: When adjusting model parameters, we defined the following ranges: GNN layers (L): { 1 , 2 , 3 , 4 } , regularization coefficient (λ): { 0 . 005 , 0 . 0005 , 0 . 00005 } , embedding dimension (d): { 4 , 8 , 16 } , and heterogeneous relation attention encoding hidden dimension (H): { 2 , 4 , 8 , 16 } . To ensure robustness, we performed 10 iterations of all tasks and present the average results. RQ1-Performance evaluation We conducted a comprehensive comparison of various baselines on three datasets. The main results are presented in Table 2. Based on the experimental findings, it is evident that CoHet4Rec surpasses the current SOTA models in predicting the Top-5 recommended items. Compared to the worst-performing model, CoHet4Rec achieved average relative improvements of 31.88% and 38.39% in HR@5 and NDCG@5, respectively. Furthermore, compared to the second-best model, we achieved average relative improvements of 3.19% in HR@5 and 3.81% in NDCG@5. Download: PPT PowerPoint slide PNG larger image TIFF original image Table 2. The performance of various models was compared using HR@K and NDCG@K metrics. https://doi.org/10.1371/journal.pone.0313772.t002 The advantages of CoHet4Rec are supported by: (1) Enhanced Interaction Learning: CoHet4Rec incorporates social and item-category relationship information, enhancing the model’s capability to learn interactions between users and items, which improves recommendation performance. Comparative analysis with models focusing primarily on social recommendations confirms the performance enhancement of CoHet4Rec. (2) Effective Semantic Integration: The heterogeneous relation attention and aggregation of HIN enable effective integration of semantic information from interactions, social connections, and item-category relationships. This facilitates better modeling of user and item features. Compared to disentangled graph RS, CoHet4Rec employs a specifically designed heterogeneous relation attention-enhanced network structure to capture the relevance between users and items, leading to improved performance. GNN-based social recommendation models outperform attention-based solutions, showcasing the effectiveness of using a multi-hop graph structure for embedding propagation to model social relationships. CoHet4Rec, with its designed heterogeneous relation attention-enhanced network structure, can learn potential connections by considering multi-hop neighbors from user and item domains. Table 2 also shows significant performance improvements of CoHet4Rec compared to other competing models in Top-K (K = 10, 20) recommendation scenarios. This further validates the exceptional ranking capabilities of our framework. As K increases, we also observe an improvement in recommendation accuracy. RQ2-Component ablation study We created three variants to analyze the roles of key components in CoHet4Rec and verify their efficacy: w/o-GNN: This variant removes the GNN layers used by CoHet4Rec to propagate information among nodes. w/o-Attention: This variant eliminates the heterogeneous collaborative neighbor attention used by CoHet4Rec to learn weights for the heterogeneous relationships among user/item/category neighbors. w/o-LN: This variant omits the layer normalization applied in each propagation layer by CoHet4Rec. The performance of various methods, including CoHet4Rec and its variants, is presented in Table 3 across three datasets. we draw the following conclusions and insights: Download: PPT PowerPoint slide PNG larger image TIFF original image Table 3. Component ablation study. https://doi.org/10.1371/journal.pone.0313772.t003 (1) Superior Performance of CoHet4Rec: CoHet4Rec consistently outperforms its three variants, emerging as the top-performing method. The ‘w/o-GNN’ variant shows the weakest performance, highlighting the significant contribution of the GNN architecture to the model’s effectiveness. This underscores the importance of incorporating GNN in CoHet4Rec for superior performance. (2) Importance of Attention Mechanism: The performance of the ‘w/o-Attention’ variant declines compared to CoHet4Rec, indicating the crucial role of the heterogeneous relationship-enhanced attention encoder in exploring potential collaborative relationships between neighbor nodes. (3) Role of Layer Normalization: Comparing with the ‘w/o-LN’ variant, it is evident that layer normalization is vital in the training process. It stabilizes gradient generation during model training, prevents premature over-smoothing in GNN, and thus significantly impacts the model’s performance. RQ3-CHIN ablation study To explore the influence of various auxiliary relationships on CoHet4Rec’s performance, we examined three variants of auxiliary relationships: w/o-S: Social matrix S is removed from CHIN. w/o-T: Item-category relationship matrix T is removed. w/o-ST: Both S and T are removed. We evaluated the performance on three datasets with varying Top-K settings (Fig 4) and derived the following key findings and conclusions: Download: PPT PowerPoint slide PNG larger image TIFF original image Fig 4. The influence of various CHIN on the performance of CoHet4Rec across three datasets. https://doi.org/10.1371/journal.pone.0313491.g004 (1) Positive impact of auxiliary relationships: Adding auxiliary heterogeneous relationships consistently improves model performance. This may be due to the benefits of integrating multiple heterogeneous semantics into the representation, allowing CoHet4Rec to explore collaborative relationships within diverse CHIN. (2) Dataset-specific observations: In Ciao and Epinions, the ‘w/o-T’ variant performs sub-optimally, highlighting the benefits of utilizing social context signals for user preference learning. In Yelp, however, ‘w/o-T’ performs the worst in most cases, even lower than ‘w/o-ST’, indicating some inconsistencies. (3) Worst performance of ‘w/o-ST’: The ‘w/o-ST’ variant consistently performs the worst in most cases, reinforcing that combining heterogeneous relationships from the user/item domain enhances recommendation accuracy. However, similar inconsistencies are observed in Yelp as noted in the second point. The inconsistencies mentioned in points (2) and (3) may be due to the varying impact of alleviating data sparsity in collaborative relationships across datasets. Adding too many collaborative relationships might introduce unnecessary noise. For example, while we assume a relationship between user interests and social connections, ConsisRec [28] highlights the often overlooked aspect of social inconsistency in existing methods. Specifically, social inconsistency indicates that the interests of users with social connections may not necessarily align. Aggregating information from inconsistent social neighbors together can undermine the ability of GNN to describe information that is beneficial for recommendation. This suggests that social relationships alone may not be sufficient for accurately predicting user ratings, and other factors need to be considered in the recommendation process. Similarly, item-category relationships may not always be relevant for link prediction. RQ4-Robustness in sparse data scenarios To investigate the robustness of CoHet4Rec in sparse data scenarios, we first ranked all users by their interaction density and social connection density in ascending order. Subsequently, we divided them into quartiles (0–25%, 25–50%, 50–75%, and 75–100%), each containing an equal number of users. Fig 5 depicts the average interactions and social connection count for each user group. Download: PPT PowerPoint slide PNG larger image TIFF original image Fig 5. Model performance under sparse data. https://doi.org/10.1371/journal.pone.0313491.g005 It is evident that CoHet4Rec performs better than the baseline in different types of sparse data scenarios, indicating the robustness of our model in handling sparse user behavior data. Coupled with the results in Table 2, it is clear that our model surpasses some SOTA models even under sparse data conditions. This further confirms the effectiveness of CoHet4Rec in mitigating cold-start problem by leveraging external knowledge from both user and item domains in situations of data scarcity. RQ5-Model efficiency study We assessed the models’ time cost from two perspectives: average training and testing runtime. The experiments compared our proposed CoHet4Rec model with the SOTA baseline DGNN [37]. Fig 6 shows the average training/testing runtime per epoch. Download: PPT PowerPoint slide PNG larger image TIFF original image Fig 6. Time cost statistics for model training and testing phases. https://doi.org/10.1371/journal.pone.0313491.g006 From the results, we observed that our non-disentangled CHIN algorithm achieves faster model training with nearly identical testing time compared to the disentangled RS DGNN. Specifically, the average training time was reduced by 21.74% across three datasets. This reduction is due to DGNN’s decoupled hierarchical message propagation approach with multiple relationships, which leads to time-consuming message propagation. Additionally, larger G results in longer training times. RQ6-Parameter selection We investigated the hyper-parameters affecting CoHet4Rec’s performance (Fig 7). By analyzing these relationships, we gain valuable insights into how hyper-parameter selection impacts CoHet4Rec’s performance, guiding effective tuning and optimization of the model’s parameters. Download: PPT PowerPoint slide PNG larger image TIFF original image Fig 7. Results of hyper-parameter ablation experiments on three datasets. https://doi.org/10.1371/journal.pone.0313491.g007 GNN layers (L): Fig 7 indicates that propagating embeddings between 2-hop neighbor nodes improves performance, highlighting the efficacy of encoding higher-order relationships. However, stacking more layers leads to performance degradation due to the over-smoothing problem. Regularization coefficient (λ): Adding a regularization term in the layer normalization process helps prevent over-fitting. The performance decreases when λ is excessively large or small, emphasizing the importance of selecting an appropriate λ range to optimize performance. Embedding dimension (d): The model performs best when d=16. Performance deteriorates as the d increases. Our model adopts smaller d while maintaining high performance, facilitating deployment in real-world scenarios. Hidden dimension of heterogeneous relationship attention encoding (H): CoHet4Rec captures the semantics between multiple heterogeneous connections by exploring collaborative factors among heterogeneous relationships. When H=8, the performance is optimal. Experimental setup Datasets: In this paper, we performed experiments and analysis using three widely-used social RS datasets (Ciao, Epinions and Yelp) which were collected from various online review systems in real-world. Ciao and Epinions are available at https://www.cse.msu.edu/ tangjili/ datasetcode/truststudy.htm, Yelp is available at https://www.yelp.com/dataset. Table 1 provides comprehensive statistical information about these datasets. Download: PPT PowerPoint slide PNG larger image TIFF original image Table 1. Statistics of datasets. https://doi.org/10.1371/journal.pone.0313772.t001 Baselines: Following [37], we comprehensively evaluate the performance of CoHet4Rec with 15 baselines from 7 different types: Graph-CF RS GCCF [41]: A linear residual GCN-based CF recommendation model. It aims to solve the problems of GCN CF models in terms of training difficulties, nonlinear activation functions, and deep stacking. NGCF [42]: This CF model utilizes graph convolution to leverage the user-item graph structure. It effectively represents and captures high-order connectivity relationships by propagating features on the graph. GNN-based RS GraphRec [8]: A GNN-based social RS algorithm that can simultaneously consider the influence of interaction and related viewpoints. DiffNet [43]: Utilizes graph information propagation to establish a hierarchical diffusion architecture for modeling users’ social relationships. It simulates dynamic social diffusion to capture recursive social influence. MHCN [31]: A hypergraph convolutional networks-based social RS that utilizes high-order user relationships to improve recommendation quality. Disentangled graph RS DGNN [37]: It explores social recommendation by learning disentangled heterogeneous factors and improves social recommendation by disentangling the heterogeneous relationships between users and item domains. DGCF [24]: It focuses particularly on user-item relationships with a finer granularity of user intent and models the intention distribution of each user-item interaction and iteratively refines the intention aware interaction graph and representation. DisenHAN [36]: A disentangled heterogeneous GAT for Top-K recommendation. It learns user/item representations for disentanglement in HINs from different perspectives. Time-aware social RS DGRec [44]: It models dynamic user behavior using RNN and contextual social impact using GAT. Knowledge-aware RS KGAT [25]: This model employs attention mechanisms to aggregate information from both interactions and the item knowledge. Attention-based social RS SAMN [30]: In this model, a two-stage attention mechanism is employed to capture user influence and identify relevant friends for preference modeling. It introduces a friend-style attention to explore the social influence between users. EATNN [45]: It utilizes attention mechanisms to integrate interactive and social information. It combines an optimization scheme to implement a multitasking learning framework. Heterogeneous graph embedding-based RS HERec [15]: A HIN embedding-based method. It learns node embeddings by generating meaningful node sequences through a random walk strategy based on meta-path. HGT [14]: In this method, specific parameters are designed to capture the heterogeneous attention on each edge, taking into account the node and edge types. This enables the model to maintain distinct embeddings for various types of nodes and edges. HAN [27]: A attention mechanism-based heterogeneous GNN method. It considers the diversity of nodes and edges in heterogeneous networks, adopts attention mechanisms at both node and semantic levels, and generates node embedding representations through hierarchical aggregation of node features. Evaluation Metrics: In our study, CoHet4Rec focuses on providing recommendations of the Top-K items for users. To evaluate and compare the models, we use two widely adopted metrics in RS: Hits Ratio (HR@K) and Normalized Discounted Cumulative Gain (NDCG@K) [13]. HR@K and NDCG@K provide performance evaluations of RS in the face of data sparsity and cold-start problem, helping to determine whether the model can still provide effective recommendations under these challenges. HR@K measures the proportion of correctly recommended items in the Top-K list that match the user’s actual interactions (positive items). It emphasizes accuracy, ranging from 0 to 1, with higher values indicating better performance. NDCG@K evaluates the balance between item relevance and ranking within the Top-K recommendation list, focusing on the orderliness of the predicted results. Its numerical range and significance are consistent with HR@K. To compute HR@K and NDCG@K, we choose 100 items as negative samples (i.e., items that the user has not interacted with) and combine them with positive instances. By evaluating these combinations, we can evaluate the model’s performance based on the metrics. The calculations and interpretations of these metrics are as follows: (12) where, N denotes the number of test users. The variable hits ( i , j ) represents whether the item j predicted by the model for the user i is positive. Specifically, hits ( i , j ) takes the value 1 if the item is positive and 0 otherwise. (13)(14) where the Discounted Cumulative Gain (DCG@K) quantifies the relevance of the items and their positions within the recommendation list. It evaluates the balance between item relevance and ranking order among the Top-K recommended items. IDCG@K refers to the Ideal situation (the possible maximum value) of DCG@K. Parameter Settings: When adjusting model parameters, we defined the following ranges: GNN layers (L): { 1 , 2 , 3 , 4 } , regularization coefficient (λ): { 0 . 005 , 0 . 0005 , 0 . 00005 } , embedding dimension (d): { 4 , 8 , 16 } , and heterogeneous relation attention encoding hidden dimension (H): { 2 , 4 , 8 , 16 } . To ensure robustness, we performed 10 iterations of all tasks and present the average results. RQ1-Performance evaluation We conducted a comprehensive comparison of various baselines on three datasets. The main results are presented in Table 2. Based on the experimental findings, it is evident that CoHet4Rec surpasses the current SOTA models in predicting the Top-5 recommended items. Compared to the worst-performing model, CoHet4Rec achieved average relative improvements of 31.88% and 38.39% in HR@5 and NDCG@5, respectively. Furthermore, compared to the second-best model, we achieved average relative improvements of 3.19% in HR@5 and 3.81% in NDCG@5. Download: PPT PowerPoint slide PNG larger image TIFF original image Table 2. The performance of various models was compared using HR@K and NDCG@K metrics. https://doi.org/10.1371/journal.pone.0313772.t002 The advantages of CoHet4Rec are supported by: (1) Enhanced Interaction Learning: CoHet4Rec incorporates social and item-category relationship information, enhancing the model’s capability to learn interactions between users and items, which improves recommendation performance. Comparative analysis with models focusing primarily on social recommendations confirms the performance enhancement of CoHet4Rec. (2) Effective Semantic Integration: The heterogeneous relation attention and aggregation of HIN enable effective integration of semantic information from interactions, social connections, and item-category relationships. This facilitates better modeling of user and item features. Compared to disentangled graph RS, CoHet4Rec employs a specifically designed heterogeneous relation attention-enhanced network structure to capture the relevance between users and items, leading to improved performance. GNN-based social recommendation models outperform attention-based solutions, showcasing the effectiveness of using a multi-hop graph structure for embedding propagation to model social relationships. CoHet4Rec, with its designed heterogeneous relation attention-enhanced network structure, can learn potential connections by considering multi-hop neighbors from user and item domains. Table 2 also shows significant performance improvements of CoHet4Rec compared to other competing models in Top-K (K = 10, 20) recommendation scenarios. This further validates the exceptional ranking capabilities of our framework. As K increases, we also observe an improvement in recommendation accuracy. RQ2-Component ablation study We created three variants to analyze the roles of key components in CoHet4Rec and verify their efficacy: w/o-GNN: This variant removes the GNN layers used by CoHet4Rec to propagate information among nodes. w/o-Attention: This variant eliminates the heterogeneous collaborative neighbor attention used by CoHet4Rec to learn weights for the heterogeneous relationships among user/item/category neighbors. w/o-LN: This variant omits the layer normalization applied in each propagation layer by CoHet4Rec. The performance of various methods, including CoHet4Rec and its variants, is presented in Table 3 across three datasets. we draw the following conclusions and insights: Download: PPT PowerPoint slide PNG larger image TIFF original image Table 3. Component ablation study. https://doi.org/10.1371/journal.pone.0313772.t003 (1) Superior Performance of CoHet4Rec: CoHet4Rec consistently outperforms its three variants, emerging as the top-performing method. The ‘w/o-GNN’ variant shows the weakest performance, highlighting the significant contribution of the GNN architecture to the model’s effectiveness. This underscores the importance of incorporating GNN in CoHet4Rec for superior performance. (2) Importance of Attention Mechanism: The performance of the ‘w/o-Attention’ variant declines compared to CoHet4Rec, indicating the crucial role of the heterogeneous relationship-enhanced attention encoder in exploring potential collaborative relationships between neighbor nodes. (3) Role of Layer Normalization: Comparing with the ‘w/o-LN’ variant, it is evident that layer normalization is vital in the training process. It stabilizes gradient generation during model training, prevents premature over-smoothing in GNN, and thus significantly impacts the model’s performance. RQ3-CHIN ablation study To explore the influence of various auxiliary relationships on CoHet4Rec’s performance, we examined three variants of auxiliary relationships: w/o-S: Social matrix S is removed from CHIN. w/o-T: Item-category relationship matrix T is removed. w/o-ST: Both S and T are removed. We evaluated the performance on three datasets with varying Top-K settings (Fig 4) and derived the following key findings and conclusions: Download: PPT PowerPoint slide PNG larger image TIFF original image Fig 4. The influence of various CHIN on the performance of CoHet4Rec across three datasets. https://doi.org/10.1371/journal.pone.0313491.g004 (1) Positive impact of auxiliary relationships: Adding auxiliary heterogeneous relationships consistently improves model performance. This may be due to the benefits of integrating multiple heterogeneous semantics into the representation, allowing CoHet4Rec to explore collaborative relationships within diverse CHIN. (2) Dataset-specific observations: In Ciao and Epinions, the ‘w/o-T’ variant performs sub-optimally, highlighting the benefits of utilizing social context signals for user preference learning. In Yelp, however, ‘w/o-T’ performs the worst in most cases, even lower than ‘w/o-ST’, indicating some inconsistencies. (3) Worst performance of ‘w/o-ST’: The ‘w/o-ST’ variant consistently performs the worst in most cases, reinforcing that combining heterogeneous relationships from the user/item domain enhances recommendation accuracy. However, similar inconsistencies are observed in Yelp as noted in the second point. The inconsistencies mentioned in points (2) and (3) may be due to the varying impact of alleviating data sparsity in collaborative relationships across datasets. Adding too many collaborative relationships might introduce unnecessary noise. For example, while we assume a relationship between user interests and social connections, ConsisRec [28] highlights the often overlooked aspect of social inconsistency in existing methods. Specifically, social inconsistency indicates that the interests of users with social connections may not necessarily align. Aggregating information from inconsistent social neighbors together can undermine the ability of GNN to describe information that is beneficial for recommendation. This suggests that social relationships alone may not be sufficient for accurately predicting user ratings, and other factors need to be considered in the recommendation process. Similarly, item-category relationships may not always be relevant for link prediction. RQ4-Robustness in sparse data scenarios To investigate the robustness of CoHet4Rec in sparse data scenarios, we first ranked all users by their interaction density and social connection density in ascending order. Subsequently, we divided them into quartiles (0–25%, 25–50%, 50–75%, and 75–100%), each containing an equal number of users. Fig 5 depicts the average interactions and social connection count for each user group. Download: PPT PowerPoint slide PNG larger image TIFF original image Fig 5. Model performance under sparse data. https://doi.org/10.1371/journal.pone.0313491.g005 It is evident that CoHet4Rec performs better than the baseline in different types of sparse data scenarios, indicating the robustness of our model in handling sparse user behavior data. Coupled with the results in Table 2, it is clear that our model surpasses some SOTA models even under sparse data conditions. This further confirms the effectiveness of CoHet4Rec in mitigating cold-start problem by leveraging external knowledge from both user and item domains in situations of data scarcity. RQ5-Model efficiency study We assessed the models’ time cost from two perspectives: average training and testing runtime. The experiments compared our proposed CoHet4Rec model with the SOTA baseline DGNN [37]. Fig 6 shows the average training/testing runtime per epoch. Download: PPT PowerPoint slide PNG larger image TIFF original image Fig 6. Time cost statistics for model training and testing phases. https://doi.org/10.1371/journal.pone.0313491.g006 From the results, we observed that our non-disentangled CHIN algorithm achieves faster model training with nearly identical testing time compared to the disentangled RS DGNN. Specifically, the average training time was reduced by 21.74% across three datasets. This reduction is due to DGNN’s decoupled hierarchical message propagation approach with multiple relationships, which leads to time-consuming message propagation. Additionally, larger G results in longer training times. RQ6-Parameter selection We investigated the hyper-parameters affecting CoHet4Rec’s performance (Fig 7). By analyzing these relationships, we gain valuable insights into how hyper-parameter selection impacts CoHet4Rec’s performance, guiding effective tuning and optimization of the model’s parameters. Download: PPT PowerPoint slide PNG larger image TIFF original image Fig 7. Results of hyper-parameter ablation experiments on three datasets. https://doi.org/10.1371/journal.pone.0313491.g007 GNN layers (L): Fig 7 indicates that propagating embeddings between 2-hop neighbor nodes improves performance, highlighting the efficacy of encoding higher-order relationships. However, stacking more layers leads to performance degradation due to the over-smoothing problem. Regularization coefficient (λ): Adding a regularization term in the layer normalization process helps prevent over-fitting. The performance decreases when λ is excessively large or small, emphasizing the importance of selecting an appropriate λ range to optimize performance. Embedding dimension (d): The model performs best when d=16. Performance deteriorates as the d increases. Our model adopts smaller d while maintaining high performance, facilitating deployment in real-world scenarios. Hidden dimension of heterogeneous relationship attention encoding (H): CoHet4Rec captures the semantics between multiple heterogeneous connections by exploring collaborative factors among heterogeneous relationships. When H=8, the performance is optimal. Conclusion and future work This paper presents CoHet4Rec, a novel social RS that leverages GNNs and CHIN. By integrating heterogeneous relationship attention units into the GNN framework, CoHet4Rec effectively captures and utilizes collaborative relationships from both the user and item domains. Extensive comparative and ablation experiments show the effectiveness of CoHet4Rec. Although CoHet4Rec demonstrates commendable performance in handling multiple collaborative relationships, deeper exploration is still needed for fine-grained processing of these relationships, such as introducing item-item relationships and more detailed forms of user-item interactions (e.g., different ratings). Additionally, exploring new neighbor sampling based message propagation and aggregation schemes will further reduce the training time of the model, which is crucial for practical deployment. Finally, further extending the model to federated learning will help in protecting user privacy. TI - CoHet4Rec: A recommendation for collaborative heterogeneous information networks JF - PLoS ONE DO - 10.1371/journal.pone.0313491 DA - 2025-04-09 UR - https://www.deepdyve.com/lp/public-library-of-science-plos-journal/cohet4rec-a-recommendation-for-collaborative-heterogeneous-information-Oarn7uAKWm SP - e0313491 VL - 20 IS - 4 DP - DeepDyve ER -