Prediction models with graph kernel regularization for network data

© 2022 Informa UK Limited, trading as Taylor & Francis Group.

Détails bibliographiques
Publié dans:Journal of applied statistics. - 1991. - 50(2023), 6 vom: 30., Seite 1400-1417
Auteur principal: Liu, Jie (Auteur)
Autres auteurs: Chen, Haojie, Yang, Yang
Format: Article en ligne
Langue:English
Publié: 2023
Accès à la collection:Journal of applied statistics
Sujets:Journal Article Graph regularization kernel function node effect prediction regression
Description
Résumé:© 2022 Informa UK Limited, trading as Taylor & Francis Group.
Traditional regression methods typically consider only covariate information and assume that the observations are mutually independent samples. However, samples usually come from individuals connected by a network in many modern applications. We present a risk minimization formulation for learning from both covariates and network structure in the context of graph kernel regularization. The formulation involves a loss function with a penalty term. This penalty can be used not only to encourage similarity between linked nodes but also lead to improvement over traditional regression models. Furthermore, the penalty can be used with many loss-based predictive methods, such as linear regression with squared loss and logistic regression with log-likelihood loss. Simulations to evaluate the performance of this model in the cases of low dimensions and high dimensions show that our proposed approach outperforms all other benchmarks. We verify this for uniform graph, nonuniform graph, balanced-sample, and unbalanced-sample datasets. The approach was applied to predicting the response values on a 'follow' social network of Tencent Weibo users and on two citation networks (Cora and CiteSeer). Each instance verifies that the proposed method combining covariate information and link structure with the graph kernel regularization can improve predictive performance
Description:Date Revised 11.04.2023
published: Electronic-eCollection
Citation Status PubMed-not-MEDLINE
ISSN:0266-4763
DOI:10.1080/02664763.2022.2028745