Prediction models with graph kernel regularization for network data
© 2022 Informa UK Limited, trading as Taylor & Francis Group.
Publié dans: | Journal of applied statistics. - 1991. - 50(2023), 6 vom: 30., Seite 1400-1417 |
---|---|
Auteur principal: | |
Autres auteurs: | , |
Format: | Article en ligne |
Langue: | English |
Publié: |
2023
|
Accès à la collection: | Journal of applied statistics |
Sujets: | Journal Article Graph regularization kernel function node effect prediction regression |
Résumé: | © 2022 Informa UK Limited, trading as Taylor & Francis Group. Traditional regression methods typically consider only covariate information and assume that the observations are mutually independent samples. However, samples usually come from individuals connected by a network in many modern applications. We present a risk minimization formulation for learning from both covariates and network structure in the context of graph kernel regularization. The formulation involves a loss function with a penalty term. This penalty can be used not only to encourage similarity between linked nodes but also lead to improvement over traditional regression models. Furthermore, the penalty can be used with many loss-based predictive methods, such as linear regression with squared loss and logistic regression with log-likelihood loss. Simulations to evaluate the performance of this model in the cases of low dimensions and high dimensions show that our proposed approach outperforms all other benchmarks. We verify this for uniform graph, nonuniform graph, balanced-sample, and unbalanced-sample datasets. The approach was applied to predicting the response values on a 'follow' social network of Tencent Weibo users and on two citation networks (Cora and CiteSeer). Each instance verifies that the proposed method combining covariate information and link structure with the graph kernel regularization can improve predictive performance |
---|---|
Description: | Date Revised 11.04.2023 published: Electronic-eCollection Citation Status PubMed-not-MEDLINE |
ISSN: | 0266-4763 |
DOI: | 10.1080/02664763.2022.2028745 |