Entropy and distance of random graphs with application to structural pattern recognition
The notion of a random graph is formally defined. It deals with both the probabilistic and the structural aspects of relational data. By interpreting an ensemble of attributed graphs as the outcomes of a random graph, we can use its lower order distribution to characterize the ensemble. To reflect t...
Veröffentlicht in: | IEEE transactions on pattern analysis and machine intelligence. - 1979. - 7(1985), 5 vom: 01. Mai, Seite 599-609 |
---|---|
1. Verfasser: | |
Weitere Verfasser: | |
Format: | Aufsatz |
Sprache: | English |
Veröffentlicht: |
1985
|
Zugriff auf das übergeordnete Werk: | IEEE transactions on pattern analysis and machine intelligence |
Schlagworte: | Journal Article |
Zusammenfassung: | The notion of a random graph is formally defined. It deals with both the probabilistic and the structural aspects of relational data. By interpreting an ensemble of attributed graphs as the outcomes of a random graph, we can use its lower order distribution to characterize the ensemble. To reflect the variability of a random graph, Shannon's entropy measure is used. To synthesize an ensemble of attributed graphs into the distribution of a random graph (or a set of distributions), we propose a distance measure between random graphs based on the minimum change of entropy before and after their merging. When the ensemble contains more than one class of pattern graphs, the synthesis process yields distributions corresponding to various classes. This process corresponds to unsupervised learning in pattern classification. Using the maximum likelihood rule and the probability computed for the pattern graph, based on its matching with the random graph distributions of different classes, we can classify the pattern graph to a class |
---|---|
Beschreibung: | Date Completed 02.10.2012 Date Revised 12.11.2019 published: Print Citation Status PubMed-not-MEDLINE |
ISSN: | 1939-3539 |