Your browser doesn't support javascript.
loading
Show: 20 | 50 | 100
Results 1 - 1 de 1
Filter
Add more filters










Database
Language
Publication year range
1.
Article in English | MEDLINE | ID: mdl-35594237

ABSTRACT

Graph classification plays an important role in a wide range of applications from biological prediction to social analysis. Traditional graph classification models built on graph kernels are hampered by the challenge of poor generalization as they are heavily dependent on the dedicated design of handcrafted features. Recently, graph neural networks (GNNs) become a new class of tools for analyzing graph data and have achieved promising performance. However, it is necessary to collect a large number of labeled graph data for training an accurate GNN, which is often unaffordable in real-world applications. Therefore, it is an open question to build GNNs under the condition of few-shot learning where only a few labeled graphs are available. In this article, we introduce a new Structure-aware Prototypical Neural Process (SPNP for short) for a few-shot graph classification. Specifically, at the encoding stage, SPNP first employs GNNs to capture graph structure information. Then, SPNP incorporates such structural priors into the latent path and the deterministic path for representing stochastic processes. At the decoding stage, SPNP uses a new prototypical decoder to define a metric space where unseen graphs can be predicted effectively. The proposed decoder, which contains a self-attention mechanism to learn the intraclass dependence between graphs, can enhance the class-level representations, especially for new classes. Furthermore, benefited from such a flexible encoding-decoding architecture, SPNP can directly map the context samples to a predictive distribution without any complicated operations used in previous methods. Extensive experiments demonstrate that SPNP achieves consistent and significant improvements over state-of-the-art methods. Further discussions are provided toward model efficiency and more detailed analysis.

SELECTION OF CITATIONS
SEARCH DETAIL
...