Movatterモバイル変換


[0]ホーム

URL:


Jump to content
WikipediaThe Free Encyclopedia
Search

Graph neural network

From Wikipedia, the free encyclopedia
Class of artificial neural networks
Part of a series on
Machine learning
anddata mining

Graph neural networks (GNN) are specializedartificial neural networks that are designed for tasks whose inputs aregraphs.[1][2][3][4][5]

One prominent example is molecular drug design.[6][7][8] Each input sample is a graph representation of a molecule, where atoms form the nodes and chemical bonds between atoms form the edges. In addition to the graph representation, the input also includes known chemical properties for each of the atoms. Dataset samples may thus differ in length, reflecting the varying numbers of atoms in molecules, and the varying number of bonds between them. The task is to predict the efficacy of a given molecule for a specific medical application, like eliminatingE. coli bacteria.

The key design element of GNNs is the use ofpairwise message passing, such that graph nodes iteratively update their representations by exchanging information with their neighbors. Several GNN architectures have been proposed,[2][3][9][10][11] which implement different flavors of message passing,[12][13] started by recursive[2] or convolutional constructive[3] approaches. As of 2022[update], it is an open question whether it is possible to define GNN architectures "going beyond" message passing, or instead every GNN can be built on message passing over suitably defined graphs.[14]

Basic building blocks of a graph neural network (GNN).(1){\displaystyle (1)} Permutation equivariant layer.(2){\displaystyle (2)} Local pooling layer.(3){\displaystyle (3)} Global pooling (or readout) layer. Colors indicatefeatures.

In the more general subject of "geometricdeep learning", certain existing neural network architectures can be interpreted as GNNs operating on suitably defined graphs.[12] Aconvolutional neural network layer, in the context ofcomputer vision, can be considered a GNN applied to graphs whose nodes arepixels and only adjacent pixels are connected by edges in the graph. Atransformer layer, innatural language processing, can be considered a GNN applied tocomplete graphs whose nodes arewords or tokens in a passage ofnatural language text.

Relevant application domains for GNNs includenatural language processing,[15]social networks,[16]citation networks,[17]molecular biology,[18] chemistry,[19][20]physics[21] andNP-hardcombinatorial optimization problems.[22]

Open sourcelibraries implementing GNNs include PyTorch Geometric[23] (PyTorch), TensorFlow GNN[24] (TensorFlow), Deep Graph Library[25] (framework agnostic), jraph[26] (Google JAX), and GraphNeuralNetworks.jl[27]/GeometricFlux.jl[28] (Julia,Flux).

Architecture

[edit]

The architecture of a generic GNN implements the following fundamentallayers:[12]

  1. Permutation equivariant: a permutation equivariant layermaps a representation of a graph into an updated representation of the same graph. In the literature, permutation equivariant layers are implemented via pairwise message passing between graph nodes.[12][14] Intuitively, in a message passing layer, nodesupdate their representations byaggregating themessages received from their immediate neighbours. As such, each message passing layer increases the receptive field of the GNN by one hop.
  2. Local pooling: a localpooling layer coarsens the graph viadownsampling. Local pooling is used to increase the receptive field of a GNN, in a similar fashion to pooling layers inconvolutional neural networks. Examples includek-nearest neighbours pooling, top-k pooling,[29] and self-attention pooling.[30]
  3. Global pooling: a global pooling layer, also known asreadout layer, provides fixed-size representation of the whole graph. The global pooling layer must be permutation invariant, such that permutations in the ordering of graph nodes and edges do not alter the final output.[31] Examples include element-wise sum, mean or maximum.

It has been demonstrated that GNNs cannot be more expressive than theWeisfeiler–Leman Graph Isomorphism Test.[32][33] In practice, this means that there exist different graph structures (e.g.,molecules with the sameatoms but differentbonds) that cannot be distinguished by GNNs. More powerful GNNs operating on higher-dimension geometries such assimplicial complexes can be designed.[34][35][13] As of 2022[update], whether or not future architectures will overcome the message passing primitive is an open research question.[14]

Non-isomorphic graphs that cannot be distinguished by a GNN due to the limitations of the Weisfeiler-Lehman Graph Isomorphism Test. Colors indicate nodefeatures.

Message passing layers

[edit]
Node representation update in a Message Passing Neural Network (MPNN) layer. Nodex0{\displaystyle \mathbf {x} _{0}} receives messages sent by all of its immediate neighboursx1{\displaystyle \mathbf {x} _{1}} tox4{\displaystyle \mathbf {x} _{4}}. Messages are computing via the message functionψ{\displaystyle \psi }, which accounts for the features of both senders and receiver.

Message passing layers are permutation-equivariant layers mapping a graph into an updated representation of the same graph. Formally, they can be expressed as message passing neural networks (MPNNs).[12]

LetG=(V,E){\displaystyle G=(V,E)} be agraph, whereV{\displaystyle V} is the node set andE{\displaystyle E} is the edge set. LetNu{\displaystyle N_{u}} be theneighbourhood of some nodeuV{\displaystyle u\in V}. Additionally, letxu{\displaystyle \mathbf {x} _{u}} be thefeatures of nodeuV{\displaystyle u\in V}, andeuv{\displaystyle \mathbf {e} _{uv}} be the features of edge(u,v)E{\displaystyle (u,v)\in E}. An MPNNlayer can be expressed as follows:[12]

hu=ϕ(xu,vNuψ(xu,xv,euv)){\displaystyle \mathbf {h} _{u}=\phi \left(\mathbf {x} _{u},\bigoplus _{v\in N_{u}}\psi (\mathbf {x} _{u},\mathbf {x} _{v},\mathbf {e} _{uv})\right)}

whereϕ{\displaystyle \phi } andψ{\displaystyle \psi } aredifferentiable functions (e.g.,artificial neural networks), and{\displaystyle \bigoplus } is apermutationinvariantaggregation operator that can accept an arbitrary number of inputs (e.g., element-wise sum, mean, or max). In particular,ϕ{\displaystyle \phi } andψ{\displaystyle \psi } are referred to asupdate andmessage functions, respectively. Intuitively, in an MPNN computational block, graph nodesupdate their representations byaggregating themessages received from their neighbours.

The outputs of one or more MPNN layers are node representationshu{\displaystyle \mathbf {h} _{u}} for each nodeuV{\displaystyle u\in V} in the graph. Node representations can be employed for any downstream task, such as node/graphclassification or edge prediction.

Graph nodes in an MPNN update their representation aggregating information from their immediate neighbours. As such, stackingn{\displaystyle n} MPNN layers means that one node will be able to communicate with nodes that are at mostn{\displaystyle n} "hops" away. In principle, to ensure that every node receives information from every other node, one would need to stack a number of MPNN layers equal to the graphdiameter. However, stacking many MPNN layers may cause issues such as oversmoothing[36] and oversquashing.[37] Oversmoothing refers to the issue of node representations becoming indistinguishable. Oversquashing refers to the bottleneck that is created by squeezing long-range dependencies into fixed-size representations. Countermeasures such as skip connections[10][38] (as inresidual neural networks), gated update rules[39] and jumping knowledge[40] can mitigate oversmoothing. Modifying the final layer to be a fully-adjacent layer, i.e., by considering the graph as acomplete graph, can mitigate oversquashing in problems where long-range dependencies are required.[37]

Other "flavours" of MPNN have been developed in the literature,[12] such as graph convolutional networks[9] and graph attention networks,[11] whose definitions can be expressed in terms of the MPNN formalism.

Graph convolutional network

[edit]

The graph convolutional network (GCN) was first introduced byThomas Kipf andMax Welling in 2017.[9]

A GCN layer defines afirst-order approximation of a localized spectralfilter on graphs. GCNs can be understood as a generalization ofconvolutional neural networks to graph-structured data.

The formal expression of a GCN layer reads as follows:

H=σ(D~12A~D~12XΘ){\displaystyle \mathbf {H} =\sigma \left({\tilde {\mathbf {D} }}^{-{\frac {1}{2}}}{\tilde {\mathbf {A} }}{\tilde {\mathbf {D} }}^{-{\frac {1}{2}}}\mathbf {X} \mathbf {\Theta } \right)}

whereH{\displaystyle \mathbf {H} } is the matrix of node representationshu{\displaystyle \mathbf {h} _{u}},X{\displaystyle \mathbf {X} } is the matrix of node featuresxu{\displaystyle \mathbf {x} _{u}},σ(){\displaystyle \sigma (\cdot )} is anactivation function (e.g.,ReLU),A~{\displaystyle {\tilde {\mathbf {A} }}} is the graphadjacency matrix with the addition of self-loops,D~{\displaystyle {\tilde {\mathbf {D} }}} is the graphdegree matrix with the addition of self-loops, andΘ{\displaystyle \mathbf {\Theta } } is a matrix of trainable parameters.

In particular, letA{\displaystyle \mathbf {A} } be the graph adjacency matrix: then, one can defineA~=A+I{\displaystyle {\tilde {\mathbf {A} }}=\mathbf {A} +\mathbf {I} } andD~ii=jVA~ij{\displaystyle {\tilde {\mathbf {D} }}_{ii}=\sum _{j\in V}{\tilde {A}}_{ij}}, whereI{\displaystyle \mathbf {I} } denotes theidentity matrix. This normalization ensures that theeigenvalues ofD~12A~D~12{\displaystyle {\tilde {\mathbf {D} }}^{-{\frac {1}{2}}}{\tilde {\mathbf {A} }}{\tilde {\mathbf {D} }}^{-{\frac {1}{2}}}} are bounded in the range[0,1]{\displaystyle [0,1]}, avoidingnumerical instabilities andexploding/vanishing gradients.

A limitation of GCNs is that they do not allow multidimensional edge featureseuv{\displaystyle \mathbf {e} _{uv}}.[9] It is however possible to associate scalar weightswuv{\displaystyle w_{uv}} to each edge by imposingAuv=wuv{\displaystyle A_{uv}=w_{uv}}, i.e., by setting each nonzero entry in the adjacency matrix equal to the weight of the corresponding edge.

Graph attention network

[edit]

The graph attention network (GAT) was introduced byPetar Veličković et al. in 2018.[11]

Graph attention network is a combination of a GNN and an attention layer.The implementation of attention layer in graphical neural networks helps provide attention or focus to the important information from the data instead of focusing on the whole data.

A multi-head GAT layer can be expressed as follows:

hu=k=1Kσ(vNuαuvWkxv){\displaystyle \mathbf {h} _{u}={\overset {K}{\underset {k=1}{\Big \Vert }}}\sigma \left(\sum _{v\in N_{u}}\alpha _{uv}\mathbf {W} ^{k}\mathbf {x} _{v}\right)}

whereK{\displaystyle K} is the number ofattention heads,{\displaystyle {\Big \Vert }} denotesvector concatenation,σ(){\displaystyle \sigma (\cdot )} is anactivation function (e.g.,ReLU),αij{\displaystyle \alpha _{ij}} are attention coefficients, andWk{\displaystyle W^{k}} is a matrix of trainable parameters for thek{\displaystyle k}-th attention head.

For the final GAT layer, the outputs from each attention head are averaged before the application of the activation function. Formally, the final GAT layer can be written as:

hu=σ(1Kk=1KvNuαuvWkxv){\displaystyle \mathbf {h} _{u}=\sigma \left({\frac {1}{K}}\sum _{k=1}^{K}\sum _{v\in N_{u}}\alpha _{uv}\mathbf {W} ^{k}\mathbf {x} _{v}\right)}

Attention in Machine Learning is a technique that mimicscognitive attention. In the context of learning on graphs, the attention coefficientαuv{\displaystyle \alpha _{uv}} measureshow important is nodeuV{\displaystyle u\in V} to nodevV{\displaystyle v\in V}.

Normalized attention coefficients are computed as follows:

αuv=exp(LeakyReLU(aT[WxuWxveuv]))zNuexp(LeakyReLU(aT[WxuWxzeuz])){\displaystyle \alpha _{uv}={\frac {\exp({\text{LeakyReLU}}\left(\mathbf {a} ^{T}[\mathbf {W} \mathbf {x} _{u}\Vert \mathbf {W} \mathbf {x} _{v}\Vert \mathbf {e} _{uv}]\right))}{\sum _{z\in N_{u}}\exp({\text{LeakyReLU}}\left(\mathbf {a} ^{T}[\mathbf {W} \mathbf {x} _{u}\Vert \mathbf {W} \mathbf {x} _{z}\Vert \mathbf {e} _{uz}]\right))}}}

wherea{\displaystyle \mathbf {a} } is a vector of learnable weights,T{\displaystyle \cdot ^{T}} indicatestransposition,euv{\displaystyle \mathbf {e} _{uv}} are the edge features (if present), andLeakyReLU{\displaystyle {\text{LeakyReLU}}} is amodified ReLU activation function. Attention coefficients are normalized to make them easily comparable across different nodes.[11]

A GCN can be seen as a special case of a GAT where attention coefficients are not learnable, but fixed and equal to the edge weightswuv{\displaystyle w_{uv}}.

Gated graph sequence neural network

[edit]

The gated graph sequence neural network (GGS-NN) was introduced byYujia Li et al. in 2015.[39] The GGS-NN extends the GNN formulation by Scarselli et al.[2] to output sequences. The message passing framework is implemented as an update rule to agated recurrent unit (GRU) cell.

A GGS-NN can be expressed as follows:

hu(0)=xu0{\displaystyle \mathbf {h} _{u}^{(0)}=\mathbf {x} _{u}\,\Vert \,\mathbf {0} }
mu(l+1)=vNuΘhv{\displaystyle \mathbf {m} _{u}^{(l+1)}=\sum _{v\in N_{u}}\mathbf {\Theta } \mathbf {h} _{v}}
hu(l+1)=GRU(mu(l+1),hu(l)){\displaystyle \mathbf {h} _{u}^{(l+1)}={\text{GRU}}(\mathbf {m} _{u}^{(l+1)},\mathbf {h} _{u}^{(l)})}

where{\displaystyle \Vert } denotesvector concatenation,0{\displaystyle \mathbf {0} } is a vector of zeros,Θ{\displaystyle \mathbf {\Theta } } is a matrix of learnable parameters,GRU{\displaystyle {\text{GRU}}} is a GRU cell, andl{\displaystyle l} denotes the sequence index. In a GGS-NN, the node representations are regarded as the hidden states of a GRU cell. The initial node featuresxu(0){\displaystyle \mathbf {x} _{u}^{(0)}} arezero-padded up to the hidden state dimension of the GRU cell. The same GRU cell is used for updating representations for each node.

Local pooling layers

[edit]

Local pooling layers coarsen the graph via downsampling. Subsequently, several learnable local pooling strategies that have been proposed are presented.[31] For each case, the input is the initial graph represented by a matrixX{\displaystyle \mathbf {X} } of node features, and the graph adjacency matrixA{\displaystyle \mathbf {A} }. The output is the new matrixX{\displaystyle \mathbf {X} '}of node features, and the new graph adjacency matrixA{\displaystyle \mathbf {A} '}.

Top-k pooling

[edit]

We first set

y=Xpp{\displaystyle \mathbf {y} ={\frac {\mathbf {X} \mathbf {p} }{\Vert \mathbf {p} \Vert }}}

wherep{\displaystyle \mathbf {p} } is a learnableprojection vector. The projection vectorp{\displaystyle \mathbf {p} } computes a scalar projection value for each graph node.

The top-k pooling layer[29] can then be formalised as follows:

X=(Xsigmoid(y))i{\displaystyle \mathbf {X} '=(\mathbf {X} \odot {\text{sigmoid}}(\mathbf {y} ))_{\mathbf {i} }}
A=Ai,i{\displaystyle \mathbf {A} '=\mathbf {A} _{\mathbf {i} ,\mathbf {i} }}

wherei=topk(y){\displaystyle \mathbf {i} ={\text{top}}_{k}(\mathbf {y} )} is the subset of nodes with the top-k highest projection scores,{\displaystyle \odot } denotes element-wisematrix multiplication, andsigmoid(){\displaystyle {\text{sigmoid}}(\cdot )} is thesigmoid function. In other words, the nodes with the top-k highest projection scores are retained in the new adjacency matrixA{\displaystyle \mathbf {A} '}. Thesigmoid(){\displaystyle {\text{sigmoid}}(\cdot )} operation makes the projection vectorp{\displaystyle \mathbf {p} } trainable bybackpropagation, which otherwise would produce discrete outputs.[29]

Self-attention pooling

[edit]

We first set

y=GNN(X,A){\displaystyle \mathbf {y} ={\text{GNN}}(\mathbf {X} ,\mathbf {A} )}

whereGNN{\displaystyle {\text{GNN}}} is a generic permutation equivariant GNN layer (e.g., GCN, GAT, MPNN).

The Self-attention pooling layer[30] can then be formalised as follows:

X=(Xy)i{\displaystyle \mathbf {X} '=(\mathbf {X} \odot \mathbf {y} )_{\mathbf {i} }}
A=Ai,i{\displaystyle \mathbf {A} '=\mathbf {A} _{\mathbf {i} ,\mathbf {i} }}

wherei=topk(y){\displaystyle \mathbf {i} ={\text{top}}_{k}(\mathbf {y} )} is the subset of nodes with the top-k highest projection scores,{\displaystyle \odot } denoteselement-wise matrix multiplication.

The self-attention pooling layer can be seen as an extension of the top-k pooling layer. Differently from top-k pooling, the self-attention scores computed in self-attention pooling account both for the graph features and the graph topology.

Heterophilic Graph Learning

[edit]

Homophily principle, i.e., nodes with the same labels or similar attributes are more likely to be connected, has been commonly believed to be the main reason for the superiority of Graph Neural Networks (GNNs) over traditional Neural Networks (NNs) on graph-structured data, especially on node-level tasks.[41] However, recent work has identified a non-trivial set of datasets where GNN's performance compared to the NN's is not satisfactory.[42]Heterophily, i.e., low homophily, has been considered the main cause of this empirical observation.[43] People have begun to revisit and re-evaluate most existing graph models in the heterophily scenario across various kinds of graphs, e.g.,heterogeneous graphs,temporal graphs andhypergraphs. Moreover, numerous graph-related applications are found to be closely related to the heterophily problem, e.g.graph fraud/anomaly detection,graph adversarial attacks and robustness, privacy,federated learning andpoint cloud segmentation,graph clustering,recommender systems,generative models,link prediction,graph classification andcoloring, etc. In the past few years, considerable effort has been devoted to studying and addressing the heterophily issue in graph learning.[41][43][44]

Applications

[edit]

Protein folding

[edit]
See also:AlphaFold

Graph neural networks are one of the main building blocks ofAlphaFold, an artificial intelligence program developed byGoogle'sDeepMind for solving theprotein folding problem inbiology. AlphaFold achieved first place in severalCASP competitions.[45][46][40]

Social networks

[edit]
See also:Recommender system

Social networks are a major application domain for GNNs due to their natural representation associal graphs. GNNs are used to develop recommender systems based on bothsocial relations and item relations.[47][16]

Combinatorial optimization

[edit]
See also:Combinatorial optimization

GNNs are used as fundamental building blocks for several combinatorial optimization algorithms.[48] Examples include computingshortest paths orEulerian circuits for a given graph,[39] derivingchip placements superior or competitive to handcrafted human solutions,[49] and improving expert-designed branching rules inbranch and bound.[50]

Cyber security

[edit]
See also:Intrusion detection system

When viewed as a graph, a network of computers can be analyzed with GNNs for anomaly detection. Anomalies within provenance graphs often correlate to malicious activity within the network. GNNs have been used to identify these anomalies on individual nodes[51] and within paths[52] to detect malicious processes, or on the edge level[53] to detectlateral movement.

Water distribution networks

[edit]
See also:Water distribution system

Water distribution systems can be modelled as graphs, being then a straightforward application of GNN. This kind of algorithm has been applied to water demand forecasting,[54] interconnecting District Metered Areas(DMAs) to improve the forecasting capacity. Other application of this algorithm on water distribution modelling is the development of metamodels.[55]

Computer Vision

[edit]
See also:Computer vision

To represent an image as a graph structure, the image is first divided into multiple patches, each of which is treated as a node in the graph. Edges are then formed by connecting each node to its nearest neighbors based on spatial or feature similarity. This graph-based representation enables the application of graph learning models to visual tasks. The relational structure helps to enhance feature extraction and improve performance on image understanding.[56]

Text and NLP

[edit]
See also:Natural language processing

Graph-based representation of text helps to capture deeper semantic relationships between words. Many studies have used graph networks to enhance performance in various text processing tasks such as text classification, question answering, Neural Machine Translation (NMT), event extraction, fact verification, etc.[57]

References

[edit]
  1. ^Wu, Lingfei; Cui, Peng; Pei, Jian; Zhao, Liang (2022)."Graph Neural Networks: Foundations, Frontiers, and Applications".Springer Singapore: 725.
  2. ^abcdScarselli, Franco; Gori, Marco; Tsoi, Ah Chung; Hagenbuchner, Markus; Monfardini, Gabriele (2009). "The Graph Neural Network Model".IEEE Transactions on Neural Networks.20 (1):61–80.Bibcode:2009ITNN...20...61S.doi:10.1109/TNN.2008.2005605.ISSN 1941-0093.PMID 19068426.S2CID 206756462.
  3. ^abcMicheli, Alessio (2009). "Neural Network for Graphs: A Contextual Constructive Approach".IEEE Transactions on Neural Networks.20 (3):498–511.Bibcode:2009ITNN...20..498M.doi:10.1109/TNN.2008.2010350.ISSN 1045-9227.PMID 19193509.S2CID 17486263.
  4. ^Sanchez-Lengeling, Benjamin; Reif, Emily; Pearce, Adam; Wiltschko, Alex (2 September 2021)."A Gentle Introduction to Graph Neural Networks".Distill.6 (9) e33.doi:10.23915/distill.00033.ISSN 2476-0757.
  5. ^Daigavane, Ameya; Ravindran, Balaraman; Aggarwal, Gaurav (2 September 2021)."Understanding Convolutions on Graphs".Distill.6 (9) e32.doi:10.23915/distill.00032.ISSN 2476-0757.S2CID 239678898.
  6. ^Stokes, Jonathan M.; Yang, Kevin; Swanson, Kyle; Jin, Wengong; Cubillos-Ruiz, Andres; Donghia, Nina M.; MacNair, Craig R.; French, Shawn; Carfrae, Lindsey A.; Bloom-Ackermann, Zohar; Tran, Victoria M.; Chiappino-Pepe, Anush; Badran, Ahmed H.; Andrews, Ian W.; Chory, Emma J. (20 February 2020)."A Deep Learning Approach to Antibiotic Discovery".Cell.180 (4): 688–702.e13.doi:10.1016/j.cell.2020.01.021.ISSN 1097-4172.PMC 8349178.PMID 32084340.
  7. ^Yang, Kevin; Swanson, Kyle; Jin, Wengong; Coley, Connor; Eiden, Philipp; Gao, Hua; Guzman-Perez, Angel; Hopper, Timothy; Kelley, Brian (20 November 2019). "Analyzing Learned Molecular Representations for Property Prediction".arXiv:1904.01561 [cs.LG].
  8. ^Marchant, Jo (20 February 2020)."Powerful antibiotics discovered using AI".Nature.doi:10.1038/d41586-020-00018-3.PMID 33603175.
  9. ^abcdKipf, Thomas N; Welling, Max (2016). "Semi-supervised classification with graph convolutional networks".IEEE Transactions on Neural Networks.5 (1):61–80.arXiv:1609.02907.Bibcode:2009ITNN...20...61S.doi:10.1109/TNN.2008.2005605.PMID 19068426.S2CID 206756462.
  10. ^abHamilton, William; Ying, Rex; Leskovec, Jure (2017)."Inductive Representation Learning on Large Graphs"(PDF).Neural Information Processing Systems.31.arXiv:1706.02216 – via Stanford.
  11. ^abcdVeličković, Petar; Cucurull, Guillem; Casanova, Arantxa; Romero, Adriana; Liò, Pietro; Bengio, Yoshua (4 February 2018). "Graph Attention Networks".arXiv:1710.10903 [stat.ML].
  12. ^abcdefgBronstein, Michael M.; Bruna, Joan; Cohen, Taco; Veličković, Petar (4 May 2021). "Geometric Deep Learning: Grids, Groups, Graphs Geodesics and Gauges".arXiv:2104.13478 [cs.LG].
  13. ^abHajij, M.; Zamzmi, G.; Papamarkou, T.; Miolane, N.; Guzmán-Sáenz, A.; Ramamurthy, K. N.; Schaub, M. T. (2022). "Topological deep learning: Going beyond graph data".arXiv:2206.00606 [cs.LG].
  14. ^abcVeličković, Petar (2022). "Message passing all the way up".arXiv:2202.11097 [cs.LG].
  15. ^Wu, Lingfei; Chen, Yu; Shen, Kai; Guo, Xiaojie; Gao, Hanning; Li, Shucheng; Pei, Jian; Long, Bo (2023)."Graph Neural Networks for Natural Language Processing: A Survey".Foundations and Trends in Machine Learning.16 (2):119–328.arXiv:2106.06090.doi:10.1561/2200000096.ISSN 1941-0093.PMID 19068426.S2CID 206756462.
  16. ^abYing, Rex; He, Ruining; Chen, Kaifeng; Eksombatchai, Pong; Hamilton, William L.; Leskovec, Jure (2018).Graph Convolutional Neural Networks for Web-Scale Recommender Systems. pp. 974–983.arXiv:1806.01973.doi:10.1145/3219819.3219890.ISBN 978-1-4503-5552-0.S2CID 46949657.
  17. ^"Stanford Large Network Dataset Collection".snap.stanford.edu. Retrieved5 July 2021.
  18. ^Zhang, Weihang; Cui, Yang; Liu, Bowen; Loza, Martin; Park, Sung-Joon; Nakai, Kenta (5 April 2024)."HyGAnno: Hybrid graph neural network-based cell type annotation for single-cell ATAC sequencing data".Briefings in Bioinformatics.25 (3) bbae152.doi:10.1093/bib/bbae152.PMC 10998639.PMID 38581422.
  19. ^Gilmer, Justin; Schoenholz, Samuel S.; Riley, Patrick F.; Vinyals, Oriol; Dahl, George E. (17 July 2017)."Neural Message Passing for Quantum Chemistry".Proceedings of Machine Learning Research:1263–1272.arXiv:1704.01212.
  20. ^Coley, Connor W.; Jin, Wengong; Rogers, Luke; Jamison, Timothy F.; Jaakkola, Tommi S.; Green, William H.; Barzilay, Regina; Jensen, Klavs F. (2 January 2019)."A graph-convolutional neural network model for the prediction of chemical reactivity".Chemical Science.10 (2):370–377.doi:10.1039/C8SC04228D.ISSN 2041-6539.PMC 6335848.PMID 30746086.
  21. ^Qasim, Shah Rukh; Kieseler, Jan; Iiyama, Yutaro; Pierini, Maurizio Pierini (2019)."Learning representations of irregular particle-detector geometry with distance-weighted graph networks".The European Physical Journal C.79 (7): 608.arXiv:1902.07987.Bibcode:2019EPJC...79..608Q.doi:10.1140/epjc/s10052-019-7113-9.S2CID 88518244.
  22. ^Li, Zhuwen; Chen, Qifeng; Koltun, Vladlen (2018). "Text Simplification with Self-Attention-Based Pointer-Generator Networks".Neural Information Processing. Lecture Notes in Computer Science. Vol. 31. pp. 537–546.arXiv:1810.10659.doi:10.1007/978-3-030-04221-9_48.ISBN 978-3-030-04220-2.
  23. ^Matthias, Fey; Lenssen, Jan E. (2019). "Fast Graph Representation Learning with PyTorch Geometric".arXiv:1903.02428 [cs.LG].
  24. ^"Tensorflow GNN".GitHub. Retrieved30 June 2022.
  25. ^"Deep Graph Library (DGL)". Retrieved12 September 2024.
  26. ^"jraph".GitHub. Retrieved30 June 2022.
  27. ^Lucibello, Carlo (2021)."GraphNeuralNetworks.jl".GitHub. Retrieved21 September 2023.
  28. ^FluxML/GeometricFlux.jl, FluxML, 31 January 2024, retrieved3 February 2024
  29. ^abcGao, Hongyang; Ji, Shuiwang Ji (2019). "Graph U-Nets".arXiv:1905.05178 [cs.LG].
  30. ^abLee, Junhyun; Lee, Inyeop; Kang, Jaewoo (2019). "Self-Attention Graph Pooling".arXiv:1904.08082 [cs.LG].
  31. ^abLiu, Chuang; Zhan, Yibing; Li, Chang; Du, Bo; Wu, Jia; Hu, Wenbin; Liu, Tongliang; Tao, Dacheng (2022). "Graph Pooling for Graph Neural Networks: Progress, Challenges, and Opportunities".arXiv:2204.07321 [cs.LG].
  32. ^Douglas, B. L. (27 January 2011). "The Weisfeiler–Lehman Method and Graph Isomorphism Testing".arXiv:1101.5211 [math.CO].
  33. ^Xu, Keyulu; Hu, Weihua; Leskovec, Jure;Jegelka, Stefanie (22 February 2019). "How Powerful are Graph Neural Networks?".arXiv:1810.00826 [cs.LG].
  34. ^Bodnar, Christian; Frasca, Fabrizio; Guang Wang, Yu; Otter, Nina; Montúfar, Guido; Liò, Pietro; Bronstein, Michael (2021). "Weisfeiler and Lehman Go Topological: Message Passing Simplicial Networks".arXiv:2103.03212 [cs.LG].
  35. ^Grady, Leo; Polimeni, Jonathan (2011).Discrete Calculus: Applied Analysis on Graphs for Computational Science(PDF). Springer.
  36. ^Chen, Deli; Lin, Yankai; Li, Wei; Li, Peng; Zhou, Jie; Sun, Xu (2020). "Measuring and Relieving the Over-Smoothing Problem for Graph Neural Networks from the Topological View".Proceedings of the AAAI Conference on Artificial Intelligence.34 (4):3438–3445.arXiv:1909.03211.doi:10.1609/aaai.v34i04.5747.S2CID 202539008.
  37. ^abAlon, Uri; Yahav, Eran (2021). "On the Bottleneck of Graph Neural Networks and its Practical Implications".arXiv:2006.05205 [cs.LG].
  38. ^Xu, Keyulu; Zhang, Mozhi;Jegelka, Stephanie; Kawaguchi, Kenji (2021). "Optimization of Graph Neural Networks: Implicit Acceleration by Skip Connections and More Depth".arXiv:2105.04550 [cs.LG].
  39. ^abcLi, Yujia; Tarlow, Daniel; Brockschmidt, Mark; Zemel, Richard (2016). "Gated Graph Sequence Neural Networks".arXiv:1511.05493 [cs.LG].
  40. ^abXu, Keyulu; Li, Chengtao; Tian, Yonglong; Sonobe, Tomohiro; Kawarabayashi, Ken-ichi;Jegelka, Stefanie (2018). "Representation Learning on Graphs with Jumping Knowledge Networks".arXiv:1806.03536 [cs.LG].
  41. ^abLuan, Sitao; Hua, Chenqing; Lu, Qincheng; Ma, Liheng; Wu, Lirong; Wang, Xinyu; Xu, Minkai; Chang, Xiao-Wen; Precup, Doina; Ying, Rex; Li, Stan Z.; Tang, Jian; Wolf, Guy; Jegelka, Stefanie (2024). "The Heterophilic Graph Learning Handbook: Benchmarks, Models, Theoretical Analysis, Applications and Challenges".arXiv:2407.09618 [cs.LG].
  42. ^Luan, Sitao; Hua, Chenqing; Lu, Qincheng; Zhu, Jiaqi; Chang, Xiao-Wen; Precup, Doina (2024)."When do We Need Graph Neural Networks for Node Classification?". In Cherifi, Hocine; Rocha, Luis M.; Cherifi, Chantal; Donduran, Murat (eds.).Complex Networks & Their Applications XII. Studies in Computational Intelligence. Vol. 1141. Cham: Springer Nature Switzerland. pp. 37–48.doi:10.1007/978-3-031-53468-3_4.ISBN 978-3-031-53467-6.
  43. ^abLuan, Sitao; Hua, Chenqing; Lu, Qincheng; Zhu, Jiaqi; Zhao, Mingde; Zhang, Shuyuan; Chang, Xiao-Wen; Precup, Doina (6 December 2022)."Revisiting Heterophily For Graph Neural Networks".Advances in Neural Information Processing Systems.35:1362–1375.arXiv:2210.07606.
  44. ^Luan, Sitao; Hua, Chenqing; Xu, Minkai; Lu, Qincheng; Zhu, Jiaqi; Chang, Xiao-Wen; Fu, Jie; Leskovec, Jure; Precup, Doina (15 December 2023)."When Do Graph Neural Networks Help with Node Classification? Investigating the Homophily Principle on Node Distinguishability".Advances in Neural Information Processing Systems.36:28748–28760.
  45. ^Sample, Ian (2 December 2018)."Google's DeepMind predicts 3D shapes of proteins".The Guardian. Retrieved30 November 2020.
  46. ^"DeepMind's protein-folding AI has solved a 50-year-old grand challenge of biology".MIT Technology Review. Retrieved30 November 2020.
  47. ^Fan, Wenqi; Ma, Yao; Li, Qing; He, Yuan; Zhao, Eric; Tang, Jiliang; Yin, Dawei (2019).Graph Neural Networks for Social Recommendation. pp. 417–426.arXiv:1902.07243.doi:10.1145/3308558.3313488.hdl:10397/81232.ISBN 978-1-4503-6674-8.S2CID 67769538.
  48. ^Cappart, Quentin; Chételat, Didier; Khalil, Elias; Lodi, Andrea; Morris, Christopher; Veličković, Petar (2021). "Combinatorial optimization and reasoning with graph neural networks".arXiv:2102.09544 [cs.LG].
  49. ^Mirhoseini, Azalia; Goldie, Anna; Yazgan, Mustafa; Jiang, Joe Wenjie; Songhori, Ebrahim; Wang, Shen; Lee, Young-Joon; Johnson, Eric; Pathak, Omkar; Nazi, Azade; Pak, Jiwoo; Tong, Andy; Srinivasa, Kavya; Hang, William; Tuncer, Emre; Le, Quoc V.; Laudon, James; Ho, Richard; Carpenter, Roger; Dean, Jeff (2021). "A graph placement methodology for fast chip design".Nature.594 (7862):207–212.Bibcode:2021Natur.594..207M.doi:10.1038/s41586-021-03544-w.PMID 34108699.S2CID 235395490.
  50. ^Gasse, Maxime; Chételat, Didier; Ferroni, Nicola; Charlin, Laurent; Lodi, Andrea (2019). "Exact Combinatorial Optimization with Graph Convolutional Neural Networks".arXiv:1906.01629 [cs.LG].
  51. ^Wang, Su; Wang, Zhiliang; Zhou, Tao; Sun, Hongbin; Yin, Xia; Han, Dongqi; Zhang, Han; Shi, Xingang; Yang, Jiahai (2022)."Threatrace: Detecting and Tracing Host-Based Threats in Node Level Through Provenance Graph Learning".IEEE Transactions on Information Forensics and Security.17:3972–3987.arXiv:2111.04333.Bibcode:2022ITIF...17.3972W.doi:10.1109/TIFS.2022.3208815.ISSN 1556-6021.S2CID 243847506.
  52. ^Wang, Qi; Hassan, Wajih Ul; Li, Ding; Jee, Kangkook; Yu, Xiao (2020)."You Are What You Do: Hunting Stealthy Malware via Data Provenance Analysis".Network and Distributed Systems Security Symposium.doi:10.14722/ndss.2020.24167.ISBN 978-1-891562-61-7.S2CID 211267791.
  53. ^King, Isaiah J.; Huang, H. Howie (2022)."Euler: Detecting Network Lateral Movement via Scalable Temporal Link Prediction"(PDF).In Proceedings of the 29th Network and Distributed Systems Security Symposium.doi:10.14722/ndss.2022.24107.S2CID 248221601.
  54. ^Zanfei, Ariele; et al. (2022)."Graph Convolutional Recurrent Neural Networks for Water Demand Forecasting".Water Resources Research.58 (7) e2022WR032299. AGU.Bibcode:2022WRR....5832299Z.doi:10.1029/2022WR032299. Retrieved11 June 2024.
  55. ^Zanfei, Ariele; et al. (2023)."Shall we always use hydraulic models? A graph neural network metamodel for water system calibration and uncertainty assessment".Water Research.242 120264.Bibcode:2023WatRe.24220264Z.doi:10.1016/j.watres.2023.120264.PMID 37393807. Retrieved11 June 2024.
  56. ^Han, Kai; Wang, Yunhe; Guo, Jianyuan; Tang, Yehui; Wu, Enhua (2022). "Vision GNN: An Image is Worth Graph of Nodes".arXiv:2206.00272 [cs.CV].
  57. ^Zhou, Jie; Cui, Ganqu; Hu, Shengding; Zhang, Zhengyan; Yang, Cheng; Liu, Zhiyuan; Wang, Lifeng; Li, Changcheng; Sun, Maosong (1 January 2020)."Graph neural networks: A review of methods and applications".AI Open.1:57–81.doi:10.1016/j.aiopen.2021.01.001.ISSN 2666-6510.

External links

[edit]
Concepts
Applications
Implementations
Audio–visual
Text
Decisional
People
Architectures
Retrieved from "https://en.wikipedia.org/w/index.php?title=Graph_neural_network&oldid=1318339179"
Categories:
Hidden categories:

[8]ページ先頭

©2009-2025 Movatter.jp