AI-CUK Joint Journal Club: V.T.Hoang, Review on "Global self-attention as a replacement for graph convolution," KDD 2022, Jan 3rd, 2023
24. Mar 2023•0 gefällt mir
0 gefällt mir
Sei der Erste, dem dies gefällt
Mehr anzeigen
•65 Aufrufe
Aufrufe
Aufrufe insgesamt
0
Auf Slideshare
0
Aus Einbettungen
0
Anzahl der Einbettungen
0
Downloaden Sie, um offline zu lesen
Melden
Technologie
The 1st AI-CUK Weekly Joint Journal Club
Presenter: Van Thuy Hoang
Date: Jan 3rd, 2023
Topic: Review on "Global self-attention as a replacement for graph convolution," KDD 2022
Schedule: https://nslab-cuk.github.io/joint-journal-club/
AI-CUK Joint Journal Club: V.T.Hoang, Review on "Global self-attention as a replacement for graph convolution," KDD 2022, Jan 3rd, 2023
KDD ’22, August 14–18, 2022, Washington, DC, USA
Thuy Hoang Van, PhD student.
Network Science Lab, The Catholic University of Korea.
https://nslab.catholic.ac.kr/
Difference: Word, image vs graph
• Words in sentences
– Tokenization & PE (absolute and relative PE)
– 1-d dimentional vectors
• Images:
– Grid, 2-d dimentional vectors
• Graphs
– Node Position (disorder)
– Edge connection, important as node
– Global , local connection
– Node Centrality
Bring a challenge to apply a graph transformer based on node self-
attention
Key notes*
• Propose a simple extension:
– Edge channels to the transformer framework
• Propose a global attention take structural
information of graphs
• Better than convolutional GNNs.
Message passing – Self Attention
• Left: It takes three stages of convolution for node 0
to aggregate node 6
• Right: With global self-attention, the model can learn
to do so in a single step (all steps in one).
– Meaning/Role of target nodes vs other nodes?
– How far attention should be?
– How much attention should be?
Conclusion
• SPD on edge channels shows a strong sampling
strategy for self-attention mechanism.
– Compare to other sampling strategy: based on
adjacency, distance, intimacy, etc. (Graph_Bert,..)
– Aggregation mechanism to avoid smoothing problem
in Message passing.
• Bettter than 1-WL
• Key: global attention: more understanding about
the meaning of graph structure:
– Edge feature
– Integrating edge to graph transformer.