- Transformer architecture is becoming a dominant choice in NLP and CV, why not in Graphs?
- Solution: Graphormer which uses special positional encoding for graphs.
- Problem not mentioned in the paper: graph "positional encoding" often requires shortest path computation between two nodes.
- Transformer is the most powerful NN in modeling sequential data such as speech and natural language processing
- It's an open question whether Transformers are suitable to model graphs - Graphormer seems to be an affirmative answer
- Most leaderboards and benchmarks in chemistry seem to do well
- Centrality encoding captures node importance in the graph
- Spatial encoding captures the structural relation between nodes - for each node pair, we assign a learnable embedding based on the spatial relation.
- GNN aim to learn representation of nodes and graphs.
- Modern GNNs mfollow a learning schema that updates the representation of an ode by aggregating representations of its neighbors.
- Aggregate-combine step:
- The task: graph classification
- A Node [VNode] is added to the graph which is then connected to every individual node (distance of 1). The point is that hte representation of the entire h_g graph is the node feature of [VNode]