Skip to content

Commit b3331bc

Browse files
committed
update hypformer
1 parent 77c6c4f commit b3331bc

File tree

1 file changed

+3
-3
lines changed

1 file changed

+3
-3
lines changed

app/projects/hypformer/page.mdx

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -114,7 +114,7 @@ By defining these operations through HRC (and HTC for linear transformations), w
114114

115115
---
116116

117-
### 2.4Framework
117+
### 2.4 Framework
118118

119119
Framework of Hypformer. Input data (text, images, graphs) are projected onto the Lorentz model, then transformed via HTC. The result passes through the hyperbolic linear attention block with positional encoding, followed by a Feedforward layer (built by HTC) and LayerNorm (built by HRC). This serves as an encoder which can optionally incorporate a GNN.
120120
For classification tasks in this study, the decoder is the fully connected layer. Dropout, activation, and residual connections are omitted for brevity.
@@ -123,7 +123,7 @@ For classification tasks in this study, the decoder is the fully connected layer
123123

124124

125125

126-
## 3.Experiments
126+
## 3. Experiments
127127
### 3.1 Experiments on Large-scale Graphs
128128

129129
We first evaluate Hypformer on diverse large-scale graphs for node classification, with node counts ranging from millions to billions, including ogbn-arxiv, ogbn-protein, and Papers100M.
@@ -155,7 +155,7 @@ We conducted additional tests on the model’s scalability regarding the number
155155
![Scalability|scale=0.5](./assets/gpucost.png)
156156

157157

158-
## 4.Conclusion
158+
## 4. Conclusion
159159

160160
In this work, we introduce a efficient hyperbolic Transformer, Hypformer. This method operates directly and fully on hyperbolic representations and employs a linear attention mechanism, enabling it to be both scalable and effective.
161161
Furthermore, this study introduces two basic blocks, HTC and HRC, which are foundational in constructing hyperbolic models. Nonetheless, the research presented is an initial exploration and numerous challenges warrant further investigation. These include the initial determination of a curvature that better reflects the data geometry, the setting of curvature at different levels for Hypformer, and the design of effective decoders for different downstream tasks. We plan to address these issues in our future work.

0 commit comments

Comments
 (0)