File tree 1 file changed +4
-4
lines changed
1 file changed +4
-4
lines changed Original file line number Diff line number Diff line change @@ -27,7 +27,7 @@ class DEQMLPMixerAttention(_DEQModule):
27
27
residual connection in the explicit MLP-Mixer architecture.
28
28
29
29
Note:
30
- To use this module, wrap it in `modules .DEQFixedPoint`.
30
+ To use this module, wrap it in `deq .DEQFixedPoint`.
31
31
32
32
Paper:
33
33
https://arxiv.org/abs/2105.02723
@@ -83,7 +83,7 @@ class DEQVanillaSoftmaxAttention(_DEQModule):
83
83
into the feed-forward self-correction term.
84
84
85
85
Note:
86
- To use this module, wrap it in `modules .DEQFixedPoint`.
86
+ To use this module, wrap it in `deq .DEQFixedPoint`.
87
87
88
88
Paper:
89
89
https://arxiv.org/abs/1706.03762
@@ -179,7 +179,7 @@ class DEQMeanFieldAttention(_DEQModule):
179
179
correction term. This all looks a lot like a transformer.
180
180
181
181
Note:
182
- To use this module, wrap it in `modules .DEQFixedPoint`.
182
+ To use this module, wrap it in `deq .DEQFixedPoint`.
183
183
184
184
Args:
185
185
num_spins (int):
@@ -317,7 +317,7 @@ class DEQAdaTAPMeanFieldAttention(_DEQModule):
317
317
first and second moments assuming a Gaussian cavity distribution.
318
318
319
319
Note:
320
- To use this module, wrap it in `modules .DEQFixedPoint`.
320
+ To use this module, wrap it in `deq .DEQFixedPoint`.
321
321
322
322
Args:
323
323
num_spins (int):
You can’t perform that action at this time.
0 commit comments