DrishtiSharma
commited on
Commit
•
4c24885
1
Parent(s):
50404f1
Update README.md
Browse files
README.md
CHANGED
@@ -10,4 +10,5 @@ This example implements the EANet model for image classification, and demonstrat
|
|
10 |
|
11 |
|
12 |
**Implement the EANet model**
|
|
|
13 |
The EANet model leverages external attention. The computational complexity of traditional self attention is O(d * N ** 2), where d is the embedding size, and N is the number of patch. the authors find that most pixels are closely related to just a few other pixels, and an N-to-N attention matrix may be redundant. So, they propose as an alternative an external attention module where the computational complexity of external attention is O(d * S * N). As d and S are hyper-parameters, the proposed algorithm is linear in the number of pixels. In fact, this is equivalent to a drop patch operation, because a lot of information contained in a patch in an image is redundant and unimportant.
|
|
|
10 |
|
11 |
|
12 |
**Implement the EANet model**
|
13 |
+
|
14 |
The EANet model leverages external attention. The computational complexity of traditional self attention is O(d * N ** 2), where d is the embedding size, and N is the number of patch. the authors find that most pixels are closely related to just a few other pixels, and an N-to-N attention matrix may be redundant. So, they propose as an alternative an external attention module where the computational complexity of external attention is O(d * S * N). As d and S are hyper-parameters, the proposed algorithm is linear in the number of pixels. In fact, this is equivalent to a drop patch operation, because a lot of information contained in a patch in an image is redundant and unimportant.
|