Skip to content

Commit 6c626d0

Browse files
committed
readme
1 parent 5b16e8f commit 6c626d0

File tree

1 file changed

+10
-8
lines changed

1 file changed

+10
-8
lines changed

README.md

Lines changed: 10 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -453,7 +453,7 @@ model = RegionViT(
453453
dim = (64, 128, 256, 512), # tuple of size 4, indicating dimension at each stage
454454
depth = (2, 2, 8, 2), # depth of the region to local transformer at each stage
455455
window_size = 7, # window size, which should be either 7 or 14
456-
num_classes = 1000, # number of output lcasses
456+
num_classes = 1000, # number of output classes
457457
tokenize_local_3_conv = False, # whether to use a 3 layer convolution to encode the local tokens from the image. the paper uses this for the smaller models, but uses only 1 conv (set to False) for the larger models
458458
use_peg = False, # whether to use positional generating module. they used this for object detection for a boost in performance
459459
)
@@ -496,6 +496,8 @@ pred = nest(img) # (1, 1000)
496496

497497
A new <a href="https://arxiv.org/abs/2111.06377">Kaiming He paper</a> proposes a simple autoencoder scheme where the vision transformer attends to a set of unmasked patches, and a smaller decoder tries to reconstruct the masked pixel values.
498498

499+
<a href="https://www.youtube.com/watch?v=LKixq2S2Pz8">DeepReader quick paper review</a>
500+
499501
You can use it with the following code
500502

501503
```python
@@ -809,13 +811,13 @@ Coming from computer vision and new to transformers? Here are some resources tha
809811
## Citations
810812
```bibtex
811813
@article{hassani2021escaping,
812-
title = {Escaping the Big Data Paradigm with Compact Transformers},
813-
author = {Ali Hassani and Steven Walton and Nikhil Shah and Abulikemu Abuduweili and Jiachen Li and Humphrey Shi},
814-
year = 2021,
815-
url = {https://arxiv.org/abs/2104.05704},
816-
eprint = {2104.05704},
817-
archiveprefix = {arXiv},
818-
primaryclass = {cs.CV}
814+
title = {Escaping the Big Data Paradigm with Compact Transformers},
815+
author = {Ali Hassani and Steven Walton and Nikhil Shah and Abulikemu Abuduweili and Jiachen Li and Humphrey Shi},
816+
year = 2021,
817+
url = {https://arxiv.org/abs/2104.05704},
818+
eprint = {2104.05704},
819+
archiveprefix = {arXiv},
820+
primaryclass = {cs.CV}
819821
}
820822
```
821823

0 commit comments

Comments
 (0)