Skip to content
New issue

Have a question about this project? # for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “#”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? # to your account

Questions about table 5 #33

Open
kaikai23 opened this issue Oct 23, 2022 · 0 comments
Open

Questions about table 5 #33

kaikai23 opened this issue Oct 23, 2022 · 0 comments

Comments

@kaikai23
Copy link

Hi,

In your paper table 5, the (G,G,G,G) uses the numbers (79.8%) from PVT paper, which uses absolution positional encoding. However, I suppose the other model variants listed in this table use CPE, so they are not directly comparable. Should the accuracy of (G,G,G,G) with CPE be 81.2% as shown in table 1?

In general, I am interested in knowing if there is a benifit of using global attention in the early layers.

Thanks.

# for free to join this conversation on GitHub. Already have an account? # to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant