patchvisiontransformer's People
Forkers
ricklentz cv-ip zz7379 lilujunai futurev peternara chaoxiang661 zhen-zohn-wang wohjyy phdfar dl-vit trellixvulnteam nastu-hopatchvisiontransformer's Issues
Absolute cosine similarity
Hey, this is an interesting work.
Just wondering why using the absolute cosine similarity
instead of the original one.
Thanks!
cosine similarity between hidden outputs
Hi, I train a vit_base_patch32 model with reslution 224 on imagenet, the valid acc comes up to 73.38%, then I dump all transformer blocks' outputs, calculate cosine similarity between them as mentioned in paper, I cannot have the same result in paper, here is my result. the 0,1,2,3,4,5,6,7,8,9,10,11 is layer depth, right is the value of cosine similarity
0 ---> 0.5428178586547845
1 ---> 0.6069238793659271
2 ---> 0.30199843167793006
3 ---> 0.26388993740273486
4 ---> 0.26132955026320265
5 ---> 0.24258930458215844
6 ---> 0.20970458839482967
7 ---> 0.21119057468517677
8 ---> 0.22155304307901189
9 ---> 0.23545575648548187
10 ---> 0.2329663067175004
11 ---> 0.22496230679589768
what does aux_class mean?
Hi, I find aux_class is setted as None in your code, so have you use patch_loss in your work?
target_lst, mask_lst, targets = aux_class
what these variables mean: target_lst, mask_lst, targets?
Thank you very much!
Difference of swin
Hi, thanks for sharing code!
I'd like to try your code with mmsegmentation.
But I can't find which part is the different with original swin.
Shortly, I don't know where the diverse part is
Sincerely
What do 'low_order' and 'high_order' represent?
Hello, I want to know what the 'low_order' and 'high_order' represent in the 'similarity' function in models.py and how to set the 'high_k'?
I also wonder which parts of the codes present the L_cos in the paper?
Thank you very much!
The training can not run
/PatchVisionTransformer/deit/models.py", line 232, in similarity
high_order = (patches.mean(dim=(2, 3), deepdim=True) - patches / (img_size ** 2)) * (img_size ** 2) / (img_size ** 2 - 1)
TypeError: mean() received an invalid combination of arguments - got (deepdim=bool, dim=tuple, ), but expected one of:
- (*, torch.dtype dtype)
didn't match because some of the keywords were incorrect: deepdim, dim - (tuple of names dim, bool keepdim, *, torch.dtype dtype)
- (tuple of ints dim, bool keepdim, *, torch.dtype dtype)
Do you need to train 600 epoch totally?
According to your instruction in readme, do you need to train 600 epoch totally to get the results shown in the paper?
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
๐ Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. ๐๐๐
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google โค๏ธ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.