You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
{{ message }}
This repository was archived by the owner on Jan 1, 2025. It is now read-only.
Thank you for your great paper.
I have a question while reading your paper and code.
In the paper, the formula for performing divided_space_time includes the class token during temporal attention, but in the actual code, the cls_token is excluded during temporal attention and included during spatial attention.
This question is a duplicate of the following question, but I did not get a satisfactory answer. https://github.com/facebookresearch/TimeSformer/issues/74
Is there a reason for this? Would it be possible to obtain experimental results by including the cls_token during temporal attention?
Thank you for your great paper.


I have a question while reading your paper and code.
In the paper, the formula for performing divided_space_time includes the class token during temporal attention, but in the actual code, the cls_token is excluded during temporal attention and included during spatial attention.
This question is a duplicate of the following question, but I did not get a satisfactory answer.
https://github.com/facebookresearch/TimeSformer/issues/74
Is there a reason for this? Would it be possible to obtain experimental results by including the cls_token during temporal attention?