We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
I am having issues aligning the variables in the code to their respective meanings, mostly N in models.py.
N
models.py
Can you provide more information / references to the paper? Here is my interpretation:
K
However, that doesn't make any sense with these lines from the Attention class:
Attention
H = self.feature_extractor_part2(H) # NxL A = self.attention(H) # NxK A = torch.transpose(A, 1, 0) # KxN A = F.softmax(A, dim=1) # softmax over N M = torch.mm(A, H) # KxL
because H = self.feature_extractor_part2(H) returns a KxL matrix ([13, 500]), no N involved.
H = self.feature_extractor_part2(H)
L
[13, 500]
Am I misinterpreting the meaning of N?
The text was updated successfully, but these errors were encountered:
No branches or pull requests
I am having issues aligning the variables in the code to their respective meanings, mostly
N
inmodels.py
.Can you provide more information / references to the paper? Here is my interpretation:
N
: number of bags (i.e., batch size)K
: number of instances in a bag (i.e., embeddings)However, that doesn't make any sense with these lines from the
Attention
class:because
H = self.feature_extractor_part2(H)
returns aK
xL
matrix ([13, 500]
), noN
involved.Am I misinterpreting the meaning of
N
?The text was updated successfully, but these errors were encountered: