Question: Suppose that the put to a self - attertion layer is a 5 - dimensional vector. In the self - attention layer, each input vector
Suppose that the put to a selfattertion layer
is a dimensional vector.
In the selfattention layer, each input vector is
first linearly projected into query key
and value where and are each dimensional.
Which of the followings could ther possibly represent
the attention matrix, given some input sequence?
Note that softmax operation is applied separately
over the columns of the alignment matrix.
o
b
d
e
h
Step by Step Solution
There are 3 Steps involved in it
1 Expert Approved Answer
Step: 1 Unlock
Question Has Been Solved by an Expert!
Get step-by-step solutions from verified subject matter experts
Step: 2 Unlock
Step: 3 Unlock
