Question: Suppose that the put to a self - attertion layer is a 5 - dimensional vector. In the self - attention layer, each input vector

Suppose that the put to a self-attertion layer
is a 5-dimensional vector.
In the self-attention layer, each input vector xi is
first linearly projected into query ?bar(q)i, key ?bar(k)i
and value ?bar(v)i, where ?bar(q)i,bar(k)i and ?bar(v)i are each 3-dimensional.
Which of the followings could ther possibly represent
the attention matrix, given some input sequence?
Note that softmax operation is applied separately
over the columns of the alignment matrix.
o)[00101201200]33
b)[001100010]33
[012010012010000000]44
d)[0010100101000000]44
55
f
e)[0000110000000000000001000]
h)[100000010000001]53
 Suppose that the put to a self-attertion layer is a 5-dimensional

Step by Step Solution

There are 3 Steps involved in it

1 Expert Approved Answer
Step: 1 Unlock blur-text-image
Question Has Been Solved by an Expert!

Get step-by-step solutions from verified subject matter experts

Step: 2 Unlock
Step: 3 Unlock

Students Have Also Explored These Related Programming Questions!