Question: [ 5 pts ] In LSTM , the activation functions of gates ( forget , input, and output ) are sigmoid functions. Explain what will

[5 pts] In LSTM, the activation functions of gates (forget, input, and output) are sigmoid functions.
Explain what will happen if we use ReLU instead.
[5 pts] What are the problems with using vocabulary indexing in text processing. Explain it using the
example given in the slides.
In TF-IDF,
[3 pts] when a 100-word document contains the term "cat" 12 times, compute TF value of 'cat'.
[3 pts] The size of the corpus is 10 million documents. If we assume there are 0.3 million documents
that contain the term "cat". Compute IDF (use the log form)
[2 pt] Compute TF-IDF of 'cat'
In CBOW,
[3 pts] Can we use a vocabulary index method (instead of a one-hot vector) as an input to CBOW
[2 pts] The activation function of the output layer should be softmax. Explain the reason.
[3 pts] After training is done in CBOW, how do we extract the word embedding vector for a certain
word.
[ 5 pts ] In LSTM , the activation functions of

Step by Step Solution

There are 3 Steps involved in it

1 Expert Approved Answer
Step: 1 Unlock blur-text-image
Question Has Been Solved by an Expert!

Get step-by-step solutions from verified subject matter experts

Step: 2 Unlock
Step: 3 Unlock

Students Have Also Explored These Related Programming Questions!