Question: grammarQuestion 3 : Efficiently Simulating Bounded Stacks with RNNs ( 2 0 pts ) As we discussed in the lecture, hierarchical structure, characterized by long
grammarQuestion : Efficiently Simulating Bounded Stacks with RNNs pts
As we discussed in the lecture, hierarchical structure, characterized by long distance and nested
dependencies, lies at the core of the human language. Indeed, this motivated our theoretical discussion
of contextfree grammars as a useful paradigm for describing language. Studying modern language
models with respect to their ability to efficiently represent hierarchical structure, therefore, provides
evidence that they are useful models for human language. This question directly addresses this point.
More precisely, we will study the ability of recurrent neural network language models to recognize a
variant of the languages.
As introduced in the lecture notes, the languages are in a way archetypal contextfree languages.
Recognizing languages is conceptually simple a system has to remember the sequence of currently
nonclosed opening brackets and make sure they are closed in the correct order, at which point the
closed bracket pairs can be "forgotten", ie popped off the stack. This means that the memory
necessary to recognize any string in is proportional to the number of nonclosed brackets at any
time. We formalize it by counting how many more open brackets than closed brackets there are at
each timestep in the string:
count:count:
where count refers to the number of times any opening bracket occurs in and count ::
the number of times any closing bracket occurs in
While contextfree languages like describe arbitrarily deep hierarchical structures, natural lan
guages exhibit bounded nesting in practice, as discussed in the lecture. Furthermore, the infinite
nesting and therefore infinitely long stacks also make it impossible to represent contextfree languages
with finite precision. In this question, we investigate how to represent languages which can only
nest up to some bounded depth We denote such languages as
Definition languages Let minN. We define the bounded Dyck language
by combining with a bound on the nesting depth:
dots,
where corresponds to the length of the string.
Due to their bounded nesting depth, languages can be recognized by stacks of bounded depth
and therefore with bounded memory this means that they are in fact finitestate. This makes them
especially wellsuited as a benchmark for finiteprecision language models.
This question is roughly divided into two parts: in the first part, you will show that an Elman RNN is
able to simulate a finitestate automaton that recognizes the for some and In the second
part of the question, you are asked to show that the RNN indeed recognizes the language as well using
a specific definition of acceptance.
We begin with a warmup question.
a pt Suppose that the current bounded stack configuration in an automaton recognizing
is
What are the new stack configurations after reading in each of the following symbols
each one starting from a stack not one after another
:
:
Use to denote an empty stack and simply state that the automaton would reject a string if the
processed string is not in
Note: You have to specify nine stack configurations altogether.
We next prove that languages are in fact finitestate by constructing an FSA recognizing
Step by Step Solution
There are 3 Steps involved in it
1 Expert Approved Answer
Step: 1 Unlock
Question Has Been Solved by an Expert!
Get step-by-step solutions from verified subject matter experts
Step: 2 Unlock
Step: 3 Unlock
