Question: True or False? [ 2 points ] The difference between planning in a known Markov Decision Process ( MDP ) and Reinforcement Learning ( RL
True or False? points The difference between planning in a known Markov Decision Process MDP and Reinforcement Learning RL is that in RL the agent doesnt know what the current state is eg doesnt know its own position when acting in a gridworld
points If the only difference between two MDPs is the value of the discount factor then they must have the same optimal policy.
points When getting to act only for a finite number of steps in an MDP the optimal policy is stationary. A stationary policy is a policy that takes the same action in a given state, independent of at what time the agent is in that state.
points As the number of particles goes to infinity, particle filtering will represent the same probability distribution you would get with exact inference.
points Consider two particle filtering implementations:
Implementation :
Initialize particles by sampling from initial state distribution and assigning uniform weights. Propagate particles, retaining weights
Resample according to weights
Weight according to observations
Implementation :
Initialize particles by sampling from initial state distribution. Propagate unweighted particles
Weight according to observations
Resample according to weights
Questions:
i Implementation will typically provide a better approximation of the estimated distribution than implementation
ii If the transition model is deterministic then both implementations provide equally good estimates of distribution
iii. If the observation model is deterministic then both implementations provide equally good estimates of the distribution.
Step by Step Solution
There are 3 Steps involved in it
1 Expert Approved Answer
Step: 1 Unlock
Question Has Been Solved by an Expert!
Get step-by-step solutions from verified subject matter experts
Step: 2 Unlock
Step: 3 Unlock
