# Assignment 7 Car solution

Problem 1: Bayesian network basics

First, let us look at a simplified version of the car tracking problem. For this problem only, let $C_t \in \{0, 1\}$ be the actual location of the car we wish to observe at time step $t \in \{1, 2, 3\}$. Let $D_t \in \{0, 1\}$ be a sensor reading for the location of that car measured at time $t$. Here's what the Bayesian network (it's an HMM, in fact) looks like:

The distribution over the initial car distribution is uniform; that is, for each value $c_1 \in \{0, 1\}$: $$ p(c_1) = 0.5. $$ The following local conditional distribution governs the movement of the car (with probability $\epsilon$, the car moves). For each $t \in \{2, 3\}$: $$ p(c_t \mid c_{t-1}) = \begin{cases} \epsilon & \text{if $c_t \neq c_{t-1}$} \\ 1-\epsilon & \text{if $c_t = c_{t-1}$}. \end{cases} $$ The following local conditional distribution governs the noise in the sensor reading (with probability $\eta$, the sensor reports the wrong position). For each $t \in \{1, 2, 3\}$: $$ p(d_t \mid c_t) = \begin{cases} \eta & \text{if $d_t \neq c_t$} \\ 1-\eta & \text{if $d_t = c_t$}. \end{cases} $$

Below, you will be asked to find the posterior distribution for the car's position at the second time step ($C_2$) given different sensor readings.

Important: For the following computations, try to follow the general strategy described in lecture (marginalize non-ancestral variables, condition, and perform variable elimination). Try to delay normalization until the very end. You'll get more insight than trying to chug through lots of equations.

Suppose we have a sensor reading for the second timestep, $D_2 = 0$. Compute the posterior distribution $\mathbb P(C_2 = 1 \mid D_2 = 0)$. We encourage you to draw out the (factor) graph.

Suppose a time step has elapsed and we got another sensor reading, $D_3 = 1$, but we are still interested in $C_2$. Compute the posterior distribution $\mathbb P(C_2 = 1 \mid D_2 = 0, D_3 = 1)$. The resulting expression might be moderately complex. We encourage you to draw out the (factor) graph.

Suppose $\epsilon = 0.1$ and $\eta = 0.2$.i. Compute and compare the probabilities $\mathbb P(C_2 = 1 \mid D_2 = 0)$ and $\mathbb P(C_2 = 1 \mid D_2 = 0, D_3 = 1)$. Give numbers, round your answer to 4 significant digits.

ii. How did adding the second sensor reading $D_3 = 1$ change the result? Explain your intuition for why this change makes sense in terms of the car positions and associated sensor observations.

iii. What would you have to set $\epsilon$ while keeping $\eta = 0.2$ so that $\mathbb P(C_2 = 1 \mid D_2 = 0) = \mathbb P(C_2 = 1 \mid D_2 = 0, D_3 = 1)$? Explain your intuition in terms of the car positions with respect to the observations.

Problem 2: Emission probabilities

In this problem, we assume that the other car is stationary (e.g., $C_t = C_{t-1}$ for all time steps $t$). You will implement a function observe that upon observing a new distance measurement $D_t = d_t$ updates the current posterior probability from $$\mathbb P(C_t \mid D_1 = d_1, \dots, D_{t-1} = d_{t-1})$$ to $$\mathbb P(C_t \mid D_1 = d_1, \dots, D_t = d_t) \propto \mathbb P(C_t \mid D_1 = d_1, \dots, D_{t-1} = d_{t-1}) p(d_t \mid c_t),$$ where we have multiplied in the emission probabilities $p(d_t \mid c_t)$ described earlier. The current posterior probability is stored as self.belief in ExactInference.

Fill in the observe method in the ExactInference class of submission.py. This method should modify self.belief in place to update the posterior probability of each tile given the observed noisy distance. After you're done, you should be able to find the stationary car by driving around it (-p means cars don't move):

Notes:

You can start driving with exact inference now.python drive.py -a -p -d -k 1 -i exactInference

You can also turn off -a to drive manually.

It's generally best to run drive.py on your local machine, but if you do decide to run it on cardinal/rice instead, please ssh into the farmshare machines with either the -X or -Y option in order to get the graphical interface; otherwise, you will get some display error message. Note: do expect this graphical interface to be a bit slow... drive.py is not used for grading, but is just there for you to visualize and enjoy the game!

Read through the code in utils.py for the Belief class before you get started... you'll need to use this class for several of the code tasks in this assignment.

Remember to normalize the posterior probability after you update it. (There's a useful function for this in utils.py).

On the small map, the autonomous driver will sometimes drive in circles around the middle block before heading for the target area. In general, don't worry too much about the precise path the car takes. Instead, focus on whether your car tracker correctly infers the location of other cars.

Don't worry if your car crashes once in a while! Accidents do happen, whether you are human or AI. However, even if there was an accident, your driver should have been aware that there was a high probability that another car was in the area.

Problem 3: Transition probabilities

Now, let's consider the case where the other car is moving according to transition probabilities $p(c_{t+1} \mid c_t)$. We have provided the transition probabilities for you in self.transProb. Specifically, self.transProb[(oldTile, newTile)] is the probability of the other car being in newTile at time step $t+1$ given that it was in oldTile at time step $t$.

In this part, you will implement a function elapseTime that updates the posterior probability about the location of the car at a current time $t$ $$\mathbb P(C_t = c_t \mid D_1 = d_1, \dots, D_t = d_t)$$ to the next time step $t+1$ conditioned on the same evidence, via the recurrence: $$\mathbb P(C_{t+1} = c_{t+1} \mid D_1 = d_1, \dots, D_t = d_t) \propto \sum_{c_t} \mathbb P(C_t = c_t \mid D_1 = d_1, \dots, D_t = d_t) p(c_{t+1} \mid c_t).$$ Again, the posterior probability is stored as self.belief in ExactInference.

Finish ExactInference by implementing the elapseTime method. When you are all done, you should be able to track a moving car well enough to drive autonomously:

python drive.py -a -d -k 1 -i exactInference

Notes:

You can also drive autonomously in the presence of more than one car:python drive.py -a -d -k 3 -i exactInference

You can also drive down Lombard:python drive.py -a -d -k 3 -i exactInference -l lombard

On Lombard, the autonomous driver may attempt to drive up and down the street before heading towards the target area. Again, focus on the car tracking component, instead of the actual driving.

Problem 4: Particle filtering

Though exact inference works well for the small maps, it wastes a lot of effort computing probabilities for every available tile, even for tiles that are unlikely to have a car on them. We can solve this problem using a particle filter. Updates to the particle filter have complexity that's linear in the number of particles, rather than linear in the number of tiles.

In this problem, you'll implement two short but important methods for the ParticleFilter class in submission.py. When you're finished, your code should be able to track cars nearly as effectively as it does with exact inference.

Some of the code has been provided for you. For example, the particles have already been initialized randomly. You need to fill in the observe and elapseTime functions. These should modify self.particles, which is a map from tiles (row, col) to the number of particles existing at that tile, and self.belief, which needs to be updated each time you re-sample the particle locations.

You should use the same transition probabilities as in exact inference. The belief distribution generated by a particle filter is expected to look noisier compared to the one obtained by exact inference.

python drive.py -a -i particleFilter -l lombard

To debug, you might want to start with the parked car flag (-p) and the display car flag (-d).

Note: The random number generator inside util.weightedRandomChoice behaves differently on different systems' versions of Python (e.g., Unix and Windows versions of Python). Please test this question (run grader.py) on rice. When copying files to rice, make sure you copy the entire folder using scp with the recursive option -r.

Problem 5: Which car is it?

So far, we have assumed that we have a distinct noisy distance reading for each car, but in reality, our microphone would just pick up an undistinguished set of these signals, and we wouldn't know which distance reading corresponds to which car. First, let's extend the notation from before: let $C_{ti} \in \mathbb R^2$ be the location of the $i$-th car at the time step $t$, for $i = 1, \dots, K$ and $t = 1, \dots, T$. Recall that all the cars move independently according to the transition dynamics as before.

Let $D_{ti} \in \mathbb R$ be the noisy distance measurement of the $i$-th car, which is now not directly observed. Instead, we observe the set of distances $D_t = \{ D_{t1}, \dots, D_{tK} \}$. (For simplicity, we'll assume that all distances are distinct values.) Alternatively, you can think of $E_t = (E_{t1}, \dots, E_{tK})$ as a list which is a uniformly random permutation of the noisy distances $(D_{t1}, \dots, D_{tK})$. For example, suppose $K=2$ and $T = 2$. Before, we might have gotten distance readings of $1$ and $2$ for the first car and $3$ and $4$ for the second car at time steps $1$ and $2$, respectively. Now, our sensor readings would be permutations of $\{1, 3\}$ (at time step $1$) and $\{2, 4\}$ (at time step $2$). Thus, even if we knew the second car was distance $3$ away at time $t = 1$, we wouldn't know if it moved further away (to distance $4$) or closer (to distance $2$) at time $t = 2$.

Suppose we have $K=2$ cars and one time step $T=1$. Write an expression for the conditional distribution $\mathbb P(C_{11}, C_{12} \mid E_1 = e_1)$ as a function of the PDF of a Gaussian $\mathcal p_{\mathcal N}(v; \mu, \sigma^2)$ and the prior probability $p(c_{11})$ and $p(c_{12})$ over car locations. Your final answer should not contain variables $d_{11}$, $d_{12}$.Remember that $\mathcal p_{\mathcal N}(v; \mu, \sigma^2)$ is the probability of a random variable, $v$, in a Gaussian distribution with mean $\mu$ and standard deviation $\sigma$.

Hint: for $K=1$, the answer would be $$\mathbb P(C_{11} = c_{11} \mid E_1 = e_1) \propto p(c_{11}) p_{\mathcal N}(e_{11}; \|a_1 - c_{11}\|, \sigma^2).$$ where $a_t$ is the position of the car at time t. You might find it useful to draw the Bayesian network and think about the distribution of $E_t$ given $D_{t1}, \dots, D_{tK}$.

Assuming the prior $p(c_{1i})$ is the same for all $i$, show that the number of assignments for all $K$ cars $(c_{11}, \dots, c_{1K})$ that obtain the maximum value of $\mathbb P(C_{11} = c_{11}, \dots, C_{1K} = c_{1K} \mid E_1 = e_1)$ is at least $K!$.You can also assume that the car locations that maximize the probability above are unique ($C_{1i} \neq c_{1j}$ for all $i \neq j$).

Note: you don't need to produce a complicated proof for this question. It is acceptable to provide a clear explanation based on your intuitive understanding of the scenario.

For general $K$, what is the treewidth corresponding to the posterior distribution over all $K$ car locations at all $T$ time steps conditioned on all the sensor readings: $$\mathbb P(C_{11} = c_{11}, \dots, C_{1K} = c_{1K}, \dots, C_{T1} = c_{T1}, \dots, C_{TK} = c_{TK} \mid E_1 = e_1, \dots, E_T = e_T)?$$ Briefly justify your answer.

(extra credit) Now suppose you change your sensors so that at each time step $t$, they return the list of exact positions of the $K$ cars, but shifted (with wrap around) by a random amount. For example, if the true car positions at time step $1$ are $c_{11} = 1, c_{12} = 3, c_{13} = 8, c_{14} = 5$, then $e_1$ would be $[1, 3, 8, 5]$, $[3, 8, 5, 1]$, $[8, 5, 1, 3]$, or $[5, 1, 3, 8]$, each with probability $1/4$. Describe an efficient algorithm for computing $p(c_{ti} \mid e_1, \dots, e_T)$ for any time step $t$ and car $i$. Your algorithm should not be exponential in $K$ or $T$.

First, let us look at a simplified version of the car tracking problem. For this problem only, let $C_t \in \{0, 1\}$ be the actual location of the car we wish to observe at time step $t \in \{1, 2, 3\}$. Let $D_t \in \{0, 1\}$ be a sensor reading for the location of that car measured at time $t$. Here's what the Bayesian network (it's an HMM, in fact) looks like:

The distribution over the initial car distribution is uniform; that is, for each value $c_1 \in \{0, 1\}$: $$ p(c_1) = 0.5. $$ The following local conditional distribution governs the movement of the car (with probability $\epsilon$, the car moves). For each $t \in \{2, 3\}$: $$ p(c_t \mid c_{t-1}) = \begin{cases} \epsilon & \text{if $c_t \neq c_{t-1}$} \\ 1-\epsilon & \text{if $c_t = c_{t-1}$}. \end{cases} $$ The following local conditional distribution governs the noise in the sensor reading (with probability $\eta$, the sensor reports the wrong position). For each $t \in \{1, 2, 3\}$: $$ p(d_t \mid c_t) = \begin{cases} \eta & \text{if $d_t \neq c_t$} \\ 1-\eta & \text{if $d_t = c_t$}. \end{cases} $$

Below, you will be asked to find the posterior distribution for the car's position at the second time step ($C_2$) given different sensor readings.

Important: For the following computations, try to follow the general strategy described in lecture (marginalize non-ancestral variables, condition, and perform variable elimination). Try to delay normalization until the very end. You'll get more insight than trying to chug through lots of equations.

Suppose we have a sensor reading for the second timestep, $D_2 = 0$. Compute the posterior distribution $\mathbb P(C_2 = 1 \mid D_2 = 0)$. We encourage you to draw out the (factor) graph.

Suppose a time step has elapsed and we got another sensor reading, $D_3 = 1$, but we are still interested in $C_2$. Compute the posterior distribution $\mathbb P(C_2 = 1 \mid D_2 = 0, D_3 = 1)$. The resulting expression might be moderately complex. We encourage you to draw out the (factor) graph.

Suppose $\epsilon = 0.1$ and $\eta = 0.2$.i. Compute and compare the probabilities $\mathbb P(C_2 = 1 \mid D_2 = 0)$ and $\mathbb P(C_2 = 1 \mid D_2 = 0, D_3 = 1)$. Give numbers, round your answer to 4 significant digits.

ii. How did adding the second sensor reading $D_3 = 1$ change the result? Explain your intuition for why this change makes sense in terms of the car positions and associated sensor observations.

iii. What would you have to set $\epsilon$ while keeping $\eta = 0.2$ so that $\mathbb P(C_2 = 1 \mid D_2 = 0) = \mathbb P(C_2 = 1 \mid D_2 = 0, D_3 = 1)$? Explain your intuition in terms of the car positions with respect to the observations.

Problem 2: Emission probabilities

In this problem, we assume that the other car is stationary (e.g., $C_t = C_{t-1}$ for all time steps $t$). You will implement a function observe that upon observing a new distance measurement $D_t = d_t$ updates the current posterior probability from $$\mathbb P(C_t \mid D_1 = d_1, \dots, D_{t-1} = d_{t-1})$$ to $$\mathbb P(C_t \mid D_1 = d_1, \dots, D_t = d_t) \propto \mathbb P(C_t \mid D_1 = d_1, \dots, D_{t-1} = d_{t-1}) p(d_t \mid c_t),$$ where we have multiplied in the emission probabilities $p(d_t \mid c_t)$ described earlier. The current posterior probability is stored as self.belief in ExactInference.

Fill in the observe method in the ExactInference class of submission.py. This method should modify self.belief in place to update the posterior probability of each tile given the observed noisy distance. After you're done, you should be able to find the stationary car by driving around it (-p means cars don't move):

Notes:

You can start driving with exact inference now.python drive.py -a -p -d -k 1 -i exactInference

You can also turn off -a to drive manually.

It's generally best to run drive.py on your local machine, but if you do decide to run it on cardinal/rice instead, please ssh into the farmshare machines with either the -X or -Y option in order to get the graphical interface; otherwise, you will get some display error message. Note: do expect this graphical interface to be a bit slow... drive.py is not used for grading, but is just there for you to visualize and enjoy the game!

Read through the code in utils.py for the Belief class before you get started... you'll need to use this class for several of the code tasks in this assignment.

Remember to normalize the posterior probability after you update it. (There's a useful function for this in utils.py).

On the small map, the autonomous driver will sometimes drive in circles around the middle block before heading for the target area. In general, don't worry too much about the precise path the car takes. Instead, focus on whether your car tracker correctly infers the location of other cars.

Don't worry if your car crashes once in a while! Accidents do happen, whether you are human or AI. However, even if there was an accident, your driver should have been aware that there was a high probability that another car was in the area.

Problem 3: Transition probabilities

Now, let's consider the case where the other car is moving according to transition probabilities $p(c_{t+1} \mid c_t)$. We have provided the transition probabilities for you in self.transProb. Specifically, self.transProb[(oldTile, newTile)] is the probability of the other car being in newTile at time step $t+1$ given that it was in oldTile at time step $t$.

In this part, you will implement a function elapseTime that updates the posterior probability about the location of the car at a current time $t$ $$\mathbb P(C_t = c_t \mid D_1 = d_1, \dots, D_t = d_t)$$ to the next time step $t+1$ conditioned on the same evidence, via the recurrence: $$\mathbb P(C_{t+1} = c_{t+1} \mid D_1 = d_1, \dots, D_t = d_t) \propto \sum_{c_t} \mathbb P(C_t = c_t \mid D_1 = d_1, \dots, D_t = d_t) p(c_{t+1} \mid c_t).$$ Again, the posterior probability is stored as self.belief in ExactInference.

Finish ExactInference by implementing the elapseTime method. When you are all done, you should be able to track a moving car well enough to drive autonomously:

python drive.py -a -d -k 1 -i exactInference

Notes:

You can also drive autonomously in the presence of more than one car:python drive.py -a -d -k 3 -i exactInference

You can also drive down Lombard:python drive.py -a -d -k 3 -i exactInference -l lombard

On Lombard, the autonomous driver may attempt to drive up and down the street before heading towards the target area. Again, focus on the car tracking component, instead of the actual driving.

Problem 4: Particle filtering

Though exact inference works well for the small maps, it wastes a lot of effort computing probabilities for every available tile, even for tiles that are unlikely to have a car on them. We can solve this problem using a particle filter. Updates to the particle filter have complexity that's linear in the number of particles, rather than linear in the number of tiles.

In this problem, you'll implement two short but important methods for the ParticleFilter class in submission.py. When you're finished, your code should be able to track cars nearly as effectively as it does with exact inference.

Some of the code has been provided for you. For example, the particles have already been initialized randomly. You need to fill in the observe and elapseTime functions. These should modify self.particles, which is a map from tiles (row, col) to the number of particles existing at that tile, and self.belief, which needs to be updated each time you re-sample the particle locations.

You should use the same transition probabilities as in exact inference. The belief distribution generated by a particle filter is expected to look noisier compared to the one obtained by exact inference.

python drive.py -a -i particleFilter -l lombard

To debug, you might want to start with the parked car flag (-p) and the display car flag (-d).

Note: The random number generator inside util.weightedRandomChoice behaves differently on different systems' versions of Python (e.g., Unix and Windows versions of Python). Please test this question (run grader.py) on rice. When copying files to rice, make sure you copy the entire folder using scp with the recursive option -r.

Problem 5: Which car is it?

So far, we have assumed that we have a distinct noisy distance reading for each car, but in reality, our microphone would just pick up an undistinguished set of these signals, and we wouldn't know which distance reading corresponds to which car. First, let's extend the notation from before: let $C_{ti} \in \mathbb R^2$ be the location of the $i$-th car at the time step $t$, for $i = 1, \dots, K$ and $t = 1, \dots, T$. Recall that all the cars move independently according to the transition dynamics as before.

Let $D_{ti} \in \mathbb R$ be the noisy distance measurement of the $i$-th car, which is now not directly observed. Instead, we observe the set of distances $D_t = \{ D_{t1}, \dots, D_{tK} \}$. (For simplicity, we'll assume that all distances are distinct values.) Alternatively, you can think of $E_t = (E_{t1}, \dots, E_{tK})$ as a list which is a uniformly random permutation of the noisy distances $(D_{t1}, \dots, D_{tK})$. For example, suppose $K=2$ and $T = 2$. Before, we might have gotten distance readings of $1$ and $2$ for the first car and $3$ and $4$ for the second car at time steps $1$ and $2$, respectively. Now, our sensor readings would be permutations of $\{1, 3\}$ (at time step $1$) and $\{2, 4\}$ (at time step $2$). Thus, even if we knew the second car was distance $3$ away at time $t = 1$, we wouldn't know if it moved further away (to distance $4$) or closer (to distance $2$) at time $t = 2$.

Suppose we have $K=2$ cars and one time step $T=1$. Write an expression for the conditional distribution $\mathbb P(C_{11}, C_{12} \mid E_1 = e_1)$ as a function of the PDF of a Gaussian $\mathcal p_{\mathcal N}(v; \mu, \sigma^2)$ and the prior probability $p(c_{11})$ and $p(c_{12})$ over car locations. Your final answer should not contain variables $d_{11}$, $d_{12}$.Remember that $\mathcal p_{\mathcal N}(v; \mu, \sigma^2)$ is the probability of a random variable, $v$, in a Gaussian distribution with mean $\mu$ and standard deviation $\sigma$.

Hint: for $K=1$, the answer would be $$\mathbb P(C_{11} = c_{11} \mid E_1 = e_1) \propto p(c_{11}) p_{\mathcal N}(e_{11}; \|a_1 - c_{11}\|, \sigma^2).$$ where $a_t$ is the position of the car at time t. You might find it useful to draw the Bayesian network and think about the distribution of $E_t$ given $D_{t1}, \dots, D_{tK}$.

Assuming the prior $p(c_{1i})$ is the same for all $i$, show that the number of assignments for all $K$ cars $(c_{11}, \dots, c_{1K})$ that obtain the maximum value of $\mathbb P(C_{11} = c_{11}, \dots, C_{1K} = c_{1K} \mid E_1 = e_1)$ is at least $K!$.You can also assume that the car locations that maximize the probability above are unique ($C_{1i} \neq c_{1j}$ for all $i \neq j$).

Note: you don't need to produce a complicated proof for this question. It is acceptable to provide a clear explanation based on your intuitive understanding of the scenario.

For general $K$, what is the treewidth corresponding to the posterior distribution over all $K$ car locations at all $T$ time steps conditioned on all the sensor readings: $$\mathbb P(C_{11} = c_{11}, \dots, C_{1K} = c_{1K}, \dots, C_{T1} = c_{T1}, \dots, C_{TK} = c_{TK} \mid E_1 = e_1, \dots, E_T = e_T)?$$ Briefly justify your answer.

(extra credit) Now suppose you change your sensors so that at each time step $t$, they return the list of exact positions of the $K$ cars, but shifted (with wrap around) by a random amount. For example, if the true car positions at time step $1$ are $c_{11} = 1, c_{12} = 3, c_{13} = 8, c_{14} = 5$, then $e_1$ would be $[1, 3, 8, 5]$, $[3, 8, 5, 1]$, $[8, 5, 1, 3]$, or $[5, 1, 3, 8]$, each with probability $1/4$. Describe an efficient algorithm for computing $p(c_{ti} \mid e_1, \dots, e_T)$ for any time step $t$ and car $i$. Your algorithm should not be exponential in $K$ or $T$.

Starting from: $30

You'll get 1 file (443.7KB)