CSE 490R Review
My solution to the homework
Lab0
- Localization: Determine the pose of a robot relative to a given map of the environment.
- Mapping: Construct a map.
- Planning: Plan a trajectory to goal based on localization and maps.
Bayes Filter
The Bayes filter algorithm includes two steps, the first step is prediction (push belief through dynamics given action), the second step is correction (apply Bayes rule given measurement).
$$ \begin{gather*} \overline{bel}\left(x_{t}\right) = \int p\left(x_{t} | u_{t}, x_{t-1}\right) bel\left(x_{t-1}\right) d x_{t-1} \\ bel\left(x_{t}\right) = \eta p\left(z_{t} | x_{t}\right) \overline{bel}\left(x_{t}\right) \end{gather*} $$Kalman Filter
Assumptions:
- Linear dynamics $p(x_t | u_t, x_{t-1}) = A_{t} x_{t-1}+B_{t} u_{t}+\varepsilon_{t}$, where $\varepsilon_t \sim \mathcal{N}(0, R_t)$
- Linear measurement model $p(z_t | x_t) = C_{t} x_{t}+\delta_{t}$, where $\delta_t \sim \mathcal{N}(0, Q_t)$
- Initial belief $bel(x_0)$ is a Gaussian distribution $$ bel\left(x_{0}\right)=p\left(x_{0}\right)=\det\left(2 \pi \Sigma_{0}\right)^{-\frac{1}{2}} \exp \left\{-\frac{1}{2}\left(x_{0}-\mu_{0}\right)^{T} \Sigma_{0}^{-1}\left(x_{0}-\mu_{0}\right)\right\} $$
Lab1
Motion Model
- Kinematic model: map wheel speeds to robot velocities.
- Dynamic model: map wheel torques to robot accelerations.
Consider only kinematic model $p(x_t | u_t, x_{t-1})$ for now (assume we can set the speed directly), then $x = [x, y, \theta]^T$, $u = [V, \delta]^T$, where $\theta$ is the heading and $\delta$ is the steering angle.
Now we will derive the motion model for rear axel. Note that a rigid body undergoing rotation and translation can be viewed as pure rotation about a instant center of rotation:
Noise
- Control signal error: $\hat{V} \sim \mathcal{N}(V, \sigma_v^2)$, $\hat{\delta} \sim \mathcal{N}(\delta, \sigma_\delta^2)$
- Incorrect physics (the Kinematic Model is inaccurate): $\hat{x} \sim \mathcal{N}\left(x, \sigma_{x}^{2}\right)$, $\hat{y} \sim \mathcal{N}\left(y, \sigma_{y}^{2}\right)$, $\hat{\theta} \sim \mathcal{N}\left(\theta, \sigma_{\theta}^{2}\right)$
Sensor Model
$p(z_t | x_t, m)$ is the probability of sensor reading $z_t$ given state $x_t$ and map $m$. Calculate simulated sensor reading $z^*$ form $x$ and $m$ and then compare with $z$.
$$ p\left(z_{t} | x_{t}, m\right)=\prod_{i=1}^{K} p\left(z_{t}^{k} | x_{t}, m\right) $$Noise
- $$ \require{mathtools} p_{\text {hit}}\left(z_{t}^{k} | x_{t}, m\right)=\begin{dcases} \eta \mathcal{N}\left(z_{t}^{k} ; z_{t}^{k *}, \sigma_{\text {hit}}^{2}\right) & \text {if } 0 \le z_{t}^{k} \le z_{\max} \\ 0 & \text { otherwise } \end{dcases} $$
- $$ p_{\text {short}}\left(z_{t}^{k} | x_{t}, m\right)=\begin{dcases} \eta \lambda_{\text {short}} e^{-\lambda_{\text {short}} z_{t}^{k}} & \text {if } 0 \le z_{t}^{k} \le z_{t}^{k *} \\ 0 & \text {otherwise } \end{dcases} $$
- $$ p_{\max}\left(z_{t}^{k} | x_{t}, m\right)=I\left(z=z_{\max}\right)=\begin{cases} 1 & \text {if } z=z_{\max } \\ 0 & \text {otherwise } \end{cases} $$
- $$
p_{\text {rand}}\left(z_{t}^{k} | x_{t}, m\right)=\begin{dcases}
\frac{1}{z_{\max}} & \text {if } 0 \le z_{t}^{k}
Particle Filter
$$ \begin{align*} &\textbf{Algorithm ParticleFilter} (\mathcal{X}_{t-1}, u_t, z_t ): \\ &\bar{\mathcal{X}}_{t}=\mathcal{X}_{t}=\varnothing \\ &\text{for } m=1 \text{ to } M \text{ do} \\ &\qquad \text{sample } x_{t}^{[m]} \sim p\left(x_{t} | u_{t}, x_{t-1}^{[m]}\right) \\ &\qquad {w_{t}^{[m]}=p\left(z_{t} | x_{t}^{[m]}\right)} \\ &\qquad {\bar{\mathcal{X}}_{t}=\bar{\mathcal{X}}_{t}+\left\langle x_{t}^{[m]}, w_{t}^{[m]}\right\rangle} \\ &\text{endfor } \\ &\text{for } m = 1 \text{ to } M \text{ do} \\ &\qquad \text{draw } i \text{ with probability } \propto w_{t}^{[i]} \\ &\qquad \text{add } x_t^{[i]} \text{ to } \mathcal{X}_t \\ &\text{endfor} \\ &\text{return } \mathcal{X}_t \end{align*} $$We first sample $x_t$ from the state transition distribution, then calculate the importance factor, $w_t$. In the next loop, we do the resampling/importance sampling, to change the distribution of particles from $\overline{bel}(x_t)$ to $bel(x_t)$.
Resample
Resampling can cause high-variance (low-entropy) problem, where particles are depleted. Possible fixes:
- If the variance of weights low, don’t resample.
- Use low-variance sampling.
Expected Pose
Calculate expected pose from particles: $x$ and $y$ can be computed directly by the weighted average. However, the weighted average of $\theta$ is not accurate. Thus we use cosine and sine averaging. (Ref)
Code
MotionModel.py
:
- Subscribe to motor topic, get control info (speed and steering angle), also save last frame info for calculation.
- Add variance to speed and steering angle. Apply motion model on particles. Add variance to states.
SensorModle.py
:
- Precompute the sensor model table and use range_libc package to achieve 2D raycasting for 2D occupancy grid.
- Subscribe to scan topic, filter out invalid and extreme scan values, downsample, and pass processed data to update weights.
ParticleFilter.py
:
- Transfer map to occupancy grid.
- Globally initialize particles and weights. Initialize motion model and sensor model.
- Subscribe to
/intialpose
topic, initialize particles and weights around initial pos. - When receiving scan infos, update weights, resample and calculate the expected pose for visualization.
Lab2
Uses Model | Stability Guarantee | Minimize Cost | |
---|---|---|---|
PID | No | No | No |
Pure Pursuit | Circular arcs | Yes - with assumptions | No |
Lyapunov | Non-linear | Yes | No |
LQR | Linear | Yes | Quadratic |
iLQR | Non-linear | Yes | Yes |
Control
Only consider cross-track error $e_{ct}$ and control steering angle for now.
PID
$$ u=-\left(K_{p} e_{c t}+K_{i} \int e_{c t}(t) d t+K_{d} \dot{e}_{c t}\right) $$where
- $K_p$: Proportional coefficient
- $K_i$: Integral coefficient
- $K_d$: Derivative coefficient
Pure-Pursuit
Key idea: The car is always moving in a circular arc.
- Find a lookahead and compute arc
- Move along the arc
- Go to step 1
Lyapunov Control
$$ V\left(e_{ct}, \theta_{e}\right)=\frac{1}{2} k_{1} e_{ct}^{2}+\frac{1}{2} \theta_{e}^{2} $$$$ \begin{split} \dot{V}\left(e_{c t}, \theta_{e}\right)&=k_{1} e_{c t} \dot{e}_{c t}+\theta_{e} \dot{\theta}_{e}\\ &=k_{1} e_{c t} V \sin \theta_{e}+\theta_{e} \frac{V}{B} \tan u \end{split} $$$$ \begin{gather*} \theta_{e} \frac{V}{B} \tan u=-k_{1} e_{c t} V \sin \theta_{e}-k_{2} \theta_{e}^{2} \\ \tan u=-\frac{k_{1} e_{c t} B}{\theta_{e}} \sin \theta_{e}-\frac{B}{V} k_{2} \theta_{e} \\ u=\tan ^{-1}\left(-\frac{k_{1} e_{c t} B}{\theta_{e}} \sin \theta_{e}-\frac{B}{V} k_{2} \theta_{e}\right) \end{gather*} $$LQR
$$ \min _{u(t)} \int_{0}^{\infty}\left(w_{1} e(t)^{2}+w_{2} u(t)^{2} \right) dt $$Given
- Linear dynamic system $$ x_{t+1} = Ax_t + Bu_t $$
- Quadratic cost $$ J=\sum_{t=0}^{T-1} x_{t}^{T} Q x_{t}+u_{t}^{T} R u_{t} $$
MPC
- Plan a sequence of control actions
- Predict the set of next states to a horizon H
- Evaluate the cost/constraint of the states and controls
- Optimize the cost
Code
controlnode.py
: main script
- Define Subscribers to pose and path infos; Publishers for visualization and Services for reset
- Enter main program when initial pose is set and controller is ready. Get pose and reference pose, get next control and publish to
/vesc/high_level/ackermann_cmd_mux/input/nav_0
topic. - Stop when path is completed
runner_script.py
:
- Loading different paths and speed
- Start the controller
controller.py
: base controller class
- Store path info
- Define utility functions: get reference pose by index, get pose and pose_ref error
Find reference pose function is the same for all controllers: Find the nearest point on the path and lookahead some distance.
pid.py
:
- Use PD equation to get next control (steering angle)
purepursuit.py
:
- Use Purepursuit equation to get next control
mpc.py
:
- Dividing steering angle $[-\pi, \pi]$ equally to $K$ rollouts
- Execute each steering angel through $T$ timesteps to collect $K * T$ poses
- Evaluate the cost of each rollout by collision cost and error cost
- Collision cost: if any pose in a trajectory is in collision with the map, add a large cost
- Error cost: norm of the distance between last pose and the reference pose, weighted by a constant
- Choose the rollout with the minimal cost and execute the first step
- Return to step 1
mpc2.py
: Similar to mpc but use scan info rather than map. The only thing change is the obstacles cost.
- Calculate $N$ obstacles pose in the map by scan info
- For $K*T$ poses, calculate the distance with every obstacles to get $K*T*N$ array
- Find the minimal distance for every pose to get $K * T$ array
- Average through timesteps to get $K$ array, then weighted by a constant to get the obstacles cost.
nonlinear.py
:
- Use Lyapunov control equation to get the control
Lab3
Steps for planing the path from one point to another:
- Random sample points on the map and construct the graph
- Use planning algorithm (A*) to search the graph for optimal path
Graph Construction
The graph we are constructing is called Random geometric graph (RGG)
- Sample a set of collision free vertices $V$
- Connect neighboring vertices to get edges $E$
Sampling
Uniform random sampling tends to clump. We want points to be spread out evenly, which can be achieved by Halton sequence
Optimal Radius
$$ r = \left(\frac{\ln|V|}{\alpha_{p,d} |V|}\right)^{1/n} $$where $\alpha_{p,d}$ is a constant. For special case of a two-dimensional space and the euclidean norm ($d=2$, $p=2$), $\alpha_{p,d} = \pi$
Dubins Path
$$ \dot{q}(t) = f(q(t), u(t)) \\ q(0) = q_1,\quad q(t) = q_2 $$where $q=(x,y,\theta)$ in our case.
$$ \{LRL, RLR, LSL, LSR, RSL, RSR\} $$where $L,R,S$ represent turn left, right, straight, respectively.
Planning Algorithm
Given start node $s_0$, goal $s_1$ and cost $c(s, s')$, create objects:
- OPEN: priority queue of nodes to be processed
- CLOSED: list of nodes already processed
- $g(s)$: estimate of the least cost from start to a given node
The pseudocode for best first search can be expressed as
- Push $s_0$ into OPEN
- While $s_1$ not expanded 3. Pop best from OPEN 4. Add best to CLOSED 5. For every successor s’ 6. If $g(s') > g(s) + c(s, s')$ 7. $g(s') = g(s) + c(s, s')$ 8. Add (update) $s'$ to OPEN
The main problem is how to choose the heuristic function $f(s)$ for step 3.
Dijkstra’s Algorithm
Choose $f(s) = g(s)$. Always pop the node with the smallest cost from the origin first.
A*
If we can pre-evaluate the cost from the node to the goal $h(s)$, then we can choose a better $f(s) = g(s) + h(s)$.
- If $h(s)$ is admissible $h(s) \le h^*(s)$, $h(\text{goal}) = 0$, then the path return by A* is optimal.
- If $h(s)$ is consistent $h(s) \le c(s, s') + h(s')$, $h(\text{goal}) = 0$, then A* is optimal and efficient (will not re-expand a node).
All consistent heuristics are admissible, not vice versa.
Weighted A*
Choose $f(s) = g(s) + \epsilon h(s)$, where $\epsilon > 1$. It is more efficient and the solution is $\epsilon$-optimal $c \le \epsilon c^*$
Lazy A*
Instead of checking edge collision for all neighbors, only check the edge to parent when expanding. OPEN list will have multiple copies of a node since we haven’t collision check.
Shortcut
After a path is found, we can randomly pick two nodes and connect them directly if the edge is collision-free.
Code
run.py
: main program
- Load map info
- Construct graph given environment, sampler, number of vertices, connection radius.
- Add start and end node
- Use A* or lazy A* algorithm to search the optimal path and visualize
MapEnvironment.py
: define utility functions associated with planning
- Check edge collision
- Compute heuristic and distance function
- Generate path on the map
- Visualization of the graph and path
Sampler.py
: create random samples for graph construction
- Use halton sequence to generate 2-D random vertices in $(0,1)$
- Scale by map info
graph_maker.py
: construct graph
- Use python package NetworkX to easily construct a graph
- Add sampled valid vertices
- Connect edges within radius and without collision (do not check collision if using lazy A*)
astar.py
: A* algorithm
- Use
heapq
package to create priority queue to store[f(s), count, node, g(s), parent]
info.count
is used to prevent comparing node when $f(s)$ are the same. - Use dict
enqueued
to store $g(s)$ and $h(s)$ for a node, and another dictexplored
to store the explored node and its parent node. - Add start node to the queue.
- While queue is not empty, pop one node and add it to
explored
if it is not already there. - For all the neighbor node $s'$, compute $g(s') = g(s) + c(s,s')$ if $s'$ is not already in
explored
. - If $s'$ is in
enqueued
, get its previous $g'(s')$ and $h'(s')$. Continue to next neighbor if $g'(s) \le g(s')$, since it is better. If not inenqueued
, compute $h(s')$ by the heuristic function. - Update $g(s')$ and $h(s')$ in
enqueued
and push it to the queue. - If reach target node, continually find parent node by
explored
and return the path.
lazy_astar.py
: lazy A* algorithm
- When expanding node, check its edge collision with parent.
- When checking neighbors, do not need to compare $g'(s')$ and $g(s')$
- Multiple nodes with different parents can be added to the queue.
runDubins.py
: similar to run.py
but use Dubins environment
DubinsMapEnvironment.py
: inherited from MapEnvironment
- Compute distances with Dubins path
- Compute heuristic with Dubins path
- Generate path by Dubins path planning
DubinsSampler.py
:
- Sample $N$ vertices same as in
Sampler.py
- Divide angles to $M$ parts, add angles to each vertices to create $M\times N$ samples
Dubins.py
: utility function for Dubins path
ROSPlanner.py
: take goal from rviz and do planning. Be careful of the frame change between map and world
- Load map info
- Construct graph and save for later use
- Subscribe to pose topic, save current pose
- Subscribe to goal topic, plan from the current pose to the goal, publish the path using service defined in lab2
- If there are multiple goals, sequentially planning through them and combine the path