MPE Mathematical Problems in Engineering 1563-5147 1024-123X Hindawi Publishing Corporation 528162 10.1155/2013/528162 528162 Research Article Guidance of Autonomous Amphibious Vehicles for Flood Rescue Support Ragi Shankarachary 1 Tan ChingSeong 2 Chong Edwin K. P. 1 Li Yongping 1 Department of Electrical and Computer Engineering Colorado State University Fort Collins, CO 80523-1373 USA colostate.edu 2 Faculty of Engineering Multimedia University Malaysia mmu.edu.my 2013 23 5 2013 2013 14 02 2013 06 05 2013 07 05 2013 2013 Copyright © 2013 Shankarachary Ragi et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

We develop a path-planning algorithm to guide autonomous amphibious vehicles (AAVs) for flood rescue support missions. Specifically, we develop an algorithm to control multiple AAVs to reach/rescue multiple victims (also called targets) in a flood scenario in 2D, where the flood water flows across the scene and the targets move (drifted by the flood water) along the flood stream. A target is said to be rescued if an AAV lies within a circular region of a certain radius around the target. The goal is to control the AAVs such that each target gets rescued while optimizing a certain performance objective. The algorithm design is based on the theory of partially observable Markov decision process (POMDP). In practice, POMDP problems are hard to solve exactly, so we use an approximation method called nominal belief-state optimization (NBO). We compare the performance of the NBO approach with a greedy approach.

1. Introduction

Various guidance algorithms for autonomous amphibious vehicles (AAVs) are being designed and tested to fight today’s global warming disasters such as flooding, typhoon, and hurricane . With this motivation, we present a guidance framework to control multiple AAVs to rescue multiple victims (henceforth called targets) in a flood situation, where the flood water (interchangeably called river) flows along a valley as shown in Figure 1. A target is said to be rescued when an AAV is within the circular region of radius ddist-thresh on the 2D plane around the target. In general, AAVs are equipped with various advanced sensors such as polarized stereo vision, laser scanning, and SONAR . The sensors onboard an AAV generate the (noisy) measurements corresponding to the targets and the river. Our goal is to design a path-planning algorithm that guides the AAVs so that every target gets rescued, while maximizing a performance measure (discussed later). The algorithm runs on a notional central fusion node, which collects the measurements from the sensors on-board each AAV, fuses them and updates the tracks on the targets and the river state (discussed later), computes the control commands for the AAVs, and sends the control commands back to the AAVs.

Flood scenario.

Guidance control methods [1, 79] for AAVs are normally based on a standard three-layered system architecture that requires human-machine interactions. We design the guidance algorithm based on the theory of partially observable Markov decision process (POMDP) [10, 11]. There are several other autonomous control methods in the literature for AAVs and underwater vehicles, for example, . Our approach differs from these existing approaches in that we place the guidance problem in the context of POMDP, wherein this approach has a look-ahead property, which trades off short-term for long-term performance.

2. Problem Specification

The AAV guidance problem is specified as follows.

2.1. Targets

In this study, we assume that there are multiple mobile targets (flood victims) located in a river, being drifted down by the flood water, as shown in Figure 1.

2.2. Autonomous Amphibious Vehicles (AAVs)

There are multiple autonomous amphibious vehicles (AAVs) located on the shore, as shown in Figure 1. An AAV is controlled by the following kinematic controls: forward acceleration and steering angle. Each AAV is equipped with on-board sensors that generate measurements of targets and the river depth. In this problem, AAVs float when moving in the river. For the purpose of this study, we assume that the number of AAVs and the number of targets are the same.

2.3. Environmental Conditions

The elevation map of the region is known a priori. The landscape for this problem is shown in Figure 1, which shows a river flowing along a valley from the north toward the south. The state of the river includes the depth dkref at a reference point on the map (lowest point in the landscape, e.g., some location at the bottom of the valley as shown in Figure 1).

2.4. River Model

Typically a river flows slowly near the coastlines (where the river is shallow) and flows quickly far from the coastlines (i.e., toward the center of the river where the river is deep). In this paper, we assume that the river flows from the north toward the south in a v-shaped channel as shown in Figure 1. We adopt the logarithmic velocity profile to model the velocity of the flow (see  for a detailed description). According to this model, the speed of the river, at the surface, at the location (p,q) at time k is given by (1)wk(p,q)=C1[log(dk(p,q))+C2], where dk(p,q) is the depth of the river at the location (p,q) at time k, and C1 (a function of the viscosity and the density of flood water) and C2 are constants (see  for more details).

2.5. Observations

The sensors onboard an AAV generate noisy observations of target locations and the depth of the river directly beneath the vehicle, that is, the sensors generate the observations of the depth of the river only when the AAV is in the river.

2.6. Objective

A target is said to be rescued if there is an AAV within a circular region of radius ddist-thresh around the target. The objective is to minimize the average rescue time, where the average is over the number of targets, and the rescue time of a target is defined as the time it takes to rescue the target.

3. Problem Formulation

We cast the AAV guidance problem into the framework of a partially observable Markov decision process (POMDP). A POMDP is a mathematical framework useful for solving resource control problems and enables us to exploit approximation methods for POMDPs to design our AAV guidance algorithm. A POMDP evolves in discrete time steps. We use k as the discrete-time index. To cast the AAV guidance problem into the POMDP framework, we need to define the following key components in terms of our guidance problem as follows.

3.1. States

Let xk represent the state of the system at time k. The state of the system includes the state of the vehicles (AAVs) sk, river state (depth of the river at a reference location) dkref, target state χk, and track states (ξkriv,Pkriv,ξktarg,Pktarg), that is, xk=(sk,dkref,χk,ξkriv,Pkriv,ξktarg,Pktarg). The vehicle state sk includes the locations and the velocities of the AAVs at time k. The river state dkref is the depth of the river at the reference point at time k. The reference point is the lowest point in the elevation map, that is, some location at the bottom of the valley in the landscape, as shown in Figure 1. Here, we assume that the flow direction of the river is the same everywhere and is known a priori. The target state χk includes the locations and the velocities of the targets at time k. The track states represent the state of the tracking algorithm, where ξkriv and Pkriv are the mean and the variance, standard in Kalman filter equations, corresponding to the river state, and, similarly, ξktarg is the mean vector and Pktarg is the covariance matrix corresponding to the target state.

3.2. Observations and Observation Law

The vehicle and the track states are assumed to be fully observable. The river and the target states are only partially observable. The observation of the river state at an AAV is given by (2)zkriv={dkref+nkrivif  AAV  is  in  river,no  measurementotherwise, where nkriv~𝒩(0,Rk), and Rk is the measurement variance. The sensors at an AAV generate the measurement of the river state only when the AAV is in the river. In practice, the sensors on an AAV measure the depth of the river exactly below the AAV. We wrote the observation model (2) as if the sensors are generating the observations of the depth of the river at the reference point. The rationale behind this assumption is that we can always calculate the depth of the river at the reference point given the elevation map and the observed depth of the river at a different location. The observation of the jth target at an AAV is given by (3)zkχj={Hχkj+nktargif  there  is  line-of-sight,no  measurementotherwise, where H is the target-state observation model, χkj is the state of jth target, and nktarg~𝒩(0,Sk), where Sk is the measurement covariance matrix. The line-of-sight between the target and the AAV is blocked sometimes, for example, whenever the target sinks in the water.

3.3. Actions

The actions include the controllable aspects of the system. In this problem, the actions include the decisions on the assignment of AAVs to targets, and kinematic control commands for AAVs. Let uk be the action tuple at time k, which is given by uk=(gk,ak), where ak represents kinematic control vectors (includes forward acceleration and steering angle for each AAV), and gk is a vector, which represents the assignment of AAVs to targets, that is, gk(i)=j means that the ith AAV is assigned to the jth target. For the purpose of this study, the number of AAVs and the targets is the same. Each AAV is assigned to only one target, and each target gets assigned only one AAV, that is, gk represents a one-to-one correspondence between the AAVs and the targets.

3.4. State-Transition Law

The state-transition law specifies the next-state distribution given the current state and the action. The transition function for the vehicle state is given by sk+1=ψ(sk,ak,ξkriv), where ψ (defined later) represents the AAV kinematic model, sk is the vehicle state, ak is the kinematic control vector (includes forward acceleration and steering angle), and ξkriv is the estimated river state at time k. The river state evolves according to the following equation: (4)dk+1ref=dkref+ok,ok~𝒩(0,Ukriv), where Ukriv is the process variance corresponding to the river state evolution. The target state evolves according to (5)χk+1=Fχk+ek,ek~𝒩(0,Uktarg), where F represents the target motion model, and Uktarg is the process covariance matrix corresponding to the target state evolution. The track states evolve according to the Kalman filter equations given the observations from the sensors onboard the AAVs. When the observations are not available, the track states evolve according to the Kalman filter equations, where only the prediction step is performed and the update step is not performed.

3.5. Cost

The cost function represents the cost of performing an action at the current state. The cost function is given by (6)C(xk,uk)=i=1N1{E[sk+1i,pos-ξk+1gk(i),targ,posxk,uk]>ddist-thresh[sk+1i,pos-ξk+1gk(i),targ,posxk,uk]}, where sk+1i,pos represents the 2D position coordinates of ith AAV, ξk+1j,targ,pos represents the estimated 2D position coordinates of the jth target at time k+1, · is the Euclidean norm (everywhere in this paper), and 1{·} is the indicator function which equals 1 when the expected distance between the AAV and the target at time k+1 is greater than some threshold distance ddist-thresh and 0 otherwise.

3.6. Belief State

The belief state bk is the posterior distribution of the state at time k. The vehicle and the track states are assumed to be fully observable, that is, the belief state corresponding to the vehicle state is given by bks(s)=δ(s-sk), where δ(·) is the Kronecker delta function. Similarly, the belief states corresponding to the track states can be written in terms of the actual track states. The belief states corresponding to the river and the target are the posterior distributions of dkref and χk, respectively, given the history of observations.

4. Objective and Optimal Policy

The goal is to find the action sequence (u0,u1,,uH-1) such that the expected cumulative cost over a time horizon H is minimized. The expected cumulative cost is given by (7)JH=E[k=0H-1C(xk,uk)]. We can write the expected cumulative cost in terms of the belief states given the initial belief state b0 (similar to the treatment in [10, 11]) as follows: (8)JH(b0)=E[k=0H-1c(bk,uk)b0], where c(bk,uk)=C(x,uk)bk(x)dx, and b0 is the belief state at time k=0. From Bellman’s principle of optimality , the optimal objective function value is given by (9)JH*(b0)=minu{c(b0,u)+E[JH-1*(b1)b0,u]}, where b1 is the random next belief state, JH-1* is the optimal cumulative cost over the horizon H-1, k=1,2,,H-1, and E[·b0,u] is the conditional expectation given the current belief state b0 and the current action u at time k=0. Let us define the Q value of taking action u given the current belief state b0: (10)QH(b0,u)=c(b0,u)+E[JH-1*(b1)b0,u]. The optimal policy (from Bellman’s principle) at time k=0 can be written as (11)π0*(b0)=argminuQH(b0,u). In general, it is hard to obtain the Q value exactly. There are several approximation methods in the literature: heuristic expected-cost-to-go (ECTG) , parametric approximation , policy rollout , hindsight optimization , and foresight optimization . In this paper, we use one such approximation method called nominal belief-state optimization (NBO), which was introduced in  along with other approximations and techniques specific to guidance problems. The rationale behind choosing NBO method over other methods to solve POMDP is that it is relatively inexpensive in terms of computation time, that is, the computational requirements are not prohibitive unlike other approximation methods. The following subsection provides a brief description of the NBO method.

4.1. NBO Approximation Method

The computational requirements of obtaining the optimal assignments of AAVs to targets (gk) over a long horizon are prohibitive. Also, we expect that the optimal assignment of AAVs to targets (gk) over a long horizon does not change with time. For these reasons, in the NBO method, we keep the assignment of AAVs to targets fixed. In other words, in approximating the expected cost-to-go in (10), gk remains fixed over the planning horizon H. Therefore, we drop the subscript k from gk in the objective function used in the planning based on (10), that is, gk=g for all k. In the NBO approximation method, we use the following objective function, written in terms of belief states: (12)JH(b0)=E[k=0H-1c(bk,ak,g)b0], where ak represents the kinematic controls for the AAVs, and g is the assignment of AAVs to the targets.

The belief states corresponding to the river state and the target state are given by (13)bkriv(d)=𝒩(d-ξkriv,Pkriv),bktarg(χ)=𝒩(χ-ξktarg,Pktarg), where (ξkriv,Pkriv,ξktarg,Pktarg) are the track states corresponding to the river and the target states, respectively, which evolve according to the Kalman filter equations. In the NBO method, we approximate the objective function as follows: (14)JH(b0)k=0H-1c(b^k,ak,g), where b^1,,b^H-1 is a nominal belief-state sequence, and the optimization is over an action sequence g,a0,,aH-1. We obtain the nominal belief states by evolving the current belief state with exactly zero-noise sequence over the horizon H (similar to the treatment in [10, 11]). Therefore, the objective function from the NBO method is given by (15)JNBO(b0)=k=0H-1i=1N  1{s^k+1i,pos-ξ^k+1g(i),targ,pos>ddist-thresh}, where s^k+1i,pos is the nominal position of the ith AAV (defined below), 𝒩(ξ^k+1j,targ,P^k+1j,targ) is the nominal belief state of the jth target at time k+1, where ξ^k+1j,targ,pos (component of ξ^k+1j,targ) represents the position estimate of the target. This nominal target belief state is obtained by evolving the track state component ξ^kj,targ with exactly zero-noise sequence as follows: (16)ξ^k+1j,targ=Fξ^kj,targ. The evolution of vehicle state depends on the river state estimate ξkriv. In the NBO method, ξkriv is replaced with ξ^kriv in the AAV kinematic model ψ(·), where (ξ^1riv,,ξ^Hriv) are the nominal track state components corresponding to the river state, and the obtained positions of the ith AAV s^k+1i,pos are called nominal positions.

Here, we adopt an approach called “receding horizon control,” according to which we optimize the action sequence for H time steps at the current time step, implement only the action corresponding to the current time step, and again optimize the action sequence for H time steps in the next time step. The length of the planning horizon H should be large enough for an AAV to receive a benefit by moving toward a target. Due to computational constraints, we cannot have an arbitrarily long horizon. Therefore, we truncate the length of the horizon to a few time steps (we set H=6 in our simulations) and append the cost function with an appropriate expected cost-to-go (ECTG). The following is a distance-based ECTG: (17)JHdist-ECTG=i=1Ns^Hi,pos-ξ^Hg(i),targ,pos, where s^Hi,pos is the nominal position of the ith AAV, and ξ^Hj,targ,pos is the estimated location of the jth target (from NBO approach) at time k=H. Therefore, the objective function from the NBO method is given by (18)JNBO(b0)=k=0H-1i=1N1{s^k+1i,pos-ξ^k+1g(i),targ,pos>ddist-thresh}+JHdist-ECTG, where JHdist-ECTG is the distance-based ECTG.

4.2. AAV Kinematics

The kinematic equations of an AAV vary depending on whether the AAV is in the river or on the land. When the AAV is in the river, we take into account the speed of the river to write the kinematic equations. The steering and thrust generation of the vehicle are modeled based on the work done by the authors of [2, 22], which is designed using single drive system. The vehicle is front-wheel driven on land. When the AAV is in the river, it is propelled using the centrifugal pump from the front wheels. The following subsections describe the kinematics of AAV on the land and in the river.

4.2.1. Kinematics of AAVs on the Land

This subsection provides the definition of ψ, which was introduced in Section 3, when the vehicle is on land. Let sk=(pk,qk,vk,θk) be the state of the vehicle at time k, where (pk,qk) represents the location of the vehicle on the 2D plane, vk represents the speed of the vehicle along the heading direction, and θk represents the heading angle of the vehicle at time k. Let ak=(fk,ϕk) represent the action vector of the vehicle, where fk represents the acceleration along the direction of the front wheels, and ϕk represents the steering angle of the front wheels. The (simplified) schematic of a basic four-wheeled vehicle is shown in Figure 2. The control variable fk lies within the interval [-fland,fland], where fland (or -fland) is the maximum acceleration (or deceleration), and the control variable ϕk lies within the interval [-δland,δland], where δland is the maximum steering angle. The function ψ can be specified by a set of nonlinear kinematic equations, as shown below: (19)pk+1=pk+vkTcos(θk),qk+1=qk+vkTsin(θk),vk+1=vk+fkTcos(ϕk),θk+1=θk-2fkT2LW2+L2sin(ϕk), where T is the length of the time step, W is the width of the vehicle, and L is the distance between the front axle and the rear axle. The derivation of the heading angle update (19) is as follows. When the front wheels of the vehicle are oriented at a particular angle ϕk with respect to the main axis of the vehicle (as shown in Figure 2), the heading direction of the vehicle at time k+1 is derived as follows: (20)α=arctan(WL),θk+1=θk+T2L2+W2(fk,1θ-fk,2θ)=θk+fkT2L2+W2[sin(α-ϕk)-sin(α+ϕk)]=θk-2fkT2L2+W2[cos(α)sin(ϕk)]=θk-2fkT2LW2+L2sin(ϕk).

Free body diagram of an AAV.

4.2.2. Kinematics of AAVs on the River

This subsection provides the definition of ψ, when the vehicle is in the river. The kinematic equations of the AAV motion are as follows: (21)pk+1=pk+vkTcos(θk)+w^kx(pk,qk)T,qk+1=qk+vkTsin(θk)+w^ky(pk,qk)T, where w^kx(pk,qk) and w^ky(pk,qk) are the estimated speeds of the river at the location (pk,qk) in x and y directions, respectively, which are obtained from the river state estimate ξ^kriv and the river model presented in Section 2. The speed and the heading angle update equations remain the same as in the case of land. When in water (or river), the control variable fk lies within the interval [-fwater,fwater], where fwater is the maximum acceleration, and ϕk lies within the interval [-δwater,δwater], where δwater is the maximum steering angle. Typically, the values of fwater and δwater are much smaller compared to that of fland and δland.

5. Simulation

We implement the NBO method in MATLAB, and we use the command fmincon (MATLAB’s optimization tool) to solve the optimization problem. For performance comparison, we also implement a greedy approach, where we optimize only the current kinematic control for the AAVs such that the following symmetric-distance-based cost is minimized: (22)JGreedy(bk)=i=1Nminjs^k+1i,pos-ξ^k+1j,targ,pos+  j=1Nminis^k+1i,pos-ξ^k+1j,targ,pos, where s^k+1i,pos and ξ^k+1j,targ,pos are the nominal positions (obtained by evolving the belief states with zero noise) of the ith AAV and the jth target at time k+1, respectively. Our simulation environment is two dimensional, that is, the AAVs, the river, and the targets move in 2D. According to the river model, the speed of the river stream wk at a location (p,q) is given by wk(p,q)=C1[log(dk(p,q))+C2], where dk(p,q) is the depth of the river at (p,q), and C1 and C2 are constants. Since the depth of the river is not fully observable, we estimate dk(p,q) as follows. The elevation map of the landscape is known a priori, that is, if we know the depth of the river at a particular location, we can obtain the depth of the river at all locations. Therefore, we estimate the depth of the river at location (p,q), that is, d^k(p,q) using the estimated depth of the river at the reference point d^kref (=ξ^kriv). Therefore, the estimated speed of the river at location (p,q) is given by w^k(p,q)=C1[log(d^k(p,q))+C2]. We set the length of the horizon H to 6 time steps, and the length of the times step T to 1 second. In the simulations, the flooded river flows along a valley in the landscape from the north toward the south as shown in Figure 1. Since the simulations are in 2D, the river flows toward the -y direction, and the river speed in x direction (toward the east) is zero at every location. Therefore, the estimated speeds of the river at location (p,q) in x and y directions are given by w^kx(p,q)=0 and w^ky(p,q)=-C1[log(d^k(p,q))+C2]. Here, we model the dynamics of the target motion by the constant velocity model (see  for the definition of the variables F and Utarg in (5)).

In the simulations, an AAV is represented by a rectangle, and the line connecting the rectangles represents the trajectory of the AAV. We define a performance metric called average rescue time—the average of the rescue times of each target (the rescue time of a target is the time elapsed after the start of the simulation until it is rescued). The POMDP cost function defined in Section 3 is reflective of this performance metric. We simulate three scenarios: Scenario I, Scenario II, and Scenario III. In Scenario I, there are two AAVs, each one located on the opposite banks of the river, and two targets are moving (being drifted by the moving water) in the river, as shown in Figure 3. Figure 3 shows a snapshot of the scenario at the end of the simulation with the NBO approach, where the average rescue time is 36 time steps. We also simulate Scenario I with the greedy approach, as shown in Figure 4, where the average rescue time is 64 time steps. In Scenario II, there are two AAVs on the left bank of the river, and two targets are moving in the river. We simulate this scenario with both the NBO and the greedy approaches. Figure 5 shows the snapshot of the scenario with the NBO approach at the end of the simulation, where the average rescue time is 45 time steps, and Figure 6 shows the simulation of the same scenario with the greedy approach, where the average rescue time is 62 time steps. In Scenario III, there are three AAVs (two on the left bank of the river and one on the right), and three targets are moving in the river. We simulate this scenario with both the NBO and the greedy approaches. Figure 7 shows the scenario with the NBO approach, where the average rescue time is 48 time steps, and Figure 8 shows the simulation of the same scenario with the greedy approach, where the average rescue time is 76 time steps. The simulation of these scenarios demonstrates that the NBO approach achieves a better coordination among the AAVs compared to the greedy approach while rescuing the targets, as evident from the average rescue times.

Simulation of Scenario I with NBO approach, average rescue time = 36 steps.

Simulation of Scenario I with greedy approach, average rescue time = 64 steps.

Simulation of Scenario II with NBO approach, average rescue time = 45 steps.

Simulation of Scenario II with the greedy approach, average rescue time = 62 steps.

Simulation of Scenario III with NBO approach, average rescue time = 48 steps.

Simulation of Scenario III with the greedy approach, average rescue time = 76 steps.

We compare the performance of the NBO approach with that of the greedy approach through Monte-Carlo simulations. We simulate the above scenarios with the NBO and the greedy approaches separately for 50 Monte-Carlo runs. In each scenario, we compute the average rescue time in every run for both the NBO and the greedy approaches. Figures 9, 10, and 11 show the plots of the cumulative frequencies of average rescue times for the NBO and the greedy approaches for Scenarios I, II, and III, respectively. Figures 9, 10, and 11 demonstrate that the NBO approach significantly outperforms the greedy approach.

Performance comparison for Scenario I: NBO approach versus greedy approach.

Performance comparison for Scenario II: NBO approach versus greedy approach.

Performance comparison for Scenario III: NBO approach versus greedy approach.

The algorithm (NBO) runtime to compute the control commands for three AAVs (in Scenario III) in any time step in MATLAB is approximately 4 seconds on a lab computer (Intel Core i7-860 Quad-Core Processor with 8 MB Cache and 2.80 GHz speed). This runtime can be greatly reduced on a better processor and by further optimizing the code. Since the algorithm runtime is not prohibitive, it can be used in real time (i.e., for practical purposes).

6. Conclusions, Remarks, and Future Scope

We designed a guidance algorithm for autonomous amphibious vehicles (AAVs) to rescue moving targets in a 2D flood scenario, where the flood water flows across the scene, and the targets move in the flood water. We designed this algorithm based on the theory of partially observable Markov decision process (POMDP). Since a POMDP problem is intractable to solve exactly, we used an approximation method called nominal belief-state optimization (NBO). We simulated a few scenarios to demonstrate the coordination among the AAVs achieved by the NBO approach. We defined a performance metric called average rescue time to compare the performance of our approach with a greedy approach. Our results show that the NBO approach outperforms the greedy approach significantly. This was expected because unlike the greedy approach the NBO approach has a lookahead property, that is, the NBO approach trades off the short-term performance for the long-term performance. Although the greedy approach achieves coordination among the AAVs in that the AAVs eventually rescue all the targets, but the performance in terms of average rescue time, which is crucial in these kinds of rescue missions, is poor compared to our NBO approach. In our future work, we would like to develop methods to further improve our NBO approach (e.g., NBO with adaptive horizon). We would also like to extend our approach to a decentralized AAV guidance problem to rescue multiple targets. In this decentralized case, we will induce coordination among the AAVs to rescue multiple targets by appropriately optimizing the communication (at the network level) between the AAVs along with the kinematic controls for the AAVs.

Acknowledgments

This work was supported in part by the Fulbright Foundation. The authors would also like to acknowledge Colorado State University’s support via the Libraries Open Access Research and Scholarship Fund (OARS).

Frejek M. Nokleby S. Design of a small-scale autonomous amphibious vehicle Proceedings of IEEE Canadian Conference on Electrical and Computer Engineering (CCECE '08) May 2008 Niagara Falls, Canada 781 786 2-s2.0-51849099151 10.1109/CCECE.2008.4564642 Papadopoulos E. Misailidis M. On differential drive robot odometry with application to path planning Proceedings of the European Control Conference July 2007 Kos, Greece 5492 5499 Tee Y. Tan Y. Teoh B. Tan E. Wong Z. A compact design of zero-radius steering autonomous amphibious vehicle with direct differential directional drive—UTAR-AAV Proceedings of IEEE International Conference on Robotics, Automation and Mechatronics (RAM '10) June 2010 Singapore 176 181 2-s2.0-77955672552 10.1109/RAMECH.2010.5513192 Ha Q. P. Tran T. H. Scheding S. Dissanayake G. Durrant-Whyte H. F. Control issues of an autonomous vehicle Proceedings of the 22nd International Symposium on Automation and Robotics in Construction (ISARC '05) September 2005 Ferrara, Italy Masayoshi W. Research and development of electric vehicles for clean transportation Journal of Environmental Sciences 2009 21 6 745 749 Brunl T. Embedded Robotics 2008 3rd Berlin, Germany Springer Tran T. H. Ha Q. P. Grover R. Scheding S. Modelling of an autonomous amphibious vehicle Proceedings of the Australasian Conference on Robotics and Automation (ACRA '04) December 2004 Canberra, Australia Manduchi R. Castano A. Talukder A. Matthies L. Obstacle detection and terrain classification for autonomous off-road navigation Autonomous Robots 2005 18 1 81 102 2-s2.0-9144265616 10.1023/B:AURO.0000047286.62481.1d Lacroix S. Mallet A. Bonnafous D. Bauzil G. Fleury S. Herrb M. Chatila R. Autonomous rover navigation on unknown terrains: functions and integration International Journal of Robotics Research 2002 21 10-11 917 942 2-s2.0-0036821235 Ragi S. Chong E. K. P. Dynamic UAV path planning for multitargte tracking Proceedings of the American Control Conference (ACC '12) June 2012 Montreal, Canada 3845 3850 Miller S. A. Harris Z. A. Chong E. K. P. A POMDP framework for coordinated guidance of autonomous UAVs for multitarget tracking EURASIP Journal on Advances in Signal Processing 2009 2009 17 2-s2.0-63749125729 10.1155/2009/724597 724597 Tran T. H. Modelling and control of unmanned ground vehicles [Ph.D. thesis] 2007 Watson S. A. Green P. N. Design considerations for micro-autonomous underwater vehicles (μAUVs) Proceedings of the IEEE International Conference on Robotics, Automation and Mechatronics (RAM '10) June 2010 Singapore 429 434 2-s2.0-77955672554 10.1109/RAMECH.2010.5513154 Watson S. A. Green P. N. Propulsion systems for micro-autonomous underwater vehicles (μAUVs) Proceedings of IEEE International Conference on Robotics, Automation and Mechatronics (RAM '10) June 2010 Singapore 435 440 2-s2.0-77955666804 10.1109/RAMECH.2010.5513155 Landau L. D. Fluid Mechanics 2000 2nd chapter IV Pergamon Press Bellman R. Dynamic Programming 1957 Princeton, NJ, USA Princeton University Press xxv+342 MR0090477 Kreucher C. Hero A. O. Kastella K. Chang D. Efficient methods of non-myopic sensor management for multitarget tracking Proceedings of the 43rd IEEE Conference on Decision and Control (CDC '04) December 2004 Paradise Island, Bahamas 722 727 2-s2.0-14344249315 Bertsekas D. P. Tsitsiklis J. N. Neuro-Dynamic Programming 1996 Belmont, Mass, USA Athena Scientific Bertsekas D. P. Castañon D. A. Rollout algorithms for stochastic scheduling problems Journal of Heuristics 1999 5 1 89 108 2-s2.0-0344672463 10.1023/A:1009634810396 Chong E. K. P. Givan R. L. Chang H. S. A framework for simulation-based network control via hindsight optimization Proceedings of the 39th IEEE Confernce on Decision and Control December 2000 Sydney, Australia 1433 1438 2-s2.0-0034439756 Bertsekas D. P. Dynamic Programming and Optimal Control 2007 2 Belmont, Mass, USA Athena Scientific Tee Y. Teoh B. Tan D. E. B. Wong Z. Tan C. Tan Y. Design considerations of autonomous amphibious vehicle (UTAR-AAV) Proceedings of IEEE Conference on Sustainable Utilization and Development in Engineering and Technology (STUDENT '10) November 2010 Petaling Jaya, Malaysia 13 18 2-s2.0-79851490727 10.1109/STUDENT.2010.5686990 Blackman S. Popoli R. Design and Analysis of Modern Tracking Systems 1999 Boston, Mass, USA Artech House