Visual Motion Estimation for an Autonomous
Underwater Reef Monitoring Robot
Matthew Dunbabin, Kane Usher, and Peter Corke
CSIRO ICT Centre, PO Box 883 Kenmore QLD 4069, Australia
Summary. Performing reliable localisation andnavigation within highly unstruc-
turedunderwater coralreef environments is adifficult task at thebestoftimes.Typ-
ical research andcommercial underwater vehicles useexpensiveacousticpositioning
andsonar systemswhich require significant external infrastructure to operateeffec-
tively.Thispaper is focusedonthe development of arobustvision-basedmotion
estimation technique usinglow-cost sensorsfor performing real-time autonomous
anduntetheredenvironmental monitoring tasksinthe GreatBarrier Reef without
theuse of acoustic positioning. Thetechnique is experimentally showntoprovide
accurate odometry andterrain profile information suitable forinput into thevehicle
controllertoperform arange of environmentalmonitoringtasks.
1Introduction
In light of recentadvances in computing andenergy storage hardware,Au-
tonomous Underwater Vehicles(AUVs)are emerging as thenextviable alter-
native to humandiversfor remote monitoring andsurveytasks.There are a
number of remotely operated (ROV)and AUVs performing variousmonitoring
tasks around theworld[17]. Thesevehicles are typically large andexpensive,
requireconsiderable external infrastructure foraccuratepositioning, andneed
more than onepersontooperateasingle vehicle.These vehiclesalsogener-
ally avoidthe highly unstructuredreef environmentssuchasAustralia’s Great
Barrier Reef,withlimited researchperformed on shallowwater applications
andreef traversing.Where surveyingatgreater depths is required, ROV’s
have been used forvideo transectsand biomass identification, however, these
vehiclesstill requirethe humanoperator in theloop.
Knowing thepositionand distanceaAUVhas movediscriticaltoensure
that correct andrepeatable measurements are being takenfor reef survey-
ing applications. It is importanttohaveaccurateodometrytoensuresurvey
transect pathsare correctly followed. Anumberoftechniquesare used to es-
timate vehicle motion.Acousticsensors such as Dopplervelocitylogs are a
commonmeans of obtainingaccuratemotioninformation. The useofvision
P. Corke and S. Sukkarieh (Eds.): Field and Service Robotics, STAR 25, pp. 31–42, 2006.
© Springer-Verlag Berlin Heidelberg 2006
32 M. Dunbabin, K. Usher, and P. Corke
for motion estimation is becoming a popular technique for underwater use
allowing navigation, station keeping, and the provision of manipulator feed-
back information [16, 12, 15]. The accuracy of underwater vision is dependent
on visibility and lighting, as well as optical distortion resulting from varying
refractive indices, requiring either corrective lenses or careful calibration[4].
Visual information is often fused with various acoustic sensors to achieve
increased sensor resolution and accuracy for underwater navigation [10]. Al-
though this fusion can result in very accurate motion estimation compared to
vision only, it is typically performed off-line and in deeper water applications.
A number of authors have investigated different techniques for odometry
estimation using vision as the primary sensor. Amidi [2] provides a detailed
investigation into feature tracking for visual odometry for an autonomous
helicopter. Another technique to determine camera motion is structure-from-
motion (SFM) with a comparison of a number of SFM techniques in terms of
accuracy and computational efficiency given by Adams[1]. Corke [7] presents
experimental results for odometry estimation of a planetary rover using om-
nidirectional vision and compares robust optic flow and SFM methods with
very encouraging results.
This research is focused on autonomously performing surveying tasks
based around the Great Barrier Reef using low-cost AUV’s and vision as the
primary sensor for motion estimation. The use of vision in this environment
is considered a powerful technique due to the feature rich terrain. However, at
the same time it can cause problems for traditional processing techniques with
highly unstructured terrain, soft swaying corals, moving biomass and lighting
ripple due to surface waves.
The focus of this paper is on the development of a robust real-time vision-
based motion estimation technique for a field deployed AUV which uses intel-
ligently fused low-cost sensors and hardware, and without the use of acoustic
positioning or artificial lighting.
2Vision System
2.1Vehicle
The vehicle developedand used in this researchwas custom designed to au-
tonomously perform theenvironmental monitoring tasks required by thereef
monitoring organisations [14].Toachievethese tasks, thevehiclemustnav-
igateoverhighly unstructured surfaces at fixed altitudes(300-500mmabove
thesea floor) andatdepthsinexcessof100m,incrosscurrentsof2knots
andknowits position during linear transectstowithin 5% of totaldistance
travelled.Itwas also consideredessentialthatthe vehicle be untethered to
reduceriskofentanglement, theneed forsupport vessels andreducing drag
imposedonthe vehicle operating in strongcurrents.
Visual Motion Estimation 33
Fig. 1 shows the hybrid vehicle design named “Starbug” developed as
part of this research. The vehicle can operate remotely or fully autonomously.
Details of the vehicle performance and system integration are given in [9].
Fig.1. The“Starbug”AutonomousUnderwater Vehicle.
2.2Sensors
The sensor platform developedfor theStarbug AUVand used in this research
hasbeen basedonpastexperience with theCSIROautonomousairborne
system [6]and enhanced to allowalow-cost navigationsuitefor thetaskof
long-termautonomousreef monitoring [8]. The primary sensing component
of theAUV is thestereocamerasystem.The AUVhas twostereoheads with
onelooking downward to estimate altitude abovethe sea-floor andodometry,
andthe otherlooking forward forobstacle avoidance(notusedinthis study).
The cameras used are acolourCMOSsensorfrom Omnivisionwith12mm
diameterscrew fit lenses whichhaveanominalfocal length of 6mm.
Each stereo pair hasthe cameras setwithabaseline of 70mmwhichallows
an effectivedistanceresolutioninthe range 0.2to1.7m. The cameras look
through 6mmthickflat glass. The twocameras are tightlysynchronizedand
line multiplexed into PALformat compositevideo signal.Fig.2showsthe
stereo camera head used in theAUV andanrepresentativeimage of the
typical terrainand visibilitythatsystem operates.
In additiontothe vision sensors,the vehicle hasamagnetic compass,
custom builtIMU (see [8]for details), pressure sensor (2.5mm resolution),a
PC/104 800MHz Crusoe computer stackrunning theLinux OS,and aGPS
which is used when surfaced.
3 Optimised Vision-Based Motion Estimation
Due to the unique characteristics of the reef environment such as highly un-
structured and feature rich terrain, relatively shallow waters and sufficient
34 M. Dunbabin, K. Usher, and P. Corke
(a)Stereocamerapair (b)Typicalreef terrain
Fig.2. Forwardlooking stereo camera system andrepresentativereefenvironment.
natural lighting, vision is considered a viable alternative to typical expensive
acoustic positioning and sonar sensors for navigation.
The system uses reasonable quality CMOS cameras with low-quality
miniature glass lenses. Therefore, it is important to have an accurate model
of the cameras intrinsic parameters as well as good knowledge of the cam-
era pair extrinsic parameters. Refraction due to the air-water-glass interface
also requires consideration as discussed in [8]. In this investigation the cam-
eras are calibrated using standard automatic calibration techniques (see e.g.
Bouguet[3]) to combine the effects of radial lens distortion and refraction.
In addition to assuming an appropriately calibrated stereo camera pair,
it is also assumed that the AUV is initialised at a known start position and
heading angle. The complete procedure for this odometry technique is outlined
in Algorithm 1.
The key components of this technique are image processing which we have
termed three-way feature matching (steps 1-7) which utilises common well
behaved procedures, and motion estimation (steps 8-10) which is the primary
contribution of this paper. These components are discussed in the following
sections.
3.1 Three-Way Feature Matching
Feature extraction
In this investigation, the Harris feature detector [5] has been implemented
due to its speed and satisfactory results. Roberts[13] compared the temporal
stability for outdoor applications and found the Harris operator to be superior
to other feature extraction methods. Only features that are matched both in
stereo (spatially) for height reconstruction, and temporally for motion recon-
struction are considered for odometry estimation. Typically, this means that
Algorithm 1 Visual motion estimation procedure.
1. Collect astereoimage.
2. Find all features in theentireimage.
3. Take the100 most dominant features as template(typicallythisnumberismore
like10-50 features).
4. Match cornersbetween stereo imagesbycalculatingthe normalized cross-
correlation ( ZNCC).
5. Storestereomatched features.
6. Usingstereomatched features at currenttimestep, matchthese with stereo
matchedfeaturesfromimagestaken at previoustimestepusing ZNCC.
7. Reconstructthose points whichhavebeen both spatially andtemporally
matchedinto3D.
8. Usingthe dualsearchoptimisation technique outlined in Algorithm 2, determine
thecameratransformation that best describesmotion from theprevioustothe
currentimage.
9. Usingmeasuredworld heading, roll andpitch angles, transformthe differential
camera motion to adifferential worldmotion.
10. Integratedifferential worldmotion to determine aworld camera displacement.
11. Go to step 1and repeat.
between tenand fifty strongfeaturesare trackedateachsampletimeand
during oceantrialswithpoorwater clarity this wasobservedtobelessthan
ten.
We are currently working on an improved robustnesstofeature extraction
that consists of acombination of this higher framerateextractionmethodwith
aslowerlooprunning amore computationally expensiveKLT (orsimilar) type
trackertotrackfeaturesoveralonger time period.This will help to alleviate
long term driftinintegrating differential motion.
Stereo matching
Stereo matching is used in this investigationtoestimatevehicle altitude,pro-
vide scaling fortemporal featuremotionand to generatecoarseterrainprofiles.
Forstereomatching,the correspondences between features in theleftand
rightimagesare found. The similarity between theregions surrounding each
corner is computed (lefttoright)using thenormalised crosscorrelation sim-
ilarity measure(ZNCC).
To reducecomputation, epipolar constraintsare used to prunethe search
spaceand only thestrongest cornersare evaluated. Once aset of matchesis
found, theresultsare then refinedwithsub-pixel interpolation. Additionally,
ratherthancorrectingthe entire image forlensdistortionand refraction ef-
fects, thecorrectionisapplied only to thecoordinate values of thetracked
features,hence saving considerable computation.
Visual Motion Estimation 35
36 M. Dunbabin, K. Usher, and P. Corke
Optic flow (motion matching)
The tracking of features temporally between image frames is similar to the spa-
tial stereo matching as discussed above. Given the full set of corners extracted
during stereo matching, similar techniques are used to find the corresponding
corners from the previous image. Differential image motion ( du, dv) is then
calculated in both the u and v directions on a per feature basis.
To maintain suitable processing speeds, motion matching is currently con-
strained by search space pruning, whereby feature matching is performed
within a disc of specified radius. The reduction of this search space size can
potentially be achieved with a motion prediction model to estimate where the
features lie in the search space.
In this motion estimation technique, temporal feature tracking currently
only has a one frame memory. This reduces problems due to significant ap-
pearance change over time. However, as stated earlier, longer term tracking
will improve integration drift problems.
3D feature reconstruction
Using the stereo matched corners, standard stereo reconstruction methods are
then used to estimate a feature’s three-dimensional position. In our previous
vision-based motion estimation involving aerial vehicles [6], the stereo data
was processed to find a consistent plane. The underlying assumption for stereo
and motion estimation was the existence of a flat ground plane. In this current
application, it cannot be assumed that the ground is flat. Hence, vehicle height
estimation must be performed on a per feature basis.
The primary purpose of 3D feature reconstruction in this investigation is
for scaling feature disparity to enable visual odometry.
3.2 Motion Estimation
The first step in the visual motion estimation process is to find a set of points
(features) which give a three-way match, that is, those points which have both
a stereo match in the current frame and a corresponding matching corner from
the previous frame as discussed in Section 3.1. Given this correspondence,
the problem is formulated as one of optimization to find at time k a vehicle
rotation and translation vector ( x
k
) which best explains the observed visual
motion and stereo reconstruction as shown in Fig. 3.
Fig. 3 shows the vehicle looking at a ground plane (not necessarily planar)
at times k − 1 and k with the features as seen in the respective image planes
shown for comparison. The basis behind this motion estimation is to optimise
the differential rotation and translation pose vector ( d x
est
) such that when
used to transform the features from the current image plane to the previous
image plane, minimises the median squared error between the predicted image
displacement ( du
, dv
) (as shown in the “reconstructed image plane”) and the
Fig.3. Motion transformation from previoustocurrent image plane.
actual image displacement(du, dv)providedfrom opticflow foreachthree-way
matchedfeature.
During theposevector optimisation,the Nelder-Meadsimplex method[11]
is employedtoupdatethe pose vector estimate.This nonlinearoptimisation
routine waschoseninthis analysisdue to itssolutionperformance andthe
fact that it does notrequirethe derivativesofthe minimised function to be
predetermined. The lack of gradientinformationallows this technique to be
‘model free’.
Theposevector optimisation consists of atwo stage processateachtime
step to best estimate vehicle motion.Since thedifferential rotations (roll,
pitch,yaw)are knownfrom IMU measurements,the first optimisationroutine
is restricted to only update thetranslationcomponentsofthe differential
pose vector withthe differential rotations held constant at their measured
values.This is aimedatkeepingthe solutionawayfrom localminima. As
theremay be errors in theIMU measurements,asecond searchisconducted
using theresultsfrom thefirstoptimisationtoseed thetranslationcomponent
of theposeestimate, with theentireposevector nowupdated during the
optimisation.This technique wasfound to provide more accurateresultsthan
asingle searchstepasithelpsinavoidingspurious localminima. Algorithm
2describes theposeoptimisationfunction used in this analysisfor thefirst
stage of themotionestimation. Note that in thesecond optimisation stage,
theprocedureisidenticaltoAlgorithm2,however, dθ, dα and dψ are also
updatedinStep3of theoptimisation.
Visual Motion Estimation 37
38 M. Dunbabin, K. Usher, and P. Corke
Algorithm 2 Pose optimisation function.
1. Seed search usingthe previoustimestep’sdifferential pose estimate such that
d x =[dx dy dz dθ dα dψ]
where dx, dy and dz arethe differential pose translationsbetween thetwo time
frames with respecttothe currentcameraframe, and dθ, dα and dψ arethe
differential roll, pitchand yawanglesrespectivelyobtained from theIMU.
2. Enteroptimisation loop.
3. Estimate thetransformation vector from theprevioustothe currentcamera
frame.
T = R
x
( dθ) R
y
( dα) R
z
( dψ)[dx dy dz]
T
4. For i =1 number of three-waymatched features,repeatsteps 5to9.
5. Displacethe observed 3D reconstructedfeature coordinates(x
i
, y
i
, z
i
)from
currentframe to estimate whereitwas in thepreviousframe ( x
e
i
, y
e
i
, z
e
i
).
[ x
e
i
y
e
i
z
e
i
]
T
= T [ x
i
y
i
z
i
]
T
6. Projectthe current3Dfeature points to theimage planetogive(u
o
i
, v
o
i
).
7. Projectthe displaced feature(step 5) to theimage planetogive(u
d
i
, v
d
i
).
8. Estimate theobservedfeature displacementonthe image plane.
[ du
i
dv
i
]
T
=[u
o
i
v
o
i
]
T
− [ u
d
i
v
d
i
]
T
9. Computethe squarederrorbetween theestimated andactualfeature displace-
ment ( du, dv)observedfromoptic flow.
e
i
=(du
i
− du
i
)
2
+(dv
i
− dv
i
)
2
10. Usingthe median square errorvalue(e
m
)fromall three-waymatched features,
update d x usingthe Nelder-Meadsimplex method.
11. If e
m
is less than apresetthreshold,end, else go to step 3and repeat usingthe
updated d x .
The resulting optimiseddifferential pose estimate at time k ( x
k
)whichis
withrespect to thecameracoordinate system attached to theAUV canthen
be transformed to aconsistentcoordinate system using theroll, pitch and
yawdatafrom theIMU.Inthis investigation, ahomogeneoustransformation
( T
H
)ofthe camera motion is performed to determine thedifferential change
in theworldcoordinate frame.
Thedifferential motion vectors are then integrated over time to obtain the
overallvehiclemotionpositionvector at time t
f
such that
x
t
f
=
t
f
k =0
T
H
k
d x
k
(1)
It wasobservedthatduring oceantrials, varying lighting and structure
could degradethe motion estimation performance due to insufficientthree-
waymatched features being extracted. Therefore,asimple constant velocity
vehicle modeland motion limit filters(basedonmeasuredvehicleperformance
limitations) were addedtoimprove motion estimation anddiscard obviously
erroneous differential optimisation solutions.Amore detailed hydrodynamic
modeliscurrently being evaluatedtofurther improvepredicted vehicle motion
andaid in pruning thesearchspace andoptimisationseeding.
4ExperimentalResults
The performance of thevisualmotionestimationtechnique describedinSec-
tion 3was evaluatedinatest tank constructedatCSIRO’sQCATsiteand
during oceantrials. The test tank hasaworking sectionof7.90x5.10m witha
depthof1.10m.The floor is lined withasand coloured matting with pebbles,
rocksofvarying sizes andlarge submerged3Dobjectstoprovideatexture
andterrainsurface forthe vision system.Fig.4showsthe AUVinthe test
tank andthe oceantestsiteoffPeel Island in Brisbane’s MoretonBay.
(a)CSIROQCATtesttank (b)Ocean test site
Fig.4. AUVduringvisual motion estimation experiments.
In thetesttankthe vehicle’s vision-based odometry system wasground
truthedusing twoverticalrodsattached to theAUV which protrudedfrom
thewater’s surface. ASICK laserrange scanner(PLS) wasthenusedtotrack
thesepointswithrespect to afixedcoordinate frame. By tracking thesetwo
points,bothpositionand vehicle heading anglecan be resolved.Fig.5shows
Visual Motion Estimation 39
40 M. Dunbabin, K. Usher, and P. Corke
the results of the vehicle’s estimated position using only vision-based motion
estimation fused with inertial information during a short survey transect in
the test tank. The ground truth obtained by the laser tracking system is shown
for comparison.
−3 −2 −1 0 1 2 3
−6
−5.5
−5
−4.5
−4
−3.5
−3
−2.5
−2
−1.5
x (m)
y (m)
Groundtruth from PLS
Vision position estimate
Fig.5. Position estimation usingonlyvision andinertial information in shortsurvey
transect.Also shownisaground truthobtained from thelasersystem.
As seen in Fig. 5, themotionestimationcomparesverywell with the
ground truthestimationwithamaximumerror of approximately2%atthe
endofthe transect.Although, this performance is encouraging,work is being
conducted to improvethe position estimation over greater transect distances.
The ground truthsystem is notconsideredperfect (asseen by thenoisy
position traceinFig.5)due to resolutionofthe laserscanner andthe size of
therodsattached to thevehiclecausing slight geometricerrors.However,the
system provides astable position estimate over time forevaluation purposes.
Apreliminary evaluation of thesystem wasconductedduring oceantests
over ahard coral androckreef in MoretonBay.The vehicle wasset off to
perform an autonomous untethered transect using theproposed visual odom-
etry technique.The vehicle wassurfaced at thestart andend of thetransect
to obtain aGPS fix andprovide aground truthfor thevehicle. Fig. 6shows
theresultsofa53m transect as measured by theGPS.
In Fig. 6, thecirclesrepresent theGPS fix locations, andthe line shows
thevehicles estimatedpositionduring thetransect.The resultsshowthatthe
vehiclespositionwas estimatedtowithin 4m of theactualend GPSgiven
location or to within8%ofthe totaldistancetravelled. Giventhe poor water
clarity andhighwaveactionexperienced during theexperiment, theresults
are extremely encouraging.
−40 −30 −20 −10 0 10
0
5
10
15
20
25
30
35
40
North (m)
East (m)
End location
(Surfaced GPS lock)
Start Location
(Start of dive)
Fig.6. Position estimation resultsfor oceantransect.
5Conclusion
Thispaper presents anew technique to estimate theegomotionand provide
feedback forthe real-timecontrol of an autonomous underwater vehicle using
only vision fusedwithlow-resolutioninertialinformation. A3Dmotionesti-
mation function wasdevelopedwiththe vehicle pose vector optimisedusing
thenonlinearNelder-Meadsimplex method to minimisethe mediansquared
error between thepredictedtoobservedcameramotionbetween consecutive
image frames.Experimentalresultsshowthatthe system performswell in
representative testswithpositionestimationaccuracyduring simple survey
transectsofapproximately 2% andinopenoceantests to 8%.The tech-
nique currently runs at betterthan4Hz sample rateonthe vehicle’sonboard
800MHz Crusoe processorwithout code optimisation.Researchiscurrently
being undertakentoimprove algorithmperformance andprocessing speed.
Otherareas of active researchfocus include improving system robustness
againstissues such as heading inaccuracies,lighting (wave“flicker”) andter-
rainstructure variations including surfacetexture compositionsuchassea-
grass, hard andsoftcoralstoallowreliable in-field deployment.
Acknowledgment
The authors would like to thankthe rest of theCSIROrobotics team:Graeme
Winstanley, JonathanRoberts,Les Overs, StephenBrosnan, ElliotDuff,Pa-
vanSikka,and JohnWhitham.
Visual Motion Estimation 41
42 M. Dunbabin, K. Usher, and P. Corke
References
1. H. Adams, S. Singh, and D. Strelow. An empirical comparison of methods for
image-based motion estimation. In Proceedings of the 2002 IEEE/IRJ Interna-
tional Conference on Intelligent Robots and Systems , October 2002.
2. O. Amidi. An Autonomous Vision-Guided Helicopter. PhD thesis, Dept of
Electrical and Computer Engineering, Carnegie Mellon University, Pittsburgh,
PA 15213, 1996.
3. J.Y. Bouguet. MATLAB camera calibration toolbox. In TR, 2000.
4. M. Bryant, D. Wettergreen, S. Abdallah, and A. Zelinsky. Robust camera cal-
ibration for an autonomous underwater vehicle. In Proceedings of the 2000
Australian Conference of Robotics and Automation, August 2000.
5. C. Charnley, G. Harris, M. Pike, E. Sparks, and M. Stephens. The droid 3d
vision system - algorithms for geometric integration. Technical Report Tech.
Rep. 72/88/N488U, Plessey Research Roke Manor, December 1988.
6. P. Corke. An inertial and visual sensing system for a small autonomous heli-
copter. Journal of Robotic Systems , 21(2):43–51, February 2004.
7. P.I. Corke, D. Strelow, and S. Singh. Omnidirectional visual odometry for a
planetary rover. In Proceedings of IROS 2004, pages 4007–4012, 2004.
8. M. Dunbabin, P. Corke, and G. Buskey. Low-cost vision-based AUV guidance
system for reef navigation. In Proceedings of the 2004 IEEE International Con-
ference on Robotics & Automation, pages 7–12, April 2004.
9. M. Dunbabin, J. Roberts, Usher K., G. Winstanley, and P. Corke. A hybrid
AUV design for shallow water reef navigation. In Proceedings of the 2005 IEEE
International Conference on Robotics & Automation, April 2005.
10. R. Eustice, O. Pizarro, and H. Singh. Visually augmented navigation in an
unstructured environment using a delay state history. In Proceedings of the
2004 IEEE International Conference on Robotics & Automation, pages 25–32,
April 2004.
11. J. Lagarias, R. Reeds, and M. Wright. Convergence properties of the nelder-
mead simplex method in low dimensions. SIAM Journal of Optimization,
9(1):112–147, 1998.
12. P. Rives and J-J. Borrelly. Visual servoing techniques applied to an underwater
vehicle. In Proceedings of the 1997 IEEE International Conference on Robotics
and Automation, pages 1851–1856, April 1997.
13. J. M. Roberts. Attentive visual tracking and trajectory estimation for dynamic
scene segmentation. PhD thesis, University of Southhampton, UK, 1994.
14. English S., C. Wilkinson, and V. Baker, editors. Survey manual for tropical
marine resources . Australian Institute of Marine Science, Townsville, Australia,
1994.
15. J. Santos-Victor and G. Sandini. Visual behaviors for docking. Computer Vision
and Image Understanding, 67(3):223–238, September 1997.
16. S. van der Zwaan, A. Bernardino, and J. Santos-Victor. Visual station keeping
for floating robots in unstructured ennvironments. Robotics and Autonomous
Systems , 39:145–155, 2002.
17. L. Whitcomb, D. Yoerger, H. Singh, and J. Howland. Advances in underwater
robot vehicles for deep ocean exploration: Navigation, control and survey op-
erations. In Proceedings of Ninth International Syposium of Robotics Research
(ISRR’99), pages 346–353, October 9-12 1999.
RoadObstacle DetectionUsingRobust Model
Fitting
Niloofar Gheissari
1
and NickBarnes
1 , 2
1
AutonomousSystems and SensingTechnologies,National ICT Australia
Locked bag8001, Canberra, ACT2601, AUSTRALIA
2
Department of Information Engineering, The Australian National University
Summary
.
Aw
areness
of
pe
destrians,o
ther
ve
hicles,a
nd
otherr
oad
obstaclesi
s
keytodriving safety, and so theirdetectionisacritical need in driverassistance
research. We propose using amodel-based approachwhichcan either directlyseg-
mentthe disparity to detect obstacles or remove the road regions from an already
segmented disparitymap. We developed two methods for segmentation: first, by
directly
segment
ing
obstacles
from
the
disparit
ym
ap;
and,
second
by
using
mor-
phological operations followedbyarobust model fitting algorithm to reject road
segmentsafter thesegmentationprocess. To test the success of our methods, we
have testedand compared them with an available methodinthe literature.
1I
nt
ro
duction
Road accidents have been considered as the thirdlargest killer after heart
disease
and
depression.
Ann
ually
ab
out
one
million
pe
oplea
re
killed
and
a
further 20 million are injured or disabled. Roadaccidents not only cause
fatalityand disability, but also they cause stress, anxietyand financialside
effects
on
pe
ople’sd
ailyl
ife.
In
the
computerv
isiona
nd
rob
otics
comm
unities,
there have been various efforts to develop systems whichassist the driver to
avoid pedestrians, cars and road obstacles.However, roadstructure, lighting,
we
ather
conditions,
and
in
teraction
be
tw
een
differen
to
bstacles
ma
ys
ignif-
icantly affect the performance of these systems. Hence, providing asystem
that is reliable in avarietyofconditionsisnecessary.
According
to
Bertozzi
et
al.,[4]
theu
se
of
visible
vision
and
image
pro-
cessing methods for obstacle detection in intelligentvehicles can be classified
as motion based [11], stereo based [12], shaped based [3] and texture based
[5] methods. Formore detailsonthe availableliterature, readers are referred
to
[10].
Among
these
differen
ta
pproac
hes,
stereo-based
vision
ha
ve
be
en
re-
ported as the mostpromising approachtoobstacle detection [7]. The recent
P. Corke and S. Sukkarieh (Eds.): Field and Service Robotics, STAR 25, pp. 43–54, 2006.
© Springer-Verlag Berlin Heidelberg 2006
44 N. Gheissari and N. Barnes
works in stereo-based obstacle detection for intelligent vehicles include the
Inverse Perspective Method (IPM) [2] and the u- and v-disparity map [9].
IPM relies on the fact that if every pixel in the image is mapped to the
ground plane, then in the projected images obstacles located on the ground
plane are distorted. This distortion generates a fringe in the image resulting
from subtracting the left and right projected images and helps us to locate
an obstacle in the image. This method requires the camera parameters and
the base line to be known as a a priori. In fact, IPM is very sensitive to cam-
era calibration accuracy. Furthermore, the existence of shadows, reflections
or markings on the road may reduce the performance of this method. The
other recent method in obstacle detection for intelligent vehicles is based on
generating u- and v-disparity maps [9], which are histograms of the disparity
map in the vertical and horizontal directions. An obstacle is represented by a
vertical line in v-disparity while by a horizontal line in u-disparity. The ground
plane can be detected as a line with a slope. Hence, techniques such as Hough
Transform can be applied to detect obstacles. Obstacle detection using u- and
v- disparity maps appear to outperform IPM [8], however they have other
shortcomings. For example, the u- and v-disparity maps are usually noisy and
unreliable. In addition, accumulating in the horizontal and vertical direction
of the disparity map causes objects behind each other (or next to each other)
be incorrectly merged. The other disadvantage of this method is that small
objects or objects which are located in a far distance from camera tend to
be undetected. This may occur due to line segments in these regions that are
either too short to detect, or too long and so easily merged with other lines
in the v- or u-disparity map.
To overcome the above problems, this paper presents two new obstacle
detection algorithms for application in intelligent vehicles. Both algorithms
segment the disparity map. The first algorithm is based on the fact that the
obstacles are located approximately parallel to the image plane, and directly
segments them using a robust model fitting method applied to the quantised
disparity space. The second algorithm incorporates some simple morphological
operations and then a robust model fitting approach to separate the road
regions from the image. As this robust fitting method is only applied to a
part of image, the computation time is low. Another advantage of our model
based approach is that we do not require calibration information, which is in
sharp contrast with methods such as IPM.
Note that for finding pedestrians and cars in a road scene, typically stereo
data is used as a first stage, then fused with other data for classification. This
paper addresses the first stage only, and is highly suitable for incorporation
with other data at a later stage, or direct fusion with other cues.
Road Obstacle Detection Using Robust Model Fitting45
2Algorithm 1: Robust Model Fitting
This algorithm relies on the idea that aconstantmodel can describethe
disparitymap associatedwith every obstacle approximately paralleltothe
imageplane.This is atrue assumption where objects:
1. have no significantrotation angle;
2. have rotation but are not tooclosetothe camera;or,
3. have rotation but have no significantwidth.
Later we willshowthat, by assuming overlapping regions in our algorithm,
we mayallowsmall rotations about the vertical or horizontal axis. In the
algorithm,
we
first
apply
ac
on
trast
filtering
metho
dt
ot
he
imagea
nd
remo
v
e
areas
of
lo
wc
on
trast
from
thed
isparit
ym
ap.
It
allow
su
st
or
emo
ve
regions
whose disparitymap, duetothe lackoftexture, is unreliable. This contrast
filtering methodisdescribed in Section 4. We then quantise the disparity
space by dividing it to anumberofoverlapping bins of length g pixels. Each
bin has g/2pixels overlap with the next bin. This overlap can help to prevent
regions being split acrosstwo successivebins. In our experimentsweset g=8
pixels.
This
quan
tisation
approac
hh
as
some
adv
an
tages;
first
we
apply
our
robust
fitting
metho
dt
oe
ac
hb
in
separately
and
hencew
ea
vo
id
exp
ensiv
e
approaches suchasrandom sampling. Second,wetakethe quantisation noise
of pixel-based disparityintoaccount. Finally, it allows an obstacle to rotate
slight
ly
around
the
ve
rtical
axis
or
ha
ve
as
omewhat
non-planar
profile(
suc
h
as apedestrian). After disparityquantisation, we fit the constant model to the
whole bin. We compute the constantparameterand the residuals. If the noise
is
Gaussian,
the
squared
residuals
will
be
sub
ject
to
a
χ square
distribution
with n-1 degrees of freedom and and thus the scale of noise will be δ =
r
2
i
n − 1
:
where r
i
is the residual of the i
th
point. We compute the scale of noise and
select the points whose corresponding residual is less than the scale of noise
mu
ltiple
by
the
significance
lev
el
T (which
can
be
lo
ok
ed
up
fromG
aussian
distribution table). These points are inliers to the constant model and thus
do not belong to the road. Nowwehave apreliminary knowledgeabout the
inliers/outliers.
In
the
next
stage
we
iterativ
ely
fit
the
mo
del
to
the
inliers,
recompute the constant parameter with more confidence and compute the final
scale of noise only using the inliers. We used 3iterationsinour experiments.
We
no
wh
ave
afi
nal
estimation
of
the
mo
del
parameter.
Ho
we
ve
ri
teration
hasshrunken the inlier space. To create larger regions and simultaneously
maintain our degree of confidence, we fit the final estimated model to the bin
(including
inliers
and
outliers)
and
reject
outliers
using
the
final
scale
of
noise.
This above taskgives us differentsets of inliers of differentdepths that
create
as
egmen
tation
map.
Ho
we
ve
r,
this
do
es
not
guaran
tee
the
lo
calit
y
of eachsegment. To enforce the localityconstraint we compute the regional
maximum of the segmentation map, assuming that we are onlyinterested in
areas whichare closertousthan the surrounding background. Finally a4-
46 N. Gheissari and N. Barnes
connected labelling operation provides us with the final segmentation map.
As a post processing stage we may apply a dilation operation to fill the holes.
Figure 1-3 show the contrast filtering result mapped on the disparity map,
the result of the 4-connected labelling operation on the segmented image (and
the dilation) and the final result for frame 243. As can be seen from this figure,
the missed white colour car (at the right side of the image) does not have
enough reliable disparity data and thus is not detected as a separate region.
100 200 300 400 500 600
50
100
150
200
250
300
350
400
450
Fig.1
.
The
con
trast
filtering
result
mapped on the disparitymap.
100 200 300 400 500 600
50
100
150
200
250
300
350
400
450
Fig. 2. The4-connected labelling op-
eration result
Fig. 3. Final results
3Algorithm 2: Basic Morphological Operations
The second segmentation algorithm presented here is asimple set of morpho-
logicaloperations, followedbyaroad separation method. We first compute
the edges of the disparitymap. Again,weapply our contrast filtering method
to the intensityimage, and from the edge map we remove areas which have
lowcontrast.Weapply adilationoperation to thickenthe edges. Then we fill
the
holes
and
small
areas.
We
apply
an
erosion
op
eration
to
createm
ore
dis-
tinct areas. To remove isolated small areas we use aclosing operation next to
an opening operation. Finally as apost processing step we dilate the resulting
region using astructuralelement of size 70 × 10. Thisstep canfill small holes
inside aregion and join closely located regions. This algorithm relies on the
removalofroad areas. An algorithm for this is explained below.
3.1R
oad
Separation
Assume that we are given thedisparitymap and an initial segmentation in
theform of aset of overlapping rectangular regions. The camera parameters
and the base line is assumed to be unknown, whichisanadvantage of our
methodoverthe existing methods. We aim at rejecting those regionswhich
be
long
to
the
road.
We
assume
ther
oad
plane
to
be
piecewise
linear.
It
can
be
easily provedthat the disparityofpixels located on the roadcan be modelled
by the followingequation[6]: d =
B
H
f
x
(
y
f
y
cos α +sin α )where yisthe image
coordinate in the horizontal direction, H is the distance of camera fromthe
road, B is the baseline and α is the tilt angle of camera with respect to the
road. The parameters f
x
and f
y
are thescaled camera focal length. Thus for
simplicity
we
can
write
d = ay+b:w
here
a and b are
some
unknow
nc
onstan
t
parameters.
That means we describethe road with aset of linear models, i.e., modelling
the road as piecewise linear (anyroad that is not smooth and piecewiselinear
certainlyisanobstacle). We fit the linear model to every segmentinthe image.
We
compute
the
parameters
a , b and
ther
esiduals.
If
the
noise
is
Gaussian,
the
squared residuals will be subject to a χ square distribution with n-2 degrees
of freedom and thus the scale of noise will be δ =
r
2
i
n − 2
:here r
i
is the residual
of the i
th
point. We compute the scale of noise and select the points whose
corresponding residual is less than the scale of noiseasinliers. Since these
points are inliers to the assumedroad model,they are not part of an obstacle.
Then
we
select
the
regions
whose
nu
mb
er
of
inliers
is
more
than
at
hreshold.
This threshold represents the maximum number of road pixelswhichcan be
located in aregion and that region be stillregarded as an obstacle region.
Again we apply thepreviously discussedrobust model fitting approachtothe
inliers to estimate the final scale of noise and model parameters. We create a
new set of inliers/outliers. We reject aregion as aroad region onlyifits sum
of squared residuals is greater thanthe scale of noise. Once we makeour final
decision, we can compute the final road parameters if we require. We alsocan
computeareliabilitymeasure for eachregion based on its scale of noise and
its number of outlierstothe road model (obstacle pixels).
4Contrast Filtering
If
an
area
do
es
not
ha
ve
sufficien
tt
exture,
then
the
disparit
ym
ap
willb
e
unreliable. To avoid suchareas we have applied acontrast filtering method
whichincludes twomedian kernels of size 5 × 5and 10 × 10. The sizes of these
Road Obstacle Detection Using Robust Model Fitting47
48 N. Gheissari and N. Barnes
pixels were chosen heuristically so that we ignore areas (smaller than 10 × 10)
in which the contrast is constant. We convolved our intensity image with both
median kernels. This results in two images I
1
and I
2
, in each of which, every
pixel is the average of the surrounding pixels (with respect to the kernel size).
We compute the absolute difference between I
1
and I
2
and construct matrix
M, so that M=I
1
-I
2
. We reject every pixel i where M
i
< FTH. The threshold
FTH is set to be 2 in all experiments.
Fig.4. If the contrast varies significantlybetween two embedded regions,then the
filterresults in ahigh value (e.g., the left embedded squares), while,for regions with
no contrast the filterresults in alow value(e.g, the rightembedded squares).
5E
xp
erimen
tal
Results
Fig. 5. Inside the ANU/NICTAintelligentvehicle. Cameras to monitorthe road
scene and findobstaclesappear in place of the rear-vision mirror.
Within the ANU/NICTAintelligentvehicles project, we have acontinu-
ing project to developpedestrian and obstacle detection, to run on our road
vehicle, see Figure 5. Forthe purposes of this paper, we
ha
ve
applied
our al-
gorithmtoanoisyimage sequence containing pedestrians, cars and buildings.
Here, we have provided afew samples of our results. We alsohave included
the results of applying the stereo-vision algorithm reported in [8] on the same
image sequence. This algorithm, which uses the u- and v-disparitymap, has
been shown to be successful in comparison with other existing methods [8].
The example frames shown here were chosen to illustratedifferent as-
pects (strength and weakness)ofboth algorithms.Wealso compared the
three methods quantitatively in figure 18.The computation time for both
propo
sed
metho
ds
is
ab
out
one
second
pe
rf
rame
in
an
on-optimized
Matlab
implementation on astandard PC. We expect it to be better thanframe rate
in aCoptimized implementation, and so comfortably real-time.
As the following results indicate all the algorithms maymiss anumberof
regions. However, it has been observed (from figure 18) that the model based
algorithm misses fewer regions and performsbetter. However, adrawback of
this
algorithm
is
that
if
thed
isparit
ym
ap
is
noisy
,a
nd
some
obstacles
ma
y
be
rejected
as
outliers
(in
the
robustfi
tting
stage).
Thisc
an
be
solv
ed
by
assum-
ingalargersignificance level T .However, it maycause under-segmentation.
In future work we plan to devise an adaptiveapproachtocompensate for a
poorand noisy disparitymap.
The morphological algorithm is onlyapplicable where the disparitymap is
sparse, otherwisefor adense disparitymap, we willhave aconsiderableunder
segment
ation.
In
this
case,u
sing
the
mo
del
based
algorithm
is
suggested.
Figures
6-8
sho
wt
hat
the
mo
del
based
algorithm
has
detected
all
the
obstacles correctly (in frame 8), while the morphologicalbased algorithm has
under-segmentedthe data, and the u- and v-disparitybased algorithm only
detected
one
obstacle.
As can be seen from figure 9-11,the model fitting based algorithm has
detected all obstacles, but failed to segment apedestrian from the white car
(in
frame
12).
The
morphological
based
algorithm
has
again
missed
the
white
car. In contrast, the u- and v- disparitybased algorithm has onlysucceeded
in detecting one of the pedestrians.
Figure
12-14,s
ho
wt
hat
all
of
the
differen
ta
lgorithm
ha
ve
successfully
ignored the rubbishand the manhole on the road. The model fitting based
algorithm has detected all obstacles except for the pedestrian close to the
camera.
The
morphologicalb
ased
algorithm
has
again
missedt
he
small
white
car while it has successfully detected the pedestrians. In contrast, the u- and
v- disparitybased algorithm has onlysucceeded in detecting the pedestrian
near to the camera.
The last example is frame 410ofthe sequence. Figure 15-17 showthat
while the model based algorithm tends to generate alarge number of different
regions, the morphologicaloperationsbased algorithm tendstodetect more
major (larger and closer) obstacles. The pedestrian has aconsiderablerotation
angleand so the model based algorithm split the pedestrian across tworegions.
This can be easily solved by apost-processing stage. Both the u- and v-
disparit
ya
nd
the
mo
del
based
algorithms
miss
the
car
at
the
righ
ts
ide
of
image.However, small obstacles at further distances, whichare ignored by
the u- andv-disparitybased algorithm, are detected by the model based one.
Furthermore, although the u- and v- disparitybased algorithm generates more
Road Obstacle Detection Using Robust Model Fitting49
50 N. Gheissari and N. Barnes
Fig. 6. Resultsofapplying model based
algorithm on frame 8.
Fig. 7. Results of applying morpholog-
ical based algorithm on frame 8.
Fig. 8. Results of applying u- and
v-disparity based algorithm in [8] on
frame 8.
precise boundaries for the pedestrian,it generates a noisy segmentation. This
may happen in all algorithms and is mainly due to noise in disparity. This is
best dealt with using other cues.
5.1 Comparison Results
In figure 18 we show the results of applying the two algorithms to 50 succes-
sive images of a road image sequence. These 50 frames were chosen because
all of them have four major obstacles, a reasonably high number in real appli-
cations. The ground truth results and also the results of applying the u- and
v-disparity based algorithm [8] have been shown in different colors. Ground
truth was labelled manually by choosing the most significant obstacles. Fig-
ure 18 clearly show that both proposed algorithms outperform the u- and
v-disparity based algorithm. More importantly the model based method for
obstacle detection has been more successful than the other two approaches.
The complete sequence is available at:
nmb/fsr/gheissaribarnesfsr.html.
Fig. 9. Resultsofapplying model based
algorithm on frame 12.
Fig. 10. Results of applying morpho-
logical based algorithm on frame 12.
Fig. 11. Results of applying u- and v-
disparitybasedalgorithm frame 12.
Note
that
with
careful
tuning
the
u-
and
v-disparit
yb
ased
algorithm
ma
y
generatebetterresults. Despiteits poorperformance, this methodhas the
advantage of generatingamore precise bounding box.
6FutureWork
Althought
he
results
of
our
algorithms
sho
wb
etter
pe
rformance
than
using
u- and v-disparitymap, stillthere is alarge spacefor improvement. The
robust model based approachcan be improvedbyusing asmarter quantization
metho
dt
han
the
curren
to
ne.
The
partitioning
metho
do
fB
ab-Hadiashar
and
Suter[1] can alsobeused to improve the results. In addition, we mayuse a
model selection criterion to decideifaregion is located on the roadoronan
obstaclea
ppro
ximately
parallel
to
the
camera
imagep
lane.
To reduce the false negatives, we willfuse the disparitysegmentation result
with other cues suchasshape, texture or motion. We alsowill include a
tracking methodtotrackobstacles overthe imagesequence suchasKalman
filtering or particlefiltering.
Road Obstacle Detection Using Robust Model Fitting51
52 N. Gheissari and N. Barnes
Fig. 12. Results of applying model
based algorithm on frame 170.
Fig. 13. Results of applying morpho-
logical based algorithm on frame 170.
Fig. 14. Results of applying u- and v-
disparitybasedalgorithm frame 170.
Finally,asthis algorithm only aims at detecting the obstacle region with no
classification, later we willinclude aclassification approach to decide whether
an
obstaclei
sac
ar,
ap
edestrian
etc.
7C
onclusion
The
main
con
tribution
of
this
pap
er
is
prop
osing
am
od
el
based
approac
hf
or
obstacledetection in driving assistantapplications. The first algorithm relies
on the fact that aconstantmodel can describethe disparitymap associated
withe
ve
ry
obstacle
appro
ximately
parallel
to
the
camera
imagep
lane.W
e
quantize the disparityspace, use arobust model fitting methodtoestimate
the constantmodel parameter, and compute the scale of noise that is used to
partition the data. In the second algorithm that only can be applied to sparse
disparitymaps, we use some basic morphologicaloperationstosegmentthe
data. Our main contribution here is not the segmentation algorithm itself but
it is the way we separate the road data from our image. This road separation
(or
detection)
metho
da
gain
is
basedo
nam
od
el-based
approac
h.
Both algorithms have been extensively tested and compared and have
shown to be more successful than the u- and v- disparitysegmentation al-
Fig. 15. Results of applying model
based algorithm on frame 410
Fig. 16. Results of applying morpho-
logical based algorithm on frame 410
Fig. 17. Results of applying u- and v-
disparitybasedalgorithm frame 410
Fig. 18. Comparison of our two algorithms withthe u- and v- disparity based
algorithm.
gorithm. The model based algorithmmisses fewer regions than the otheral-
gorithms.
This
indicates
that
the
mo
del
based
approach
is
effe
ctiv
ef
or
obstacle
detection, and is worthyoffurther study to improve its performance further.
Road Obstacle Detection Using Robust Model Fitting 53
54 N. Gheissari and N. Barnes
Acknowledgment
National ICT Australia is funded by the Australian Department of Commu-
nications,Information Technology and the Arts and the Australian Research
Council through Backing Australia’s ability and the ICT Centre of Excellence
Program.
References
1. Bab-Hadiashar A.,Suter D.,Robust Segmentation of Visual Data Using
Rank
ed
Un
biasedS
cale
Estimator,
In
ternational
Journal
of
Information,
Educa-
tion and Research in Robotics and Artificial Intelligence, ROBOTICA, volume
17, 649-660,1999.
2. Bertozzi, M., BroggiA., Fascioli A., Stereo Inverse PerspectiveMapping: The-
ory and Applications,Image and Vision Computing Journal, 16(8),pp. 585-590,
1998.
3. Bertozzi, M., Broggi A., Fascioli A., and Sechi, M., Shape-based Pedestrian
Detection,
Proc
eedingso
fI
EEE
In
telligen
tV
ehicles
Sympo
sium,
pp.2
15-220,
Oct. 2000.
4. Bertozzi, M., Broggi,A., Grisleri, P.,Graf, T., andMeinecke,M., Pedestrain
Detection
in
InfraredI
mages,
Pro
ceedingso
fI
EEE
In
telligen
tV
ehicles
Sympo
-
sium, pp.662-667, June 2003.
5.
Curio,
C.,
Edelbrunner
J.,K
alink
eT
.,
Tzomak
as,
C.,
and
Seelen
W.
vo
n,
Wa
lk-
ing Pedestrian Recognition, IEEETransactions on IntelligentTransportation
Systems, vol. 1, pp. 155-163,Sep, 2000.
6.
Franke, U. and Kutzbach, Fast Stereo based Object Detection for Stopand Go
Traffic, Intelligent Vehicles Symposium, pp. 339-344 ,1996.
7.
Gavrila, D. M., Giebel, J., and Munder, S., Vision-Based Pedestrian Detection:
The PROTECTORSystem, pp. 13-18, 2004.
8.
Grubb Grant, Alexander Zelinsky,Lars Nilsson, Magnus Rible,3DVision Sens-
ing for ImprovedPedestrian Safety, Intelligent Vehicles Symposium(2004), pp.
19- 24, Parma Italy, June2004.
9.
Labayrade, R., Aubert, D.,and Tarel, J P.,Real Time Obstacle Detection in
Stereo
vision
on
NonF
lat
Road
Geometry
Through
”V-disparit
y”
Represen
ta-
tion, pp. 646-651,June 2002.
10.
Sun,Z
.,
Bebis,
G.,
and
Miller,
R.,
On-RoadV
ehic
le
Detection
UsingO
ptical
Sensors:AReview, ProceedingsofIEEE Intelligenct Transportation Systems
Conference
,W
ashington,
D.C.
USA,
pp.
585-590,2002.
11. Viola, P.,Jones, M., and Snow, D.,DetectingPedestriansUsing Patterns of
Motion
and
App
earance,
Pro
ceedings
of
the
In
ternational
Conference
on
Com-
puter Vision (ICCV), pp. 734-741, Oct. 2003.
12. Zhao, L. and ThorpeCharles E., Stereo- and Neural Network-BasedPedestrian
Detection IEEETransactions on IntelligentTransportationSystemes, vol. 1, pp.
148-154,S
ep,
2000.
Real-Time Regular Polygonal Sign Detection
Nick Barnes
1
and Gareth Loy
2
1
National ICT Australia, Locked Bag 8001, Canberra, ACT 2601,
Department Of Information Engineering, The Australian National University
2
Computer Vision and Active Perception Laboratory, Royal Institute of
Technology (KTH), Stockholm, Sweeden
Summary. In this paper, we present a new adaptation of the regular polygon de-
tection algorithm for real-time road sign detection for autonomous vehicles. The
method is robust to partial occlusion and fading, and insensitive to lighting con-
ditions. We experimentally demonstrate its application to the detection of various
signs, particularly evaluating it on a sequence of roundabout signs taken from the
ANU/NICTA vehicle. The algorithm runs faster than 20 frames per second on a
standard PC, detecting signs of the size that appears in road scenes, as observed
from a camera mounted on the rear-vision mirror. The algorithm uses the symmetric
nature of regular polygonal shapes, we also use the constrained appearance of such
shapes in the road scene to the car in order to facilitate their fast, robust detection.
1 Introduction
Improving safety is a key goal of road vehicle technology. Driver support
systems aim to improve safety by helping drivers react to changing road con-
ditions. Although full automation of road vehicles may be achievable in the
future, our research focusses on systems that can assist drivers immediately.
Rather than replacing the driver, we aim to keep the driver in the loop, while
supporting them in controlling the car.
Road signs present information to alert drivers to changes in driving con-
ditions. Critical information signs, such as speed, give-way, roundabout, and
stop give information that a driver must react to, as opposed to informational
and directional signs. These signs appear clearly in the road scene, and are
well distinguished. However, drivers may sometimes miss such signs due to
distractions or a lack of concentration. This makes detecting critical informa-
tion signs and making the driver aware of any they may have missed a key
target for improving driver safety. The lack of driver awareness of a sign may
be detected through a lack of response.
We have previously demonstrated the application of the radial symmetry
operator [1] to detecting speed signs, demonstrating real-time performance [2].
P. Corke and S. Sukkarieh (Eds.): Field and Service Robotics, STAR 25, pp. 55–66, 2006.
© Springer-Verlag Berlin Heidelberg 2006