機械手機器人外文翻譯-割草機器人多傳感器融合與導航技術的研究【中文5220字】【PDF+中文WORD】
機械手機器人外文翻譯-割草機器人多傳感器融合與導航技術的研究【中文5220字】【PDF+中文WORD】,中文5220字,PDF+中文WORD,機械手,機器人,外文,翻譯,割草,傳感器,融合,導航,技術,研究,中文,5220,PDF,WORD
Proceedings of the 2007 IEEE International Conference on Robotics and Biomimetics
December 15 -18, 2007, Sanya, China
Multisensor Fusion and Navigation for Robot Mower*
Ming Cong and Bo Fang
School of Mechanical Engineering, Dalian University of Technology Dalian, 116024, China
congm@dlut.edu.cn
Abstract - This paper presents a multisensor system for combining measurements from ultrasonic sensors and navigation for robot mowers. The proposed sensing system enables robot mowers to mapping unknown environments. It is important for an autonomous robot mower to explore its surroundings in performing the task of localization and navigation for mowing. Because of the complexity of the environment, one simple kind of sensors is not sufficient for robot mower to accomplish these tasks. We develop a robot mower equipped with DSP TMS320F2812 as its CPU. The sensing system integrates with ultrasonic sensors, infrared sensors, collision sensors, encoders, a temperature sensor and an electronic compass. A method of high accuracy ultrasonic ranging technology based on wavelet transform is reported to improve the measurement precision of ultrasonic sensors. Simulation studies show that the proposed multisensor fusion method is very effective for the navigation of robot mowers. Experimental results indicate that this sensing system based on generalized auto-correlation method for obstacle detection and localization shows great potential for providing a high performance-to-price ratio and robust solution for robot mowers in dynamic working condition.
Index Terms - multisensor fusion, ultrasonic sensors, robot mower, mapping, navigation
I. INTRODUCTION
Lawn mowing is considered by many to be one of the most boring and tiring routine tasks. The environmental robots are needed urgently to perform the task. Some predictions indicate that the robot mowers will be one of the most promising personal robot applications and have substantial market in the world. Therefore, the concept of Intelligent Robot Mower (IRM) had been proposed for the first time in 1997’s annual conference of the OPEI (Outdoor Power Equipment Institute) [1]. The robots mainly face to the general families to help the busy people and the hypodynamic old folks save the payments for hiring labours, also remove people from noise, pollen and danger of mowing blade. The robot mowers serve for home care as the outdoor mobile robots, actually kind of intelligent mechatronics devices for environment clean-up [2][3]. The important thing is that the robot mowers are representative of some area-covering environmental robots used not only for indoor floor cleaning as in [4] but also in hazardous environments such as removing landmines, cleaning up radiant points and prospecting for resources etc. The robot mowers get great challenges differing from indoor mobile robots.
The robot mowers use sensors to understand environments as well as their real-time states for obstacle avoidance, map building, location and navigation in the whole work area. Because of the complexity of the environment, one simple kind of sensors is not sufficient for robot mower to accomplish these tasks. It is necessary to combine the observed sensor data coming from different sensors to reduce the uncertainties of the robot in any working environment. To merge the information from the various sensors, robust and real-time sensor fusion is required [5]. In cases of sensor error or failure, multisensor fusion can also reduce uncertainty in the information and increase its reliability.
A sensing system of low cost, low power consumption, high performance is described. The detecting range of ultrasonic sensors is 0.3m~5m, they provide good range information. However, uncertainties in ultrasonic sensors caused by the specular reflection from environments make them less attractive. The detecting range of infrared sensors is 0.02m~1m, they can detect the obstacles within the ultrasonic sensor’s blind zone.
In order to satisfy the needs of robot mowers for the low cost and high accuracy ranging technology, the research on the high accuracy ultrasonic ranging technology based on wavelet transform (WT) is reported to improve the measurement precision of ultrasonic sensors. Measurement data gathered from the sensing system are integrated to avoid the robot mower from unknown obstacles and plan an optimum, reliable and realizable plan completely coverage of entire working area.
Finally, simulation studies and experimental results show the effectiveness of the sensing system for the navigation, obstacle detection and localization of robot mowers.
II. SYSTEM HARDWARE OF IRM
The IRM uses DSP TMS320F2812 as its CPU, including four units: vehicle system, cutting system, sensing system and control system. The sensing system is used to collect the external dynamic information of the working environment for obstacle avoidance, map building, navigation and localization. It is also used to detect vehicle system’s movement parameters and cutting mechanism’s working status. The controller compares the acquired information with the database, and then sends out revisory and accurate command to the robot to perform its tasks. The hardware of the IRM is shown in Fig. 1.
* This work is supported by national natural science fund #50675027to Ming Cong
978-1-4244-1758-2/08/$25.00 ? 2008 IEEE. 417
based on TOF measurement, which calculates the distance of the nearest reflector using the speed of sound in air and the emitted pulse and echo arrival times. The distance d to a reflected object is calculated by
d = (c ′ t ) 2 (1)
422
Fig. 1 Hardware overview of IMR
The robot must be physically strong, computationally fast, behaviourally accurate and safety. It should have the ability to perform on its own, and required no human intervention during the whole or most part of the mowing period. The IRM is modularized designed and each unit of the IRM is relatively independent. Modularized design makes the maintenance much easier. Any broken unit of the IRM can be replaced directly without influencing the functions of other units.
III. SENSING SYSTEM
A. Ultrasonic Sensor Unit
Because ultrasonic sensors can provide good range information based on the time of the flight (TOF) principle, mainly due to their simplicity and relatively low cost, they have been widely used in mobile robots for obstacle avoidance, map building and so on. This type of external sensor is very good in obstacles distance measurement. The main lobe of the sensitivity function is contained within an angle of 20 degrees, as shown in Fig. 2 [6]. A number of tests showed that the range accuracy of the sensors is in the order of
±2cm.
where c is the speed of sound, and t is the time-of-flight. The TOF method produces a range value when the echo amplitude first exceeds the threshold level after transmitting, ignoring a second echo from a further reflector.
The ultrasonic sensor unit includes a trigger pulse generation unit, a multi-channel selection unit and an echo receiving unit. A sensor interface circuitry designed to send and receive ultrasonic sound pulses catches always the first returning echo. The range data related to an object is considered to be on the conic axes even if it is located off the axes.
The ultrasonic wave typically has a frequency between 40 and 180 kHz, and the frequency of the ultrasonic sensors used in the system is 40 kHz. The beam angle is 20 degrees. The 40 kHz PWM pulse is generated by the general-purpose timer unit of DSP. To drive the transmitter effectively and not to bring much vibration, an 8 cycle burst of ultrasound at 40 kHz is sent out once a time. When the ultrasonic pulse is emitted, the sensor will experience “ringing”. Ringing caused by the transmitted pulse can cause the receiver to detect a false echo. This problem is solved by not enabling the capture interrupt of DSP until a delay interval has passed. This means that the ranger can not detect an object whose distance from the sensor is less than half the distance that sound travels during the delay interval. This is the blind zone of the ultrasonic sensor, as shown in Fig. 3.
Trigger pulse
Emitted signal
Received signal
Echo
Blind zone
Fig. 2 Typical intensity distribution of an ultrasonic sensor
On IRM, we set up a sensor array which consists of 12 ultrasonic sensors spaced 30 degrees apart. The ultrasonic signals can cover all the space around and satisfy the space requirement about which robot can detect the environmental signals. Classical techniques used in ultrasonic transducers are
TOF
Fig. 3 The sketch map of ultrasonic transmission and reception
B. Infrared Sensor Unit and Other Sensors
To overcome the ultrasonic sensor’s blind zone, infrared sensors are added. The infrared sensors can detect obstacles within 20cm, which patch up the problem caused by the blind zone problem of ultrasonic sensors.
This unit has 16 infrared sensors. Each infrared range finder has a conic view of 6 degrees which is the main lobe of the sensitivity function. This sensor has a useful measuring range of a target up to about one meter with high accuracy. A number of tests showed that the range accuracy of the sensors is in the order of ±lcm.
In order to save the DSP’s resource, 16 infrared sensors are connected with DSP TMS320F2812’s data interface
y
d
R
y
。
R
d
| ?
| ?
x
x
instead of the IO interface. This kind of architecture can also read the sensors’ status at the same time, ensuring the real- time capability of the system. A sensor interface circuitry designed to send and receive infrared pulses catches always the first retuning echo to process its amplitude.
Robot mower works in an outdoor environment, where the temperature changes rapidly. The changing of temperature will affect the speed of sound. Therefore, a temperature sensor is used to guarantee the precision of the ultrasonic sensor. Collision sensor is a group of sensitive swatches, which used to prevent the damage caused by unexpected collision. Because moist environment do harm to the circuit of the IRM, humidity sensors are introduced to detect the humidity of the environment. Although these sensors are not absolutely necessary for an autonomous robot mower, they can provide helpful functions to make the work availability and safety.
IV. SENSOR-BASED NAVIGATION
A. Mapping O
As seen in Fig. 4, a reference direction x is defined and
Fig. 4 The robot position on x-y section
the robot coordinates are shown as xR , yR . By the help of an
The equations involving the robot due to the object can be written as
electronic compass built in on the robot [7], the angle qi ,
R2 = (x - (r + d ) cos(b ))2 + ( y
- (r + d ) sin(b ))2
(6)
which is the ith sensor’s angle from the 1st sensor, can be
i i i i i i
R2 = x 2 + y 2 + (r + d )2 - 2(r + d )( x cos(b ) + y sin(b ))
easily measured. Actually if only the angle qS
(heading angle
i i i i i i i i
T T
of the robot) is measured, other sensor angles can be found as
If we define the positions as: Pi = [ p1, p1 , pn ]
= [ xi , yi ] ,
bi = qS + qi
(2)
then we have
where bi is the angle to the our world coordinate center. The
i i i [
i i i ] i
R2 = R 2 + (r + d )2 - 2(r + d ) cos(b ), y sin(b ) P
(7)
number of maximum sensor group on the ultrasonic ring is n,
After the inserting the 2 2
and the radius is r (in our system n=12 and r=0.25m). The distance between the origin and the center of the ring is R, and
R i in R ,
[(r + di ) + R cos(Ω - bi )] = [cos(bi ), yi sin(bi )] Pi
(8)
reference angle to the center is Ω . The reference position of
the robot's center is ( xR , yR ). The distance from the origin to
Here again we have n such equations. And we write them in matrix form
object which is detected by the ith sensor data on the two dimensional plane is called Ri .
[m] = [ A] Pi
And if we introduce new matrix as
(9)
Now let dmi denote measured value which is combined data from the ultrasonic and infrared sensors, for the exact
[ L(bi )] = [cos(bi ), sin(bi )] Pi
can be written as
and [f ] = [0, 0] , then (10),
distance Ri . There will be an error di between these values as
r + dm1 + R cos(ΩR - b1 ) L(b1 )
f f ×
f p1
| × | | f
L(b ) f ×
f | | × |
dmi = di + di . (3)
| | | 2
| | |
| × | = | f f
× × ×
| | × |
In this work we naturally assume that di
is a uniform | ×
| | × × × ×
f | | × |
| | | | | |
random variable in the range of (-W, W). Here W denotes the
|r + dm
+ R cos(Ω - b )| | f f
× f L(b )| | p |
maximum distance measurement error.
L n R n 」 L
n 」 L n 」
Here the problem is, given xR , yR , r, q1 ,q2 , ,qn , and
Here if we perform the least squares estimate for Pi , we obtain
dm1 , dm2 , , dmn , to estimate the coordinates of the occupied
(P ) = ( AT A)-1
AT m
(11)
lsq i
cells xi and
yi (or equivalently Ri ) in most efficient way.
Thus we find the best squares estimate of the positions.
The equations involving the detected object can be written as
B. Simulation Studies
Sensor-based navigation has been tested with simulation
R2 = (x
+ (r + d ) cos(b ))2 + ( y
+ (r + d ) sin(b ))2
(4)
to shown the usefulness of this sensor fusion method in the
i R i i R i i
i i i i i
R2 = R2 + (r + d )2 + 2(r + d )(x cos(b ) + y sin(b ))
i i i R i
R2 = R2 + (r + d )2 + 2(r + d ) cos(Ω - b )
(5)
two environments respectively as shown in Fig. 5 and Fig. 6. The mower has been primarily tested in a structured laboratory as shown in Fig. 5. Start at (0.3m, 0.5m, 0degree), a virtual
robot was driven around a virtual square corridor one time. The walls in the artificial environment are denoted by the real map.
The entire vehicle is self-contained. It has a maximum travel speed on 0.4 m/s. The laboratory area was surveyed out to a 10cm grid with accuracy better than about 1cm. To extract the mapping, a start and goal points were presented. The robot position and orientation were established by the electronic compass [8].
Fig. 5 Data collection and navigation result in structured environment
The result in Fig. 5 demonstrates the mapping quality and the usefulness of this sensor fusion method. In the tests, we
find that the average error ( e ) in estimating the position of the
obstacles in the environment was in the range of [-0.2, 0.2]m. In the simulations we see that (Plsq )i in (11), obtained does
Wall following was selected for the initial problem domain because it is a fairly simple problem to set up and evaluate [9]. It also lays the groundwork for more complex problem domains, such as maze traversal, mapping and complete coverage path planning which is used on lawn mowing and vacuuming. The simulation result of wall- following behavior shown in Fig. 6, and the experimental result in Fig. 6 demonstrate that the IRM have the capability to perform its mowing task in unstructured environment.
The program of sensor-based navigation simulation in Fig. 5 is given below.
Sub Main
Dim PI,Fcr,Fct,X_target,Y_target,X,Y As Single Dim X_grid, Y_grid, i, j, C As Integer
Dim Frx,Fry,d, dist_targ, rot, Fx, Fy As Single Dim Fcx,Fcy, Rx,Ry As Single
PI=3.1415927
Fcr=1 Fct=1
X_target=GetMarkX(0) Y_target=GetMarkY(0)
SetCellSize(0,0.1) 'Set cell size 10 cm x 10 cm
SetTimeStep(0.1) 'Set simulation time step of 0.1 seconds
Do ' Start main loop
X=GetMobotX(0) 'Present mobot coordinates (in meters)
Y=GetMobotY(0)
X_grid=CoordToGrid(0,X) ' indexes of cells where the Y_grid=CoordToGrid(0,Y) ' mobot center is MeasureRange(0,-1,3) ' Perform a range scan and update
' the Certainty Grid (max. cell value=3)
not satisfy Ri = (Plsq )i
which actually should. In the case a
Frx=0
better estimate for the positions can be given as
Fry=0
' Each ocuppied cell inside the windows of 33 x 33 cells
Ri
(Plsq )i
(P )
= (P )
(12)
' applies a repulsive force to the mobot.
e i lsq i
In this case, estimate for the angle Ωi does not change but the estimate for distance Ri is scaled to it best estimate. Therefore for the position, the distance estimate Ri remains the same as before, while the least squares estimate works only for the angle Ωi . Simulations show that this way produces more accurate results.
For i=X_grid-16 To X_grid+16
For j=Y_grid-16 To Y_grid+16 C=GetCell(0,i,j)
If C>0 Then
d=Sqr((X_grid-i)^2+(Y_grid-j)^2) If d>0 Then
Frx=Frx+Fcr*C/d^2*(X_grid-i)/d Fry=Fry+Fcr*C/d^2*(Y_grid-j)/d
End If End If
Next Next
dist_targ=Sqr((X-X_target)^2+(Y-Y_target)^2) Fcx=Fct*(X_target-X)/dist_targ Fcy=Fct*(Y_target-Y)/dist_targ Rx=Frx+Fcx
Ry=Fry+Fcy rot=RotationalDiff(0,X+Rx,Y+Ry)
'shortest rotational difference between 'current direction of travel and direction of vector R
SetSteering(0,0.5,3*rot)'mobot turns into the direction of R
'at constant speed and steering rate 'proportional to the rotational difference
Fig. 6 The simulation result of wall-following behavior
StepForward
Loop Until dist_targ<0.1 'Loop until mobot reaches the target
End Sub
V. ULTRASONIC RANGING TECHNOLOGY BASED ON WT
Unfortunately, the practical received multi-echoes has time-varying property and is a typical non-stationary signal because the influence of the environmental complexity and the noise. Furthermore, the noise mixed in the ultrasonic pulse- echo is Non-Gaussian white noise but colored noise, and correlated with the target echo. The TOF method can not be used directly in such conditions. Referencing the generalized correlation method for estimation of time delay [10], we put forward the generalized auto-correlation method for estimation of time-of-flight based on wavelet transform [11] and present in Fig. 7.
Fig. 7 Delay estimation of generalized auto-correlation based on WT
Where j (t) is the mother wavelet and jat (t) is the daughter wavelet. The coefficient a is the scale (or scaling factor) andt is the time displacement. The wavelet transform
of the signal x(t) is y(t) . Actually this is a filtering process of the ultrasonic echo using a multitude of bandpass filters of equal Q , which is equivalent to the whitening filter of the generalized correlation method for estimation of time delay, in order to eliminate the input noise which can influence the following processing. Ryy (t ) can be found as
Ryy (t ) = E[ y(t) y(t - t )] = Rxx (t) *[jat (t) *jat (-t)]
As there has the relationship of Fourier transform
Fig. 9 Denoised echo using WT
Fig. 10 Auto-correlation function Ryy (t )
between auto-correlation function spectrume:
Ryy (t )
*
and his power
2
Fig. 11 Peak detection
The noisy ultrasonic echo is shown in Fig. 8, and the
Gyy (w) = F[Ryy (t )] = Gxx (w)Y(aw)Y (aw) = Gxx (w) Y(aw)
We obtain the generalized auto-correlation function as
denoised ultrasonic echo by wavelet transform is shown in
Fig. 9. It is obvious that the noise mixed in the ultrasonic echo is effectively eliminated after WT operation. The auto-
( g ) 1 ¥
jwt 1 ¥
jwt
Ryy
(t ) =
2p
-¥ Gyy (w)e dw =
2p
y g (w)Gyy (w )e dw
-¥
correlation operation Ryy (t ) of the denoised ultrasonic echo is
shown in Fig. 10. Fig. 11 shows the envelope of Ryy (t ) through
Last, the peak values of Ryy (t ) are detected to accomplish
the estimation of TOF and calculate the real ultrasonic velocity.
Fig. 8 Noisy ultrasonic echo
Hilbert transform. As
收藏