999精品在线视频,手机成人午夜在线视频,久久不卡国产精品无码,中日无码在线观看,成人av手机在线观看,日韩精品亚洲一区中文字幕,亚洲av无码人妻,四虎国产在线观看 ?

Autonomous Landing of Small Unmanned Aerial Rotorcraft Based on Monocular Vision in GPS-denied Area

2015-08-11 11:57:29CunxiaoMiaoandJingjingLi
IEEE/CAA Journal of Automatica Sinica 2015年1期

Cunxiao Miao and Jingjing Li

Autonomous Landing of Small Unmanned Aerial Rotorcraft Based on Monocular Vision in GPS-denied Area

Cunxiao Miao and Jingjing Li

—Focusing on the low-precision attitude of a current smallunmanned aerialrotorcraftat the landing stage,the present paper proposes a new attitude controlmethod for the GPS-denied scenario based on the monocular vision.Primarily,a robust landmark detection technique is developed which leverages the well-documented merits of supporting vector machines(SVMs) to enable landmark detection.Then an algorithm of nonlinear optimization based on Newton iteration method for the attitude and position of camera is put forward to reduce the projection error and get an optimized solution.By introducing the wavelet analysis into the adaptive Kalman filter,the high frequency noise of vision is filtered out successfully.At last,automatic landing tests are performed to verify the method's feasibility and effectiveness.

Index Terms—Automatic landing,monocular vision,attitude, wavelet filter.

I.INTRODUCTION

W ITH the ability to take off and land vertically,along with hovering ability,the smallunmanned aerialrotorcraft(SUAR)has an irreplaceable role in civilapplications[1]. Thus,it has been widely used in many regions,such as road traffic monitoring,data acquisition,mapping and surveillance, etc[2?4].And it allows us to easily access environments to which no humans or other vehicles can get access.

Now most navigation systems for SUAR are based on inertial sensors(gyroscopes and accelerometer)and GPS[5?6]. However,buildings and other obstacles in urban environments can easily disturb the GPS signal and even cause the signal to get lost due to electromagnetic noise from active emitters and multi-path effects.And it directly leads to the poor performance of SUAR as a result of the inertial navigation errors generated by the gyroscope drifts and accelerometer biases without GPS.So there is a requirementthatthe vehicle is able to navigate without using GPS to guarantee the safety of operations in unknown,unstructured and GPS-denied environments.

Considering the problem mentioned above,a feasible solution is to navigate the vehicle using a system based on vision.Small cameras which are low-lost,light-weight and passive are attractive sensors for SUAR.Furthermore,a vision system is self-suffi cient,and provides position values which are much more accurate than the standard GPS's(cm accuracy).Off-board cameras for motion capture systems are well studied[7?8].However,the cameras can be used only in small environments that are accessible physically by humans,and the cameras's field of view directly constrains the operation of the helicopter.Therefore,it is necessary to install the camera onboard.Saripalli used vision for the precise target detection and recognition and the navigation successfully in combination with GPS[9].Moreover,Hermansson established an EKF model to fuse the measurements of vision,GPS and compass,and realized a landing within 0.5 m in the horizontal direction[10].Obviously,these systems still relies on the GPS signal.The stereo vision is also used to detectthe safe landing area and to achieve soft landing[11?12].However,the range data become inaccurate when the distance to the scene is much larger than the baseline[13].

In this paper,we propose an autonomous landing method for SUAR to tackle the problem of poor navigation in GPS-denied environment by using a single camera and onboard inertial sensors.The position and attitude of the SUAR are estimated by the vision system using a specially designed landmark firstly.Then a wavelet-adaptive KF is developed by fusing the position estimated by the vision system with inertial data to improve the performance of SUAR.

The paper is organized as follows:The image processing algorithm is described in Section II.In Section III,an adaptive-KF based on wavelet filter is proposed to get high-precision estimated values.Simulation and test results confirm the effectiveness of the proposed method in Section IV.Finally, conclusions are drawn in Section V.

II.THE DESCRIPTION OF VISION SYSTEM

To land successfully,two basic stages are required by the visual navigation system for SUAR:the landmark detection and the accurate estimation of position and attitude for SUAR.

A.Landmark Detection Based on SVM

As a basic step of a visualnavigation system,usually there are three methods for landmark detection.Special landmarksare identified easily,but this approach cannot be applied commonly.Besides,pattern matching[14]is a mostly used way, too.However,the main shortcomings are the large calculation and vast experiments are needed when dividing the similarity threshold.Machine learning[15?16]is recognized as the most intelligent method,and can be used commonly to detect any landmark.

In order to significantly improve the effect of landmark detection,an innovative landmark detection technique was studied and finally specified.Its support vector machine (SVM)takes advantage of the Hu invariant moments[17].

A series of Hu invariant moments of several images of landmark are used as inputto the SVM,which now is widely used for detection purposes and is based on a training set of elements.Basically,the SVM technique aims to geometrically separate the training set represented in a n space,with n standing for the number of criteria taken into account for classification,using a hyper-plane or some more complex surface if necessary.The SVM training algorithm finds out the best frontier in order to maximize the margin,defined as a symmetric zone centered on the frontierwith no training points included,and to minimize the number of wrong classification occurrences,as which can be shown in Fig.1.

Fig.1.SVM detection in the two-dimensional space.

The output of SVM training stage is thus the hyper-plane equation[18]

where→x is the n components vector representing the image to be classified.It should be noted that the normal vector→w is a linear combination of a reduced setof training vectors→x, located nearby the two parallelhyper-planes defining the margin:points located far from the margin have no contribution in the hyper-plane definition.

Then the classification algorithm is straight forward:if→yi(〈→w,→xi〉+b)?1≥ 0,then the corresponding mark is classified as landmark-free.

When the linear separation is not efficient,it often proves to become linear when applying some non-linear transform on the coordinate,and SVM technique would be applied to achieve efficiently this non-linear transform using kernels.

In addition,the SVM convex programming problem can be converted to a dual problem using Lagrange method. That is,a conditional extremum can be constructed as min‖w‖2/2,s.t. →yi(〈→w,→xi〉+b)?1≥0.Here,we use thedefined as minw,bmaxαi≥0L(w,b,αi).Then,the expression above can be transformed through dualistic transformation as maxαi≥0minw,bL(w,b,αi).Then the problem can be resolved using a series of numerical method[19].

As the input to SVM,the first,second and third moments of inertia are testified to be sufficient to distinguish between the landing target and other objects presented in the image. The moments of inertia can be shown as

whereφ1,φ2,φ3are the three lower order invariantmoments, andηpqis the normalized central moment which can be defined as

whereγ=p+2q+1,p+q=2,3,···andμpqrepresents the central moment of an object,which is given by

where(ˉx,ˉy)represents the center of the gravity of the object, f(x,y)represents a two-dimensional object as a continuous function with p,q=0,1,2,···.

B.State Estimation Based on Vision

We have chosen black triangles on white background as they are fast to detect and provide accurate image features. When the landmark is detected in the current image,corners are extracted by the Kitchen and Rosenfild algorithm.Then from the projection ofthe cornerpoints ofthe targetlandmark, the attitude and position of the SUAR is uniquely determined through the following equations,if allintrinsic camera parameters are known.

Suppose that the landmark is detected,and the normalized coordinates of the corner in the navigation system can be represented as qqqi(i=1,2,···,12)(12 points are used in this research),and the corresponding coordinates in the image plane is xxxi=[u v 1]T(i=1,2,···,12);then the simplified form of(5)can be expressed as follows

whereλiis equal to zc,and thus we can obtain thatλi=where eeeT3=0 0 1.

If yyyi=AAA?1xxxi,then we can define Formula(6)as

and rrr3=rrr1×rrr2.Finally,the attitude of the SUAR can be obtained from RRR.

Through the linear algorithm above, 9 variables (rrr1,rrr2,ppp)composed of attitude and position parameters ?,θ,ψ,p1,p2,p3can be estimated.It is clear that the 9 variables contain much noise,but they can be used as initial values of the non-linear algorithm below.

The estimation from the non-linear algorithm is optimized by minimizing the reprojection error G as below

Therefore,the attitude and position parameters based on Newton method can be expressed as

III.SENSOR FUSION

The attitude and position information from the vision system cannot be fed back to the controller directly because of their lack of robustness.

Therefore,a filter based on Kalman filter(KF)is developed to fuse highly accurate position estimated from vision system with inertial data from the inertial measurement unit(IMU, including angular rate gyroscope and accelerometers).Not only can the filter filter out most of the noise,but it can also provide sufficient information to complete the task when the vision system is disturbed.

A.The State and Observation Model of SUAR

The KF is done using the error state space as follows

We can conclude the discrete state equation XXXk= Φk,k?1XXXk?1+WWWk?1.In(12),δrrr,δVVV andΨare the error vectors of the position,velocity and attitude,ωieis the Earth rotation rate,fff is the accelerometers outputs,▽is the bias of accelerometer,εis the zero driftof gyroscope,XXXkis the state vector,Φk,k?1is the state transition matrix and WWWkrepresents the system noise.

The three observations are obtained by the difference between the position from vision system and the position from INS.Then the observational equation can be given as

where ZZZkis the observation vector and VVVkis the observation noise.

B.The Wavelet-adaptive KF

1)The wavelet filter method

Sensor noises are always treated as zero mean Gaussian white noise in traditional Kalman filter.However,to get high precision position and attitude information,it's necessary to get rid of high frequency noises in the sensors(SINS and vision system).The common ways to deal with the high frequency noises include the data smoothing filter,infinite impulse response(IIR)filter,finite impulse response(FIR) filter and wavelet filter.The data smoothing can eliminate the measurement outlier and noise for the high frequency data, but it has high requirement for the sensor data collection system.Although the IIR can eliminate the high frequency noise,itcan cause a phase delay.The FIR can reduce the noise signal energy,but it has limitation in the high frequency noise suppression.Wavelet filter is a time domain and frequency domain method,having good representation for the partial signal characteristic.In the low frequency region,it has a higher frequency resolution and lower time resolution.On the other hand,ithas higher time resolution and a lower frequency resolution in the high frequency region.Therefore,wavelet filter is used as a tool to reduce high frequency noises in the sensor information.

We can suppose the data with noise as

where f(t)and g(t)represent the real signal and noise respectively.Besides,they are independent of each other,and ε(t)∈N?0,σ2¢.

The process of wavelet decomposition and reconstruction scheme includes three steps:

a)Wavelet decomposition

b)Prediction:defining the detailed representation characteristics by choosing a predictor.

c)Update:averaging the signal of rough representation against the original signal.

2)AKF description

The Kalman filter is a setof mathematicalequations thatuse an underlying process model to estimate the current system state and correct the estimated value.Using this predictorcorrectormechanism,itcan approximate an optimalestimation from the linearization ofthe process and measurementmodels.

Then the filter consists of the following stages:

a)State prediction

b)Measurement prediction

c)Updated state estimation

d)Gain of the filter

e)Error covariance

f)Variance matrix of the observation noise

Considering that the noise structure has changed after wavelet filtering,we cannot simply use experiential value or the statistics of partialnoise as the observation noise variance; here the observation noise variance is estimated using the maximum by a posteriori adaptive method.

IV.SIMULATION AND EXPERIMENT

A.System Description

To test the effectiveness of the proposed method,an autopilot with embedded system based on 2 pieces of highspeed DSPs and microcontroller were developed,to realize the complex algorithms of image processing,navigation and control,as shown in Fig.2.In addition,it consists of a horizontal main board,housing 3-axis rate gyroscopes,3-axis accelerometers and a barometer.The gyro employs the LCG50 produced by System Donner Inertial Company in British.The accelerometer is the Model 1221 manufactured by Silicon Designs Company in Japan.Besides,the camera used here is a RICOR aerial camera,which is light(about 188 g),with a fixed focal length about 6 mm.In view of the time limit, we set the frame frequency at 5 f/s.For flight stability,the navigation and control cycle are set at 50 Hz.Besides,the data fusion method is conducted in 1 Hz of data frequency and the coefficient w is determined by a series of images of the landmark before landing.

Fig.2.The onboard system and the SUAR used in the experiment.

B.Autonomous Landing Test

The landing pad used here is a 1.2 m×1.2 m specially made landmark.The flight trajectory of the SUAR in the landing process is shown in Fig.3.When the SUAR arrives above the landing area,it starts to search for the landmark and begin to land.With the constant adjustment for the planed hovering altitude,the SUAR descends the altitude with hovering station. When the altitude is less than 2 m,the system reaches to the land disturbance region.Since there exists land disturbance,the pilot control model[20]is used to control the collective motor.The pilot control model provides a safe control mode for the SUAR.When the altitude is less than 0.5 m shown in Fig.4,the gun is decreased with constantspeed.Thus,the lift force has been decreased correspondingly.Finally,the SUAR lands the ground at the(0.26,0.42,0).Figs.5 and 6 show the attitude and horizontal velocity in the landing process, respectively.

Fig.3.The 3D trajectory of the SUAR.

Fig.4.The vertical positioning results.

Fig.5.The attitude estimated during the landing process.

Fig.6.The velocities in two directions.

To furtherdemonstrate the proposed method,severallanding experiments have been conducted using the wavelet-adaptive KF and the EKF method used before.

Fig.7 shows the landing results using the wavelet-adaptive KF and the EKF,separately.The SUAR enters the landing mode roughly ten meters above the landing platform and then slowly descends until it has landed.The average Euclidean distances from the landing target are 63 cm and 101 cm, respectively.

Fig.7.The result of landing tests.

V.CONCLUSION

Focusing on the low measurementperformance of the sensor for the low cost SUAR,a navigation method based on vision was proposed to improve the accuracy and reliability of the measurement information.The nonlinear optimization based on Newton iteration method for camera attitude and position was put forward to reduce the projection error and get an optimized solution.With the wavelet filter method,the high frequency noises in the SINS and vision have been eliminated effectively.By using the adaptive Kalman filter,the system fuses the output from SINS and vision to get high precision navigation information.Finally,the landing tests show thatthe compound navigation system based on the proposed method can provide high performance navigation information.With the high performance position and attitude information,theSUAR can realize stable autonomous landing based on the altitude information.

REFERENCES

[1]Zhong Li-Na,Wang Jun-Hao,Wang Rong.Auto flight algorithm of quad-rotor helicopter based on magnetic sensor.Journal of Chongqing University of Technology(Natural Science),2013,27(12):86?90(in Chinese)

[2]Jiang Bin,Sun Zhi-Feng.The realization of autonomous navigation for quadrotor aircraft.Electronic Technology,2012,39(2):10?12,9(in Chinese)

[3]Najib M,Tarek H.A UAR for bridge inspection:visualservoing control law with attitude limits.AutomationinConstruction,2007,17(1):3?10

[4]Weiss S,Scaramuzza D,Siegwart R.Monocular-SLAM-based navigation for autonomous micro helicopters in GPS-denied environments. Journal of Field Robotics,2011,28(6):854?874

[5]Wendel J,Meister O,Schlaile C,Trommer G F.An integrated GPS/MEMS-IMU navigation system for an autonomous helicopter. Aerospace Science and Technology,2006,10(6):527?533

[6]Shin E H,El-Sheimy N.Accuracy improvement of low cost INS/GPS for land applications.In:Proceedings of the 2002 U.S.Inst.Navigation, National Technical Meeting.San Diego,CA USA:ION NTM,2002. 146?158

[7]Michael N,Fink J,Kumar V.Cooperative manipulation and transportation with aerial robots.Autonomous Robots,2011,30(1):73?86

[8]How J P,Bethke B,Frank A,Dale D,Vian J.Real-time indoor autonomous vehicle test environment.IEEE Control Systems Magazine, 2008,28(2):51?64

[9]Saripalli S,Sukhatme G S,Montgomery J F.AnExperimental Study of the Autonomous Helicopter Landing Problem—Experimental Robotics VIII.Berlin:Springer-Verlag,2003.466?475

[10]Hermansson J,Gising A,Skoglund M,Sch¨on.Autonomous Landing of an Unmanned Aerial Vehicle,Technical Report,Automatic Control, Linkopings Universitet,Sweden,2010.

[11]Johnson A E,Montgomery J F,Matthies L.Vision guided landing of an autonomous helicopter in hazardous terrain.In:Proceedings of the 2005 IEEE International Conference on Robotics and Automation.Barcelona, Spain:IEEE,2005.3966?3971

[12]Yu Z Y,NonamiK,Shin J,Demian C.3D vision based landing controlof a small scale autonomous helicopter.International Journal of Advanced Robotic Systems,2007,4(1):51?56

[13]Press W H,Teukolsky S A,Vetterling W T,Flannery B P.Numerical Recipes in C:the Art of Scientifi c Computing(Second edition).Cambridge:Cambridge University Press,1992.

[14]Sun Yan-Yue,He Xiao-Hai,Song Hai-Ying,Chen Wei-Long.A blockmatching image registration algorithm for video super-resolution reconstruction.Acta Automatica Sinica,2011,37(1):37?43(in Chinese)

[15]Wan Yu-Chai,Liu Xia-Bi,Han Fei-Fei,Tong Kun-Qi,Liu Yu.Online learning of binary classifiers for improving google image search results. ActaAutomatica Sinica,2014,40(8):1699?1708

[16]Chong Yan-Wen,Kuang Hu-Lin,Li Qing-Quan.Two-stage pedestrian detection based on multiple features and machine learning.Acta AutomaticaSinica,2012,38(3):375?381(in Chinese)

[17]Hu M K.Visual pattern recognition by moment invariants.IRE Transactions onInformationTheory,1962,8(2):179?187

[18]Latry C,Panem C,Dejean P.Cloud detection with SVM technique. In:Proceedings of the 2007 IEEE International Conference on Geoscience and Remote Sensing Symposium.Barcelona,Spain:IEEE,2007. 448?451

[19]Cristiani N,Shawe-Taylor J.AnIntroductiontoSupportVectorMachines and Other Kernel-based LearningMethods.Cambridge:Cambridge University Press,2000.

[20]Lei Xu-Sheng,Fang Jian-Cheng,Bai Lang,etc.Autonomous Taking Off and Landing Technology for the Small Unmanned Rotorcraft Vehicle Based on the Pilot Model,C.N.Patent 102289714B,December 2011. (in Chinese)

Cunxiao Miao Lecturer at the School of Mechanical Engineering,University of Science and Technology Beijing.He received the B.S.degree from Northeast University,China in 2006 and the Ph.D.degree from Beihang University,China in 2013.His research interests include navigation,identification,and control of unmanned aerial vehicles. Corresponding author of this paper.

Jingjing Li Assistant engineer at Beijing Aerospace Times Optical-electronic Technology Co., Ltd..She received her master degree from Beihang University in 2013.Her research interests include inertial navigation and visual navigation of unmanned aerial vehicles.

t

October 10,2013;accepted July 18,2014.This work was supported by China Postdoctoral Science Foundation(2013M540857) and Fundamental Research Funds for the Central Universities(FRF-TP-14-019A1).Recommended by Associate Editor Bin Xian

:Cunxiao Miao,Jingjing Li.Autonomous landing of small unmanned aerial rotorcraft based on monocular vision in GPS-denied area. IEEE/CAA Journal of AutomaticaSinica,2015,2(1):109?114

Cunxiao Miao is with the School of Mechanical Engineering,University of Science and Technology Beijing,Beijing 100083,China(e-mail: miao cunxiao@163.com).

Jingjing Li is with Beijing Aerospace Times Optical-electronic Technology Co.,Ltd(ATOT),Beijing 100094,China(e-mail:001rose001@sina.com).


登錄APP查看全文

主站蜘蛛池模板: 最新国语自产精品视频在| 国产激情在线视频| 国产乱子伦一区二区=| 亚洲国产日韩欧美在线| 亚洲男人的天堂视频| 无码国产伊人| 污污网站在线观看| 国产菊爆视频在线观看| 久热中文字幕在线| 国产精品女同一区三区五区| a级毛片免费在线观看| 免费一极毛片| 99偷拍视频精品一区二区| 欧美在线精品怡红院| 91精品专区国产盗摄| 亚洲性视频网站| 国产视频入口| 91亚洲免费| 九九热视频精品在线| 亚洲国产成人麻豆精品| 国产亚洲精品自在久久不卡| 亚洲成A人V欧美综合| 中文字幕日韩久久综合影院| 精品偷拍一区二区| 久久黄色小视频| 一本无码在线观看| 女人18一级毛片免费观看| 亚洲永久免费网站| 欧美a级完整在线观看| 无遮挡国产高潮视频免费观看| 欧美日韩在线亚洲国产人| 久久综合干| 成人毛片免费在线观看| 青青操视频免费观看| 无码专区国产精品一区| 国产精品第一区在线观看| 国产成人狂喷潮在线观看2345 | 亚洲色图另类| jizz国产在线| 在线99视频| 91探花在线观看国产最新| 天天做天天爱夜夜爽毛片毛片| 在线免费亚洲无码视频| 嫩草在线视频| 人妻无码中文字幕第一区| 午夜精品区| 国产丝袜第一页| 久久精品亚洲专区| 精品国产美女福到在线不卡f| 尤物在线观看乱码| 国产玖玖玖精品视频| 午夜视频在线观看免费网站| 中文字幕亚洲精品2页| 国产在线视频自拍| 特黄日韩免费一区二区三区| 精品午夜国产福利观看| 一级毛片免费观看久| 久久亚洲国产视频| 少妇精品在线| 99视频只有精品| 中文字幕乱妇无码AV在线 | 女人一级毛片| 国产97视频在线| 国产99欧美精品久久精品久久| 二级特黄绝大片免费视频大片| 欧美精品色视频| 欧美日在线观看| 最新加勒比隔壁人妻| 呦女亚洲一区精品| 老司机久久精品视频| 免费无码一区二区| 久久国产精品嫖妓| 午夜丁香婷婷| 四虎影视库国产精品一区| 国产小视频在线高清播放| 国产午夜精品一区二区三区软件| 亚洲国产成人精品无码区性色| 伊人蕉久影院| 91久久性奴调教国产免费| 极品国产一区二区三区| 97人人做人人爽香蕉精品| 97超级碰碰碰碰精品|