1. Trang chủ
  2. » Kỹ Thuật - Công Nghệ

Mobile Robots -Towards New Applications 2008 Part 11 pps

40 265 0

Đang tải... (xem toàn văn)

Tài liệu hạn chế xem trước, để xem đầy đủ mời bạn chọn Tải xuống

THÔNG TIN TÀI LIỆU

Thông tin cơ bản

Định dạng
Số trang 40
Dung lượng 0,98 MB

Nội dung

An Active Contour and Kalman Filter for Underwater Target Tracking and Navigation 391 Kalman State Estimation 0,00 50,00 100,00 150,00 200,00 250,00 5 30 55 80 105 130 155 180 205 230 255 280 305 330 355 380 405 430 Frames Pixels Actual state Predicted State Updated state Fig. 11. Comparison of actual and predicted and updated position of underwater pipeline using the Kalman tracking algorithm. 9. References Whitcomb, L.L. (2000). Underwater robotics: out of the research laboratory and into the field, IEEE International Conference on Robotics and Automation, ICRA '00, Vol.1, 24-28 April 2000, pp. 709 – 716. Asakawa, K., Kojima, J., Kato, Y., Matsumoto, S. and Kato, N. (2000). Autonomous underwater vehicle AQUA EXPLORER 2 for inspection of underwater cables, Proceedings of the 2000 International Symposium on Underwater Technology, 2000, UT 00. 23-26 May 2000, pp. 242 – 247. Ortiz, A., Simo, M., Oliver, G. (2002). A vision system for an underwater cable tracker, Machine vision and application 2002, Vol.13 (3), July 2002, pp. 129-140. Griffiths, G. and Birch, K. (2000). Oceanographic surveys with a 50 hour endurance autonomous underwater vehicle, Proceeding of the Offshore Technology Conference, May 2000, Houston, TX. Asif, M. and Arshad, M.R. (2006). Visual tracking system for underwater pipeline inspection and maintenance application, First International Conference on Underwater System Technology, USYS06. 18 – 20 July 2006, pp 70-75. Cowls, S. and Jordan, S. (2002). The enhancement and verification of a pulse induction based buried pipe and cable survey system. Oceans '02 MTS/IEEE. Vol. 1, 29-31 Oct. 2002, pp. 508 – 511. Petillot, Y.R., Reed, S.R. and Bell, J.M. (2002). Real time AUV pipeline detection and tracking using side scan sonar and multi-beam echo-sounder, Oceans '02 MTS/IEEE. Vol. 1, 29-31 Oct. 2002, pp. 217 - 222. Evans, J., Petillot, Y., Redmond, P., Wilson, M. and Lane, D. (2003). AUTOTRACKER: AUV 392 Mobile Robots, Towards New Applications embedded control architecture for autonomous pipeline and cable tracking, OCEANS 2003, Proceedings, Vol. 5, 22-26 Sept. 2003, pp. 2651 – 2658. Balasuriya, A. & Ura, T. (1999) Multi-sensor fusion for autonomous underwater cable tracking, Riding the Crest into the 21st Century OCEANS '99 MTS/IEEE, Vol. 1, 13-16 Sept. 1999, pp. 209 – 215. Foresti, G.L. (2001). Visual inspection of sea bottom structures by an autonomous underwater vehicle Systems, IEEE Transactions on Man and Cybernetics, Part B, Vol. 31 (5), Oct. 2001, pp. 691 – 705. Matsumoto, S. & Ito, Y. (1997). Real-time vision-based tracking of submarine-cables for AUV/ROV, MTS/IEEE Conference Proceedings of OCEANS '95, 'Challenges of Our Changing Global Environment, Vol. 3, 9-12 Oct. 1995, pp. 1997 – 2002. Balasuriya, B.A.A.P., Takai, M., Lam, W.C., Ura, T. & Kuroda, Y. (1997). Vision based autonomous underwater vehicle navigation: underwater cable tracking MTS/IEEE Conference Proceedings of OCEANS '97, Vol. 2, 6-9 Oct. 1997, pp. 1418 – 1424. Zanoli, S.M. & Zingretti, P. (1998). Underwater imaging system to support ROV guidance, IEEE Conference Proceedings of OCEANS '98, Vol. 1, 28 Sept 1 Oct. 1998, pp. 56 – 60. Perona, P. & Malik, J. (1990). Scale-space and edge detection using anisotropic diffusion, IEEE Transactions on Pattern Analysis and Machine Intelligence, Vol. 12(7), July 1990, pp. 629 – 639. Weickert, J. (2001). Applications of nonlinear diffusion in image processing and computer vision, Proceedings of Algoritmy 2000, Acta Math. University Comenianae Vol. LXX, 2001, pp. 33 – 50. Blake, A. & Isard, M. (1998). Active Contour, Springer, Berlin, 1998. Cootes, T., Cooper, D. Taylor, C. & Graham, J. (1995). Active shape models – their training and application, Computer Vision and Image Understanding, Vol. 61(1), 1995, Pages 38 – 59. MacCormick, J. (2000). Probabilistic modelling and stochastic algorithms for visual localization and tracking. Ph.D. thesis, Department of Engineering Science, University of Oxford 2000. 19 Robotics Vision-based Heuristic Reasoning for Underwater Target Tracking and Navigation Kia Chua and Mohd Rizal Arshad USM Robotics Research Group, Universiti Sains Malaysia Malaysia 1. Introduction Most of underwater pipeline tracing operations are performed by remote operated vehicles (ROVs) driven by human operators. These tasks often require continued attention and knowledge/experience of human operators to maneuver the robot (Foresti. G.L. and Gentili.,2000). In these operations, human operators does not require an exact measurement from the visual feedback, but based on the reasoning. For these reasons, it is desirable to develop robotics vision system with the ability to mimic the human mind (human expert’s judgement of the terrain traverse) as a translation of human solution. In this way, human operators can be reasonably confident that decisions made by the navigation system are reliable to ensure safety and mission completion. To achieve such confidence, the system can be trained by expert (Howard. A. et al, 2001). In order to enable robots to make autonomous decision that guide them through the most traversable regions of the terrain, fuzzy logic techniques can be developed for classifying traverse using computer vision-based reasoning. Computing with words is highly recommended either when the available information is too imprecise to use numbers, or when there is a tolerance for imprecision which can be exploited to get tractability and a suitable interface with the real world (Zadeh. L, 1999). Current position based navigation techniques cannot be used in object tracking because the measurement of the position of the interested object is impossible due to its unknown behavior (Yang Fan and Balasuriya, A, 2000). The current methods available in realizing target tracking and navigation of an AUV, used optical, acoustic and laser sensors. These methods have some problems, mainly, in terms of complicated processing requirement and hardware space limitation on AUVs (Yang Fan and Balasuriya, A, 2000). Other relevant research consists of neural-network terrain-based classifier in Foresti. et.al (2000) and Foresti, G. L. and Gentili (2002). Also, existing method using Hough transform and Kalman filtering for image enhancement has also been very popular (Tascini, G. et al, 1996), (Crovatot, D. et al, 2000), (El-Hawary, F. and Yuyang, Jing, 1993) , (Fairweather, A. J. R. et al, 1997) and (El-Hawary, F. and Yuyang, Jing, 1995). 2. Research Approach Visible features of underwater structure enable humans to distinguish underwater pipeline from seabed, and to see individual parts of pipeline. A machine vision and image processing 394 Mobile Robots, Towards New Applications system capable of extracting and classifying these features is used to initiate target tracking and navigation of an AUV. The aim of this research is to develop a novel robotics vision system at conceptual level, in order to assist AUV’s interpretation of underwater oceanic scenes for the purpose of object tracking and intelligent navigation. Underwater images captured containing object of interest (Pipeline), simulated seabed, water and other unwanted noises. Image processing techniques i.e. morphological filtering, noise removal, edge detection, etc, are performed on the images in order to extract subjective uncertainties of the object of interest. Subjective uncertainties became multiple input of a fuzzy inference system. Fuzzy rules and membership function is determined in this project. The fuzzy output is a crisp value of the direction for navigation or decision on the control action. 2.1 Image processing operations For this vision system, image analysis is conducted to extract high-level information for computer analysis and manipulation. This high-level information is actually the morphological parameter for the input of a fuzzy inferences system (linguistic representation of terrain features). When an RGB image is loaded, it is converted into gray scale image. RGB image as shown in Fig. 1. Gray-level thresholding is then performed to extract region of interest (ROI) from the background. The intensity levels of the object of interest are identified. The binary image B[i,j], is obtained using object of interest’s intensity values in the range of [T 1 , T 2 ] for the original gray image F[i,j]. That is, ¯ ®  = 0 1 ],[ jiB otherwise TjiFT 21 ],[ ≤< (1) The thresholding process producing a binary image with a large region of connected pixels (object of interest) and large amount of small region of connected pixels (noise). Each region is labeled and the largest connected region is identified as object of interest. In the labeling process, the connected pixels are labeled as either object of interest or unwanted objects by examining their connectivity’s (eight-connectivity) to neighboring pixels. Label will be assigned to the largest connected region that represents the object of interest. At this stage, feature extraction is considered completed. The object of interest is a pipeline laid along the perspective view of the camera. The image is segmented into five segments and processed separately for terrain features as multiple steps of inputs for the fuzzy controller. In order to investigate more closely each specific area within the image segment, each segment is further divided into six predefined sub segments in the image. Each sub segment (as illustrated by Fig. 2) is defined as follows. • Sub segment 1 = Upper left segment of the image • Sub segment 2 = Upper right segment of the image • Sub segment 3 = Lower left segment of the image • Sub segment 4 = Lower right segment of the image • Sub segment 5 = Upper segment of the image • Sub segment 6 = Lower segment of the image. A mask image with constant intensity is then laid on the image as shown in Fig. 3. This is actually an image addition process whereby it will produce a lighter (highest intensity value) area when intersects the region of interest. The remaining region with highest Robotics Vision-Based Heuristic Reasoning for Underwater Target Tracking and Navigation 395 intensity value then be calculated its coverage area in the image as shown in Fig. 4. The area, A of the image is determined by. ¦¦ == = m j n i jiBA 11 ],[ (2) Sub segment 5-6 are being determined its location relative to the image center. Coverage area and location of object of interest in each sub segment is finally be accumulated as multiple input of the fuzzy inference system. Fig. 1. Typical input image (RGB). Fig. 2. Show image sub segment. Fig. 3. Mask on threshold, removed noise image. 396 Mobile Robots, Towards New Applications Fig. 4. Acquired area information. 2.2 The fuzzy inference system The fuzzy controller is designed to automate how a human expert, who is successful at this task, would control the system. The multiple inputs to the controller are variables defining the state of the camera with respect to the pipeline, and the single output is the steering command set point. Consider the situation illustrated by Fig. 5. The fuzzy logic is used to interpret this heuristic in order to generate the steering command set point. In this case, the set point of AUV has a certain amount (ƦX) to the right. Basically, a human operator does not require a crisp / accurate visual input for mission completion. There are total of six inputs based on the image processing algorithm. • Input variable 1, x 1 = Pipeline area at upper left segment in the image Input variable 1 fuzzy term set, T(x 1 ) = {Small, Medium, Large} Input variable 1 universe of discourse, U(x 1 ) = [0.1 -1.0] • Input variable 2, x 2 = Pipeline area at upper right segment in the image Input variable 2 fuzzy term set, T(x 2 ) = {Small, Medium, Large} Input variable 2 universe of discourse, U(x 2 ) = [0.1 -1.0] • Input variable 3, x 3 = Pipeline area at lower left segment in the image Input variable 3 fuzzy term set, T(x 3 ) = {Small, Medium, Large} Input variable 3 universe of discourse, U(x 3 ) = [0.1 -1.0] • Input variable 4, x 4 = Pipeline area at lower right segment in the image Input variable 4 fuzzy term set, T(x 4 ) = {Small, Medium, Large} Input variable 4 universe of discourse, U(x 4 ) = [0.1 -1.0] • Input variable 5, x 5 = End point of pipeline relative to image center point Input variable 5 fuzzy term set, T(x 5 ) = {Left, Center, Right} Input variable 5 universe of discourse, U(x 5 ) = [0.1 -1.0] • Input variable 6, x 6 = Beginning point of pipeline relative to image center point Input variable 6 fuzzy term set, T(x 6 ) = {Left, Center, Right} Input variable 6 universe of discourse, U(x 6 ) = [0.1 -1.0] The only fuzzy output. • Output variable 1, y 1 = AUV steering command set point Output variable 1 fuzzy term set, T(y 1 ) = {Turn left, Go straight, Turn right} Output variable 1 universe of discourse, V(y 1 ) = [0 -180] The input vector, x is. Robotics Vision-Based Heuristic Reasoning for Underwater Target Tracking and Navigation 397 x = (x 1 , x 2 , x 3 , x 4 , x 5 , x 6 ) T (3) The output vector, y is. y = (y 1 ) T (4) Gaussian and Ǒ-shaped membership functions are selected in this case to map the input to the output. Gaussian curves depend on two parameters ǔ and c and are represented by. 2 () (; ,) exp 2 2 xc fx c σ σ −− = ªº «» ¬¼ (5) Ǒ-shaped membership function are represented by. ( ; , / 2, ) (;,) 1(;, /2, ) Sxcbcb c forx c fxbc s xccb cb forx c −− ≤ = −++ >  ® ¯ (6) where S(x; a, b, c) represents a membership function defined as. 0 2 2( ) 2 () (; , ,) 2 2( ) 1 2 () 1 for x a xa f or a x b ca S xabc xa f or b x c ca for x c < − ≤< − = − −≤≤ − >  ° ° ° ° ® ° ° ° ° ¯ (7) In the above equation, ǔ, a, b and c are the parameters that are adjusted to fit the desired membership data. Typical input variable and output variable membership function plot are shown in Fig. 6 and Fig. 7. Fig. 5. Illustration of tracking strategy. 398 Mobile Robots, Towards New Applications Fig. 6. Typical input variable membership function plot. Fig. 7. Typical output variable membership function plot. There are totally 13 fuzzy control rules. The rule base as shown in Fig. 8. Fig. 8 Rule viewer for fuzzy controller. In order to obtain a crisp output, the output fuzzy set is then aggregated and fed into a centroid (center of gravity) method defuzzification process. The defuzzifier determines the actual actuating signal, 'y as follows. () () ¦ ¦ = = = 13 1 13 1 ' i iB iB i i y yy y μ μ (8) Robotics Vision-Based Heuristic Reasoning for Underwater Target Tracking and Navigation 399 3. Simulation and Experimental Results The simulation procedures are as follows: a. Define the envelope curve (working area) of prototype. b. Give the real position and orientation of pipeline defined on a grid of coordinates. c. Predefine the AUV drift tolerance limit (±8.0cm) away from the actual pipeline location. d. Initiate the algorithm. e. AUV navigating paths are recorded and visualized graphically. The algorithm has been tested on computer and prototype simulations. For comparative pur poses, the results before and after fuzzy tuning are presented. Typical examples of results before fuzzy tuning are shown in Fig. 9 and Table 1. Fig. 9. AUV path (no proper tuning). AUV path Actual location x-axis (cm) Simulated result x-axis (cm) Drift (cm) Percentage of Drift (%) 5 47.5 69.5 +22.0 275.0 4 58.5 71.7 +13.2 165.0 3 69.6 73.3 +3.7 46.3 2 80.8 78.3 -2.5 31.3 1 91.9 75.7 -16.2 202.5 Table 1. Data recorded (without proper tuning). 400 Mobile Robots, Towards New Applications Typical examples of results after fuzzy tuning are shown in Fig. 10 and Table 2. Fig. 10. AUV path (with proper tuning). AUV path Actual location x-axis (cm) Simulated result x-axis (cm) Drift (cm) Percentage of Drift (%) 5 47.5 55.2 +7.7 96.3 4 58.5 57.4 -1.1 13.8 3 69.6 68.5 -1.1 13.8 2 80.8 88.1 +7.3 91.3 1 91.9 85.9 -6.0 75.0 Table 2. Data recorded (with proper tuning). The simulation results show that the drift within tolerance limit is achievable when proper tuning (training) is applied to the fuzzy system. The percentage of drift is considered acceptable , as long as it is less than 100%, since this implies the path is within the boundary. The effectiveness of the system has been further demonstrated with different target orientation and lighting conditions. 4. Conclusions This paper introduces a new technique for AUV target tracking and navigation. The image processing algorithm developed is capable of extracting qualitative information of the terrain required by human operators to maneuver ROV for pipeline tracking. It is interesting to note [...]... drive-axis and belts, the position encoders will be positioned at the driven part as originally planed The next generation of our C-arm will integrate the experiences we have obtained and allow for further evaluation 408 Mobile Robots, Towards New Applications 5 Kinematics Now that the C-arm joints can be moved motor driven, some of the applications such as weight compensation can already be realized But... calculated from d and l using the theorem of Pythagoras Now the value of (13) 2 can be derived a sin a tan 2 d l (14) px py (15) sign 4 (16) 412 Mobile Robots, Towards New Applications z0 mz i l z0 x0 Fig 11 Special case: 5 m3 d45 Oz 4 d d3 y0 m3 d 90 6 Applications When this project started, the major goal was to simplify the handling of the C-arm The user should be able to reach the intended target... Intraoperative soft tissue 3D reconstruction with a mobile C-arm International Congress Series, 2003, 1256:200-6, doi:10.1016/S05315131(03)00243-7 Siegert, HJ and Bocionek, S., Robotik: Programmierung intelligenter Roboter SpringerVerlag, 1996 418 Mobile Robots, Towards New Applications Toft, P., “The Radon transform - Theory and Implementation,” Ph.D Thesis, Department of Mathematical Modelling, Section... was defined by the difference between the result of the input face and mean face If this system rejects the result, it is replaced by the corresponding facial parts of the mean face and fitted it as the facial parts 424 Mobile Robots, Towards New Applications 4 Demonstrations in EXPO2005 4.1 352 facial data We demonstrated successfully that COOPER system could be exhibited at the Prototype Robot Exhibition... We, therefore, have to develop such new module that could detect the blinking automatically (2) ROI extraction Facial parts of ROI couldn’t be occasionally extracted at the proper location We must try to improve the method for extracting ROI involving the major part of the respective facial parts at least (3) Contour extraction of facial parts The contour of facial parts couldn't be frequently detected... vectors are defined: m3 z Part of m3 perpendicular to m z mz 3 Part of m z perpendicular to m3 m3 mz O3 3 z O5 The Surgeon's Third Hand - An Interactive Robotic C-Arm Fluoroscope 411 Now starting at O5, O4 and O3 can be calculated: O4 O5 a 5 m3 z (9) O3 O4 a4 m z 3 (10) The armlength is then: O3 x 2 d3 O3 y 2 (11) As defined at beginning of this section, 3 gv , z gv , and all their parts are functions of... both requirements: increase precision and reduce the duration of an intervention Fig 1 Robotized C-arm 404 Mobile Robots, Towards New Applications We want to introduce another type of robot which assists the surgeon by simplifying the handling of everyday OR equipment Main goal is to integrate new features such as enhanced positioning modes or guided imaging while keeping the familiar means of operation... radiograph If an image is not centered properly and the ROI is on the edge of the radiograph, important information might be cut off Fig 12 Selecting the new center on an misplaced radiograph for automatic repositioning 414 Mobile Robots, Towards New Applications Nevertheless, the picture can still be used to obtain a corrected image After marking the desired center in the old image, the robotic C-arm... Although the proof of concept study of the mechanical version is completed now, the simulation is still a valuable device for testing new concepts and ideas Algorithms for collision detection and path planning are now developed on this base 406 Mobile Robots, Towards New Applications Fig 3 Screenshot of the simulation GUI Fig 4 Comparison of real radiograph and simulation for a hip joint 4 Realisation... future of computer facial caricaturing As a result of 11 days demonstration, COOPER generated 352 caricatures in total and presented a shrimp rice cracker to the respective visitor This system generated successful output from the standpoint of wide distribution of visitors in generation, race and sexuality at EXPO site 420 Mobile Robots, Towards New Applications In section 2, the outline of COOPER system . underwater pipeline from seabed, and to see individual parts of pipeline. A machine vision and image processing 394 Mobile Robots, Towards New Applications system capable of extracting and classifying. expected to bring the system into practical use. Fig. 11. AUV path and its image capturing procedure. 402 Mobile Robots, Towards New Applications 5. References Arjuna Balasuriya & Ura,. valuable device for testing new concepts and ideas. Algorithms for collision detection and path planning are now developed on this base. 406 Mobile Robots, Towards New Applications Fig. 3. Screenshot

Ngày đăng: 11/08/2014, 16:22

Nguồn tham khảo

Tài liệu tham khảo Loại Chi tiết
[1] Robot Project: Prototype Robot Exhibition: http://www.expo2005.or.jp/jp/C0/C3/C3.8/C3.8.2/C3.8.2.6 Link
[3] Caricaturing robot, EXPO'85, Matsushita pavilion: http://www.mid.co.jp/mid/shirahama/nigaoe.html Link
[2] Koshimizu, H. (1997) Computer Facial Caricaturing, Trans. The Institute of Image Information and Television Engineers, Vol.51, No.8, pp. 1140-1146 Khác
[4] Takigawa, E. Kishiba, H. and Kawade, M. (2002) @face Automatic Gender and Age Estimation with Face, Proc. of the Conf. on Japanese Academy of Facial Studies 2002, p171, 1346-8081 Khác
[5] Teranishi, K. Kotani, N. and Shinya, M. (2000) Chara-Face: Aportrait Caricaturing System, Proc. of General Conf. on IEICE, A-14-5 Khác