In order to develop the steering control for collision avoidance behaviour, robot must be able to determine its heading orientation with respect to environment. Orientation can be measured by dedicated sensors or through visual features perception. In vision-based orientation estimation problem, most of approaches are making use of a matching process between pair of frames. This paper proposes a method of estimating robot's heading orientation by using only a single-frame of fish-eye image. CIE-LAB colour space is applied to handle colour and illumination intensity change. Straight line segments are extracted from thresholded CIE-LAB image take advantage of Progressive Probabilistic Hough Transform. Angle of the corresponding line segment is measured using combination of Law of Cosines and quadrant principle. Heading orientation in yaw angle is estimated by implementing voting mechanism based on region grouping and length of perpendicular line. Some experiments are made in robot soccer field environment to compare orientation estimation system against IMU's measurement. Discussion about the performance and limitation of the system are included in this paper.
|Journal of Physics: Conference Series
|Published - 23 Jul 2020
|3rd International Conference on Science and Technology 2019, ICST 2019 - Surabaya, Indonesia
Duration: 17 Oct 2019 → 18 Oct 2019