Latest Issue
Empowering Education with Online Khmer Handwritten Text Recognition for Teaching and Learning Assistance
Published: August 30,2025Undergraduate Student Dropout Prediction with Class Balancing Techniques
Published: August 30,2025Status of Seawater Quality at Koh Rong Island, Sihanoukville, Cambodia
Published: August 30,2025Low-Complexity Detection of Primary Synchronization Signal for 5G New Radio Terrestrial Cellular System
Published: August 30,2025Word Spotting on Khmer Printed Documents
Published: August 30,2025Tuning Hyperparameters Learning Rate and Gamma in Gym Environment Inverted Pendulum
Published: August 30,2025Examining Passenger Loyalty in Phnom Penh Public Bus System: A Structural Equation Modelling Approach
Published: August 30,2025Prediction on Load model for future load profile of Electric Vehicle charging demand in Phnom Penh
Published: August 30,2025Economic Study on Integrating PV-DG with Grid-Tie: Case Study in Cambodia
Published: August 30,2025Mobile Robot Localization using Extended Kalman Filter with Kinematic Model
-
1. Dynamics and Control Laboratory, Department of Industrial and Mechanical Engineering, Institute of Technology of Cambodia, Russian Federation Blvd., P.O. Box 86, Phnom Penh, Cambodia.
Academic Editor:
Received: July 19,2021 / Revised: / Accepted: November 19,2021 / Available online: December 30,2021
In mobile robotic field, one of the most important tasks in mobile robot navigation is Robot Localization. Localization is the task of determining the location of the robot inside the environment at a specific time step. To obtain the position and the orientation of the robot, various sensors have been utilized along with implementation of numerous kinds of algorithms. In this paper, the sensor fusion based on Extended Kalman Filter (EKF) algorithm is proposed. Differential drive wheeled mobile robot kinematic model is derived. The robot is equipped with an inertial measurement unit (IMU), wheel encoders, and a light detecting and ranging sensor (Lidar). The noise of the sensor data is assumed to be gaussian white noise. In the prediction step, the robot's linear velocity and angular velocity are determined by using the information from wheel encoders and subsequently is used to compute the robot pose. In the correction step, the robot pose is updated from the information from IMU and Scan Matching Lidar. To control the robot movement, a trajectory control algorithm based on a backstepping controller is used. The robot is controlled to move in two trajectories: circular and “8” shape. The robot position and orientation are represented in two dimensional cartesian coordinate system. The numerical experiment is conducted inside the simulation software Gazebo and Robotic Operating System (ROS) framework. Finally, the numerical experiment results showed that the sensor fusion algorithm is effective in estimating the robot pose against the desired trajectory.