PPO Reinforcement learning for smooth control of two-wheeled mobile

Authors

  • Thi Minh Tam Le Hung Yen University of Technology and Education
  • Duc-Hung Pham Hung Yen University of Technology and Education
  • Viet Ngu Nguyen Hung Yen University of Technology and Education

Keywords:

Two-wheeled mobile robot, Reinforcement learning, Trajectory tracking, Proximal Policy Optimization, Sliding mode control, Intelligent control

Abstract

This paper proposes a reinforcement learning (RL) approach to improve trajectory tracking in two-wheeled mobile robots, which are difficult to control due to nonlinear dynamics and nonholonomic constraints. Unlike traditional methods such as sliding mode control, the proposed strategy uses the proximal policy optimization (PPO) algorithm to map robot state position, orientation, and tracking error directly to velocity commands. The reward function encourages accuracy, smooth motion, and energy efficiency. Simulation results show that the RL controller matches the accuracy of a baseline sliding mode controller (SMC) while producing smoother inputs and avoiding chattering. It also generalizes well across various trajectories without retuning. This demonstrates RL as a robust, adaptive alternative to model-dependent methods, with future work aimed at hardware testing and hybrid RL-classical control designs.

Downloads

Download data is not yet available.

Downloads

Published

28-03-2026

How to Cite

Le, T. M. T., Pham, D.-H., & Nguyen, V. N. (2026). PPO Reinforcement learning for smooth control of two-wheeled mobile. Journal of Measurement, Control and Automation, 30(1), 57–65. Retrieved from https://mca-journal.org/index.php/mca/article/view/390

Issue

Section

Article

Similar Articles

1 2 3 4 5 6 7 8 9 10 > >> 

You may also start an advanced similarity search for this article.