Safe Reinforcement Learning for Real-World Engine Control

21 Pages Posted: 24 Feb 2025

See all articles by Julian Bedei

Julian Bedei

RWTH Aachen University

Lucas Koch

RWTH Aachen University

Kevin Badalian

RWTH Aachen University

Alexander Winkler

RWTH Aachen University

Patrick Schaber

RWTH Aachen University

Jakob Andert

RWTH Aachen University

Abstract

This work introduces a toolchain for applying Reinforcement Learning (RL), specifically the Deep Deterministic Policy Gradient (DDPG) algorithm, in safety-critical real-world environments. As an exemplary application, transient load control is demonstrated on a single-cylinder internal combustion engine testbench in Homogeneous Charge Compression Ignition (HCCI) mode, that offers high thermal efficiency and low emissions. However, HCCI poses challenges for traditional control methods due to its nonlinear, autoregressive, and stochastic nature.RL provides a viable solution, however, safety concerns – such as excessive pressure rise rates – must be addressed when applying to HCCI. A single unsuitable control input can severely damage the engine or cause misfiring and shut down. Additionally, operating limits are not known a priori and must be determined experimentally. To mitigate these risks, real-time safety monitoring based on the k-nearest neighbor algorithm is implemented, enabling safe interaction with the testbench.The feasibility of this approach is demonstrated as the RL agent learns a control policy through interaction with the testbench. A root mean square error of 0.1374 bar is achieved for the indicated mean effective pressure, comparable to neural network-based controllers from the literature. The toolchain’s flexibility is further demonstrated by adapting the agent’s policy to increase ethanol energy shares, promoting renewable fuel use while maintaining safety.This RL approach addresses the longstanding challenge of applying RL to safety-critical real-world environments. The developed toolchain, with its adaptability and safety mechanisms, paves the way for future applicability of RL in engine testbenches and other safety-critical settings.

Keywords: Reinforcement Learning, Deep Deterministic Policy Gradient, Safe Learning, transfer learning, Homogeneous Charge Compression Ignition, Renewable Fuels

Suggested Citation

Bedei, Julian and Koch, Lucas and Badalian, Kevin and Winkler, Alexander and Schaber, Patrick and Andert, Jakob, Safe Reinforcement Learning for Real-World Engine Control. Available at SSRN: https://ssrn.com/abstract=5152158 or http://dx.doi.org/10.2139/ssrn.5152158

Julian Bedei

RWTH Aachen University ( email )

Templergraben 55
52056 Aachen, 52056
Germany

Lucas Koch

RWTH Aachen University ( email )

Templergraben 55
52056 Aachen, 52056
Germany

Kevin Badalian

RWTH Aachen University ( email )

Templergraben 55
52056 Aachen, 52056
Germany

Alexander Winkler

RWTH Aachen University ( email )

Templergraben 55
52056 Aachen, 52056
Germany

Patrick Schaber

RWTH Aachen University ( email )

Templergraben 55
52056 Aachen, 52056
Germany

Jakob Andert (Contact Author)

RWTH Aachen University ( email )

Templergraben 55
52056 Aachen, 52056
Germany

Do you have a job opening that you would like to promote on SSRN?

Paper statistics

Downloads
15
Abstract Views
91
PlumX Metrics