TY - JOUR
T1 - A Deep Reinforcement Learning Framework for Fast Charging of Li-Ion Batteries
AU - Park, Saehong
AU - Pozzi, Andrea
AU - Whitmeyer, Michael
AU - Perez, Hector
AU - Kandel, Aaron
AU - Kim, Geumbee
AU - Choi, Yohwan
AU - Joe, Won Tae
AU - Raimondo, Davide M.
AU - Moura, Scott
PY - 2022
Y1 - 2022
N2 - One of the most crucial challenges faced by the Li-ion battery community concerns the search for the minimum time charging without damaging the cells. This goal can be achieved by solving a large-scale constrained optimal control problem, which relies on accurate electrochemical models. However, these models are limited by their high computational cost, as well as identifiability and observability issues. As an alternative, simple output-feedback algorithms can be employed, but their performance strictly depends on trial and error tuning. Moreover, particular techniques have to be adopted to handle safety constraints. With the aim of overcoming these limitations, we propose an optimal-charging procedure based on deep reinforcement learning. In particular, we focus on a policy gradient method to cope with continuous sets of states and actions. First, we assume full state measurements from the Doyle-Fuller-Newman (DFN) model, which is projected to a lower dimensional feature space via the principal component analysis. Subsequently, this assumption is removed, and only output measurements are considered as the agent observations. Finally, we show the adaptability of the proposed policy to changes in the environment's parameters. The results are compared with other methodologies presented in the literature, such as the reference governor and the proportional-integral-derivative approach.
AB - One of the most crucial challenges faced by the Li-ion battery community concerns the search for the minimum time charging without damaging the cells. This goal can be achieved by solving a large-scale constrained optimal control problem, which relies on accurate electrochemical models. However, these models are limited by their high computational cost, as well as identifiability and observability issues. As an alternative, simple output-feedback algorithms can be employed, but their performance strictly depends on trial and error tuning. Moreover, particular techniques have to be adopted to handle safety constraints. With the aim of overcoming these limitations, we propose an optimal-charging procedure based on deep reinforcement learning. In particular, we focus on a policy gradient method to cope with continuous sets of states and actions. First, we assume full state measurements from the Doyle-Fuller-Newman (DFN) model, which is projected to a lower dimensional feature space via the principal component analysis. Subsequently, this assumption is removed, and only output measurements are considered as the agent observations. Finally, we show the adaptability of the proposed policy to changes in the environment's parameters. The results are compared with other methodologies presented in the literature, such as the reference governor and the proportional-integral-derivative approach.
KW - Actorcritic
KW - approximate dynamic programming (ADP)
KW - electrochemical model (EM)
KW - fast charging
KW - reinforcement learning (RL)
KW - Actorcritic
KW - approximate dynamic programming (ADP)
KW - electrochemical model (EM)
KW - fast charging
KW - reinforcement learning (RL)
UR - http://hdl.handle.net/10807/214127
U2 - 10.1109/TTE.2022.3140316
DO - 10.1109/TTE.2022.3140316
M3 - Article
SN - 2332-7782
VL - 8
SP - 2770
EP - 2784
JO - IEEE Transactions on Transportation Electrification
JF - IEEE Transactions on Transportation Electrification
ER -