Ephrem et al.
[16][67] adopted LSTM to train the lithium battery SOC model under fixed and varying ambient temperatures in the dataset
[6][46]. In the fixed ambient temperature SOC model, the training dataset included the data under eight mixed drive cycles, and the two discharge test cases were used as the validation dataset; the test dataset was the charging test case; in the varying ambient temperature SOC model, the training dataset with 27 drive cycles included three sets of nine drive cycles recorded at 0 °C, 10 °C, and 25 °C. The test dataset included the data of another mixed-drive cycle. Both models’ input variables are voltage, current, and temperature. After evaluation, the model achieved the lowest MAE of 0.573% at 10 °C and an MAE of 1.606% with ambient temperature from 10 to 25 °C. Cui et al.
[17][68] used LSTM with an encoder–decoder
[18][69] structure in the dataset
[19][43]; the input was “I
t, V
t, I
avg, V
avg”, and the test result was an RMSE of 0.56% and MAE of 0.46% in US06, which was higher than that using only LSTM and GRU in that paper. Wong et al.
[20][70] used the undisclosed ‘UNIBO Power-tools Dataset’ as a training dataset and dataset
[21][51] as a test dataset in the LSTM structure; the input variables were current, voltage, and temperature, and the MAE was 1.17% at 25 °C. Du et al.
[22][71] tested two LR1865SK Li-ion battery cells at room temperature and used the dataset in
[23][45] as the comparative case to test the model trained by LSTM; the input variables were current, voltage, temperature, cycles, energy, power, and time; the MAE was 0.872% at an average level. YANG et al.
[24][72] used the LSTM to build a model for lithium battery SOC estimation; the data were obtained from the A123 18560 lithium battery under three drive cycles, i.e., DST, US06, and FUDS; the input vectors were current, voltage, and temperature. In addition, the model robustness was tested in the unknown initial state of the lithium battery, with the Unscented Kalman Filter
[25][73] (UKF) method for comparison; the test results showed that the RMS of LSTM was significantly smaller than that of UKF.
1.4. Recurrent Type—GRU
YANG et al.
[26][74] trained the model by using GRU, and the dataset was tested using three LiNiMnCoO2 batteries with DST and FUDS drive cycles; the input vectors were the current, voltage, and temperature. Then, the trained model was tested in a dataset of another material; it obtained 3.5% of max. RMS. The authors of studies
[27][28][29][75,76,77] all used GRU as the neural network for model training; the dataset was the INR 18650-20R and A123 18650 lithium battery from the CALCE dataset
[6][46] with inputs of voltage, current, and temperature, and the RMS error obtained from the test dataset was not significantly different. Kuo et al.
[30][78] tested a 18650 Li-ion battery cell and used GRU with an encoder–decoder structure, in which the input vectors were current, voltage, and temperature; further, they compared this with LSTM, GRU, and a sequence-to-sequence structure, and the result showed that the MAE of their proposed neural network was lower than that of other methods at three different drive cycles and temperatures.
2. Hybrid Structure
The main idea of the hybrid neural network in the estimation of the SOC of a lithium battery is to improve the prediction accuracy of the model by combining the advantages of various types of neural networks. The current common architecture in the lithium battery SOC estimation problem is a 1D-CNN as a feature extraction layer to extract deeper features of the input data, and a recurrent neural network (LSTM or GRU is used more often) as a model building layer to construct a model between the SOC and the input variables. Some scholars also added the fully connected layer (FC) before the final output layer to improve the accuracy of the model. The architecture of 1D-CNN + X + Y in lithium battery SOC estimation is depicted in
Figure 47.
Figure 47.
1D-CNN + X + Y architecture diagram.
2.1. 1D-CNN + LSTM
SONG et al.
[31][79] used a neural network combination of “1D-CNN + LSTM” to build a model with inputs of voltage, current, temperature, average voltage, and average current, for the dataset, and the 1.1 Ah A123 18650 lithium battery was tested at seven different temperatures with drive cycles of US06, FUDS. The results showed that the error of the “1D-CNN + LSTM” method was significantly smaller than that of the method that only used one neural network when tested in the test dataset and compared with the 1D-CNN and LSTM methods.
2.2. 1D-CNN + GRU + FC
HUANG et al.
[32][80] used a “1D-CNN + GRU + FC” neural network architecture with inputs of voltage, current, and temperature; the dataset was obtained from the BAK 18650 lithium battery at seven different temperatures with drive cycles of DST and FUDS. Compared with the method of one neural network such as RNN, GRU, and a support vector machine, it achieved the lowest RMS.
2.3. NN + Filter Algorithm
The NN + filter algorithm type uses a neural network and filter algorithm for improving Li-ion SOC estimation performance,
Figure 58 is a case of that structure, which is the combination of LSTM and the adaptive H-infinity filter that can be found in
[33][81] in more detail.
Figure 58. A case of the NN + filter algorithm architecture diagram (LSTM + AHIF, reprinted with permission from Ref. [33]. Copyright 2021, Elsevier.). A case of the NN + filter algorithm architecture diagram (LSTM + AHIF, reprinted with permission from Ref. [81]. Copyright 2021, Elsevier.).
YANG et al.
[34][82] tried to combine the advantages of both LSTM and UKF. They used LSTM and an offline training neural network to obtain a pre-trained model with the data obtained; then, the real-time data obtained were inputted into UKF and the pre-trained model, whose data input occurred after normalization. The UKF filters out the noise and improves the model performance. After this, combinations of LSTM and filtering class algorithms appear as “LSTM + CKF (Cubature Kalman Filter)”
[35][83], “LSTM + EKF (Extended Kalman Filter)”
[36][84], and “LSTM + AHIF (Adaptive H-infinity Filter)
[33][81], through the test dataset, and their model performance was better than the models only trained by LSTM.
3. Trans Structure
Trans structure is mainly used to transfer the knowledge of source data to target data and in this chapter includes the section on transfer learning and transformers.
3.1. Transfer Learning
As depicted in
Figure 69, the knowledge is utilized from the learning task trained by source data and that of the target data, which can improve the robustness of the model to achieve higher performance. Some researchers applied transfer learning to enhance the performance of SOC estimation.
Figure 69.
Schematic of transfer learning.
Bian et al.
[37][85] added a fully connected layer after bidirectional LSTM on this basis with inputs of voltage, current, and temperature; the datasets were three different lithium battery datasets, A123 18650, INR 18650-20R from the CALCE dataset
[6][46] as the target dataset, and the Panasonic lithium battery dataset
[3][49] as the pre-trained dataset. Then, they used transfer learning to transfer features from the model trained with the pre-trained dataset to the model trained with the target dataset. Compared with the method of one neural network such as RNN, LSTM, and GRU, the model of the transfer learning method achieved the lowest RMS.
Liu et al.
[38][86] applied TCN to two different types of lithium battery data and migrated the trained model for lithium battery SOC estimation as a pre-trained model to another battery dataset by transfer learning
[39][87]. The training dataset of the pre-trained model was “US06, HWFET, UDDS, LA92, Cycle NN”, corresponding to 25 °C, 10 °C, and 0 °C in the dataset
[3][49], and the test set was “Cycle 1–Cycle 4”; the input vectors were current, voltage, and temperature. The model trained under 25 °C was migrated to the new lithium battery SOC model as a pre-trained model by transfer learning, the training dataset of the new lithium battery SOC model included the data measured under two mixed driving cycles in the dataset
[40][50], the test dataset was “US06, HWFET, UDDS, LA92” in the dataset
[40][50], and its RMS range was 0.36–1.02%.
3.2. Transformer
Transformer
[41][88] is based on the encoder–decoder structure and attention mechanism, which is multi-head attention. It can enhance the connection and relation of data, and hence the transformer is applied in the natural language process, image detection, and segmentation, etc. In recent years, some scholars tried to use the structure based on the transformer for SOC estimation. The diagram of the transformer is shown in
Figure 710.
Figure 710.
Structure of the transformer.
Hannan et al.
[42][89] used the structure based on the encoder of the transformer
[41][88] to estimate SOC, and the dataset was used in
[21][51]; the input variables were current, voltage, and temperature, and compared with different methods including DNN, LSTM, GRU, and other deep learning methods, the test performance was 1.19% for RMSE and 0.65% for MAE.
Shen et al.
[43][90] used two encoders and one decoder of the transformer, in which the input variables were the current–temperature and voltage–temperature sequences; the dataset was obtained from
[6][46], in which the ‘DST’ and ‘FUDS’ were used as the training dataset, and the ‘US06′ was used as test dataset. Further, they added a closed loop to improve the performance of SOC estimation; then, compared with LSTM and LSTM + UKF, the test results showed that the RMSE of their proposed method was lower than that of other methods.