Energies 16 08102
Energies 16 08102
Article
Optimization of Electric Vehicles Charging Scheduling Based on
Deep Reinforcement Learning: A Decentralized Approach
Imen Azzouz 1 and Wiem Fekih Hassen 2, *
1 Higher School of Communication of Tunis (Sup’Com), University of Carthage, 2083 Ariana, Tunisia;
[email protected]
2 Chair of Distributed Information Systems, University of Passau, Innstraße 41, 94032 Passau, Germany
* Correspondence: [email protected]
Abstract: The worldwide adoption of Electric Vehicles (EVs) has embraced promising advancements
toward a sustainable transportation system. However, the effective charging scheduling of EVs is not
a trivial task due to the increase in the load demand in the Charging Stations (CSs) and the fluctuation
of electricity prices. Moreover, other issues that raise concern among EV drivers are the long waiting
time and the inability to charge the battery to the desired State of Charge (SOC). In order to alleviate
the range of anxiety of users, we perform a Deep Reinforcement Learning (DRL) approach that
provides the optimal charging time slots for EV based on the Photovoltaic power prices, the current
EV SOC, the charging connector type, and the history of load demand profiles collected in different
locations. Our implemented approach maximizes the EV profit while giving a margin of liberty to the
EV drivers to select the preferred CS and the best charging time (i.e., morning, afternoon, evening, or
night). The results analysis proves the effectiveness of the DRL model in minimizing the charging
costs of the EV up to 60%, providing a full charging experience to the EV with a lower waiting time
of less than or equal to 30 min.
Keywords: smart EV charging; day-ahead planning; deep Q-Network; data-driven approach; waiting
time; cost minimization; real dataset
the optimization of various EV charging problems. The authors in [4,5] showed the effec-
tiveness of DRL in optimizing the charging duration and the cost. The authors suggested
a model-free method that can handle charging problems like the uncertainty of arrival
time and charging price fluctuation, which cannot be considered using traditional methods’
optimization approaches. The proposed approach [4] can potentially result in high demand
during charging at lower prices. A smart scheduling solution based on DRL [6] developed
a solution to plan long trips for EVs using generated scenarios. This approach depicted the
scarcity of the real data that may lead to incorrect decisions, showcasing the importance of
considering real-world data in the optimization scheduling problems.
A model-free reinforcement learning algorithm called Q-learning was proposed in [7]
to optimize the charging cost for Vehicle-to-Grid (V2G) and Grid-to-Vehicle (G2V) charging
behavior. Another use of Q-learning is the routing plan, presenting an energy-efficient solu-
tion for the EVs online. This approach assumed the model limitations in terms of handling a
high-dimensional Q-table, which is addressed by employing Deep Q-Learning (DQL) meth-
ods. The authors in [8] propose a smart scheduling solution to coordinate the EV charging
control of the voltage in the grid regarding the uncertainties of the charging process and
load demand. Another model-free reinforcement learning is the Deep Deterministic Policy
Gradient (DDPG), which has proven its prominence in solving the charging scheduling task
specifically in continuous action–space environments. A real-time scheduling strategy was
proposed based on the price signals to tackle the problem of a substantial number of EVs
powered by the distribution network. In [9], the authors use Multi-Modal Approximate
Dynamic Programming to optimize the scheduling of charging and discharging EVs while
considering a single objective, which is the pricing schema [10]. The literature review
indicated a remarkable focus on DRL techniques to address the charging scheduling prob-
lem. The authors in [6] proposed a DRL approach based on the Deep Q-Network (DQN)
model that recommends the optimal charging station at the beginning of the trip based
on the following features: State of Charge (SOC), trip distance, average speed, charging
prices, and energy demand. However, this model presents some limitations, which are the
long training time and the used dataset, which was not collected for real-world use cases.
The challenge of the determination of the optimal charging scheduling-based model-free
DRL model is tackled in previous research while considering the randomness of pricing
and user preferences [11].
The majority of previous studies have used the DRL to solve different scheduling
problems, and they achieved good results [8,9]. But the developed models present some
shortcomings, such as the lack of findings evaluation and the use of simulated data that are
far from the truth.
The overall goal of this work consists of the day-ahead prediction of the best EV
charging time slots, with a focus on reducing the energy cost and the waiting time inside
the Charging Station (CS), as well as guaranteeing a fast and full charging of EV batteries.
Additionally, the EV’s driver preferences for the charging location, the charging connector
type, and the desired charging part of the day are taken into consideration in the charging
scheduling problem.
The main contributions of this paper are as follows:
1. The prior study maintains a major focus on a single objective, especially reducing the
energy cost [9], whereas our proposed method tackles two main objectives simultane-
ously: (1) the reduction in the charging cost and (2) the minimization of the waiting
time, while giving a margin of flexibility to the EV to select a preferred location and a
time of the day (i.e., morning, afternoon, evening, and night).
2. Our model proposes a flexible choice regarding the number and the weight of
constraints, cost, power, and load profiles.
3. Some surveys used a multi-agent RL algorithm which does not imply a convergence
toward the optimal policy, while our approach relies on one agent that is capable of
resolving multiple objectives at the same time.
4. We have developed a new algorithm to evaluate our proposed model.
Energies 2023, 16, 8102 3 of 18
5. Our implemented model stands out by involving individual user preference in the
charging process rather than previous work that significantly focuses on the CS profit.
6. Our novel approach is based on real data rather than simulated scenarios which
completely differ from real-world cases.
The remainder of this paper is organized as follows: Section 2 is devoted to intro-
ducing the system’s overall architecture, specifying the adopted variables and constraints,
and detailing the mathematical formulation of the scheduling problem using the Markov
Decision Process (MDP) framework. Section 3 describes the proposed solution based on
the deep Q-Network (DQN) algorithm. Section 4 showcases the training phase and the
simulation results. The last section is dedicated to conclusions and future works.
SOC,
Connector
PV
Recommanded
DRL Agent
timeslots
Grid
Power flow
Charging stations Electric vehicles
Information flow
Our developed model deviates from being considered provisional distributed storage,
and it explicitly focuses on the Grid-to-Vehicle (G2V) direction. We have made this decision
since the dataset we use does not contain key features such as the EV battery size and SOC.
We implemented a DQN agent that makes optimal decisions on the basis of the ob-
served charging environment that comprises the history of charging prices, the charging
load demand, and the amount of power allocated to the EV battery. Additionally, we con-
sider the amount of power delivered to the EV battery for an efficient charging experience
Energies 2023, 16, 8102 4 of 18
as well as the specific charging connector type of the EV. Our proposed architecture is
illustrated in Figure 1. We have opted for a decentralized optimized charging scheduling
approach, i.e., the recommendation of the optimal charging time slots to all vehicles; hence,
each decision is taken separately considering the individual preferences of each EV. We
mean by the decentralized approach that the decision making regarding the optimum
charging time slots for all vehicles is made separately, taking into consideration each EV
driver’s preferences. In our system, the actor is the EV, i.e., if we implement the agent in an
EV, it will make an individual decision based on the current SOC of each EV and the type
of charging connector. The role of the CS is to provide data to train the DQN model. In our
proposed architecture, we have decided on one agent able to solve multiple goals at the
same time based on the designed reward function.
where SOCmin is the minimum SOC and SOCmax is the maximum SOC.
We adopt an SOCmin of 20%. Below this threshold, the damage could occur and the
lifespan of the battery could be minimized. Moreover, stopping the charging at 80%
(SOCmax ) is recommended rather than achieving 100% of SOC, since full charging can
result in high voltage and heat that may deteriorate the EV’s battery material.
2. Charging time: It is the major constraint regarding the charging scheduling problem.
It is important to take into account the preferences of an EV driver concerning the
charging time. For instance, if the EV driver wants to charge in the morning, a restric-
tion on available charging times is considered to ensure convenient charging in the
preferred time slots.
3. Power Capacity: It is an important factor that affects the charging planning. Given
the user’s preference to fully charge the battery of his vehicle, it is imperative to meet
the user’s expectations and ensure that the battery reaches its maximum. Based on
the load demand, the availability of the charging station can provide an idea for the
EV driver about the waiting time and help him improve his charging experience.
the waiting time. Thus, considering the variation of the charging prices, the load demand
during different parts of the day, and the EV patterns (e.g., the SOC and the charging time)
is crucial to meet the system’s goals. To address these complexities, we have adopted the
mathematical formulation of the EV charging scheduling problem based on an MDP with a
5-tuple (S, A, P, R, γ), where the following apply:
• S represents the set of states.
• A is a finite set of all possible actions.
• P is the state transition probability where P: S × A × S → [0, 1].
• R denotes the reward signal for a given state and action.
• The discount factor γ represents the importance of future rewards compared to imme-
diate rewards.
2.3.1. State
The state is defined as a set of relevant information that the agent needs to make a
decision. It is the environmental variables that affect directly the charging process of the
EV. For each time step t, the state st is given as follows:
st = (SOCt , Ct , loc, t arr , pday , prk : prk+6 , pok : pok+6 , dk : dk+6 ) (2)
2.3.2. Action
The action refers to the agent’s decision regarding a given state that the agent tries
to optimize throughout the learning process. We have divided the day A into one-hour
intervals ({ T1 , T2 , . . . , T24 }) as follows:
A = { T1 , T2 , . . . , T24 } (3)
2.3.3. Reward
In the environment simulation, at a time step t, a reward signal rt is sent to the agent
after taking an action at in a state st to improve the evaluation of the agent decision process.
In the charging scheduling environment, all the rewards are formulated as penalties.
The design reward function involves multiple stakeholders from the points of view of the
EV. The reward rt is the sum of the charging cost reward rcost , the charging power reward
r power and the reward of charging at high demand rload and is outlined by Equation (4).
The various parts of the reward function are expressed in the following:
1. The charging cost reward rcost
Energies 2023, 16, 8102 6 of 18
We design the charging cost reward in the form of a penalty multiplying the charging
cost (kWh/€) by the charging price. rcost is given by the following Equation:
po at
SOCt = SOCt−1 + · 100 (7)
Ca
rload = −λ · d at (8)
given a state. We opted for the DQN algorithm for many reasons. First, DQN is intended
to be a discrete action space precisely tailored to our scenario (i.e., each action represents a
one-time slot). Second, the DQN algorithm converges faster compared to other algorithms;
for example, the Soft Actor–Critic (SAC) algorithm, which is computationally much more
expensive in training time and resources. Moreover, the DQN algorithm relies on the
experience replay, which plays a major role in optimizing the decision process and hence
leads to a faster convergence of the model. In essence, the combined advantages of the
DQN including the fast convergence and the adaptability to discrete action spaces make it
a compelling choice to solve our problem.
It overcomes the limitation of the state space dimensionality in Q-learning by incorpo-
rating Neural Networks (NNs) as a function approximation to calculate the optimal Q-value
or the action-value function Qπ (s, a), which refers to the expected return for performing an
action at in a state st following a policy π. It represents how good it is to take an action in a
particular state. The formula is presented in Equation (9):
" #
L
Qπ (s, a) = Eπ [ Gt | st = s, at = a] = Eπ ∑ γk Rt+k+1 | st = s, at = a (9)
k =0
Qπ (s, a) is equal to the sum of the discounted cumulative reward Gt and L denotes
the maximum of the time step. During the training of the DQN, the Q-value is updated
according to the Bellman optimality equation expressed as:
Figure 3. Average load demand per time slot across different locations.
We generated 2000 test scenarios where in each testing scenario, we compared the
agent’s chosen time slots to the human decision based on the observed state. We made use
of the success rate, which is calculated by dividing the number of times the agent correctly
chooses the desired action in each scenario by the total number of testing use cases.
Energies 2023, 16, 8102 11 of 18
The result of the algorithm is a signification success rate percentage, which is about
80% of true decisions. The achieved success rate not only demonstrates the agent’s ability
to generalize the learned policy but also a meaningful proficiency if applied to real-world
scenarios, since the trained data (experiences) are based on the dataset.
To sum up, over 2000 test scenarios, we achieved that in 80% of cases, the agent
successfully chooses the optimal time slots.
The coefficients of the reward function were set to α = 0.6, β = 0.2 and λ = 0.2.
The choice was based on prioritizing the optimization of the charging expenses. The training
parameters are summarized in Table 1. The training phase takes about 5 h, which is quite
fast, since we work on a powerful machine. In some cases, the running time of a DRL
algorithm is not a critical factor, which is our case, since we run on a powerful machine
instance with a GPU of 200 GB.
4.6. Results
This section is depicted to evaluate the agent’s performance in providing the opti-
mal time slots to charge the EV. The assessment turns around the main objective of the
model: (1) minimize the waiting time, (2) deliver maximum power to the EV battery, and
(3) optimize the charging expenses.
• Waiting time minimization
First, we calculated the average waiting time for each location based on the history of
the load demand dataset.
We opted for the M/M/1 queue, which is a stochastic process based on two assump-
tions [17]: (1) Markovian Arrival Process and (2) Markovian Service Times.
1. M: Markovian Arrival Process: In the EV scheduling context, M signifies that the
arrival time of the EVs at the CS follows a Poisson process defined by:
λn
f (n ) = e−λ (11)
n!
where λ is the average arrival rate per hour and n is the occurrence of the charging
event. Figure 6 depicts an example of the arrival time per hour. Based on this graph,
we can conclude that the EV arrival times at the CS are independent.
2. Markovian Service Times: There is an assumption here that the charging duration
follows an exponential distribution, which is confirmed in our dataset for the five
locations (see Figure 7).
The waiting time calculation process based on the M/M/1 pursues the following steps:
(a) First, we calculate the average charging duration of hourly charging events
denoted µ, which is defined as the average service rate.
(b) Second, we calculate the EV arrival rate λ.
(c) Then, we introduce the traffic intensity defined as ρ = λµ .
(d) Finally, we obtain the waiting time for each location and for each time of the
ρ2
day denoted q where q = 1− ρ · µ1 .
Energies 2023, 16, 8102 13 of 18
To assess the agent performance regarding the recommendation of the optimal time
slots, we conducted a benchmark between the hourly waiting time calculated using the
M/M/1 queuing model and the average waiting time of the recommended time slots by
our DQN agent. The M/M/1 queuing model presents a reference theory to calculate the
waiting time. We supposed that each location is an independent CS that considers the
EVs arrival rate and traffic intensity for each time slot. Regarding the computation of the
average waiting time, we calculated the agent’s performance over the average waiting time
based on the load demand dataset for each location and for each specific time slot using
the following formula:
∑ Charging duration
Waiting time = (12)
Number of EV
By considering this evaluation process, we gain valuable insights about the application
and potential performance of our proposed method in real-world scenarios.
We conducted the described method on 10,000 scenarios to effectively evaluate the
agent performance. The graphs in Figure 8 outline the agent’s recommended time slots
demonstrating lower waiting time compared to the baseline model for the majority of
the locations and during different times of the day. In the highway, city and countryside
locations, a significant optimization in the waiting time is observed, especially at high
demand hours between 10 a.m. and 8 p.m., compared to the baseline model M/M/1
queuing model. This disparity highlights the effectiveness of our proposed approach in
bringing significant benefits to both rural and urban citizens as well as commuters. Further
detailed analyses are summarized in Table 2.
We computed the average waiting time for the model-recommended time slots, and
we benchmarked this with the average waiting time using the M/M/1 model. Notably,
a significant reduction in the waiting time emerged along the highway with an average of
55 min and in the city with an average of 1 h of waiting time. Whereas outside the city and
in the countryside, the waiting time is reduced by some minutes. These results showcase a
significant reduction in the average waiting time, contributing to an enhanced charging
experience for the EV driver.
Energies 2023, 16, 8102 14 of 18
Location Waiting Time DQN Model (min) Waiting Time M/M/1 Model (min)
Highway 31.03 86.08
In the City 26.01 86.11
Countryside 29.29 62.93
Outside the City 26.74 29.80
Near the Highway 19.77 29.80
Energies 2023, 16, 8102 15 of 18
Based on the charging schedule results illustrating the average charging cost profit
using the proposed charging scheduling method, we calculated the daily average cost profit
presented in Table 3, while comparing the agent’s recommended time slot cost with the
daily average charging cost data. We notice a significant reduction in the average charging
time ranging from 60% in the Near the Highway location to 6% in the Highway location for
the red initial EV SOC. This charging cost reduction is favorable for the EV driver if she/he
follows the recommendation and opts for charging at lower prices. It is worth noting that
the agent makes decisions based on the historical data of PV charging prices. To assess the
performance of our proposed model, we have considered different key evaluation metrics.
In contrast, most of the previous studies have assessed their work on the basis of one metric,
which is cost minimization [18,19]. Our approach achieved a daily average cost profit of
approximately 360 € annually. Previous research has resulted in 18.45% cost profit, while
our approach has led to 60% of the profit [19].
• Prioritize full charging
One of the primary targets of the scheduling problem is to provide the EV with a full
charging experience. We set random initial values of the SOC and observed the agent’s
behavior. We evaluate its performance to select the optimal time slot that guarantees full
charging for the EV, as described in Figure 9.
According to Figure 9, we analyzed the agent performance for different SOC values
during different hours of the day. The gained SOC refers to the SOC after testing the agent
online (given a single state). The initial SOC is set randomly. We notice that the agent
succeeds in selecting the optimal charging time slots that output a final SOC equal to 80%
most of the time.However, the graph showcases some instances where the SOC did not
reach its maximum and where the CS could not deliver power to the battery. Based on the
load demand dataset, the server could have stopped, meaning that no current data are sent
from the CS or the power provided to the EV is equal to 0.
Energies 2023, 16, 8102 16 of 18
5. Conclusions
As the industry endeavors to transition from fuel cars to electric ones, it is mandatory
to include the vehicle driver’s preferences and habits in the charging process. In this
context, we propose a model-free DRL solution that recommends the optimal time slots in
the context of the day-ahead charging planning for the EV based on the preferred location
and part of the day. We also consider the initial SOC and the connector type. To assess
that the decision taken in the scheduling process is close enough to the real-word usage,
we trained the model on real-world data that include the arrival times and the amount
of power delivered to the EV for different locations. Furthermore, the testing algorithm
showed that our model achieved a high level of success (80% of time) in recommending
the optimal time slots. The experimental analysis proves the effectiveness of our novel
approach in minimizing the charging costs up to 60%, reducing the waiting time in the
charging station by an average of 30 min compared to a baseline queuing model (M/M/1).
Finally, we validate the effectiveness of our method to provide a full charging experience to
the EV.
Added to the cost and waiting time optimization, our proposed mechanism proved
that we effectively recommend the optimal charging time slot for the EV owner that
guarantees a full charging level to the EVB, which is set to 80%. In fact, our proposed
model is capable of handling unseen use cases; for instance, if the initial SOC is less than
20%, it will still recommend the time slot that guarantees, as depicted in Figure 9, even
if it is trained on an SOC which is between 20% and 80%, showing its ability to handle
unseen charging states. Despite the tremendous results achieved by our novel approach, it
presents a few limitations. The precise location of the CS is not specified in our model, as
our proposed approach operates independently from any exact charging station location.
Additionally, the battery capacity parameter is set at 30 kWh, and any deviation from this
value would require retraining the model. These limitations, although present, do not
decrease the effectiveness of our proposed method in recommending the optimal charging
time slot for the EV.
We aim to consider the exact locations of the charging stations and the distance traveled
to reach the charging point to boost the reliability and efficiency of our solution in real-world
scenarios. Additionally, we aim to include an external factor that dynamically affects the
Energies 2023, 16, 8102 17 of 18
charging process such as the weather and traffic conditions. These enhancements provide
a more comprehensive insight into the charging process. Another future improvement
could be the incorporation of the V2G direction, since only the case of G2V is treated in this
proposed method.
Author Contributions: Conceptualization, I.A. and W.F.H.; methodology, I.A. and W.F.H.; software,
I.A. and W.F.H.; validation, I.A. and W.F.H.; formal analysis, I.A. and W.F.H.; investigation, I.A.
and W.F.H.; resources, I.A. and W.F.H.; data curation, I.A. and W.F.H.; writing—original draft
preparation, I.A. and W.F.H.; writing—review and editing, I.A. and W.F.H.; visualization, I.A. and
W.F.H.; supervision, W.F.H.; project administration, W.F.H.; funding acquisition, W.F.H. All authors
have read and agreed to the published version of the manuscript.
Funding: We acknowledge support by the Open Access Publication Fund of University Library
Passau. This research is also funded by the German Federal Ministry for Digital and Transport
with the Project OMEI (Open Mobility Elektro-Infrastruktur FK: 45KI10A011) https://omei.bayern
(accessed on 5 November 2023).
Data Availability Statement: Data are contained within the article.
Conflicts of Interest: The authors declare no conflicts of interest.
References
1. IEA. Global EV Outlook 2023. Available online: https://www.iea.org/energy-system/transport/electric-vehicles (accessed on
30 September 2023).
2. Thiel, C.; Alemanno, A.; Scarcella, G.; Zubaryeva, A.; Pasaoglu, G. Attitude of European car drivers towards electric vehicles: A
survey. In JRC Report; Publications Office of the European Union: Luxembourg, 2012.
3. Abdullah, H.M.; Gastli, A.; Ben-Brahim, L. Reinforcement learning based EV charging management systems—A review. IEEE
Access 2021, 9, 41506–41531. [CrossRef]
4. Li, H.; Wan, Z.; He, H. Constrained EV charging scheduling based on safe deep reinforcement learning. IEEE Trans. Smart Grid
2019, 11, 2427–2439. [CrossRef]
5. Zhang, C.; Liu, Y.; Wu, F.; Tang, B.; Fan, W. Effective charging planning based on deep reinforcement learning for electric vehicles.
IEEE Trans. Intell. Transp. Syst. 2020, 22, 542–554. [CrossRef]
6. Viziteu A. Furtună D. Robu A. Senocico S. Cioată P. Remus Baltariu M. Filote, C.R.M. Smart Scheduling of Electric Vehicles Based
on Reinforcement Learning. Sensors 2022, 22, 3718. [CrossRef] [PubMed]
7. Dang, Q.; Wu, D.; Boulet, B. A q-learning based charging scheduling scheme for electric vehicles. In Proceedings of the 2019 IEEE
Transportation Electrification Conference and Expo (ITEC), Detroit, MI, USA, 19–21 June 2019; pp. 1–5.
8. Liu, D.; Zeng, P.; Cui, S.; Song, C. Deep Reinforcement Learning for Charging Scheduling of Electric Vehicles Considering
Distribution Network Voltage Stability. Sensors 2023, 23, 1618. [CrossRef] [PubMed]
9. Korkas, C.D.; Baldi, S.; Yuan, S.; Kosmatopoulos, E.B. An adaptive learning-based approach for nearly optimal dynamic charging
of electric vehicle fleets. IEEE Trans. Intell. Transp. Syst. 2017, 19, 2066–2075. [CrossRef]
10. Lee, J.; Lee, E.; Kim, J. Electric vehicle charging and discharging algorithm based on reinforcement learning with data-driven
approach in dynamic pricing scheme. Energies 2020, 13, 1950. [CrossRef]
11. Wan, Z.; Li, H.; He, H.; Prokhorov, D. Model-Free Real-Time EV Charging Scheduling Based on Deep Reinforcement Learning.
IEEE Trans. Smart Grid 2019, 10, 5246–5257. [CrossRef]
12. Schubert, C.; Hassen, W.F.; Poisl, B.; Seitz, S.; Schubert, J.; Usabiaga, E.O.; Gaudo, P.M.; Pettinger, K.H. Hybrid Energy Storage
Systems Based on Redox-Flow Batteries: Recent Developments, Challenges, and Future Perspectives. Batteries 2023, 9, 211.
[CrossRef]
13. Erdogan, G.; Fekih Hassen, W. Charging Scheduling of Hybrid Energy Storage Systems for EV Charging Stations. Energies 2023,
16, 6656. [CrossRef]
14. Mnih, V.; Kavukcuoglu, K.; Silver, D.; Rusu, A.A.; Veness, J.; Bellemare, M.G.; Graves, A.; Riedmiller, M.; Fidjeland, A.K.;
Ostrovski, G.; et al. Human-level control through deep reinforcement learning. Nature 2015, 518, 529–533. [CrossRef] [PubMed]
15. Fedus, W.; Ramachandran, P.; Agarwal, R.; Bengio, Y.; Larochelle, H.; Rowland, M.; Dabney, W. Revisiting fundamentals
of experience replay. In Proceedings of the International Conference on Machine Learning, PMLR, Virtual, 12–18 July 2020;
pp. 3061–3071.
16. Paraskevas, A.; Aletras, D.; Chrysopoulos, A.; Marinopoulos, A.; Doukas, D.I. Optimal management for EV charging stations: A
win–win strategy for different stakeholders using constrained Deep Q-learning. Energies 2022, 15, 2323. [CrossRef]
17. Kaul, S.K.; Yates, R.D. Timely updates by multiple sources: The M/M/1 queue revisited. In Proceedings of the 2020 54th Annual
Conference on Information Sciences and Systems (CISS), Princeton, NJ, USA, 18–20 March 2020; pp. 1–6.
Energies 2023, 16, 8102 18 of 18
18. Janjic, A.; Velimirovic, L.; Stankovic, M.; Petrusic, A. Commercial electric vehicle fleet scheduling for secondary frequency control.
Electr. Power Syst. Res. 2017, 147, 31–41. [CrossRef]
19. Aljafari, B.; Jeyaraj, P.R.; Kathiresan, A.C.; Thanikanti, S.B. Electric vehicle optimum charging-discharging scheduling with
dynamic pricing employing multi-agent deep neural network. Comput. Electr. Eng. 2023, 105, 108555.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual
author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to
people or property resulting from any ideas, methods, instructions or products referred to in the content.