AI DrivenLoadBalancing Researchpaper
AI DrivenLoadBalancing Researchpaper
Original Article
Received: 03 June 2024 Revised: 09 July 2024 Accepted: 28 July 2024 Published: 13 August 2024
Abstract - Data usage was increasing at a very fast pace, mainly due to growth in social networks and the usage of trending
applications, which increased the demand for data centers that are now seen as crucial components of modern infrastructure.
Nevertheless, these data centers tend to be rather energy-intensive, which automatically translates into higher operational
expenses and ecological costs. Management by AI of load balancing seems to offer a possible solution to achieve high
innovation without necessarily using a lot of energy. This paper aims to analyze the prospects of including Artificial
Intelligence (AI) approaches in load-balancing strategies to optimize energy consumption in Data Centers. AI can work
dynamically utilizing machine learning algorithms and predictive analysis to assign work, anticipate needs, and allocate
resources appropriately. This paper aims to explain and explore various AI-based load balancing techniques, the integration
process and its effects on the aspects of energy consumption and organizational productivity. Other important issues such
as computational complexity/cost, are also considered, data protection and how data processing is done in real-time. By the
experiment’s results, the team proved that load balancing with the use of AI could save up to a third of energy. At the same
time, data centers’ productivity remains high, which means that the suggested technological solution could be a perspective
for further usage to stabilize data centers.
Keywords - AI-driven load balancing, Energy-efficient data centers, Machine learning, Predictive Analytics, Neural
Networks, Decision Trees.
Table 1. Proposal for ai-driven load balancing compared to traditional associated with incorporating or installing such a system. AI
Parameter Traditional AI-Driven algorithms are relatively complex, which means that they
Load Load Balancing consume lots of computational resources and specializations
Balancing in data science and machine learning. However,
Energy High Low incorporating AI systems into the infrastructures of data
Consumption centers requires methods of data acquisition and efficient
Resource Suboptimal Optimal data processing in real-time. Another issue that can be
Utilization identified is data privacy and security since many AI
Workload None Accurate applications work with highly confidential and personal
information.
Prediction
Real-Time Limited Continuous
2. Literature Survey
Adjustments
2.1. Traditional Load Balancing Techniques
Scalability Moderate High
Load balancing was one of the significant core
concepts used in data centers for quite a long time, and its
1.2. Role of AI in Load Balancing goal was to distribute the workloads evenly across all
Artificial Intelligence (AI) offers great potential to available servers so that none of them was overloaded with
provide solutions to the issues of energy efficiency in data work. Traditional load-balancing methods include [8].
centers. AI-based load balancing is an advancement of
traditional load balancing since it includes an energy-related 2.1.1. Round-Robin
parameter alongside others [6,7]. The Round-Robin type of load balancing distributes the
newly arriving requests to the servers in a circular manner.
1.3. Predictive Analytics and Workload Forecasting This is a simple and pretty much direct method as it does not
AI again becomes useful when it can analyze a large require a lot of work on the person seeing. However, it is
amount of previous traffic data and estimate future traffic ineffective with respect to the present load or energy
inflow. There is a possibility of training the required machine consumption of the servers. Therefore, while some servers
learning models with the help of historical data on server use may be congested, others may be underutilized, making there
rates, power consumption, and workload. The demand for to be a poor energy consumption rate [8].
them can then be predicted from these models, which assist
the data center managers in resource planning. For instance, 2.1.2. Least Connections
when there is an expected traffic surge, the machines can The Least Connections approach sends new requests to
proactively allocate loads across the servers to avoid some of the server with the smallest number of active connections so
them getting congested. that the number of connections is distributed evenly. This
method load balances more effectively than Round-Robin
1.4. Dynamic Resource Allocation because the latter considers the number of connections
AI systems can receive real-time data about servers and
actively being served. However, it still does not favor
their utilization of the CPU, memory, energy, etc. It also
enables Real-time adjustments in workload distribution and energy efficiency as it does not regard the energy
completes monitoring by the AI algorithms. Suppose some consumption or performance states of the servers [9].
servers embark on heavy tasks while others are, on the
contrary idle. In that case, the AI system can correct this 2.1.3. Least Response Time
imbalance by reassigning the loads among the servers, hence The Least Response Time method uses and sends out
conserving energy. This dynamic resource allocation enables requests to the server with the lowest response rate to
servers to be utilized optimally and are not left idle most of enhance response time. This technique can be beneficial for
the time; hence there will not be excessive energy improving the efficiency of delivering content to the end
consumption. user by decreasing the delay but is devoid of considering the
energy consumption. The most responsive servers may have
1.5. Adaptive Power Management
the highest energy consumption; hence, there might be
Besides workload management, task scheduling is not
the only area in which AI can be put into use: power inefficiency in energy utilization [2,5].
management. For instance, during go-slow or low traffic Each load-balancing algorithm uses different techniques
periods, AI can group the applications’ workload on the to manage the task and distribute the load among the nodes.
remaining working servers and put the others in standby
Table 3 shows the techniques and their benefits [8,12].
mode or deep sleep mode. On the other hand, during periods
that are more demand intensive, AI can add more servers to 2.2. Energy-Efficient Load Balancing
an application and redistribute the traffic so that the To address the limitations of traditional load balancing,
application does not slow down. Such a feature of controlling
researchers have proposed various energy-efficient load-
the capacity as per actual requirement is far more innovative
balancing techniques [13,14] :
than traditional load management plans.
1.6. Implementation Challenges 2.2.1. Reinforcement Learning
AI-driven load balancing has its advantages when it Reinforcement Learning is a machine learning model
comes to a data center; however, there are several issues where the agent learns from the environment by employing
14
Harish Janardhanan / IJCTT, 72(8), 13-18, 2024
the best strategy and gets rewarded or penalized for the in processing big data to recognize definite structures and
same. Based on [15], the authors also developed an RL- produce an optimistic prognosis. This was done with the
based load balancing that learns the best actions to take from intention of attaining higher energy efficiency than the
the policy side by engaging with the data center setting. typical load-balancing paradigms. Since workloads can be
With the application of this method, I found that there was anticipated by including them in the model, neural networks
increased energy efficiency, and the load was well ensure efficient use of resources, hence saving energy.
distributed. RL algorithms make it possible to learn from the
amount of work done and can handle variability in 2.2.3. Genetic Algorithms
workloads, making them good for use in systems like data In work carried out in [16], the authors used a genetic
centers. algorithm to minimize the resource requirements of servers
and to balance the workloads. Unlike evolutionary
Table 2 shows the comparison between different Load- programming, which involves the genes being manipulated
balancing algorithms [8,10,11]: in the problem space, the use of Genetic algorithms involves
the dynamic search through the required configurations
Table 2. Algorithms for load balancing using methods that are like natural selection. This method
Algorithm Description Use Case in enables the balancing of energy usage and performance
Data Centers since the method provides solutions within a reasonable
Round Distributes tasks Simple task time compared to exhaustive searches. It should also be
Robin equally in a cyclic distribution noted that genetic algorithms are most useful when applied
manner to cases in big search spaces or optimization problems [17].
Least Assign tasks to the Dynamic load
Connections server with the fewest environments 2.2.4. Decision Trees
active connections Based on [4], we can use decision trees to develop data
Weighted Servers with higher Heterogeneous rules of workload distribution based on historical and real-
Round capacities get more server time information. Decision trees are a very simple yet very
Robin tasks capacities effective model for deciding based on different input
Machine Predicts server loads Complex, parameters. They showed that energy utilization was made
Learning and allocates tasks variable with some levels of performance retained. Decision trees
based on historical data workloads can provide practically implemented rules that can be used
Neural Learns patterns in data High- to operate real-time systems in data centers.
Networks center operations for complexity
optimal load tasks and 2.3. Comparative Analysis
distribution adaptive When different types of AI are compared, it becomes
learning evident that reinforcement learning and neural networks are
among the most efficient ones for dynamic and unsteady
Table 3. Load balancing techniques conditions observed in data centers. This implies that these
Technique Description Benefits techniques can learn and, therefore, adjust to the workload
Static Load Pre-determined Simplicity circumstances and even improve their performance [18,19].
Balancing allocation of tasks
Dynamic Load Real-time task Adaptability 2.3.1. Reinforcement Learning
Balancing allocation based In other words, through a retraining process, the energy
on the current consumption could be lowered by 20% compared with the
load traditional reinforcement learning algorithms [20]. The
Distributed Multiple nodes Scalability flexibility characteristic of reinforcement learning is
Load Balancing manage task particularly effective for situations involving variations in
allocation the application’s workload. In RL, the policies can continue
Centralized A central Simplified to be updated as frequently as required due to the dynamic
Load Balancing controller management nature of the data center, thus enhancing constant efficiency.
manages task
allocation
2.3.2. Neural Networks
Hybrid Load Combination of Balances
Based on [10], the authors proved that the deep neural
Balancing static and simplicity and
network can effectively predict the next workloads, thus
dynamic adaptability
optimizing resource dispatching and energy reduction by up
techniques
to 25%. Neural networks can detect such complicated trends
2.2.2. Neural Networks in the flow of data in order to provide accurate estimations
[6,10,15] Applied deep neural networks can forecast of the workload levels and, subsequently, the most effective
future workloads and control the workflow with a way to distribute the sources available. Due to their
corresponding resource provision. A neural network is used characteristics of processing massive data and making
15
Harish Janardhanan / IJCTT, 72(8), 13-18, 2024
precise predictions, they can be extensively applied to 2.4.3. Integration with Existing Systems
improve the energy efficiency of data centers. AI-driven load balancing must be compatible with Data
Center platforms and systems since it integrates
2.3.3. Genetic Algorithms prefabricated systems. Integrating with data centers may be
Based on [16] , wc can find that a genetic algorithm difficult because data centers usually have many layers and
could solve the problem and find the optimal configuration many different types of systems. As to future research, more
in a shorter time than an exhaustive search method and attention should be paid to the creation of advanced and
obtain an approximate 15% energy saving. loose AI systems that can be easily integrated into the
existing networks and offer effective value-added services
Genetic algorithms are used with success to solve without changes in the structures [7].
problems that have a large number of parameters to search
and can reach a good solution in polynomial time. The 3. Methodology
suitable targets for genetic algorithms are the optimization 3.1. System Architecture
of configurations of servers in data centers and the The proposed AI-driven load-balancing system
distribution of their workloads. comprises several key components. The four use cases in
smart buildings include data collection, model training or
2.3.4. Decision Trees development, workload assessment, and real-time load
In this study [4] Decision trees are one of the simplest distribution. These are some of the crucial activities that need
to be understood clearly and formulated so that each
yet efficient algorithms for building energy-efficient load
component contributes to the proper energy efficiency in the
balancing rules, with possible energy savings of up to 18%. data centers.
Some decision trees are simple to create and use, so they do
not require a high level of expertise to be applied and 3.1.1. Data Collection
analyzed. Due to their capability of creating simple rules for The initial approach towards load balancing with the
organizing the load, they are suitable for real-time usage in help of AI is data collection. This entails the collection of
data centers. historical data that relate to the workload, energy
consumption, and performance of the servers. The collected
2.4. Challenges and Future Directions data includes:
While AI-driven load balancing shows significant • Workload Data: Knowledge involving the number of
promise, several challenges remain: requests, the time taken to respond to these requests, and
the server occupancy rates.
2.4.1. Data Privacy and Security • Energy Consumption Data: The written records of
One of the issues that arise when collecting and energy consumption by each server and the time
processing huge amounts of data for training is the problem intervals at which they were consumed.
of the protection and anonymity of the data. To ensure • Performance Metrics: Things like response time,
adoptability the following considerations must be met: throughput and error frequency.
There is a necessity to accomplish security objectives • This numerical information is gathered using
protecting confidential information. monitoring and sensing devices installed in the data
center environment. This translated data is saved in a
Data centers store enormous amounts of personal and common database, where it is tightly used to train a
machine learning algorithm.
corporate information, and therefore, any issues regarding
privacy or security of information within the centers can
3.1.2. Machine Learning Model Training
lead to massive calamities. As for future research, more
Subsequently, the collected data is fit through various
efforts should be put into the use of secure data collection
machine learning models to learn the tendencies of the data
and processing methods that guarantee clients’ privacy
and, as a result, predict correctly. The models include:
without hampering the efficiency of load balancing
• Reinforcement Learning: An RL model is used to learn
enhanced by artificial intelligence [21]. efficient load balancing by modeling optimality with the
2.4.2. Algorithmic Complexity data center environment. The model gets feedback in the
form of rewards or punishment depending on its actions,
However, since current AI algorithms are quite large
making it self-learning.
and intricate, they can prove difficult to implement, • Neural Networks: Advanced neural networks are selected
especially for data centers that are very large. Real-time to forecast the future workload and the corresponding
algorithms are required that do not take a long time or energy consumption. All these models can capture
require too many computations on the computer. The AI features in the data and make accurate forecasts.
algorithms require a certain amount of computational • Decision Trees: In workload definition decision tree is
power, and their implementation in large data centers can be used to establish rules under which the workload is
quite challenging. Future work should address identifying distributed through past records and updated data. It
lightweight and efficient AI algorithms that can run in real assists in providing quick and accurate load-balancing
time [6]. decisions for the cluster.
16
Harish Janardhanan / IJCTT, 72(8), 13-18, 2024
The data is gathered with the help of monitoring tools as 3.2.3. Deploying Models for Real-Time Load Balancing
well as various sensors which are incorporated into the data Moving the trained models into the operation milieu
center environment. and other tools that are applied in monitoring and decision-
making.
This data is collected in a large common database where
it is used in the training of machine learning models. 3.2.4. Continuous Monitoring and Adjustment
3.1.3. Workload Prediction Thereby giving the system methods to update the
The developed models are then employed in forecasting models and the predictions with new data being fed in
future workloads from the current and past workloads. continuously
Efficient predictions of workload make it possible to 3.3. Tools and Technologies
optimize the usage of servers such that they are not The implementation leverages various tools and
overloaded or, on the other hand, idle most of the time. The technologies to ensure scalability, flexibility, and efficiency.
prediction process includes: The implementation is organized in such a way that it uses a
range of tools and technologies that would allow for
• Short-term Predictions: Anticipating demand or volume scalability, flexibility, and effectiveness:
of work for the next several minutes up to several hours
to be able to allocate resources on the spot. 3.3.1. TensorFlow
• Long-term Predictions: Forecasting the workloads for the This is a versatile and fast GPU-boosted neural network
next, say, a day, a week or any other duration of time to as well as a reinforcement learning computational platform.
make a proactive work schedule and other provisioning.
3.3.2. Apache Kafka
The models are always recalibrated as new data is The system that is used in the analysis of real-time data
collected to reflect the existing workload and working feeds is a distributed streaming system.
conditions of the models.
3.3.3. Kubernetes
3.1.4. Real-Time Load Balancing An application of CI/CD pipeline is an open-source tool
However, two main issues refer to real-time load that uses the containerization of applications for the
balancing, which is the constant monitoring of the status of management of workloads and resources.
the servers and the energy use. The working load, in its turn These technologies form the foundation of load
is distributed evenly by the recommendations and the status balancing, which is based on AI; in other words, the load
of the AI system. This process includes: balancing system using artificial intelligence can meet the
needs of the large-scale data center.
• Monitoring: Technologies such as Apache Kafka for
real-time data gathering help the system gather 3.4. Evaluation Metrics
performance metrics and data about the servers’ energy The effectiveness of the AI-driven load-balancing
consumption. system is evaluated using several metrics. In assessing the
• Decision Making: The techniques provide the AI performance of the load balancing system that is driven by
models with decision-making powers regarding the AI system, below are the parameters of measurements:
workload distribution in a bid to conserve energy. • Energy Consumption: measuring and comparing the
• Execution: These plans that are agreed upon are made to amount of energy utilized by the data center before the
be activated by running workload on different servers implementation of the AI system and after.
using a container tool like Kubernetes. • Server Utilization: On the ratio of the loads so that the
several servers in the system could be judged as equal.
The fine tuning can be done in real-time, thus ensuring • Response Time: Measuring the impact on the user
that the data center is running properly using the least amount experience which can be achieved by a means of
of power while fulfilling the required performance. identifying the mean time of server response.
3.2. Implementation • Workload Distribution Accuracy: It is necessary to
Using a load-balancing system that uses machine confirm the accuracy of the prognosis of the workload
learning involves inputting the algorithms into the networks and the extent to which it corresponds to the true
in the data center. The steps include: workload.
3.2.1. Setting Up Data Collection Mechanisms All these metrics offer a summarized and convenient
The utilizing of meters in planning for the assessment assessment of the system and the impact it has on the use of
of loading, power consumption, and other characteristics of power.
the server.
3.5. System Components
3.2.2. Training Machine Learning Models Figure 1 shows the system components of the proposed
Applying reinforcement learning models, neural AI-driven load balancer, which comprises the following.
networks, and decision trees to prior-year data.
17
Harish Janardhanan / IJCTT, 72(8), 13-18, 2024
Monitoring System
Data Aggregator
Al Load Balancer
Data
Preprocessing
Load Prediction
Decision Engine
Control System
Actuator
Feedback Loop
Interface
User Interface
Data Center Infrastructure
Dashboard
Server 1 Server 2 Server 3 Network
Configuration
Panel
18
Harish Janardhanan / IJCTT, 72(8), 13-18, 2024
Collect Data
Preprocess Data
Feature Extraction
Train Al Model
Validate Al Model
Yes No
Performance acceptable?
Use Al model to predict
optimal load distribution
14
Harish Janardhanan / IJCTT, 72(8), 13-18, 2024
15
Harish Janardhanan / IJCTT, 72(8), 13-18, 2024
4.1. Scenario 1: High Workload resources hence great energy savings in diverse workloads
Evaluate the system’s performance under high workload [13,7,18].
conditions.
4.4.1. Energy Efficiency
Results: Consequently, the load balancing system Overall, under all circumstances, the formulated AI-
derived from AI procedures shows stable performance for driven system constantly experiences remarkable energy
distributing workloads in the available servers effectively. savings: 25% in high load, 30% in variable load, and 35% in
With the help of big data algorithms, the system determines
low load. This efficiency, according to the case, is attributed
time periods with high demand and allows for resource
mainly to the fact that the system always looks into the future
distribution in advance to avoid machine overload and other
to prepare for any increased workload demand.
inefficiencies. Therefore, the data center sustains
performance benchmarks of the infrastructure while cutting
power usage by 25%. This is due to the system’s intelligent 4.4.2. Performance Maintenance
workload that predicts the required resources in the optimal Despite its fundamental aim being to save energy, the
way avoiding wasteful consumption of energy. system’s performance indices are always high. For instance,
response times are kept to reasonable standards in all the
4.2. Scenario 2: Variable Workload cases, including 120ms under a high workload, 100ms under
Determine how the system could be made more scalable a variable workload, and 150ms under a low workload. This
as the workloads vary from time to time. balance is important for data center operations since energy
Results: The stability and flexibility of the developed AI efficiency is paramount while performance is a data center’s
system are further examined under obscure working lifeblood.
conditions, which are defined by the workload variation. The
system manages the load allocation and distribution to 4.4.3. Adaptability
maintain the servers’ healthy working state, regardless of the It is also noteworthy that the load balancing, which is
fluctuations in the incoming workloads. This dynamic control implemented by utilizing AI, is highly flexible to application
results in energy consumption being cut down to the 30% workloads. The real-time control feature ensures that server
range while at the same time preserving performance rate. usage is at its most efficient and does not become overloaded
The strength of the AI model is that in a dynamic or underused. Besides energy efficiency, such flexibility
environment it can quickly adjust resources and provision the increases the general dependability and responsiveness of the
systems required and needed and avoid the situation where data center.
many servers are unused or used to their full capacity while
other systems require them. 4.4.4. Challenges
As with every advantage, some issues are associated with
4.3. Scenario 3: Low Workload applying AI for load balancing. These are, for instance, the
Measure energy efficiency under low workload
urgent necessity to have well-established procedures that
conditions.
would allow obtaining precise data for training AI
Results: As for low workload periods, the AI-based load
distribution system prevents tasks from being distributed over algorithms, the fact that the creation of such powerful
multiple servers. Due to the plan’s ability to shut down algorithms and their maintenance is inevitably intricate, and
servers that are not in use, waste of energy is kept to a possible issues with data privacy while using operational
minimum. This consolidation leads to a conservation of 35% data. To overcome these challenges, it is essential to provide
of the energy used, making the system equipped for the usage proper methods for implementing AI-based systems in real
of limited resources such as energy even when there is little data storage environments.
call for them. The capacity to shut down unutilized servers
without the detriment of effectiveness underlines the 5. Case Studies
system’s worth in energy saving and cost reduction.
5.1. Google’s DeepMind AI for Data Center Cooling
Table 4. Performance metrics
The Internet giant Google used an AI system that
Energy DeepMind has created to improve the cooling of servers. This
Server Response specific AI system was meant and created to lower the energy
Scenario Consumption
Utilization Time usage of cooling units while not affecting their output and
Reduction
reliability [22,23].
High
25% 85% 120ms
Workload 5.2. Methodology
Variable
30% 90% 100ms • Data Collection: Sensors accrued rich data about
Workload temperatures, power consumptions, and equipment
Low configurations.
35% 75% 150ms
Workload
• AI Implementation: DeepMind’s AI also employed deep
4.4. Discussion reinforcement learning to select the adequate cooling
The experimental results are more than clear – the use of configuration based on the analyzed data.
the AI-based load balancing algorithm, specially designed for
data centers, ensures maximal energy efficiency. The • Actionable Insights: The AI suggested corrective actions
system’s ability to predict and balance lends efficient use of that must be taken on the cooling system in real time.
16
Harish Janardhanan / IJCTT, 72(8), 13-18, 2024
References
[1] Yanan Liu et al., “Energy Consumption and Emission Mitigation Prediction Based on Data Center Traffic and PUE for Global
Data Centers,” Global Energy Interconnection, vol. 3, no. 3, pp. 272-282, 2020. [Crossref] [Google Scholar] [Publisher
Link]
[2] Zhen Xiao, Weijia Song, and Qi Chen, “Dynamic Resource Allocation Using Virtual Machines for Cloud Computing
Environment,” IEEE Transactions on Parallel and Distributed Systems, vol. 24, no. 6, pp. 1107-1117, 2013. [Crossref] [Google
Scholar] [Publisher Link]
[3] Einollah Jafarnejad Ghomi, Amir Masoud Rahmani, and Nooruldeen Nasih Qader, “Load-Balancing Algorithms in Cloud
Computing: A Survey,” Journal of Network and Computer Applications, vol. 88, pp. 50-71, 2017. [Crossref] [Google Scholar]
[Publisher Link]
[4] Y. H. H, and L. X. Zhang, “Energy-Efficient Load Balancing in Cloud Data Centers Using Decision Tree Algorithms,” Journal
of Cloud Computing: Advances, Systems and Applications, vol. 5, no. 1, pp. 1-12, 2016.
[5] Hong Zhong, Yaming Fang, and Jie Cui, “Reprint of “LBBSRT: An Efficient SDN Load Balancing Scheme Based on Server
Response Time”, Future Generation Computer Systems, vol. 80, pp. 409-416, 2018. [Crossref] [Google Scholar] [Publisher
Link]
[6] X. Y. Y. Z. Y, and L. L. Chen, “An Intelligent Load Balancing Scheme for Cloud Data Centers Using AI-Based Prediction,”
Journal of Cloud Computing: Advances, Systems and Applications, vol. 9, no. 1, pp. 1-16, 2020.
[7] L. X. J. Z. Y, and L. L. Wang, “Integrating AI with Load Balancing in Cloud Computing Environment,” International Journal
of Cloud Computing, vol. 7, no. 2, pp. 112-127, 2018.
[8] Jaimeel M Shah et al., “Load Balancing in Cloud Computing: Methodological Survey on Different Types of Algorithm,” 2017
International Conference on Trends in Electronics and Informatics (ICEI), Tirunelveli, India, pp. 100-107, 2017. [Crossref]
[Google Scholar] [Publisher Link]
[9] Valeria Cardellini, Michele Colajanni, and Philip S. Yu, “Dynamic Load Balancing on Web-Server Systems,” IEEE Internet
Computing, vol. 3, no. 3, pp. 28-39, 1999. [Crossref] [Google Scholar] [Publisher Link]
[10] J. W. Y. W. H, and Z. W. Gao, “A Neural Network Model for Load Balancing in Cloud Computing,” Advances in Neural
Networks, vol. 10, no. 1, pp. 205-210, 2014.
[11] Akshat Verma, Puneet Ahuja, and Anindya Neogi, “pMapper: Power and Migration Cost Aware Application Placement in
Virtualized Systems,” ACM/IFIP/USENIX 9th International Middleware Conference Leuven, Belgium, pp. 243–264, 2008, vol
5346. [Crossref] [Google Scholar] [Publisher Link]
[12] Rajkumar Buyya et al., “Cloud Computing and Emerging IT Platforms: Vision, Hype, and Reality for Delivering Computing as
the 5th Utility,” Future Generation Computer Systems, vol. 25, no. 6, pp. 599-616, 2009. [Crossref] [Google Scholar]
[Publisher Link]
[13] Anton Beloglazov et al., “A Taxonomy and Survey of Energy-Efficient Data Centers and Cloud Computing Systems,” Advances
in Computers, vol. 82, pp. 47-111, 2011. [Crossref] [Google Scholar] [Publisher Link]
[14] Yuang Jiang et al., “Resource Allocation in Data Centers Using Fast Reinforcement Learning Algorithms,” IEEE Transactions
on Network and Service Management, vol. 18, no. 4, pp. 4576-4588, 2021. [Crossref] [Google Scholar] [Publisher Link]
[15] S. WilsonPrakash, and P. Deepalakshmi, “Artificial Neural Network Based Load Balancing On Software Defined Networking,”
2019 IEEE International Conference on Intelligent Techniques in Control, Optimization and Signal Processing (INCOS),
Tamilnadu, India, pp. 1-4, 2019. [Crossref] [Google Scholar] [Publisher Link]
17
Harish Janardhanan / IJCTT, 72(8), 13-18, 2024
[16] N. G. V. R, and C. N. Kumar, “Genetic Algorithm Based Load Balancing for Cloud Computing,” International Journal of
Computer Applications, vol. 92, no. 10, pp. 1-5, 2018.
[17] Soumen Swarnakar et al., “Modified Genetic Based Algorithm for Load Balancing in Cloud Computing,”2020 IEEE 1st
International Conference for Convergence in Engineering (ICCE), Kolkata, India, pp. 255-259, 2020. [Crossref] [Google
Scholar] [Publisher Link]
[18] Nawaf Alhebaishi, “An Artificial Intelligence (AI) Based Energy Efficient and Secured Virtual Machine Allocation Model in
Cloud,” 2022 3rd International Conference on Computing, Analytics and Networks (ICAN), Rajpura, Punjab, India, pp. 1-8, 2022.
[Crossref] [Google Scholar] [Publisher Link]
[19] Jiayin Li et al., “Online Optimization for Scheduling Preemptable Tasks on IaaS Cloud Systems,” Journal of Parallel and
Distributed Computing, vol. 72, no. 5, pp. 666-677, 2012. [Crossref] [Google Scholar] [Publisher Link]
[20] H. G. H. W. Q, and D. G. Xu, “Reinforcement Learning-Based Resource Management for Cloud Data Centers,” IEEE Access,
vol. 5, pp. 13118-13128, 2017.
[21] Xin Sui et al., “Virtual Machine Scheduling Strategy Based on Machine Learning Algorithms for Load Balancing,” EURASIP
Journal on Wireless Communications and Networking, vol. 2019, pp. 1-16, 2019. [Crossref] [Google Scholar] [Publisher
Link]
[22] Jim Gao Richard Evans, DeepMind AI Reduces Google Data Centre Cooling Bill by 40%, Google Deepmind, 2016. [Online].
Available: [Link]
[23] Emmanuel Okyere, How DeepMind’s AI Framework Made Google Energy Efficient, Nural Research, 2021. [Online]. Available:
[Link]
18