CC Unit 1 - 2 notes
CC Unit 1 - 2 notes
It turns traditional computing services into a self-service utility, just like electricity or water.
Users can access technology on demand, without needing to understand or manage the
underlying infrastructure.
🌐 Cloud Types
To better understand cloud computing, we divide it into two main categories:
1. Deployment Models
Define where the cloud infrastructure is located and who manages it.
● Public Cloud: Open to the public or a large group. Owned by service providers (e.g.,
AWS, Azure).
● Private Cloud: Exclusively used by a single organization. Managed internally or by a
third party.
● Hybrid Cloud: A combination of two or more clouds (public, private, community) that
remain separate but are linked.
● Community Cloud: Shared infrastructure for a specific group or organization with
shared concerns (e.g., government agencies).
🔍 Example: The U.S. Government’s Apps.gov is a community cloud serving federal agencies.
2. Service Models
🧱
Provides virtual machines, storage, and infrastructure.
📌
You manage the OS and applications.
Examples: Amazon EC2, Linode, RackSpace.
🛠️
Provides OS, runtime, and development tools.
📌
You deploy your apps; the provider manages the platform.
Examples: Google App Engine, Microsoft Azure, Force.com.
📋
Delivers ready-to-use apps via a browser.
You just use the software; everything else is handled.
📌 Examples: Google Workspace, SalesForce.com, QuickBooks Online.
This layered model is also called the SPI Model (Software, Platform, Infrastructure).
Initially, the NIST model didn't require virtualization or multi-tenancy, but newer versions
include both. It also doesn't fully cover service brokers, provisioning, or integration services,
which are becoming more important in modern cloud computing.
📦 XaaS – Everything as a Service
Beyond IaaS, PaaS, and SaaS, many new service models are emerging:
○ Users can automatically access computing resources like storage and processing
power without human intervention from the provider.
2. Broad Network Access
○ Services are accessible over the network via standard platforms (e.g., phones,
tablets, laptops, etc.).
3. Resource Pooling
○ Cloud resources are pooled to serve multiple users using a multi-tenant model.
Resources are dynamically assigned and reassigned based on demand.
4. Rapid Elasticity
○ Cloud systems automatically control and optimize resource use through metering
(e.g., bandwidth, storage, processing). Users are billed based on usage.
🛠️ Additional Features
● Lower Costs: Efficient operations lead to reduced costs for users.
● Ease of Use: Services are typically plug-and-play.
● Quality of Service (QoS): Guaranteed performance levels.
● Reliability: Redundancy and failover systems ensure high availability.
● Outsourced IT: Management and maintenance are handled by the provider.
● Simplified Maintenance: Centralized software updates and patching.
● Low Barrier to Entry: Minimal upfront investment needed.
○ Teams can access, edit, and share documents in real time, from anywhere.
7. Environmentally Friendly
○ Most providers offer 99.9% uptime and robust disaster recovery options.
○ Users may have less control over infrastructure and services compared to
on-premise systems.
2. Security and Privacy Risks
○ Regulations like GDPR, HIPAA, and SOX may be difficult to comply with due to
data crossing borders.
6. Downtime
○ Even top providers can experience outages, affecting availability.
7. Vendor Lock-In
🧠 Key Concept
● Virtualization allows multiple virtual systems to run on a single physical system.
● Users access cloud services through virtualized interfaces, not the actual physical
machines.
Logical Naming Physical resources are given logical names and accessed through
pointers.
Dynamic The link between virtual and physical resources is flexible and
Mapping responsive to load changes.
Facile Changes Mapping can be updated instantly without service interruption.
Access Users can access cloud services from anywhere via virtual interfaces.
Application Multiple instances of an application run in the cloud and requests are routed
based on load.
CPU Physical CPUs are divided into virtual machines or workloads are distributed
using load balancing.
Storage Data is distributed across multiple storage devices and replicated for
availability.
Pattern Meaning
D2D Datacenter to
Datacenter
🔁 Session Persistence
Maintains user sessions across load-balanced systems using:
📌 Purpose of Hypervisors
Hypervisors play a central role in virtualization, which is a foundational technology in cloud
computing. They allow cloud providers to:
🔹 Types of Hypervisors
Hypervisors are primarily categorized into two types based on how they interact with hardware
and host operating systems.
Use Case Data centers, servers, cloud infra Development, testing, personal use
VM Type Description
System Virtual Emulates an entire hardware system with its own OS and
Machine applications.
Process Virtual Designed to run a single process or application (e.g., JVM, .NET
Machine CLR).
🔹 Virtualization Techniques
Hypervisors implement different virtualization methods to manage guest operating systems:
✅ Full Virtualization
● Emulates the complete hardware environment.
● Guest OS runs without modification.
● Allows running multiple OS types on the same hardware.
● Common in Type 1 hypervisors.
✅ Paravirtualization
● Guest OS is modified to interact with the hypervisor via an API (para-API).
● Requires support from both the host and guest OS.
● Offers better performance than full virtualization.
✅ Emulation
● Software completely simulates hardware.
● Guest OS does not need to match host hardware.
● Useful for cross-platform compatibility.
● Typically slower due to overhead.
Virtualization Guest OS Performance Use Case
Type Modification
For example, Amazon Web Services (AWS) uses Xen and KVM hypervisors for their Amazon
Machine Instances (AMIs), while Microsoft Azure uses Hyper-V.
vSphere is the successor to VMware Infrastructure and includes both infrastructure services
(like ESXi hypervisor and vCenter Server) and application services (like High Availability,
DRS, etc.).
○ A clustered file system optimized for storing virtual machine disk images.
○ Supports concurrent access by multiple ESXi hosts.
4. VMotion:
○ Enables live migration of VMs from one physical server to another with zero
downtime.
○ Maintains VM state and memory contents during transfer.
5. Storage VMotion:
○ Moves a VM’s virtual disks from one datastore to another while the VM
remains active.
6. vNetwork Distributed Switch (DVS):
These VMs can be dynamically moved and scaled according to business needs without being
tied to a specific piece of hardware.
Network Virtualization:
● Uses virtual NICs (vNICs) and virtual switches to mimic physical network interfaces.
● Allows network policies (like security, QoS) to be enforced virtually.
● External virtualization can include VLANs and network hardware abstraction using
software-defined networking (SDN) principles.
In cloud computing, this is often referred to as a server image, machine image, or VM image.
✅ Scalability: Easily scale up services by launching more instances from the same
configuration drift.
●
image.
✅ Disaster Recovery: Recover systems quickly using stored machine images.
✅ Automation: Integral part of DevOps and Infrastructure-as-Code (IaC).
●
●
🔸 Key Terms
Term Explanation
AMI (Amazon Machine AWS-specific image format used to launch EC2 instances.
Image)
Base Image Clean OS install with minimal Start fresh with custom setup
configuration
✅ Azure
● Uses Managed Images and Shared Image Gallery
● Support for image versioning, regions, and replication
✅ Google Cloud
● Uses Custom Images
● Can be stored and used in multiple regions
✅ VMware
● Create VM templates or OVF (Open Virtualization Format) exports
● Used in vCenter to deploy cloned VMs or deploy via automation
🔸 Best Practices
✅ Use golden images for production environments.
✅ Automate image creation with scripts (e.g., Packer).
●
🔸 Real-World Example
Suppose you’re deploying a web app that runs on Ubuntu with Apache, MySQL, and PHP.
Rather than configuring each server manually:
🔄 Scalable on demand
💰 Cost-optimized
●
Utilization Metrics CPU, memory, disk, and network usage statistics used for
decision-making
○ Set rules to add/remove instances based on metrics (CPU > 70%, etc.)
7. Continuously Monitor and Adjust
○ Use tools like AWS CloudWatch, Azure Monitor, Google Cloud Operations Suite
○ Adapt based on real-time and predictive metrics
🔸 Common Challenges
Challenge Impact
Inaccurate Forecasting Leads to over- or under-provisioning
🔸 Example Scenario
Let’s say you're running an e-commerce website. During regular days, 4 VMs are enough. But
during a festival sale:
● Traffic spikes by 4×
● You need to scale up to 16 VMs
● After the sale, scale back to 4 VMs
🔸 Best Practices
✅ Use Auto Scaling and Elastic Load Balancing
✅ Perform load testing before major events
●
System Metrics Track hardware and infrastructure CPU usage, memory usage,
performance disk I/O
Custom Metrics User-defined metrics for specific needs Queue depth, job processing
time
Cloud platforms like AWS CloudWatch, Azure Monitor, and Google Cloud Monitoring collect
both default and custom metrics for analysis and visualization.
🔸 What Is a Baseline?
A baseline is a reference pattern or average measurement that reflects “normal” system
behavior over time. It acts as a benchmark for comparing real-time data to detect anomalies or
abnormal performance.
Time-Dependent Baselines vary by time of day, week, or season (e.g., higher usage on
Mondays)
Dynamic or Baselines can be fixed (static) or adapt over time using machine learning
Static
For instance, if average CPU usage during peak hours is consistently 60–70%, that range
becomes the CPU baseline for those hours.
Capacity Planning Use trends to forecast resource needs and scale appropriately
SLA Monitoring Ensure services meet the agreed Service Level Agreements
These observed ranges become baselines. If on a Tuesday afternoon the CPU spikes to 95%,
an alert is triggered — indicating a possible system overload or abnormal traffic.