Thanks to visit codestin.com
Credit goes to www.scribd.com

0% found this document useful (0 votes)
3 views18 pages

Notes

Uploaded by

konasindhura
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as DOCX, PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
3 views18 pages

Notes

Uploaded by

konasindhura
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as DOCX, PDF, TXT or read online on Scribd
You are on page 1/ 18

1.

Workload Distribution Architecture in Cloud Computing


Workload distribution architecture in cloud computing refers to the way computational tasks,
application components, and service requests are allocated, balanced, and executed across
multiple cloud resources (servers, storage systems, networks) to ensure high performance,
scalability, and fault tolerance.
Purpose of Workload Distribution
 Optimize resource utilization – prevent overloading of a single server.
 Improve performance – distribute workload for faster response.
 Enhance availability – reroute tasks if one server fails.
 Scalability – add/remove resources based on demand.
 Cost efficiency – avoid underutilization or over-provisioning.
Components of a Workload Distribution Architecture:
 Load Balancers: Distribute traffic to multiple instances of an application or service.
 Monitors: Track resource utilization, performance metrics, and identify potential
issues.
 Hypervisors: Manage virtual machines and their workload distribution.
 Resource Clusters: Groups of resources that work together to handle workloads.
 Replication: Creating copies of resources to handle increased demand or failover.
 Resource Pool: Group of computing resources (VMs, containers, storage) where
workloads run.
 Workload Scheduler / Orchestrator: Decides which resource should run which
task.
Examples: Kubernetes (for containers), AWS Auto Scaling.

Workload Distribution Architecture Types


1. Horizontal Distribution
o Multiple identical servers handle workload in parallel.
o Example: Web servers behind a load balancer.
2. Vertical Distribution
o Workload divided based on task type.
o Example: Separate servers for frontend, backend, and database.
3. Hybrid Distribution
o Combination of horizontal and vertical distribution.

2. Resource pooling architecture in cloud


Resource pooling is a core concept in cloud architecture where a shared pool of computing,
storage, and networking resources is dynamically allocated to multiple tenants (users or
organizations) as per demand.
It enables multi-tenancy, efficient resource utilization, and cost-effectiveness in cloud
environments.
Purpose of Resource Pooling
 Share resources among multiple users securely.
 Improve utilization by avoiding idle hardware.
 Dynamically assign and reassign resources based on demand.
 Enable scalability (add/remove) and elasticity(grow/shrink).
Audit Monitor – This mechanism monitors resource pool usage to ensure compliance with
privacy and regulation requirements, especially when pools contain cloud storage devices or
data loaded into memory
Cloud Usage Monitor – Various cloud usage monitors are involved in the runtime tracking
and synchronization that are required by the pooled IT resources and any underlying
management systems.
Hypervisor – The hypervisor mechanism is responsible for providing virtual serv ers with
access to resource pools, in addition to hosting the virtual servers and sometimes the resource
pools themselves.
Logical Network Perimeter – The logical network perimeter is used to logically orga nize
and isolate resource pools.
Pay-Per-Use Monitor – The pay-per-use monitor collects usage and billing informa tion on
how individual cloud consumers are allocated and use IT resources from various pools.
Remote Administration System – This mechanism is commonly used to interface with
backend systems and programs in order to provide resource pool administration features via a
front-end portal.
Resource Management System – The resource management system mechanism sup plies
cloud consumers with the tools and permission management options for administering
resource pools.
Resource Replication – This mechanism is used to generate new instances of IT resources
for resource pools
3. Dynamic Scalability Architecture in Cloud Computing
Dynamic scalability architecture in cloud computing allows systems to automatically adjust
resources (CPU, memory, storage, network bandwidth) in response to fluctuating demands,
optimizing performance and cost.
Key Principles
1. Real-Time Monitoring
o Continuous tracking of metrics (CPU usage, memory load, request latency).
2. Auto-Scaling Policies
o Predefined rules that trigger scaling actions (up or down).
3. Automation
o Scaling decisions implemented without manual admin intervention.
4. Elastic Resource Pool
o Backed by virtualization and resource pooling for quick provisioning.

Types of Scaling in Dynamic Scalability


1. Vertical Scaling (Scale Up / Down)
o Increasing or decreasing resources (CPU, RAM) for an existing VM or
container.
o Example: Upgrading AWS EC2 instance type.
2. Horizontal Scaling (Scale Out / In)
o Adding or removing instances to handle load.
o Example: Adding more web servers behind a load balancer.
Dynamic Horizontal Scaling
 Definition:
The number of IT resource instances (e.g., servers, containers, VMs) is scaled out
(increased) or scaled in (decreased) automatically to handle fluctuating workloads.
 How it works:
An automatic scaling listener monitors workload metrics (e.g., number of requests,
CPU usage).
When thresholds are crossed, it triggers resource replication to create or remove
instances.
 Example:
A web application behind a load balancer automatically adds more server instances
during peak hours and removes them during low traffic periods.
2. Dynamic Vertical Scaling
 Definition:
Adjusting the processing capacity of a single IT resource instance by increasing or
decreasing its hardware capabilities.
 How it works:
The system upgrades or downgrades the CPU, RAM, or storage capacity of a virtual
machine without replacing it.
 Example:
If a VM running an application is overloaded, it can have additional CPU cores or
more memory allocated dynamically.
3. Dynamic Relocation
 Definition:
Moving an IT resource instance from one host environment to another with higher
capacity or better performance.
 How it works:
The system migrates the workload to another host with greater processing power,
faster storage, or better network throughput.
 Example:
Moving a database from a tape-based SAN (4 GB/s I/O) to a disk-based SAN (8
GB/s I/O) for faster data access.
4. Elastic Resource Capacity Architecture in Cloud Computing
Elastic resource capacity architecture refers to a cloud design pattern in which the available
IT resources (compute, storage, and network) can be automatically and rapidly expanded
or reduced to meet fluctuating workload demands.

It is one of the core principles of cloud elasticity, ensuring that the system always has just
the right amount of resources without overprovisioning or underutilizing.

1. Purpose

 Maintain consistent performance during varying workloads.


 Reduce operational costs by releasing unused resources.
 Provide flexibility to meet unpredictable demand.
 Enable automation in resource management.

2. Key Principles

1. On-Demand Allocation
o Resources can be provisioned instantly when demand increases.
2. Automatic Deallocation
o Resources are released when demand decreases.
3. Scalability
o Supports both horizontal (adding/removing instances) and vertical
(upgrading/downgrading instance capacity) scaling.
4. Pay-as-You-Go
o Users are billed only for the capacity they consume.

3. Components of Elastic Resource Capacity Architecture

Component Function

Workload Monitor Tracks performance metrics and detects load changes.

Elasticity Manager Decides when to add/remove capacity based on policies.

Resource Pool Set of available resources (VMs, storage, bandwidth).

Provisioning Engine Automates creation, configuration, and release of resources.

Load Balancer Distributes requests to new or existing instances.

4. Working Mechanism

1. Monitor: The system continuously monitors workloads.


2. Trigger: If workload exceeds or drops below thresholds, an elasticity trigger is
activated.
3. Scale Out/In or Up/Down: The provisioning engine allocates or deallocates
resources.
4. Balance: Load balancer ensures optimal request distribution.
5. Optimize: After scaling, performance and costs are re-evaluated.

5. Types of Elasticity in Resource Capacity

1. Reactive Elasticity
o Scaling happens after the workload changes are detected.
o Example: Adding new VM instances after CPU usage exceeds 80%.
2. Proactive Elasticity
o Scaling happens before workload changes, based on prediction.
o Example: Increasing capacity before a scheduled online sale.
5. Server load balancing architecture
In cloud computing, server load balancing architecture distributes network traffic across
multiple servers to prevent any single server from becoming overwhelmed, ensuring optimal
performance, high availability, and scalability. This is achieved by using a load balancer,
which acts as a traffic controller, directing incoming requests to available servers based on
various factors like server load, health, and geographical location.
Key Functions of Server Load Balancing
1. Traffic Distribution
o Balances requests between multiple servers.
2. Health Monitoring
o Checks server availability and removes unresponsive servers from the pool.
3. Failover Management
o Redirects traffic to healthy servers during failures.
4. Scalability Support
o Integrates with auto-scaling to handle changing workloads.
Components of Server Load Balancing Architecture
Component Function
Clients Users sending requests to the cloud service.
Central component distributing requests based on configured
Load Balancer
algorithms.
Server Pool Collection of physical or virtual servers running applications/services.
Health Monitor Checks server status at regular intervals.
Application Servers Process client requests and deliver responses.
Data Storage Layer Stores application and user data, accessed by the servers.
Load Balancing Methods in Cloud
1. Static Algorithms
o Round Robin: Requests assigned sequentially to each server.
o Weighted Round Robin: Servers with higher capacity get more requests.
o IP Hash: Client IP determines which server handles the request.
2. Dynamic Algorithms
o Least Connections: Directs traffic to the server with the fewest active
sessions.
o Least Response Time: Chooses the server responding fastest.
o Resource-Based: Assigns requests based on CPU/memory usage.
6. Cloud Bursting Architecture in Cloud Computing
Cloud bursting is a hybrid cloud deployment model where an application normally runs in a
private cloud (or on-premises data center), but bursts into a public cloud when the local
resources reach their capacity limits.

It’s designed to handle unexpected spikes in demand without over-provisioning local


infrastructure.

1. Purpose

 Provide extra computing capacity during peak demand.


 Avoid investing in additional permanent hardware.
 Maintain application performance during load surges.
 Enable cost-efficient scaling by paying only for extra resources when needed.
2. Key Characteristics

1. Hybrid Environment
o Primary workload runs on private cloud/on-premises.
o Overflow workload moves to public cloud.
2. Automatic Triggering
o Monitoring system detects resource threshold breaches.
3. Elastic Expansion
o Public cloud resources are provisioned on demand.
4. Temporary Usage
o Public cloud capacity is released when demand subsides.

3. Components of Cloud Bursting Architecture

Component Function

Private Cloud / Data


Primary environment for hosting workloads.
Center

Public Cloud Overflow capacity for peak demand.

Burst Controller Decides when to initiate cloud bursting.

Load Balancer Routes traffic between private and public clouds.

High-speed secure link between private and public clouds (VPN,


Network Connection
Direct Connect, ExpressRoute).

Monitoring & Threshold


Tracks resource usage and triggers bursting.
Engine

4. Working Mechanism

1. Application runs in private cloud.


2. Monitoring system detects resource saturation (e.g., CPU > 80%).
3. Burst controller provisions additional instances in the public cloud.
4. Load balancer redirects overflow requests to public cloud instances.
5. When load decreases, public cloud instances are de-provisioned.

5. Types of Cloud Bursting

1. Manual Cloud Bursting


o Admin manually triggers the burst into public cloud.
2. Automated Cloud Bursting
o System automatically detects and triggers bursting.
7. Elastic Disk Provisioning Architecture in Cloud
1. Introduction
 Elastic disk provisioning in cloud computing refers to the ability to dynamically
allocate, expand, or shrink disk storage capacity for virtual machines (VMs) or
applications based on demand.
 This ensures pay-as-you-go storage usage and avoids under- or over-provisioning.
 Common in IaaS platforms like AWS EBS, Azure Managed Disks, Google Persistent
Disks.

2. Key Characteristics
1. On-demand scaling – Storage can be increased (or decreased in some cases) without
service downtime.
2. Automated provisioning – Triggered by user request or automated policies.
3. Seamless integration – Works with compute services, databases, and file systems.
4. Performance tiers – Choice of SSD, HDD, throughput, IOPS.

3. Architecture Overview
a) Main Components
1. User Interface / API Layer
o Cloud portal or REST APIs where the user requests storage.
o Defines size, type, IOPS requirements.
2. Provisioning Service
o Orchestrates the creation and allocation of storage volumes.
o Interfaces with backend storage pools.
3. Storage Pool / Backend Infrastructure
o Physical storage hardware (SSDs, HDDs) pooled and virtualized.
o Uses Software Defined Storage (SDS) for flexibility.
4. Elastic Controller
o Monitors usage metrics and triggers scaling.
o Manages volume resizing without VM downtime.
5. Data Path Layer
o Responsible for attaching disks to VMs over a network (iSCSI, NVMe-oF).
o Maintains data consistency and performance.
6. Monitoring & Billing
o Tracks capacity usage, IOPS, and duration for pay-per-use billing.

b) Workflow
1. Request Initiation
o User/admin requests new or extended storage via console or API.
2. Validation
o Provisioning service checks quota, type, and compatibility.
3. Dynamic Allocation
o Elastic controller allocates from the available storage pool.
o Logical volume created on SDS.
4. Attachment
o Disk is attached to VM or application instance.
o If resizing, changes are applied live (in most cases).
5. Monitoring & Scaling
o Performance metrics monitored for auto-scaling triggers.
o If thresholds are reached, storage is expanded automatically.
8. Redundant Storage Structure
Redundant Storage Structure in cloud computing refers to storing the same data in
multiple physical or logical locations so that if one storage location fails, the data remains
accessible. This redundancy improves fault tolerance, high availability, and disaster
recovery.

1. Purpose of Redundant Storage


 Data Reliability – Avoids loss during hardware/software failures.
 High Availability – Keeps applications running without downtime.
 Disaster Recovery – Ensures data can be restored after outages or disasters.
 Performance Optimization – In some configurations, can improve read speeds.

A logical unit number (LUN) is a logical drive that represents a partition of a physical drive.

The storage service gateway is a component that acts as the external interface to cloud storage
services, and is capable of automatically redirect ing cloud consumer requests whenever the
location of the requested data has changed
Storage replication is a variation of the resource replication mecha nisms used to
synchronously or asynchronously replicate data from a primary storage device to a secondary
storage device. It can be used to replicate partial and entire LUNs.

You might also like