CLOUD COMPUTING (Professional Elective – IV) (CS714PE)
UNIT – I
Computing Paradigms
1. High-Performance Computing
In high-performance computing systems, a pool of processors (processor machines or central
processing units [CPUs]) connected (networked) with other resources like memory, storage, and
input and output devices, and the deployed software is enabled to run in the entire system of
connected components.
The processor machines can be of homogeneous or heterogeneous type.
Thus, examples of HPC include a small cluster of desktop computers or personal computers (PCs)
to the fastest supercomputers.
HPC systems are normally found in those applications where it is required to use or solve scientific
problems.
Most of the time, the challenge in working with these kinds of problems is to perform suitable
simulation study, and this can be accomplished by HPC without any difficulty.
Scientific examples such as protein folding in molecular biology and studies on developing models
and applications based on nuclear fusion are worth noting as potential applications for HPC.
2. Parallel Computing
Parallel computing is also one of the facets of HPC.
Here, a set of processors work cooperatively to solve a computational problem.
These processor machines or CPUs are mostly of homogeneous type.
In parallel computing, since there is simultaneous use of multiple processor machines, the
following apply:
It is run using multiple processors (multiple CPUs).
A problem is broken down into discrete parts that can be solved concurrently.
Each part is further broken down into a series of instructions.
Instructions from each part are executed simultaneously on different processors.
An overall control/coordination mechanism is employed.
3. Distributed Computing
Distributed computing is also a computing system that consists of multiple computers or
processor machines connected through a network, which can be homogeneous or heterogeneous,
but run as a single system.
The connectivity can be such that the CPUs in a distributed system can be physically close
together and connected by a local network, or they can be geographically distant and connected
by a wide area network.
The heterogeneity in a distributed system supports any number of possible configurations in the
processor machines, such as mainframes, PCs, workstations, and minicomputers.
The goal of distributed computing is to make such a network work as a single computer.
Distributed computing systems are advantageous over centralized systems, because there is a
support for the following characteristic features:
Prepared By: Mr. Vinod V. Kulkarni Dept. of CSE,SVIT Secunderabad
CLOUD COMPUTING (Professional Elective – IV) (CS714PE)
Scalability: It is the ability of the system to be easily expanded by adding more machines as
needed, and vice versa, without affecting the existing setup.
Redundancy or replication: Here, several machines can provide the same services, so that even if
one is unavailable (or failed), work does not stop because other similar computing supports will be
available.
4. Cluster Computing
A cluster computing system consists of a set of the same or similar type of processor machines
connected using a dedicated network infrastructure.
All processor machines share resources such as a common home directory and have a software
such as a message passing interface (MPI) implementation installed to allow programs to be run
across all nodes simultaneously.
This is also a kind of HPC category.
The individual computers in a cluster can be referred to as nodes.
The reason to realize a cluster as HPC is due to the fact that the individual nodes can work together
to solve a problem larger than any computer can easily solve.
The nodes need to communicate with one another in order to work cooperatively and meaningfully
together to solve the problem in hand.
If we have processor machines of heterogeneous types in a cluster, this kind of clusters become a
subtype and still mostly are in the experimental or research stage.
5. Grid Computing
The computing resources in most of the organizations are underutilized but are necessary for
certain operations.
The idea of grid computing is to make use of such nonutilized computing power by the needy
organizations, and thereby the return on investment (ROI) on computing investments can be
increased.
Thus, grid computing is a network of computing or processor machines managed with a kind of
software such as middleware, in order to access and use the resources remotely.
The managing activity of grid resources through the middleware is called grid services.
Grid services provide access control, security, access to data including digital libraries and
databases, and access to large-scale interactive and long-term storage facilities.
Grid computing is more popular due to the following reasons:
Its ability to make use of unused computing power, and thus, it is a cost-effective solution
(reducing investments, only recurring costs).
As a way to solve problems in line with any HPC-based application.
Enables heterogeneous resources of computers to work cooperatively and collaboratively to
solve a scientific problem
Prepared By: Mr. Vinod V. Kulkarni Dept. of CSE,SVIT Secunderabad
CLOUD COMPUTING (Professional Elective – IV) (CS714PE)
Electrical Power Grid and Grid Computing
Electrical Power Grid Grid Computing
Never worry about where the electricity that we are Never worry about where the computer power that
using comes from; that is, whether it is from coal in we are using comes from; that is, whether it is from
Australia, from wind power in the United States, or a supercomputer in Germany, a computer farm in
from a nuclear plant in France, one can simply plug India, or a laptop in New Zealand, one can simply
the electrical appliance into the wall-mounted plug in the computer and the Internet and it will get
socket and it will get the electrical power that we the application execution done.
need to operate the appliance.
The infrastructure that makes this possible is called The infrastructure that makes this possible is called
the power grid. It links together many different the computing grid. It links together computing
kinds of power plants with our home, through resources, such as PCs, workstations, servers, and
transmission stations, power stations, transformers, storage elements, and provides the mechanism
power lines, and so forth. needed to access them via the Internet.
The power grid is pervasive: electricity is available The grid is also pervasive in the sense that the
essentially everywhere, and one can simply access it remote computing resources would be accessible
through a standard wall-mounted socket. from different platforms, including laptops and
mobile phones, and one can simply access the grid
computing power through the web browser.
The power grid is a utility: we ask for electricity and The grid computing is also a utility: we ask for
we get it. We also pay for what we get. computing power or storage capacity and we get it.
We also pay for what we get.
6. Cloud Computing
In cloud computing, the IT and business resources, such as servers, storage, network, applications,
and processes, can be dynamically provisioned to the user needs and workload.
In addition, while a cloud can provision and support a grid, a cloud can also support nongrid
environments, such as a three-tier web architecture running on traditional or Web 2.0 applications.
The potential difference between grid and cloud is that grid computing supports leveraging several
computers in parallel to solve a particular application, while cloud computing supports leveraging
multiple resources, including computing resources, to deliver a unified service to the end user.
7. Bio computing
Biocomputing systems use the concepts of biologically derived or simulated molecules (or models)
that perform computational processes in order to solve a problem.
Prepared By: Mr. Vinod V. Kulkarni Dept. of CSE,SVIT Secunderabad
CLOUD COMPUTING (Professional Elective – IV) (CS714PE)
The biologically derived models aid in structuring the computer programs that become part of the
application.
Biocomputing provides the theoretical background and practical tools for scientists to explore
proteins and DNA.
DNA and proteins are nature’s building blocks.
The function of the final molecule rather strongly depends on the order of these blocks.
Thus, the biocomputing scientist works on inventing the order suitable for various applications
mimicking biology.
Biocomputing shall, therefore, lead to a better understanding of life and the molecular causes of
certain diseases.
8. Mobile Computing
In mobile computing, the processing (or computing) elements are small (i.e., handheld devices) and
the communication between various resources is taking place using wireless media.
Mobile communication for voice applications (e.g., cellular phone) is widely established
throughout the world and witnesses a very rapid growth in all its dimensions including the increase
in the number of subscribers of various cellular networks.
An extension of this technology is the ability to send and receive data across various cellular
networks using small devices such as smartphones.
There can be numerous applications based on this technology; for example, video call or
conferencing is one of the important applications that people prefer to use in place of existing voice
(only) communications on mobile phones.
Mobile computing–based applications are becoming very important and rapidly evolving with
various technological advancements as it allows users to transmit data from remote locations to
other remote or fixed locations.
9. Quantum Computing
Manufacturers of computing systems say that there is a limit for cramming more and more
transistors into smaller and smaller spaces of integrated circuits (ICs) and thereby doubling the
processing power about every 18 months.
This problem will have to be overcome by a new quantum computing–based solution, wherein the
dependence is on quantum information, the rules that govern the subatomic world.
Quantum computers are millions of times faster than even our most powerful supercomputers
today.
Since quantum computing works differently on the most fundamental level than the current
technology, and although there are working prototypes, these systems have not so far proved to be
alternatives to today’s silicon-based machines.
Prepared By: Mr. Vinod V. Kulkarni Dept. of CSE,SVIT Secunderabad
CLOUD COMPUTING (Professional Elective – IV) (CS714PE)
10. Optical Computing
Optical computing system uses the photons in visible light or infrared beams to perform digital
computations.
An electric current flows at only about 10% of the speed of light.
This limits the rate at which data can be exchanged over long distances and is one of the factors
that led to the evolution of optical fiber.
By applying some of the advantages of visible and/or IR networks at the device and component
scale, a computer can be developed that can perform operations 10 or more times faster than a
conventional electronic computer.
11. Nano computing
Nanocomputing refers to computing systems that are constructed from nanoscale components.
The transistors based on carbon nanotubes in traditional computers.
The successful realization of nanocomputers relates to the scale and integration of these nanotubes
or components.
The issues of scale relate to the dimensions of the components.
They are, at most, a few nanometers in at least two dimensions.
The issues of integration of the components are twofold: first, the manufacture of complex arbitrary
patterns may be economically infeasible, and second, nanocomputers may include massive
quantities of devices.
Researchers are working on all these issues to bring nanocomputing a reality.
Prepared By: Mr. Vinod V. Kulkarni Dept. of CSE,SVIT Secunderabad