?
:- , “ ” “ ”.
- , &
‘ ,
.
OR
, . ., ,
, , .
,
- , , , , .
:-
, :
,
▪ : –
▪ :
▪ : :
,
▪ ,
▪ ,
▪ :
.
.
’ /
’ . ‘ ’ .
:-
-
, & ,
( )
( ).
, ?
, . ( )
’ ; ;
,
. -
.
( ) :-
( ).
/ .
( )
. ,
.
.
.
,
.
. , ( )
( ) ( ),
.
.
, .
, .
, :
,
, / .
. .
, ,
.
, , ,
.
.
, ,
, .
.
- - .
.
:- - / , , .
, ,
.
,
( )
, , ,
.
:- , , .
, ,
, .
, ’
. /
.
70% 3
.
.
: .
. .
’
.
,
’ . ,
. , ’
.
(
).
/ , /
.
.
,
. . ,
.
, .
, .
.
, .
,
.
, .
,
.
.
,
.
. ,
.
:
1.
’ . .
. ,
.
, .
.
,
.
, , .
,
.
2.
,
. , -
. ,
. ,
, .
’ ,
. ,
.
Consciously or Subconsciously our mind makes up plans for every task which we have to
accomplish which is why things become clearer in our mind. Similarly, if we have to develop an AI
project, the AI Project Cycle provides us with an appropriate framework which can lead us towards
the goal. The AI Project Cycle mainly has 5 stages:-
Starting with Problem Scoping, you set the goal for your AI project by stating the problem
which you wish to solve with it. Under problem scoping, we look at various parameters
which affect the problem we wish to solve so that the picture becomes clearer.
To proceed,
You need to acquire data which will become the base of your project as it will help
you in understanding what the parameters that are related to problem scoping are.
You go for data acquisition by collecting data from various reliable and authentic
sources. Since the data you collect would be in large quantities, you can try to give it
a visual image of different types of representations like graphs, databases, flow
charts, maps, etc.
This makes it easier for you to interpret the patterns which your acquired data follows.
● After exploring the patterns, you can decide upon the type of model you would build to achieve
the goal. For this, you can research online and select various models which give a suitable output.
● You can test the selected models and figure out which is the most efficient one.
● The most efficient model is now the base of your AI project and you can develop your algorithm
around it.
● Once the modelling is complete, you now need to test your model on some newly fetched data.
The results will help you in evaluating your model and improving it.
● Finally, after evaluation, the project cycle is now complete and what you get is your AI project.
PROBLEM SCOPING
It is a fact that we are surrounded by problems. They could be small or big, sometimes
ignored or sometimes even critical. Many times, we become so used to a problem that it
becomes a part of our life. Identifying such a problem and having a vision to solve it, is
what Problem Scoping is about.
SUSTAINABLE DEVELOPMENT GOALS
17 goals have been announced by the United nations which are termed as the Sustainable
Development Goals. The aim is to achieve these goals by the end of 2030. A pledge to do so
has been taken by all the member nations of the UN.
Scoping a problem is not that easy as we need to have a deeper understanding around it so
that the picture becomes clearer while we are working to solve it. Hence, we use the 4’Ws
Problem Canvas to help us out.
4’Ws Problem Canvas
The 4’Ws Problem canvas helps in identifying the key elements related to the problem.
Who?
The “Who” block helps in analysing the people getting
affected directly or indirectly due to it.
Under this, we find out who the ‘Stakeholders’ to
this problem are and what we know about them.
Stakeholders are the people who face this problem and
would be benefitted with the solution.
What?
Under the “What” block, you need to look into what you have
on hand. At this stage, you need to determine the nature of
the problem. What is the problem and how do you know
that it is a problem?
Under this block, you also gather evidence to prove that the
problem you have selected actually exists. Newspaper
articles, Media, announcements, etc are some examples.
Where?
Now that you know who is associated with the problem and what the problem actually is;
you need to focus on the context/situation/location of the problem.
This block will help you look into the situation in which the problem arises, the context of
it, and the locations where it is prominent.
Why?
You have finally listed down all the major elements that affect the problem directly. Now it is
convenient to understand who the people that would be benefitted by the solution are; what
is to be solved; and where will the solution be deployed. These three canvases now become
the base of why you want to solve this problem.
Thus, in the “Why” canvas, think about the benefits which the stakeholders would get from
the solution and how it will benefit them as well as the society.
Problem Solving Canvas (4‘Ws)
Data Acquisition
As we move ahead in the AI Project Cycle, we come across the second element which is :
Data Acquisition. As the term clearly mentions, this stage is about acquiring data for the
project.
Data can be a piece of information or facts and statistics collected together for reference or
analysis. Whenever we want an AI project to be able to predict an output, we need to train
it first using data.
Training Data
The observations in the training set form the experience that the algorithm uses to learn. In
supervised learning problems, each observation consists of an observed output variable and
one or more observed input variables.
Testing Data
The test set is a set of observations used to evaluate the performance of the model using
some performance metric. It is important that no observations from the training set are
included in the test set. If the test set does contain examples from the training set, it will be
difficult to assess whether the algorithm has learned to generalize from the training set or
has simply memorized it.
Over Fitting
Overfitting is a problem where the evaluation of machine learning algorithms on training
data is different from unseen data. (Memorizing the training set is called over-fitting.)
For better efficiency of an AI project, the Training data needs to be relevant and authentic.
For any AI project to be efficient, the training data should be authentic and relevant to the
problem statement scoped.
Data Features
Data features refer to the type of data you want to collect. There can be various ways in
which you can collect data. Some of them are:
Surveys
Web Scrapping
Sensors
API (Application Program Interface)
Cameras
Observations
Data Exploration
Analysing the Data and finding patterns and relations in the data is Termed As Data
Exploration. Thus, to analyse the data, you need to visualise it in some user-friendly format
so that you can:
Quickly get a sense of the trends, relationships and patterns contained within the
data.
Define strategy for which model to use at a later stage.
Communicate the same to others effectively.
To visualise data, we can use various types of visual representations.
Modelling
When it comes to machines accessing and analysing data, it needs the data in the most
basic form of numbers (which is binary – 0‘s and 1‘s) and when it comes to discovering
patterns and trends in data, the machine goes in for mathematical representations of the
same. The ability to mathematically describe the relationship between parameters is the
heart of every AI model. Thus, whenever we talk about developing AI models, it is the
mathematical approach towards analysing data which we refer to.
Generally, AI models can be classified as follows:
Machine
Learning Learning
Based Deep
AI Models Learning
Rule
Based
Rule Based Approach
Refers to the AI modelling where the rules are defined by the developer. The machine follows
the rules or instructions mentioned by the developer and performs its task accordingly.
This is known as a rule-based approach because we fed the data along with rules to the
machine and the machine after getting trained on them is now able to predict answers for
the same.
A drawback/feature for this approach is that the learning is static. The machine once
trained, does not take into consideration any changes made in the original training dataset.
That is, if you try testing the machine on a dataset which is different from the rules and data
you fed it at the training stage, the machine will fail and will not learn from its mistake.
Once trained, the model cannot improvise itself on the basis of feedbacks. Thus, machine
learning gets introduced as an extension to this as in that case, the machine adapts to
change in data and rules and follows the updated path only, while a rule-based model
does what it has been taught once.
Learning Based Approach
Refers to the AI modelling where the machine learns by itself. Under the Learning Based
approach, the AI model gets trained on the data fed to it and then is able to design a model
which is adaptive to the change in data. That is, if the model is trained with X type of data
and the machine designs the algorithm around it, the model would modify itself according to
the changes which occur in the data so that all the exceptions are handled in this case.
This is the machine learning approach which introduces the dynamicity in the model.
The learning-based approach can further be divided into three parts:
Learning Based
Supervised Unsupervised Reinforcement
Learning Learning Learning
Supervised Learning
In a supervised learning model, the dataset which is fed to the machine is labelled. In other
words, we can say that the dataset is known to the person who is training the machine only
then he/she is able to label the data. A label is some information which can be used as a
tag for data.
There are two types of Supervised Learning models:
Supervised
Learning
Classification Regression
Classification
In classification the data is classified
according to the labels. This model
works on discrete dataset which means
the data need not be continuous.
Regression
Such models work on continuous data.
Here, the data which has been fed to
the machine is continuous.
Unsupervised Learning
An unsupervised learning model works on unlabelled dataset. This means that the data
which is fed to the machine is random and there is a possibility that the person who is
training the model does not have any information regarding it. The unsupervised learning
models are used to identify relationships, patterns and trends out of the data which is fed
into it. It helps the user in understanding what the data is about and what are the major
features identified by the machine in it.
Unsupervised learning models can be further divided into two categories:
Unsupervised
Learning
Dimensionality
Clustering
Reduction
Clustering
Refers to the unsupervised learning algorithm
which can cluster the unknown data according
to the patterns or trends identified out of it.
The patterns observed might be the ones which
are known to the developer or it might even
come up with some unique patterns out of it.
Dimensionality Reduction
We humans are able to visualise upto 3-Dimensions only but according to a lot of theories
and algorithms, there are various entities which exist beyond 3-Dimensions. For example, in
Natural language Processing, the words are considered to be N-Dimensional entities. Which
means that we cannot visualise them as they exist beyond our visualisation ability.
Hence, to make sense out of it, we need to reduce their dimensions. Here, dimensionality
reduction algorithm is used. As we reduce the dimension of an entity, the information which
it contains starts getting distorted.
Hence, to reduce the dimensions and still be able to make sense out of the data, we use
Dimensionality Reduction.
Evaluation
Once a model has been made and trained, it needs to go through proper testing so that one
can calculate the efficiency and performance of the model.
Hence, the model is tested with the help of Testing Data (which was separated out of the
acquired dataset at Data Acquisition stage) and the efficiency of the model is calculated on
the basis of the parameters mentioned below:
Neural Networks
Neural networks are loosely modelled after how neurons in the human brain behave. The key
advantage of neural networks are that they are able to extract data features automatically
without needing the input of the programmer. A neural network is essentially a system of
organizing machine learning algorithms to perform certain tasks. It is a fast and efficient
way to solve problems for which the dataset is very large, such as in images.
As seen in the figure given, the larger Neural Networks tend to perform better with larger
amounts of data whereas the traditional machine learning algorithms stop improving after a
certain saturation point.
This is a representation of how neural networks work. A Neural Network is divided into
multiple layers and each layer is further divided into several blocks called nodes. Each node
has its own task to accomplish which is then passed to the next layer. The first layer of a
Neural Network is known as the input layer. The job of an input layer is to acquire data and
feed it to the Neural Network. No processing occurs at the input layer. Next to it, are the
hidden layers.
Hidden layers are the layers in which the whole processing occurs. Their name essentially
means that these layers are hidden and are not visible to the user. Each node of these
hidden layers has its own machine learning algorithm which it executes on the data
received from the input layer. The processed output is then fed to the subsequent hidden
layer
.
Script bots are easy to make Smart-bots are flexible and powerful
Script bots work around a script which Smart bots work on bigger databases
is programmed in them and other resources directly
Mostly they are free and are easy to Smart bots learn with more data
integrate to a messaging platform
No or little language processing skills Coding is required to take this up on
board
Limited functionality Wide functionality
2+3 = 3+2
5.
2/3 2.7 ≠ 2/3 3
2.7, 1 3
1.7
6
1
: &
?
:
:
( ) ( ) ( )
( )
’
( %)
( %)