Unit 4 Reasoning
Unit 4 Reasoning
UNIT-IV REASONING
Uncertainty
Artificial intelligence (AI) uncertainty is when there’s not enough information or ambiguity in
data or decision-making. It is a fundamental concept in AI, as real-world data is often noisy
and incomplete. AI systems must account for uncertainty to make informed decisions.
AI deals with uncertainty by using models and methods that assign probabilities to different
outcomes. Managing uncertainty is important for AI applications like self-driving cars and
medical diagnosis, where safety and accuracy are key.
Sources of Uncertainty in AI
There are several sources of uncertainty in AI that can impact the reliability and effectiveness
of AI systems. Here are some common sources of uncertainty in AI:
1. Data Uncertainty: AI models are trained on data, and the quality and accuracy of the data
can affect the performance of the model. Noisy or incomplete data can lead to uncertain
predictions or decisions made by the AI system.
2. Model Uncertainty: AI models are complex and can have various parameters and
hyperparameters that need to be tuned. The choice of model architecture, optimization
algorithm, and hyperparameters can significantly impact the performance of the model,
leading to uncertainty in the results.
3. Algorithmic Uncertainty: AI algorithms can be based on different mathematical
formulations, leading to different results for the same problem. For example, different
machine learning algorithms can produce different predictions for the same dataset.
4. Environmental Uncertainty: AI systems operate in dynamic environments, and changes
in the environment can affect the performance of the system. For example, an autonomous
vehicle may encounter unexpected weather conditions or road construction that can impact
its ability to navigate safely.
5. Human Uncertainty: AI systems often interact with humans, either as users or as part of
the decision-making process. Human behavior and preferences can be difficult to predict,
leading to uncertainty in the use and adoption of AI systems.
6. Ethical Uncertainty: AI systems often raise ethical concerns, such as privacy, bias, and
transparency. These concerns can lead to uncertainty in the development and deployment
of AI systems, particularly in regulated industries.
7. Legal Uncertainty: AI systems must comply with laws and regulations, which can be
ambiguous or unclear. Legal challenges and disputes can arise from the use of AI systems,
leading to uncertainty in their adoption and implementation.
8. Uncertainty in AI Reasoning: AI systems use reasoning techniques to make decisions or
predictions. However, these reasoning techniques can be uncertain due to the complexity
of the problems they address or the limitations of the data used to train the models.
9. Uncertainty in AI Perception: AI systems perceive their environment through sensors
and cameras, which can be subject to noise, occlusion, or other forms of interference. This
can lead to uncertainty in the accuracy of the data used to train AI models or the
effectiveness of AI systems in real-world applications.
10. Uncertainty in AI Communication: AI systems communicate with humans through
natural language processing or computer vision. However, language and visual cues can be
ambiguous or misunderstood, leading to uncertainty in the effective communication
between humans and AI systems.
To mitigate these sources of uncertainty, developers, and users of AI systems need to invest in
better data quality, model interpretability, and transparency, as well as engage in open
dialogue about ethical and legal considerations.
Types of Uncertainty in AI
Uncertainty in artificial intelligence (AI) refers to the lack of complete information or the
presence of variability in data and models. Understanding and modeling uncertainty is crucial
for making informed decisions and improving the robustness of AI systems. There are several
types of uncertainty in AI, including:
1. Aleatoric Uncertainty: This type of uncertainty arises from the inherent randomness or
variability in data. It is often referred to as “data uncertainty.” For example, in a
classification task, aleatoric uncertainty may arise from variations in sensor measurements
or noisy labels.
2. Epistemic Uncertainty: Epistemic uncertainty is related to the lack of knowledge or
information about a model. It represents uncertainty that can potentially be reduced with
more data or better modeling techniques. It is also known as “model uncertainty” and
arises from model limitations, such as simplifications or assumptions.
3. Parameter Uncertainty: This type of uncertainty is specific to probabilistic models, such
as Bayesian neural networks. It reflects uncertainty about the values of model parameters
and is characterized by probability distributions over those parameters.
4. Uncertainty in Decision-Making: Uncertainty in AI systems can affect the decision-
making process. For instance, in reinforcement learning, agents often need to make
decisions in environments with uncertain outcomes, leading to decision-making
uncertainty.
5. Uncertainty in Natural Language Understanding: In natural language processing (NLP),
understanding and generating human language can be inherently uncertain due to language
ambiguity, polysemy (multiple meanings), and context-dependent interpretations.
6. Uncertainty in Probabilistic Inference: Bayesian methods and probabilistic graphical
models are commonly used in AI to model uncertainty. Uncertainty can arise from the
process of probabilistic inference itself, affecting the reliability of model predictions.
7. Uncertainty in Reinforcement Learning: In reinforcement learning, uncertainty may
arise from the stochasticity of the environment or the exploration-exploitation trade-off.
Agents must make decisions under uncertainty about the outcomes of their actions.
8. Uncertainty in Autonomous Systems: Autonomous systems, such as self-driving cars or
drones, must navigate uncertain and dynamic environments. This uncertainty can pertain to
the movement of other objects, sensor measurements, and control actions.
9. Uncertainty in Safety-Critical Systems: In applications where safety is paramount, such
as healthcare or autonomous vehicles, managing uncertainty is critical. Failure to account
for uncertainty can lead to dangerous consequences.
10. Uncertainty in Transfer Learning: When transferring a pre-trained AI model to a new
domain or task, uncertainty can arise due to domain shift or differences in data
distributions. Understanding this uncertainty is vital for adapting the model effectively.
11. Uncertainty in Human-AI Interaction: When AI systems interact with humans, there
can be uncertainty in understanding and responding to human input, as well as uncertainty
in predicting human behavior and preferences.
Addressing and quantifying these various types of uncertainty is an ongoing research area in
AI, and techniques such as probabilistic modeling, Bayesian inference, and Monte Carlo
methods are commonly used to manage and mitigate uncertainty in AI systems.
We’ve just discussed the different types of uncertainty in AI. Now, let’s switch gears and
learn techniques for addressing uncertainty in AI. It’s like going from understanding the
problem to finding solutions for it.
Probabilistic logic programming (PLP) is a way to mix logic and probability to handle
uncertainty in computer programs. This is useful for computer programmers when they are
not completely sure about the facts and rules they are working with. PLP uses probabilities to
help them make decisions and learn from data. They can use different techniques, like
Bayesian logic programs or Markov logic networks, to put PLP into action. PLP is handy in
various areas of artificial intelligence, like making guesses when we’re not sure, planning
when there are risks involved, and creating models with pictures and symbols.
To deal with uncertainty in logic programming, there’s a method called fuzzy logic
programming (FLP). FLP combines regular logic with something called “fuzzy” logic. This
helps programmers express things that are a bit unclear or not black and white. FLP also helps
them make decisions and learn from this uncertain information. They can use different ways
to do FLP, like fuzzy Prolog, fuzzy answer set programming, and fuzzy description logic.
FLP is useful in various areas of artificial intelligence, like understanding language, working
with images, and making decisions when things are not very clear.
Hybrid logic programming (HLP) is a way to handle situations where things are unclear or
don’t quite match up in logic programming. HLP brings together different styles of logic
programming to help programmers work with complicated information. It lets them use
various kinds of logic to express rules and complex facts and make sense of them. There are
different methods to make HLP work, like using a mix of probabilities, fuzzy reasoning, and
handling situations where things might not always follow the same rules. HLP is useful in
various areas of artificial intelligence, like managing interactions between different computer
systems and organizing information on the internet. It can also be used for creating structured
knowledge systems.
Bayes’ Rule
Bayes’ rule is an important tool in probability that lets us adjust our best guesses when we
learn new stuff. It’s a way to use what we already know and mix it with new information to
make better guesses about what might happen. People use Bayes’ rule a lot in artificial
intelligence for things like sorting things into groups, making guesses about the future, and
deciding what to do when things are uncertain.
Here,
The posterior probability, represented by P(A|B), is the chance of event A happening when
event B has happened.
P(B|A) shows how likely event B is when event A has already happened.
The prior probability, P(A), is the initial chance of event A happening before any new
information is considered.
P(B) is the probability of event B happening, whether or not event A has happened.
In AI, Bayes’ theorem updates probabilities of hypotheses or predictions with new data or
evidence. It is helpful for dealing with uncertainty and making decisions with incomplete or
unclear information.
Bayesian Statistics
Bayesian statistics is a type of statistics that uses probability to analyze data. The framework
helps us make inferences and estimate probabilities using data and prior knowledge. Bayesian
statistics has been used in different fields to handle uncertainty and make informed choices. It
has been applied in environmental modeling, social sciences, and medical research.
Example:
Let’s consider an example of a financial risk assessment system that utilizes probabilistic
reasoning to handle uncertainty when deciding if loan applicants are creditworthy. This
system is designed to determine whether an individual or a business is a suitable candidate for
a loan based on various financial and personal factors, but these factors can be subject to
uncertainty and ambiguity.
The system uses probabilistic reasoning techniques to address uncertainty in the following
ways:
Prior Probabilities: The system assigns prior probabilities to different creditworthiness
categories based on historical data and market conditions. These prior probabilities
represent the initial beliefs about the likelihood of an applicant falling into each
creditworthiness category before taking into account the applicant’s specific financial and
personal information.
Likelihoods: The system employs statistical models to estimate the likelihood of
observing certain financial behaviors and personal characteristics given an applicant’s
creditworthiness category. For instance, it considers factors such as income, credit history,
outstanding debt, and employment status. These likelihoods may be modeled with
probabilistic distributions to account for the uncertainty inherent in the data.
Bayesian Updating: Bayes’ rule is applied to update the probabilities of different
creditworthiness categories based on the prior probabilities and the observed financial and
personal information of the applicant. The updated probabilities, referred to as posterior
probabilities, represent the revised beliefs about the likelihood of the applicant belonging
to each creditworthiness category.
Decision-Making: The system uses the posterior probabilities of different creditworthiness
categories to make a final lending decision. The decision can be based on a predetermined
threshold, a decision-making rule, or a combination of factors. For example, if the
posterior probability of an applicant being in the “low credit risk” category exceeds a
certain threshold, the system may approve the loan. Alternatively, the system may generate
a ranked list of creditworthiness categories, allowing the financial institution to decide the
terms and conditions of the loan based on the level of risk they are willing to accept.
Uncertainty:
Till now, we have learned knowledge representation using first-order logic and propositional
logic with certainty, which means we were sure about the predicates. With this knowledge
representation, we might write A→B, which means if A is true then B is true, but consider a
situation where we are not sure about whether A is true or not then we cannot express this
statement, this situation is called uncertainty.
So to represent uncertain knowledge, where we are not sure about the predicates, we need
uncertain reasoning or probabilistic reasoning.
Causes of uncertainty:
Following are some leading causes of uncertainty to occur in the real world.
Probabilistic reasoning:
o Bayes' rule
o Bayesian Statistics
Probability: Probability can be defined as a chance that an uncertain event will occur. It is
the numerical measure of the likelihood that an event will occur. The value of probability
always remains between 0 and 1 that represent ideal uncertainties.
Sample space: The collection of all possible events is called sample space.
Random variables: Random variables are used to represent the events and objects in the real
world.
Prior probability: The prior probability of an event is probability computed before observing
new information.
Posterior Probability: The probability that is calculated after all evidence or information has
taken into account. It is a combination of prior probability and new information.
Conditional probability:
Conditional probability is a probability of occurring an event when another event has already
happened.
Let's suppose, we want to calculate the event A when event B has already occurred, "the
probability of A under the conditions of B", it can be written as:
If the probability of A is given and we need to find the probability of B, then it will be given
as:
It can be explained by using the below Venn diagram, where B is occurred event, so sample
space will be reduced to set B, and now we can only calculate event A when event B is
already occurred by dividing the probability of P(A⋀ B) by P( B ).
Example:
In a class, there are 70% of the students who like English and 40% of the students who likes
English and mathematics, and then what is the percent of students those who like English also
like mathematics?
Solution:
Filtering is a technique used by Artificial Intelligence (AI) algorithms to sift through and
analyze massive data sets to find specific information, correlations, or patterns. Filtering
techniques help to eliminate irrelevant, duplicate data and present useful data to the user in
the desired format. With the great potential of AI, filtering has expanded its range of
applications to different areas, from internet data mining to social network
analysis, recommendation systems, and many more.
Filtering techniques in AI come in different forms, each with specific strengths and
applications. The following are the main types of data filtering techniques:
Rule-Based Filtering: This technique works with a set of predefined rules or patterns to
analyze and categorize data. Rule-based filtering is useful in areas that require highly
structured data, such as finance and accounting.
Content-Based Filtering: This technique filters data by analyzing the specific attributes
of the items or products. Content-based filtering is often used in recommendation systems
to predict users' taste based on their previous behavior or preferences.
Collaborative Filtering: This technique uses data similarity to recommend products or
services. It is based on the assumption that users with similar preferences will like the
same items or products.
Hybrid Filtering: This technique combines the strengths and weaknesses of different
filtering techniques to provide more accurate results. Hybrid filtering systems are often
used when there is missing data or when data is highly complex.
Filtering techniques in AI have a wide range of applications in different areas. The following
are some of the popular applications of filtering techniques:
Although filtering techniques in AI provide several benefits, there are also some challenges
that need to be addressed. The following are some of the main challenges with filtering
techniques:
Data Bias: Filtering techniques can suffer from data bias, where the data used to train the
algorithm is not representative of the target population.
Data Privacy: Filtering techniques often rely on personal data, which can raise concerns
about data privacy and security.
Overfitting: Filtering techniques can suffer from overfitting, where the model is trained
to fit the training data too well and fails to generalize to unseen data.
Unbalanced Data: Filtering techniques can suffer from unbalanced data, where the target
class is rare in the data set, leading to poor classification performance.
Assumes:
Used when system states are hidden and evolve over time.
Common in:
Speech recognition
Bioinformatics
Natural language processing
Applications of Filtering:
What is Prediction?
Prediction in AI refers to forecasting future values or states based on current and historical
data. It’s at the core of many machine learning tasks.
Purpose:
1. Supervised Learning
Algorithms:
Linear Regression
Decision Trees
Support Vector Machines
Neural Networks
Techniques:
Applications of Prediction:
Spam detection
Predictive maintenance
Healthcare diagnosis
Recommendation systems (e.g., Netflix, Amazon)
Natural language generation (e.g., text completion)
Hidden Markov Models (HMMs) are a type of probabilistic model that are commonly used in
machine learning for tasks such as speech recognition, natural language processing, and
bioinformatics. They are a popular choice for modelling sequences of data because they can
effectively capture the underlying structure of the data, even when the data is noisy or
incomplete. In this article, we will give a comprehensive overview of Hidden Markov Models,
including their mathematical foundations, applications, and limitations.
What are Hidden Markov Models?
Now, we will explore some of the key applications of HMMs, including speech recognition,
natural language processing, bioinformatics, and finance.
Speech Recognition
One of the most well-known applications of HMMs is speech recognition. In this field,
HMMs are used to model the different sounds and phones that makeup speech. The
hidden states, in this case, correspond to the different sounds or phones, and the
observations are the acoustic signals that are generated by the speech. The goal is to
estimate the hidden state sequence, which corresponds to the transcription of the speech,
based on the observed acoustic signals. HMMs are particularly well-suited for speech
recognition because they can effectively capture the underlying structure of the speech,
even when the data is noisy or incomplete. In speech recognition systems, the HMMs are
usually trained on large datasets of speech signals, and the estimated parameters of the
HMMs are used to transcribe speech in real time.
Natural Language Processing
Another important application of HMMs is natural language processing. In this field,
HMMs are used for tasks such as part-of-speech tagging, named entity recognition,
and text classification. In these applications, the hidden states are typically associated
with the underlying grammar or structure of the text, while the observations are the
words in the text. The goal is to estimate the hidden state sequence, which corresponds to
the structure or meaning of the text, based on the observed words. HMMs are useful in
natural language processing because they can effectively capture the underlying structure
of the text, even when the data is noisy or ambiguous. In natural language processing
systems, the HMMs are usually trained on large datasets of text, and the estimated
parameters of the HMMs are used to perform various NLP tasks, such as text
classification, part-of-speech tagging, and named entity recognition.
Bioinformatics
HMMs are also widely used in bioinformatics, where they are used to model sequences
of DNA, RNA, and proteins. The hidden states, in this case, correspond to the different
types of residues, while the observations are the sequences of residues. The goal is to
estimate the hidden state sequence, which corresponds to the underlying structure of the
molecule, based on the observed sequences of residues. HMMs are useful in
bioinformatics because they can effectively capture the underlying structure of the
molecule, even when the data is noisy or incomplete. In bioinformatics systems, the
HMMs are usually trained on large datasets of molecular sequences, and the estimated
parameters of the HMMs are used to predict the structure or function of new molecular
sequences.
Finance
Finally, HMMs have also been used in finance, where they are used to model stock
prices, interest rates, and currency exchange rates. In these applications, the hidden states
correspond to different economic states, such as bull and bear markets, while the
observations are the stock prices, interest rates, or exchange rates. The goal is to estimate
the hidden state sequence, which corresponds to the underlying economic state, based on
the observed prices, rates, or exchange rates. HMMs are useful in finance because they
can effectively capture the underlying economic state, even when the data is noisy or
incomplete. In finance systems, the HMMs are usually trained on large datasets of
financial data, and the estimated parameters of the HMMs are used to make predictions
about future market trends or to develop investment strategies.
Now, we will explore some of the key limitations of HMMs and discuss how they can impact
the accuracy and performance of HMM-based systems.
The Kalman Filter is an iterative process that estimates the state of a dynamic system from a
series of incomplete and noisy measurements. It is recursive, meaning that it can run in real-
time using only the current input measurements and the previously calculated state and its
uncertainty matrix; no additional past information is required.
The Kalman Filter operates in two steps: the "predict" or "time update" phase and the
"update" or "measurement update" phase.
Predict Phase
In the predict phase, the Kalman Filter uses the state from the previous time step to produce
estimates of the current state. This prediction includes the estimation of the system's state
variables and the uncertainty of the estimate. The uncertainty is often expressed as a
covariance matrix, which is a measure of the "spread" or the expected accuracy of the
prediction.
Update Phase
During the update phase, the current prediction is combined with the current observation to
refine the state estimate. This step adjusts the predicted state by a factor proportional to the
difference between the actual measurement and the prediction. The Kalman Filter uses the
covariance matrix to weigh the accuracy of the prediction against the accuracy of the new
measurement, thus updating the state estimate and its uncertainty.
These two phases are repeated in a loop, with each iteration refining the estimates. This
process allows the filter to react to new measurements and improve the estimate over time,
which is why it is particularly useful for systems where the measurements are uncertain or
vary over time.
The Kalman Filter is advantageous because it is a linear estimator that is optimal under the
assumption that the errors are Gaussian. It is computationally efficient, which allows it to run
in real-time applications, and it can handle cases where the noise statistics are not fully known.
However, the Kalman Filter has limitations. It assumes that the process and measurement
noise are both Gaussian and white, and that the system dynamics are linear. For systems that
do not meet these assumptions, extensions to the Kalman Filter, such as the Extended Kalman
Filter (EKF) and the Unscented Kalman Filter (UKF), have been developed to handle non-
linear systems.
1. Introduction
Artificial Intelligence (AI) has made significant strides by integrating probabilistic reasoning
into decision-making and inference systems. One of the most powerful tools in this domain is
the Bayesian Network. While static Bayesian Networks are effective for modeling
probabilistic relationships between variables at a single point in time, they fall short when
dealing with sequences or processes that evolve over time. This is where Dynamic Bayesian
Networks (DBNs) come into play.
This article explores the theoretical underpinnings, structure, learning, inference methods, and
applications of Dynamic Bayesian Networks in the field of AI.
Before diving into DBNs, it's important to understand Bayesian Networks (BNs).
3.1 Definition
4. Structure of a DBN
Each time slice contains a subset of variables, usually representing observable and hidden
(latent) states.
Determine the most probable sequence of states given the entire sequence of observations.
6. Learning in DBNs
Learning a DBN involves estimating both the structure and parameters from data.
Assuming the structure is known, parameter learning estimates the conditional probabilities
using:
Structure learning is more complex, especially when dealing with hidden variables or
incomplete data. It involves:
Score-based methods (e.g., Bayesian Information Criterion, Akaike Information Criterion)
For example, HMMs are often used in speech recognition, but DBNs can model complex
multimodal interactions in video analysis or robotics.
8.3 Robotics
8.5 Finance
Models stock prices, market trends, and customer behavior.
The way people interact with digital gadgets and systems has changed dramatically in recent
years due to noteworthy developments in speech recognition technology. Speech recognition
is a crucial component of artificial intelligence (AI) that helps close the communication gap
between people and machines. Automation, accessibility features, virtual assistants,
transcription services, and other uses for machine understanding and interpretation of spoken
language are made possible by this technology. The intriguing field of voice recognition in
artificial intelligence, along with its services, difficulties, and prospects, will all be covered in
this article.
Speech recognition technology, also known as Automatic Speech Recognition (ASR), makes
it possible for computers and artificial intelligence (AI) systems to translate spoken words
into text. There are several steps in this process:
1. Decoding: Based on the data obtained in the above processes, the last step includes
choosing the most probable translation for the spoken words.
2. Feature extraction: In this stage, the audio input is processed to extract characteristics
such as Mel-frequency cepstral coefficients (MFCCs), which give the system the
necessary information to recognize the sound.
3. Acoustic Analysis: The audio signal is captured by the system, which then dissects it into
its constituent elements, such as prosody and phonemes.
4. Language Modeling: To increase recognition accuracy, language models are used to
comprehend the semantics and grammatical structure of spoken words.
5. Acoustic Modeling: To link the retrieved characteristics with recognized phonetic
patterns and language context, the system applies statistical models.
Many companies use cutting-edge technology like artificial intelligence, machine learning,
and neural networks to develop voice recognition software. Technologies like Cortana, Alexa,
Siri and Google Assistant have altered how people use electronics and technology. They
include automobiles, cell phones, home security systems, and more.
Recall that speech and voice recognition are two different things. Speech recognition
translates spoken words into text by first identifying them in an audio recording of a speaker.
On the other hand, speech recognition can only identify pre-programmed spoken instructions.
The sole commonality between these two approaches is the conversion of sound to text.
1. Audio Input: A microphone is usually used to record the audio input, which starts the
process. Any spoken human speech, including commands and conversations, can be used
as this audio input.
2. Preprocessing: To enhance its quality and prepare it for analysis, the raw audio signal is
preprocessed. This might be signal amplification, noise reduction, or other methods to
improve the audio data.
3. Language Modeling: Language models are used to comprehend the semantics and
grammatical structure of spoken words. By assisting the system in understanding the
context and connections between words, these models increase the accuracy of word
recognition. When it comes to managing homophones?words that sound identically but
have distinct meanings-and the order of words and sentence structure changes, language
modelling is incredibly crucial.
4. Decoding: By integrating the data from the acoustic and linguistic models, the system
decodes the spoken words. It assesses several word combinations and determines which
transcription is more plausible based on statistical probability.
5. Output: The recognized language or a command that may be applied to several different
situations is the ultimate output. This output can be utilized for transcription, operating a
device, giving instructions to a virtual assistant, and other similar tasks.
Recognition of Speech Machines can now comprehend and interpret human language thanks
to the closely connected sciences of artificial intelligence (AI) and natural language
processing (NLP). NLP covers various applications, such as language translation, sentiment
analysis, and text summarization, whereas voice recognition AI concentrates on translating
spoken words into digital text or commands.
Making it possible for robots to comprehend and interpret human language similarly to how
humans do is one of the main objectives of natural language processing (NLP). This entails
knowing the broader context and meaning of the words and recognizing them individually.
For instance, depending on the situation, "I saw a bat" might mean several things. Either the
animal or a piece of athletic gear might be the subject.
Deep Neural Networks (DNNs): Used widely in voice recognition artificial intelligence,
DNNs are a machine learning model. DNNs represent intricate links between the speech
input and the associated text output by employing a hierarchy of layers.
Hidden Markov Models (HMMs): AI voice recognition uses Hidden Markov Models
(HMMs), which are statistical models. To match input speech to the most likely sound
sequence, HMMs first model the probability distribution of speech sounds.
Convolutional Neural Networks (CNNs): Artificial Intelligence for speech recognition
has also made use of Convolutional Neural Networks (CNNs), a class of machine
learning model that is frequently employed in image recognition. To find pertinent
characteristics, CNNs process incoming speech signals by applying filters.
Among the most recent developments in AI voice recognition are:
End-to-end models: These models are made to translate speech impulses directly into
text, eliminating the need for any intermediary stages. These models have demonstrated
the potential for raising voice recognition AI's precision and effectiveness.
Multimodal models: These enable more intuitive and natural interactions between
machines and humans by fusing voice recognition Intelligence with other modalities,
including vision or touch.
Transformer-based models: BERT and GPT are two examples of transformer-based
models that have shown great success in tasks related to natural language processing and
are now being used in artificial intelligence for voice recognition.
Data augmentation: Increasing the data used for training for speech recognition AI
models will increase their accuracy and resilience. Data augmentation strategies include
introducing background noise and modifying the speaking tempo.
Even though voice recognition technology has advanced significantly, several issues still exist:
In many domains and uses, artificial intelligence is used as a commercial solution for speech
recognition. Voice-activated audio content assistants, call centres, ATMs, and more benefit
from AI's more natural user interactions with hardware and software and its increased
accuracy in data transcription.
Decisions Making
Machine learning takes AI decision-making a step further. By learning from past experiences
(data and outcomes), AI systems can continuously improve their decision-making abilities.
This AI-powered automation helps businesses make faster, more accurate choices in various
fields, from retail and finance to telecom & media.
Compare this to traditional human and computer processes. These require complete data,
formatted a certain way, and carefully aggregated. Traditional data analysis relies heavily on
human expertise and manual processes. Analysts must collect data, clean it, and then use
statistical methods to draw conclusions.
It’s very hard for people to make quick decisions when faced with varying opinions from
different sources, or when they have a personal stake in the decision. People are also creatures
of habit and don’t always learn from past mistakes.
AI can easily manage messy and contradictory input. It can process more data than a person
could possibly take in and easily identify the signals in the noise. With machine learning,
every decision’s outcome trains the system to make better decisions in the future.
Are you curious about how artificial intelligence helps in decision-making? Here are seven
ways:
Today, business moves fast, and the ability to make quick decisions is crucial. AI systems can
process data and generate insights in real time, enabling organizations to respond swiftly to
changing market conditions and customer needs.
2. Supercharge productivity
Let’s face it, even the most dedicated employees can only work so many hours in a day.
That’s where AI comes in. It’s like having a tireless assistant working around the clock to
process data and provide insights. Need real-time analysis or rapid solutions? AI has got your
back.
By leveraging AI’s ability to work 24/7, people can supercharge their productivity and free up
time for strategic thinking and decision-making. Of course, AI is no substitute for human
judgment – it’s a tool that enhances our capabilities. Think of it as a productivity booster that
helps you work smarter, not harder.
3. Improve accuracy
AI algorithms excel at processing and analyzing vast amounts of structured and unstructured
data. By identifying complex patterns, correlations, and anomalies that humans might
overlook, AI can provide more accurate insights and predictions.
AI plays a crucial role in identifying and mitigating potential risks and threats. By analyzing
vast amounts of historical data, AI algorithms can detect patterns and anomalies that indicate
potential risks, such as fraudulent activities, market fluctuations, or supply chain disruptions.
This early detection allows organizations to proactively mitigate risks and avoid costly
mistakes. AI can also run countless scenario simulations and predict potential outcomes,
enabling decision-makers to assess the impact of different strategies and make informed
choices that minimize risk exposure.
5. Increase efficiency
AI can also identify inefficiencies in existing processes and suggest improvements, leading to
increased productivity and operational efficiency. Furthermore, AI’s ability to process data
24/7 ensures that decision-making processes continue uninterrupted, even outside regular
business hours.
6. Create consistency
Picture a world where every decision, from the mundane to the monumental, adheres to a
meticulously standardized playbook. By leveraging machine learning algorithms and
predefined rules, AI promises to ensure that decisions are made consistently across the board,
eliminating the fluctuations that can plague human judgment.
Whether approving loans, managing supply chains, or executing repetitive tasks, AI acts as
the great equalizer, applying the same criteria and logic every single time. It’s an unwavering
consistency that streamlines processes and promotes fairness and impartiality. With AI at the
helm, organizations can trust that decisions are made objectively, based on data and
established best practices, rather than individual biases or whims.
By leveraging AI’s infinite institutional memory, organizations can make smarter, more
informed decisions that build upon past success while avoiding the pitfalls of yesteryear. AI
becomes the keeper of corporate wisdom, ensuring that valuable knowledge is preserved and
applied to drive the business forward.
Forms of learning
Machine learning algorithms fall into five broad categories: supervised learning, unsupervised
learning, semi-supervised learning, self-supervised and reinforcement learning.
Supervised machine learning is a type of machine learning where the model is trained on a
labeled dataset (i.e., the target or outcome variable is known). For instance, if data scientists
were building a model for tornado forecasting, the input variables might include date, location,
temperature, wind flow patterns and more, and the output would be the actual tornado activity
recorded for those days.
Supervised learning is commonly used for risk assessment, image recognition, predictive
analytics and fraud detection, and comprises several types of algorithms.
The most common unsupervised learning method is cluster analysis, which uses clustering
algorithms to categorize data points according to value similarity (as in customer
segmentation or anomaly detection). Association algorithms allow data scientists to identify
associations between data objects inside large databases, facilitating data visualization and
dimensionality reduction.
K-means clustering—assigns data points into K groups, where the data points
closest to a given centroid are clustered under the same category and K represents
clusters based on their size and level of granularity. K-means clustering is commonly
used for market segmentation, document clustering, image segmentation and image
compression.
Hierarchical clustering—describes a set of clustering techniques, including
agglomerative clustering—where data points are initially isolated into groups and
then merged iteratively based on similarity until one cluster remains—and divisive
clustering—where a single data cluster is divided based on the differences between
data points.
Probabilistic clustering—helps solve density estimation or “soft” clustering
problems by grouping data points based on the likelihood that they belong to a
particular distribution.
Unsupervised ML models are often behind the “customers who bought this also bought…”
types of recommendation systems.
Self-supervised learning (SSL) enables models to train themselves on unlabeled data, instead
of requiring massive annotated and/or labeled datasets. SSL algorithms, also called predictive
or pretext learning algorithms, learn one part of the input from another part, automatically
generating labels and transforming unsupervised problems into supervised ones. These
algorithms are especially useful for jobs like computer vision and NLP, where the volume of
labeled training data needed to train models can be exceptionally large (sometimes
prohibitively so).
4. Reinforcement learning
Reinforcement learning, also called reinforcement learning from human feedback (RLHF), is
a type of dynamic programming that trains algorithms using a system of reward and
punishment. To deploy reinforcement learning, an agent takes actions in a specific
environment to reach a predetermined goal. The agent is rewarded or penalized for its actions
based on an established metric (typically points), encouraging the agent to continue good
practices and discard bad ones. With repetition, the agent learns the best strategies.
Reinforcement learning algorithms are common in video game development and are
frequently used to teach robots how to replicate human tasks.
5. Semi-supervised learning
The fifth type of machine learning technique offers a combination between supervised and
unsupervised learning.
Semi-supervised learning algorithms are trained on a small labeled dataset and a large
unlabeled dataset, with the labeled data guiding the learning process for the larger body of
unlabeled data. A semi-supervised learning model might use unsupervised learning to identify
data clusters and then use supervised learning to label the clusters.
Generative adversarial networks (GANs)—deep learning tool that generates unlabeled data by
training two neural networks—are an example of semi-supervised machine learning.
Regardless of type, ML models can glean data insights from enterprise data, but their
vulnerability to human/data bias make responsible AI practices an organizational imperative.
Knowledge in learning
In AI, knowledge refers to the information an agent/system has about the world, which can
include:
3. Reinforcement Learning
Knowledge is gained through trial and error by interacting with the environment.
Example: A robot learns how to navigate a maze by bumping into walls and getting rewards.
Type of knowledge gained: Policies (what actions to take in different situations), value
functions.
4. Transfer Learning
An AI system uses knowledge from one domain to learn faster in a related domain.
Statistical learning refers to a framework where statistics and machine learning overlap. It
involves using algorithms that apply statistical theory to:
Statistical learning methods are typically used for supervised or unsupervised learning tasks
and provide the foundation for many machine learning algorithms.
1. Linear Models
Linear models are some of the simplest and most widely used statistical learning methods.
They assume that the relationship between input variables (features) and the target (output) is
linear.
Pros:
Cons:
2. Bayesian Methods
Bayesian methods apply Bayes' Theorem to infer probabilities and update beliefs
based on new data.
Example Models:
Pros:
Cons:
SVM is a supervised learning method used for classification and regression. It aims to find
the hyperplane that best separates different classes in the feature space, maximizing the
margin between them.
Example:
For binary classification, the algorithm finds a hyperplane that separates data into two classes
with the largest margin.
Uses the kernel trick to handle non-linear boundaries by transforming the feature space.
Pros:
Cons:
4. Decision Trees
Decision trees split the data into smaller subsets based on the most significant feature at each
step. They are particularly useful for classification and regression tasks.
Example Models:
CART (Classification and Regression Trees): Builds binary trees based on Gini impurity or
mean squared error (MSE) to split nodes.
Pros:
Cons:
Can be unstable (small changes in data may cause large changes in the tree).
KNN is a non-parametric method that classifies a data point based on the majority class of
its neighbors.
How it works: For a new data point, KNN finds the K nearest neighbors (based on distance
metrics like Euclidean) and assigns the most common class among those neighbors.
Pros:
Cons:
6. Ensemble Methods
Ensemble methods combine multiple models to improve performance and reduce overfitting.
Example Models:
Pros:
Cons:
Agent(): An entity that can perceive/explore the environment and act upon it.
Environment(): A situation in which an agent is present or surrounded by. In RL, we
assume the stochastic environment, which means it is random in nature.
Action(): Actions are the moves taken by an agent within the environment.
State(): State is a situation returned by the environment after each action taken by the
agent.
Reward(): A feedback returned to the agent from the environment to evaluate the action
of the agent.
Policy(): Policy is a strategy applied by the agent for the next action based on the current
state.
Value(): It is expected long-term retuned with the discount factor and opposite to the
short-term reward.
Q-value(): It is mostly similar to the value, but it takes one additional parameter as a
current action (a).
Goal -- define probability distribution over set of strings
Unigram, bigram, n-gram
Count using corpus but need smoothing:
1. add-one
2. Linear interpolation
PCFGs
Learning PCFGs
Information Retrieval
IR 2
IR3
For this, need model of how queries are related to docs. Bag of words: freq of
words in doc., naïve Bayes.
Good example pp 842-843.
Evaluating IR
Case
Stems
Synonyms
Spelling correction
Metadata --keywords
IR Presentation
Agglomerative
K-means
IR Implementation
CSC172!
Lexicon with “stop list”,
“inverted” index: where words occur
Match with vectors: vectorof freq of words dotted with query terms.
Let’s break down perception in AI, including the technologies used and real-world
applications:
Computer Vision allows AI to understand images or videos, simulating human vision. The
goal is to extract useful information from visual data like images or videos.
Semantic Segmentation: Dividing an image into regions with distinct meanings (e.g.,
labeling pixels as part of the road, pedestrians, or vehicles in an image).
Optical Flow: Analyzing the motion of objects between consecutive frames in a video to
estimate their movement.
Applications:
Self-driving cars use computer vision to detect road signs, pedestrians, other vehicles, and
obstacles.
AI can perceive audio signals through speech recognition, sound classification, and audio
processing.
Speech Recognition: Converts spoken language into text, allowing AI systems to understand
verbal commands.
Applications:
Virtual assistants (like Siri or Alexa) use speech recognition to understand and respond to
voice commands.
Call center systems use speech emotion recognition to detect the mood of customers.
Force Sensing: AI uses sensors to detect the pressure or force applied by or on an object.
Tactile Feedback: AI can simulate the sense of touch, allowing robots or systems to feel their
environment.
Applications:Robotic surgery: Surgeons control robotic arms, and the system provides
haptic feedback, mimicking the sense of touch.
Virtual reality (VR) and gaming: Haptic gloves and controllers provide tactile feedback for
more immersive experiences.
Sensor Fusion
Sensor fusion refers to combining data from multiple sensors (e.g., cameras, LiDAR, radar,
microphones) to create a more complete and accurate perception of the environment.
It enables AI systems to overcome the limitations of individual sensors, such as poor visibility
or low resolution, by integrating data from different sources.
Example:
Autonomous vehicles combine data from LiDAR (light detection and ranging), radar, and
cameras to create a 360-degree understanding of their environment. Each sensor has
strengths and weaknesses, so fusing data from all of them allows for more accurate perception.
Applications:
Self-driving cars use sensor fusion to combine data from LiDAR, radar, and cameras for
accurate perception and decision-making.
Drones use multiple sensors (camera, GPS, IMU) to navigate and map environments.
In addition to visual, auditory, and tactile data, some AI systems perceive environmental
factors like temperature, humidity, and air quality.
This type of perception can also be used for context awareness (understanding the
environment's current state) to adjust behavior.
Key Techniques:
Contextual Awareness: Analyzing data from various sensors and understanding the broader
context of the environment (e.g., a robot might adjust its behavior when it senses that the floor
is wet).
Environmental Monitoring: Using AI for monitoring things like air pollution, temperature
changes, or natural disasters.
Applications:
Smart homes use sensors to monitor air quality, temperature, and humidity, adjusting
conditions based on occupant behavior and preferences.
Environmental monitoring systems use AI to predict and analyze pollution levels, climate
change, and natural disaster events.
Perception in Robotics
Perception is one of the key components in robotics, enabling robots to sense their
environment and take action accordingly.
Visual Perception: Robots with cameras or vision systems can "see" their surroundings,
similar to how humans use their eyes.
Example: A robot arm uses a camera to identify and pick up objects from a table.
Auditory Perception: Robots can use microphones and speech recognition to "hear" sounds
and respond appropriately.
Example: A robot uses audio sensors to recognize a person’s voice command and perform
tasks like opening a door.
Tactile Perception: Robots with touch sensors can detect pressure, temperature, or texture,
and adjust their actions accordingly.
Example: A robot uses tactile sensors to feel the texture of an object before determining how
to handle it.
Example: A robot in a warehouse uses visual, tactile, and proximity sensors to navigate, avoid
obstacles, and pick items.
Computer Vision: Detects pedestrians, road signs, traffic lights, and other vehicles.
Autonomous vehicles integrate all this sensor data to understand their environment in real-
time, make decisions, and navigate autonomously.
Future of Perception in AI
More Robust Sensory Fusion: As more sensors (e.g., 5G, advanced radar) become available,
AI systems will better integrate data to create richer, more accurate perceptions.
Improved Multimodal Perception: Combining multiple forms of sensory data (e.g., visual,
auditory, and tactile) for more comprehensive understanding.
Greater Contextual Awareness: AI systems will increasingly be able to perceive and react
to complex, changing environments with better context awareness (e.g., adapting to weather
changes or human behavior).