Unit 1 Bda
Unit 1 Bda
1|Page
UNIT I
UNDERSTANDING BIG DATA
Introduction to big data – convergence of key trends – unstructured data – industry examples of
big data – web analytics – big data applications– big data technologies – introduction to Hadoop –
open source technologies – cloud and big data – mobile business intelligence – Crowd sourcing
analytics – inter and trans firewall analytics.
Big data definitions may vary slightly, but it will always be described in terms of volume, velocity,
and variety. These big data characteristics are often referred to as the “3 Vs of
Volume
As its name suggests, the most common characteristic associated with big data is its high
volume. This describes the enormous amount of data that is available for collection and
produced from a variety of sources and devices on a continuous basis.
Velocity
Big data velocity refers to the speed at which data is generated. Today, data is often
produced in real time or near real time, and therefore, it must also be
2|Page
processed, accessed, and analyzed at the same rate to have any meaningful impact.
Variety
Data is heterogeneous, meaning it can come from many different sources and can be
structured, unstructured, or semi-structured. More traditional structured data (such as data in
spreadsheets or relational databases) is now supplemented by unstructured text, images,
audio, video files, or semi- structured formats like sensor data that can’t be organized in a
fixed data schema. big data” and were first defined by Gartner in 2001.
In addition to these three original Vs, three others that are often mentioned in relation to harnessing the
power of big data: veracity, variability, and value.
Veracity:
Big data can be messy, noisy, and error-prone, which makes it difficult to control the quality
and accuracy of the data. Large datasets can be unwieldy and confusing, while smaller
datasets could present an incomplete picture. The higher the veracity of the data, the more
trustworthy it is.
Variability:
The meaning of collected data is constantly changing, which can lead to inconsistency over
time. These shifts include not only changes in context and interpretation but also data
collection methods based on the information that companies want to capture and analyze.
Value:
It’s essential to determine the business value of the data you collect. Big data must contain
the right data and then be effectively analyzed in order to yield insights that can help drive
decision-making.
o Social networking sites: Facebook, Google, LinkedIn all these sites generate huge amount of
data on a day to day basis as they have billions of users worldwide.
o E-commerce site: Sites like Amazon, Flipkart, Alibaba generates huge number of logs from
which users buying trends can be traced.
o Weather Station: All the weather station and satellite gives very huge data which are stored
and manipulated to forecast weather.
o Telecom company: Telecom giants like Airtel, Vodafone study the user trends and
accordingly publish their plans and for this they store the data of its million users.
o Share Market: Stock exchange across the world generates huge amount of data through its
daily transaction.
3|Page
Making big data work requires three main actions:
1. Integration:
Big data collects terabytes, and sometimes even petabytes, of raw data from many sources that
must be received, processed, and transformed into the format that business users and analysts
need to start analyzing it.
2. Management:
Big data needs big storage, whether in the cloud, on-premises, or both. Data must also be stored in
whatever form required. It also needs to be processed and made available in real time.
Increasingly, companies are turning to cloud solutions to take advantage of the unlimited compute
and scalability.
3. Analysis:
The final step is analyzing and acting on big data—otherwise, the investment won’t be worth it.
Beyond exploring the data itself, it’s also critical to communicate and share insights across the
business in a way that everyone can understand. This includes using tools to create data
visualizations like charts, graphs, and dashboards.
1. Descriptive analytics
Descriptive analytics refers to data that can be easily read and interpreted. This data helps create
reports and visualise information that can detail company profits and sales.
Example: During the pandemic, a leading pharmaceutical company conducted data analysis on its
offices and research labs. Descriptive analytics helped them identify consolidated unutilised spaces
and departments, saving the company millions of pounds.
4|Page
2. Diagnostics analytics
Diagnostics analytics helps companies understand why a problem occurred. Big data technologies and
tools allow users to mine and recover data that helps dissect an issue and prevent it from happening in
the future.
Example: An online retailer’s sales have decreased even though customers continue to add items to
their shopping carts. Diagnostics analytics helped to understand that the payment page was not
working correctly for a few weeks.
3. Predictive analytics
Predictive analytics looks at past and present data to make predictions. With artificial intelligence
(AI), machine learning, and data mining, users can analyse the data to predict market trends.
Example: In the manufacturing sector, companies can use algorithms based on historical data to
predict if or when a piece of equipment will malfunction or break down.
4. Prescriptive analytics
Prescriptive analytics solves a problem, relying on AI and machine learning to gather and use data for
risk management.
Example: Within the energy sector, utility companies, gas producers, and pipeline owners identify
factors that affect the price of oil and gas to hedge risks.
1. Structured data
Structured data is data whose elements are addressable for effective analysis. It has been
organized into a formatted repository that is typically a database. It concerns
5|Page
all data which can be stored in database in a table with rows and columns. They have relational keys
and can easily be mapped into pre-designed fields. Today, those data are most processed in the
development and simplest way to manage information. Example: Relational data.
2. Semi-Structured data
Semi-structured data is information that does not reside in a relational database but that has
some organizational properties that make it easier to analyze. With some processes, you can store
them in the relation database (it could be very hard for some kind of semi-structured data), but Semi-
structured exist to ease space. Example: XML data.
3. Unstructured data
Unstructured data is a data which is not organized in a predefined manner or does not have a
predefined data model, thus it is not a good fit for a mainstream
relational database. So for Unstructured data, there are alternative platforms for storing and managing,
it is increasingly prevalent in IT systems and is used by organizations in a variety of business
intelligence and analytics applications. Example: Word, PDF, Text, Media logs.
Unstructured data is the data which does not conforms to a data model and has no easily
identifiable structure such that it can not be used by a computer program easily. Unstructured data is
not organised in a pre-defined manner or does not have a pre-defined data model, thus it is not a good
fit for a mainstream relational database.
From 80% to 90% of data generated and collected by organizations is unstructured, and its
volumes are growing rapidly — many times faster than the rate of growth for structured databases.
However, unstructured data has historically been very difficult to analyze. With the help of AI
and machine learning, new software tools are emerging that can search through vast quantities of it to
uncover beneficial and actionable business intelligence.
Fits into designated fields (e.g., zip codes, phone No predefined data model or
Organization
numbers, credit cards) consistent structure
6|Page
Aspect Structured Data Unstructured Data
Email: Email message fields are unstructured and cannot be parsed by traditional analytics
tools. That said, email metadata affords it some structure, and explains why email is
sometimes considered semi-structured data.
Text files: This category includes word processing documents, spreadsheets, presentations,
email, and log files.
Social media and websites: data from social networks like Twitter, LinkedIn, and Facebook,
and websites such as Instagram, photo-sharing sites, and YouTube.
Mobile and communications data: For this category, look no further than text messages,
phone recordings, collaboration software, chat, and instant messaging.
Media: This data includes digital photos, audio, and video files. Here are
some examples of unstructured data generated by machines:
Scientific data: This includes oil and gas surveys, space exploration, seismic imagery, and
atmospheric data.
Digital surveillance: This category features data like reconnaissance photos and videos.
Satellite imagery: This data includes weather data, land forms, and military movements.
le business intelligence.
8|Page
BIG DATA INDUSTRY APPLICATIONS
Here are some of the sectors where Big Data is actively used:
Ecommerce - Predicting customer trends and optimizing prices are a few of the ways e-
commerce uses Big Data analytics
Marketing - Big Data analytics helps to drive high ROI marketing operations, which result in
improved sales
Education - Used to develop new and improve existing courses based on market requirements
Healthcare - With the help of a patient’s medical history, Big Data analytics is
used to predict how likely they are to have health issues
Media and entertainment - Used to understand the demand of shows, movies, songs, and more
to deliver a personalized recommendation list to its users
Banking - Customer income and spending patterns help to predict the likelihood of choosing
various banking offers, like loans and credit cards
Telecommunications - Used to forecast network capacity and improve customer experience
Government - Big Data analytics helps governments in law enforcement, among other things
In today’s world, there are a lot of data. Big companies utilize those data for their business growth. By
analyzing this data, the useful decision can be made in various cases as discussed below:
2. Recommendation:
By tracking customer spending habit, shopping behavior, Big retails store provide a
recommendation to the customer. E-commerce site like Amazon, Walmart, Flipkart does product
recommendation. They track what product a customer is searching, based on that data they
recommend that type of product to that customer.
As an example, suppose any customer searched bed cover on Amazon. So, Amazon got data
that customer may be interested to buy bed cover. Next time when that customer will go to any google
9|Page
page, advertisement of various bed covers will be seen. Thus, advertisement of the right product to the
right customer can be sent.
YouTube also shows recommend video based on user’s previous liked, watched video type.
Based on the content of a video, the user is watching, relevant advertisement is shown during video
running. As an example suppose someone watching a tutorial video of Big data, then advertisement of
some other big data course will be shown during that video.
7. IoT:
Manufacturing company install IOT sensor into machines to collect operational
data. Analyzing such data, it can be predicted how long machine will work without any problem when
it requires repairing so that company can take action before the situation when machine facing a lot of
issues or gets totally down. Thus, the cost to replace the whole machine can be saved.
In the Healthcare field, Big data is providing a significant contribution. Using big data tool,
data regarding patient experience is collected and is used by doctors to give better treatment. IoT
device can sense a symptom of probable coming disease in the human body and prevent it from giving
advance treatment. IoT Sensor placed near-patient, new-born baby constantly keeps track of various
health condition like heart bit rate, blood presser, etc. Whenever any parameter crosses the safe limit,
an alarm sent to a doctor, so that they can take step remotely very soon.
10 | P a g e
8. Education Sector:
Online educational course conducting organization utilize big data to search candidate,
interested in that course. If someone searches for YouTube tutorial video on a subject, then online or
offline course provider organization on that subject send ad online to that person about their course.
9. Energy Sector:
Smart electric meter read consumed power every 15 minutes and sends this read data to the
server, where data analyzed and it can be estimated what is the time in a day when the power load is
less throughout the city. By this system manufacturing unit or housekeeper are suggested the time
when they should drive their heavy machine in the night time when power load less to enjoy less
electricity bill.
Big data technologies can be categorized into four main types: data storage, data mining, data
analytics, and data visualization [2]. Each of these is associated with certain tools, and you’ll want to
choose the right tool for your business needs depending on the type of big data technology required.
1. Data storage
Big data technology that deals with data storage has the capability to fetch, store, and manage big data.
It is made up of infrastructure that allows users to store the data so that it is convenient to access.
Most data storage platforms are compatible with other programs. Two commonly used tools are
Apache Hadoop and MongoDB.
Apache Hadoop: Apache is the most widely used big data tool. It is an open- source software
platform that stores and processes big data in a distributed computing environment across
hardware clusters. This distribution allows for faster data processing. The framework is
designed to reduce bugs or faults, be scalable, and process all data formats.
MongoDB: MongoDB is a NoSQL database that can be used to store large volumes of data.
Using key-value pairs (a basic unit of data), MongoDB categorizes documents into
collections. It is written in C, C++, and JavaScript, and is one of the most popular big data
databases because it can manage and store unstructured data with ease.
2. Data mining
Data mining extracts the useful patterns and trends from the raw data. Big data technologies such as
Rapidminer and Presto can turn unstructured and structured data into usable information.
11 | P a g e
Rapidminer: Rapidminer is a data mining tool that can be used to build predictive models. It
draws on these two roles as strengths, of processing and preparing data, and building machine
and deep learning models. The end-to- end model allows for both functions to drive impact
across the organization [3].
Presto: Presto is an open-source query engine that was originally developed by Facebook to
run analytic queries against their large datasets. Now, it is available widely. One query on
Presto can combine data from multiple sources within an organization and perform analytics
on them in a matter of minutes.
3. Data analytics
In big data analytics, technologies are used to clean and transform data into information that can be
used to drive business decisions. This next step (after data
mining) is where users perform algorithms, models, and predictive analytics using tools such as Apache
Spark and Splunk.
Apache Spark: Spark is a popular big data tool for data analysis because it is fast and
efficient at running applications. It is faster than Hadoop because it uses random access
memory (RAM) instead of being stored and processed in batches via MapReduce . Spark
supports a wide variety of data analytics tasks and queries.
Splunk: Splunk is another popular big data analytics tool for deriving insights from large
datasets. It has the ability to generate graphs, charts, reports, and dashboards. Splunk also
enables users to incorporate artificial intelligence (AI) into data outcomes.
4. Data visualization
Finally, big data technologies can be used to create stunning visualizations from the data. In data-
oriented roles, data visualization is a skill that is beneficial for presenting recommendations to
stakeholders for business profitability and operations—to tell an impactful story with a simple graph.
Tableau: Tableau is a very popular tool in data visualization because its drag- and-drop
interface makes it easy to create pie charts, bar charts, box plots, Gantt charts, and
more. It is a secure platform that allows users to share visualizations and dashboards in real
time.
Looker: Looker is a business intelligence (BI) tool used to make sense of big data analytics
and then share those insights with other teams. Charts, graphs, and dashboards can be
configured with a query, such as monitoring weekly brand engagement through social media
analytics.
OPEN SOURCE TECHNOLOGIES / BIG DATA ANALYTICS TOOLS
There are hundreds of data analytics tools out there in the market today but the selection of the right
tool will depend upon your business NEED, GOALS, and VARIETY to get business in the right
direction. Now, let’s check out the top 10 analytics tools in big data.
1. APACHE Hadoop
It’s a Java-based open-source platform that is being used to store and process big data. It is
built on a cluster system that allows the system to process data efficiently and let the data run parallel.
It can process both structured and unstructured data from one server to multiple computers. Hadoop
also offers cross-platform support for its users. Today, it is the best big data analytic tool and is
popularly used by many tech giants such as Amazon, Microsoft, IBM, etc.Features of Apache
Hadoop:
12 | P a g e
Highly flexible and can be easily implemented with MySQL, and JSON.
Highly scalable as it can distribute a large amount of data in small segments.
It works on small commodity hardware like JBOD or a bunch of disks.
2. Cassandra
APACHE Cassandra is an open-source NoSQL distributed database that is used to fetch large
amounts of data. It’s one of the most popular tools for data analytics and has been praised by many
tech companies due to its high scalability and availability without compromising speed and
performance. It is capable of delivering thousands of operations every second and can handle
petabytes of resources with almost zero downtime. It was created by Facebook back in 2008 and was
published publicly.
3. Qubole
It’s an open-source big data tool that helps in fetching data in a value of chain using ad-hoc
analysis in machine learning. Qubole is a data lake platform that offers end-to-end service with
reduced time and effort which are required in moving data pipelines. It is capable of configuring
multi-cloud services such as AWS, Azure, and Google Cloud. Besides, it also helps in lowering the
cost of cloud computing by 50%.
Features of Qubole:
Supports ETL process: It allows companies to migrate data from multiple sources in one
place.
Real-time Insight: It monitors user’s systems and allows them to view real-time insights
Predictive Analysis: Qubole offers predictive analysis so that companies can take actions
accordingly for targeting more acquisitions.
Advanced Security System: To protect users’ data in the cloud,
4.Qubole uses an advanced security system and also ensures to protect any future breaches.
Xplenty
It is a data analytic tool for building a data pipeline by using minimal codes in it. It offers a
wide range of solutions for sales, marketing, and support. With the help of its interactive graphical
interface, it provides solutions for ETL, ELT, etc. The best part of using Xplenty is its low investment
in hardware & software and its offers support via email, chat, telephonic and virtual meetings.
Xplenty is a platform to process data for analytics over the cloud and segregates all the data together.
Features of Xplenty:
Rest API: A user can possibly do anything by implementing Rest API
Flexibility: Data can be sent, and pulled to databases, warehouses, and salesforce.
13 | P a g e
Data Security: It offers SSL/TSL encryption and the platform is capable of verifying
algorithms and certificates regularly.
Deployment: It offers integration apps for both cloud & in-house and supports deployment to
integrate apps over the cloud.
4. Spark
APACHE Spark is another framework that is used to process data and perform numerous
tasks on a large scale. It is also used to process data via multiple computers with the help of
distributing tools. It is widely used among data analysts as it offers easy-to-use APIs that provide easy
data pulling methods and it is capable of handling multi-petabytes of data as well. Recently, Spark
made a record of processing 100 terabytes of data in just 23 minutes which broke the previous world
record of Hadoop (71 minutes). This is the reason why big tech giants are moving towards spark now
and is highly suitable for ML and AI today.
5. Mongo DB
Came in limelight in 2010, is a free, open-source platform and a document- oriented (NoSQL)
database that is used to store a high volume of data. It uses collections and documents for storage and
its document consists of key-value pairs which are considered a basic unit of Mongo DB. It is so
popular among developers due to its availability for multi-programming languages such as Python,
Jscript, and Ruby.Features of Mongo DB:
Written in C++: It’s a schema-less DB and can hold varieties of documents inside.
Simplifies Stack: With the help of mongo, a user can easily store files without any
disturbance in the stack.
Master-Slave Replication: It can write/read data from the master and can be called back for
backup.
6. Apache Storm
A storm is a robust, user-friendly tool used for data analytics, especially in small companies.
The best part about the storm is that it has no language barrier (programming) in it and can support
any of them. It was designed to handle a pool of large data in fault-tolerance and horizontally scalable
methods. When we talk about real-time data processing, Storm leads the chart because of its
distributed real-time big data processing system, due to which today many tech giants are using
APACHE Storm in their system. Some of the most notable names are Twitter, Zendesk, NaviSite, etc.
Features of Storm:
Data Processing: Storm process the data even if the node gets disconnected
Highly Scalable: It keeps the momentum of performance even if the load increases
Fast: The speed of APACHE Storm is impeccable and can process up to 1 million messages of
100 bytes on a single node.
7. SAS
Today it is one of the best tools for creating statistical modeling used by data
analysts. By using SAS, a data scientist can mine, manage, extract or update data in different variants
from different sources. Statistical Analytical System or SAS allows a user to access the data in any
14 | P a g e
format (SAS tables or Excel worksheets). Besides that it also offers a cloud platform for business
analytics called SAS Viya and also to get a strong grip on AI & ML, they have introduced new tools
and products.
Features of SAS:
Flexible Programming Language: It offers easy-to-learn syntax and has also vast libraries
which make it suitable for non-programmers
Vast Data Format: It provides support for many programming languages which also include
SQL and carries the ability to read data from any format.
Encryption: It provides end-to-end security with a feature called SAS/SECURE.
8. Data Pine
Datapine is an analytical used for BI and was founded back in 2012 (Berlin, Germany). In
a short period of time, it has gained much popularity in a number of countries and it’s mainly used
for data extraction (for small-medium companies fetching data for close monitoring). With the help of
its enhanced UI design, anyone can visit and check the data as per their requirement and offer in 4
different price brackets, starting from $249 per month. They do offer dashboards by functions, industry,
and platform.
Features of Datapine:
Automation: To cut down the manual chase, datapine offers a wide array of AI assistant and
BI tools.
Predictive Tool: datapine provides forecasting/predictive analytics by using historical and
current data, it derives the future outcome.
Add on: It also offers intuitive widgets, visual analytics & discovery, ad hoc reporting,
etc.
9. Rapid Miner
It’s a fully automated visual workflow design tool used for data analytics. It’s a no-code
platform and users aren’t required to code for segregating data. Today, it is being heavily used in
many industries such as ed-tech, training, research, etc. Though it’s an open-source platform but has a
limitation of adding 10000 data rows and a single logical processor. With the help of Rapid Miner,
one can easily deploy their ML models to the web or mobile (only when the user interface is ready to
collect real-time figures).
1. Big Data:
Big data refers to the data which is huge in size and also increasing rapidly with respect to
time. Big data includes structured data, unstructured data as well as semi-structured data. Big
data cannot be stored and processed in traditional data management tools it needs specialized
big data management tools. It refers to complex and large data sets having 5 V’s volume,
velocity, Veracity, Value and variety information assets. It includes data storage, data analysis,
data mining and data visualization.
Examples of the sources where big data is generated includes social media data, e- commerce
data, weather station data, IoT Sensor data etc.
15 | P a g e
Characteristics of Big Data :
Variety of Big data – Structured, unstructured, and semi structured data
Velocity of Big data – Speed of data generation
Volume of Big data – Huge volumes of data that is being generated
Value of Big data – Extracting useful information and making it valuable
ariability of Big data – Inconsistency which can be shown by the data at times.
2. Cloud Computing :
Cloud computing refers to the on demand availability of computing resources over internet. These
resources includes servers, storage, databases, software, analytics, networking and intelligence over
the Internet and all these resources can be used as per requirement of the customer. In cloud
computing customers have to pay as per use. It is very flexible and can be resources can be scaled
easily depending upon the requirement. Instead of buying any IT resources physically, all resources
can be availed depending on the requirement from the cloud vendors. Cloud computing has three
service models i.e Infrastructure as a Service (IaaS), Platform as a Service (PaaS) and Software as a
Service (SaaS).
Examples of cloud computing vendors who provides cloud computing services are Amazon Web
Service (AWS), Microsoft Azure, Google Cloud Platform, IBM Cloud Services etc.
Characteristics of Cloud Computing :
On-Demand availability
Accessible through a network
Elastic Scalability
Pay as you go model
Multi-tenancy and resource pooling.
Big data refers to the data which is huge in Cloud computing refers to the on
size and also increasing rapidly with respect demand availability of computing
01.
to time. resources over internet.
Value of data are considered as the 5 most elasticity and measured service are considered
important characteristics of Big data. as the main characteristics of cloud
computing.
17 | P a g e
Some of the challenges of big data are variety Some of the challenges of cloud computing are
of data, data storage and integration, data availability, transformation, security concern,
processing and resource management. charging model.
07.
Big data refers to huge volume of data, its Cloud computing refers to remote IT resources
management, and useful information and different internet service models.
08.
extraction.
18 | P a g e
WEB ANALYTICS
Web Analytics or Online Analytics refers to the analysis of quantifiable and measurable data of your
website with the aim of understanding and optimizing the web usage.
Web Analytics is the methodological study of online/offline patterns and trends. It is a technique that
you can employ to collect, measure, report, and analyze your website data. It is normally carried out to
analyze the performance of a website and optimize its web usage.
web analytics used to track key metrics and analyze visitors’ activity and traffic flow. It is a tactical
approach to collect data and generate reports. It is an ongoing process that helps in attracting more
traffic to a site and thereby, increasing the Return on Investment.
Web analytics improves online experience for your customers and elevates your business prospects.
There are various Web Analytics tools available in the market. For example, Google Analytics,
Kissmetrics, Optimizely, etc.
19 | P a g e
Assess web content problems so that they can be rectified
Have a clear perspective of website trends
Monitor web traffic and user flow
Demonstrate goals acquisition
Figure out potential keywords
Identify segments for improvement
Find out referring sources
Count
It is most basic metric of measurement. It is represented as a whole number or a fraction. For
example,
Number of visitors = 12999, Number of likes = 3060, etc.
Total sales of merchandise = $54,396.18.
Ratio
It is typically a count divided by some other count. For example, Page views per visit.
20 | P a g e
Micro Level Analysis
It pertains to an individual or a small group of individuals. For example, number of times job
application submitted, number of times print this page was clicked, etc.
Macro Level Analysis It is concerned with the primary business objectives with huge groups
of people such as communities, nation, etc. For example, number of conversions in a particular
demographic.
Engagement Rate
It shows how long a person stays on your web page. What all pages he surf. To make your
web pages more engaging, include informative content, visuals, fonts and bullets.
Bounce Rate
If a person leaves your website within a span of 30 sec, it is considered as a bounce. The rate
at which users spin back is called the bounce rate. To minimize bounce rate include related
posts, clear call-to-action and backlinks in your webpages.
Dashboards
Dashboard is single page view of information important to user. You can create your own
dashboards keeping in mind your requirements. You may keep only frequently viewed data
on dashboard.
Event Tracking
Event tracking allows you to track other activities on your website. For example, you can
track downloads and sign-ups through event tracking.
Traffic Source
You can overview traffic sources. You can even filter it further. Figuring out the key areas
can help you learn about the area of improvement.
Annotations
It allows you to view a traffic report for past time. You can click on graph and type in to
save it for future study.
Visitor Flow
It gives you a clear picture of pages visited and the sequence of the same. Understanding
users’ path may help you in re-navigation in order to give customer a hassle-free navigation.
Content
It gives you insight about website’s content section. You can see how each
page is doing, website loading speed, etc.
Conversions
Analytics lets you track goals and path used to achieve these goals. You can get details
regarding, product performances, purchase amount, and mode of billing. Web Analytics
offer you more than this. All you need is to analyze things minutely and keep patience.
21 | P a g e
Behavior
Behavior lets you know page views and time spent on website. You can find out how
customer behaves once he is on your website.
Advantages of mobile BI
1. Simple access
Mobile BI is not restricted to a single mobile device or a certain place. You can view your data at
any time and from any location. Having real-time visibility into a firm improves production and the
daily efficiency of the business. Obtaining a company's perspective with a single click simplifies the
process.
22 | P a g e
2. Competitive advantage
Many firms are seeking better and more responsive methods to do business in order to stay ahead of
the competition. Easy access to real-time data improves company opportunities and raises sales
and capital. This also aids in making the necessary decisions as market conditions change.
3. Simple decision-making
As previously stated, mobile BI provides access to real-time data at any time and from any
location. During its demand, Mobile BI offers the information. This assists consumers in obtaining
what they require at the time. As a result, decisions are made quickly.
4. Increase Productivity
By extending BI to mobile, the organization's teams can access critical company data when they
need it. Obtaining all of the corporate data with a single click frees up a significant amount of time
to focus on the smooth and efficient operation of the firm. Increased productivity results in a
smooth and quick-running firm.
Disadvantages of mobile
1. Stack of data
The primary function of a mobile BI is to store data in a systematic manner and then present it to
the user as required. As a result, Mobile BI stores all of the information and does end up with
heaps of earlier data. The corporation only needs a small portion of the previous data, but they
need to store the entire information, which ends up in the stack
2. Expensive
Mobile BI can be quite costly at times. Large corporations can continue to pay for their expensive
services, but small businesses cannot. As the cost of mobile BI is not sufficient, we must
additionally consider the rates of IT workers for the smooth operation of BI, as well as the
hardware costs involved. However, larger corporations do not settle for just one Mobile BI
provider for their organisations; they require multiple. Even when doing basic commercial
transactions, mobile BI is costly.
3 Time consuming
Businesses prefer Mobile BI since it is a quick procedure. Companies are not patient enough to
wait for data before implementing it. In today's fast-paced environment, anything that can produce
results quickly is valuable. The data from the warehouse is used to create the system, hence the
implementation of BI in an enterprise takes more than 18 months.
4 Data breach
The biggest issue of the user when providing data to Mobile BI is data leakage. If you handle
sensitive data through Mobile BI, a single error can destroy your data as well as make it public,
which can be detrimental to your business.
Many Mobile BI providers are working to make it 100 percent secure to protect their potential users'
data. It is not only something that mobile BI carriers must consider, but it is also something that
we, as users, must consider when granting data access authorization.
23 | P a g e
analysed by Mobile BI is irrelevant or completely useless. This can speed down the entire procedure. This
requires you to select the data that is important and may be required in the future.
Sisense is a flexible business intelligence (BI) solution that includes powerful analytics,
visualisations, and reporting capabilities for managing and supporting corporate data. Businesses
can use the solution to evaluate large, diverse databases and generate relevant business insights.
You may easily view enormous volumes of complex data with Si Sense's code-first, low-code, and
even no-code technologies. Si Sense was established in 2004 with its headquarters in New York.
Since then, the team has only taken precautionary steps in their investigation. Once the company
had received $ 4 million in funding from investors, they began to pace its research.
1. Enterprise
2. IT Marketing
3. Education
4. Finance
5. Science and Health
How to Crowdsource?
1. For scientific problem solving, a broadcast search is used where an organization mobilizes a
crowd to come up with a solution to a problem.
2. For information management problems, knowledge discovery and
management is used to find and assemble information.
3. For processing large datasets, distributed human intelligence is used. The organization
mobilizes a crowd to process and analyze the information.
Examples of Crowdsourcing
1. Doritos: It is one of the companies which is taking advantage of crowdsourcing for a long
time for an advertising initiative. They use consumer-created ads for one of their 30-Second
Super Bowl Spots(Championship Game of Football).
2. Starbucks: Another big venture which used crowdsourcing as a medium for idea generation.
Their white cup contest is a famous contest in which customers need to decorate their
Starbucks cup with an original design and then take a photo and submit it on social media.
3. Lays:” Do us a flavor” contest of Lays used crowdsourcing as an idea-generating medium. They
asked the customers to submit their opinion about the next chip flavor they want.
4. Airbnb: A very famous travel website that offers people to rent their houses or apartments by
listing them on the website. All the listings are crowdsourced by people.
25 | P a g e
Crowdsourced Marketing
As discussed already crowdsourcing helps grow businesses grow a lot. May it be a business idea
or just a logo design, crowdsourcing engages people directly and in turn, saves money and
energy. In the upcoming years, crowdsourced marketing will surely get a boost as the world is
accepting technology faster.
Crowdsourcing Sites
Here is the list of some famous crowdsourcing and crowdfunding sites.
1. Kickstarter
2. GoFundMe
3. Patreon
4. RocketHub
Advantages of Crowdsourcing
1. Evolving Innovation: Innovation is required everywhere and in this advancing world
innovation has a big role to play. Crowdsourcing helps in getting innovative ideas from
people belonging to different fields and thus helping businesses grow in every field.
2. Save costs: There is the elimination of wastage of time of meeting people and convincing
them. Only the business idea is to be proposed on the internet and you will be flooded with
suggestions from the crowd.
3. Increased Efficiency: Crowdsourcing has increased the efficiency of business models as
several expertise ideas are also funded.
Disadvantages of Crowdsourcing
1. Lack of confidentiality: Asking for suggestions from a large group of people can bring the
threat of idea stealing by other organizations.
2. Repeated ideas: Often contestants in crowdsourcing competitions submit repeated, plagiarized
ideas which leads to time wastage as reviewing the same ideas is not worthy.
26 | P a g e
INTER AND TRANS FIREWALL ANALYTICS
27 | P a g e
28 | P a g e
Inter-firewall analytics
Focus: Analyzes traffic flows between different firewalls within a network.
Methodology: Utilizes data collected from multiple firewalls to identify anomalies
and potential breaches.
Benefits: Provides a comprehensive view of network traffic flow and helps identify
lateral movement across different security zones.
Limitations: Requires deployment of multiple firewalls within the network and efficient data
exchange mechanisms between them.
29 | P a g e
30 | P a g e
31 | P a g e
Trans-firewall analytics
Focus: Analyzes encrypted traffic that traverses firewalls, which traditional security solutions may not be
able to decrypt and inspect.
Methodology: Uses deep packet inspection (DPI) and other advanced techniques to analyze the content
of encrypted traffic without compromising its security.
Benefits: Provides insight into previously hidden threats within encrypted traffic and helps detect
sophisticated attacks.
Limitations: Requires specialized hardware and software solutions for DPI, and raises concerns regarding
potential data privacy violations.