Module 2 - Introduction To BA
Module 2 - Introduction To BA
Module 2 - Introduction To BA
• Business metrics are decided. First, metrics are created that will effectively evaluate
performance against business goals, such as improving operational efficiency or
increasing revenue. the success of descriptive analytics heavily relies on KPI (key
performance indicator) governance.
• The data required is identified. Data is sourced from repositories such as reports
and databases. ‘To measure accurately against KPIs, ‘companies must catalogue
and prepare the correct data sources to extract the needed data and calculate
metrics based on the current state of the business.
• The data is collected and prepared. Data preparation –transformation and cleansing,
for example – takes place before the analysis stage and is a critical step to ensure
accuracy; it is also one of the most time-consuming steps for the analyst.
• The data is analysed. Summary statistics, clustering, pattern tracking and regression
analysis are used to find patterns in the data and measure performance.
• The data is presented. Finally, charts and graphs are used to present findings in a
way that non-analytics experts can understand.
Descriptive analytics :Usage
• Descriptive analytics is frequently used in the day-to-day operations of an
organisation.
• Company reports – such as those on inventory, workflow, sales and
revenue – are all examples of descriptive analytics that provide a historical
review of an organisation’s operations.
• Data collected by these kinds of reports can be easily aggregated and
used to create snapshots of an organisation’s operations.
• Social analytics are almost always an example of descriptive analytics.
The number of followers, likes and posts can be used to determine the
average number of replies per post, the number of page views and the
average response time, for example. The comments that people post on
Facebook or Instagram are also examples of descriptive analytics and can
be used to better understand user attitudes.
advantages and disadvantages of descriptive analytics
• Predictive analytics has its roots in the ability to “predict” what might
happen.
• These analytics are about understanding the future.
• Predictive analytics provides companies with actionable insights
based on data.
• Predictive analytics provides estimates about the likelihood of a future
outcome
• use of predictive analytics to produce a credit score. These scores are
used by financial services to determine the probability of customers
making future credit payme
Predictive Analytics
• The healthcare industry, as an example, is a key beneficiary of predictive analytics. In 2019, RMIT
University partnered with Digital Health Cooperative Research Centre to
develop clinical decision support software for aged care that will reduce emergency
hospitalisations and predict patient deterioration by interpreting historical data and developing
new predictive analytics techniques. The goal is that predictive analytics will allow aged-care
providers, residents and their families to better plan for the end of life.
• Other examples of industries in which predictive analysis can be used,, include the following:
• E-commerce – predicting customer preferences and recommending products to customers
based on past purchases and search history
• Sales – predicting the likelihood that customers will purchase another product or leave the store
• Human resources – detecting if employees are thinking of quitting and then persuading them to
stay
• IT security – identifying possible security breaches that require further investigation
• Healthcare – predicting staff and resource needs
Prescriptive Analytics: Advise on possible outcomes
• The relatively new field of prescriptive analytics allows users to “prescribe” a number of
different possible actions and guide them towards a solution.
• In a nutshell, these analytics are all about providing advice.
• Prescriptive analytics attempts to quantify the effect of future decisions in order to advise
on possible outcomes before the decisions are actually made.
• At their best, prescriptive analytics predicts not only what will happen, but also why it
will happen, providing recommendations regarding actions that will take advantage of the
predictions.
• Prescriptive analytics use a combination of techniques and tools such as business rules,
algorithms, machine learning and computational modelling procedures. These techniques
are applied against input from many different data sets including historical and
transactional data, real-time data feeds, and big data.
• Descriptive analytics tells you what has happened and predictive analytics tells
you what could happen, then prescriptive analytics tells you what should be done
Prescriptive Analytics
• Prescriptive analytics takes what has been learned through
descriptive and predictive analysis and goes a step further by
recommending the best possible courses of action for a business.
• This is the most complex stage of the business analytics process,
requiring much more specialised analytics knowledge to perform, and
for this reason it is rarely used in day-to-day business operations.
• predictive analytics looks at historical data using statistical
techniques to make predictions about the future, machine learning, a
subset of artificial intelligence, refers to the ability of a computer
system to understand large – often huge – amounts of data, without
explicit directions, and while doing so adapt and become increasingly
smarter.
Prescriptive analytics
• Before enterprises can initiate Data Governance, they must identify the
regulations and frameworks relevant to their businesses.
• Data Governance programs could benefit from ethics frameworks from the
government and/or the wider public sector.
• The UK's Department for Digital, Culture, Media & Sport, for example,
formulated one in service of its National Data Strategy.
• It outlines principles on how data should be used in the public sector,
emphasising the importance of collective standards and ethical
frameworks.
• Such frameworks serve as guidelines on understanding the effects of
technology, data workflows and data sharing, as well as their ethical
and real-world consequences
Ethics in data management
Ethical principles for using data provide a high-level and wide context
for resolving ethical predicaments, namely:
• Secure vulnerable humans who could be impaired by the activities in
their professions;
• Enhance and protect the trust and reputation for the profession;
• Give a basis for public evaluation and expectations of the profession;
• Make the profession as a diverse moral public worthy of self-
sufficiency from external regulation and control;
• Serve as a guide for adjudicating disputes among organizations,
both non-members and members;
• Make institutions buoyant in the face of external burdens; and
• https://www2.deloitte.com/us/en/insights/industry/public-sector/chief-
data-officer-government-playbook/managing-data-ethics.html
• India : The new law, the Personal Data Protection Bill (PDP), is
currently in front of parliament and was proposed to effect a
comprehensive overhaul of India's current data protection regime,
which today is governed by the Information Technology Act, 2000.
• PDP Bill proposes that the processing of personal data must
comply with seven principles for processing, namely: (i) processing
of personal data has to be fair and reasonable; (ii) it should be for a
specific purpose; (iii) only personal data necessary for the
purpose should be collected; (iv) it should be lawful;
UK GDPR sets out seven key principles:
• Lawfulness, fairness and transparency.
• Purpose limitation.
• Data minimisation.
• Accuracy.
• Storage limitation.
• Integrity and confidentiality (security)
• Accountability.
• https://iclg.com/practice-areas/data-protection-laws-and-regulations/
india
• https://www.youtube.com/watch?v=RbxdkTixxLo
• https://www.youtube.com/watch?v=q_okDS2RtzY
(Crisp DM)
Machine learning
• Machine learning is a branch of artificial intelligence (AI) and computer science
which focuses on the use of data and algorithms to imitate the way that
humans learn, gradually improving its accuracy.
• Arthur Samuel, is credited for coining the term, “machine learning”
• Robert Nealey, the self-proclaimed checkers master, played the game on an
IBM 7094 computer in 1962, and he lost to the computer.
• Compared to what can be done today, this feat almost seems trivial, but it’s
considered a major milestone within the field of artificial intelligence.
• Over the next couple of decades, the technological developments around
storage and processing power will enable some innovative products that we
know and love today, such as Netflix’s recommendation engine or self-driving
cars.
ML and Deep Learning
• Machine learning, deep learning, and neural networks are all sub-fields of artificial
intelligence.
• However, deep learning is actually a sub-field of machine learning, and neural
networks is a sub-field of deep learning.
• The way in which deep learning and machine learning differ is in how each algorithm
learns.
• Deep learning automates much of the feature extraction piece of the process,
eliminating some of the manual human intervention required and enabling the use of
larger data sets.
• Classical, or "non-deep", machine learning is more dependent on human intervention to
learn.
• Human experts determine the set of features to understand the differences between data
inputs, usually requiring more structured data to learn.
ML and Deep Learning
• The acronym
SEMMA stands
for Sample,
Explore, Modify,
Model, Assess,
and refers to the
process of
conducting a
data mining
project.
STAGES : Crisp Semma
• Sample — a portion of a large data set is taken that is big enough to extract
significant information and small enough to manipulate quickly.
• Explore — data exploration can help in gaining understanding and ideas as
well as refining the discovery process by searching for trends and anomalies.
• Modify — data modification stage focuses on creating, selecting and
transformation of variables to focus model selection process. This stage may
also look for outliers and reducing the number of variables.
• Model — there are different modeling techniques present and each type of
model has its strengths and is appropriate for a specific goal for data mining.
• Assess — this final stage focuses on the evaluation of the reliability and
usefulness of findings and estimates the performance.
1.Determine Business Objectives
- Background
- Business Objectives
- Business Success Criteria
2.Assess the Situation
- Inventory of Resources
- Requirements, Assumptions, and Constraints
- Risks and Contingencies
- Terminology
- Costs and Benefits
3.Determine Goals
- Data Mining Goals
- Data Mining Success Criteria
4.Produce Project Plan
- Project Plan
- Initial Assessment of Tools and Techniques
Data Understanding
• The second stage consists of collecting and exploring the input dataset.
The set goal might be unsolvable using the input data, you might need
to use public datasets, or even create a specific one for the set goal.
1.Collect Initial Data
- Initial Data Collection Report
2.Describe Data
- Data Description Report
3.Explore Data
- Data Exploration Report
4.Verify Data Quality
- Data Quality Report
Data Preparation
• As we all know, bad input inevitably leads to bad output. Therefore no matter what
you do in modeling — if you made major mistakes while preparing the data —
you will end up returning to this stage and doing it over again.
1.Select Data
- The rationale for Inclusion/Exclusion
2.Clean Data
- Data Cleaning Report
3.Construct Data
- Derived Attributes
- Generated Records
4.Integrate Data
- Merged Data
5.Format Data
- Reformatted Data
Modelling
1.Evaluate Results
- Assessment of Data Mining Results w.r.t. Business Success Criteria
- Approved Models
2.Review Process
- Review of Process
3.Determine Next Steps
- List of Possible Actions
- Decision
• Deployment
Deployment
1.Plan Deployment
- Deployment Plan
2.Plan Monitoring and Maintenance
- Monitoring and Maintenance Plan
3.Produce Final Report
- Final Report
- Final Presentation
4.Review Project
- Experience Documentation