Machine Learning Seminar Report

Download as pdf or txt
Download as pdf or txt
You are on page 1of 26

CHAPTER 1

INTRODUCTION

Machine Learning (ML) is the computerized approach to analyzing


computational work that is based on both a set of theories and a set of
technologies. And, being a very active area of research and
development, there is not a single agreed-upon definition that would
satisfy everyone, but there are some aspects, which would be part of
any knowledgeable person’s definition. The definition mostly offers
is:

Definition: Ability of a machine to improve its own performance


through the use of a software that employs artificial intelligence
techniques to mimic the ways by which humans seem to learn, such as
repetition and experience.
➢ Machine Learning (ML) is a sub-field of Artificial Intelligence
(AI) which concerns with developing computational theories of
learning and building learning machines.
➢ The goal of machine learning, closely coupled with the goal of
AI, is to achieve a thorough understanding about the nature of
learning process (both human learning and other forms of
learning), about the computational aspects of learning
behaviors, and to implant the learning capability in computer
systems.

1
➢ Machine learning has been recognized as central to the success
of Artificial Intelligence, and it has applications in various areas
of science, engineering and society.

1.1 Why Machine Learning?

To answer this question, we should look at two issues:

(1). What are the goals of machine learning;

(2). Why these goals are important and desirable.

1.1.1 The Goals of Machine Learning

The goal of ML, in simples words, is to understand the nature of


(human and other forms of) learning, and to build learning capability
in computers. To be more specific, there are three aspects of the goals
of ML.

➢ To make the computers smarter, more intelligent. The more


direct objective in this aspect is to develop systems (programs)
for specific practical learning tasks in application domains.
➢ To develop computational models of human learning process
and perform computer simulations. The study in this aspect is
also called cognitive modeling.
➢ To explore new learning methods and develop general learning
algorithms independent of applications.

2
1.1.2 Why the goals of ML are important and desirable.

It is self-evident that the goals of ML are important and desirable.


However, we still give some more supporting argument to this issue.

First of all, implanting learning ability in computers is practically


necessary. Present day computer applications require the
representation of huge amount of complex knowledge and data in
programs and thus require tremendous amount of work.

➢ Our ability to code the computers falls short of the demand for
applications. If the computers are endowed with the learning
ability, then our burden of coding the machine is eased (or at
least reduced).
➢ This is particularly true for developing expert systems where the
"bottle-neck" is to extract the expert’s knowledge and feed the
knowledge to computers.
➢ The present day computer programs in general (with the
exception of some ML programs) cannot correct their own
errors or improve from past mistakes, or learn to perform a new
task by analogy to a previously seen task.
➢ In contrast, human beings are capable of all the above. ML will
produce smarter computers capable of all the above intelligent
behavior.

Second, the understanding of human learning and its computational


aspect is a worthy scientific goal.

3
➢ We human beings have long been fascinated by our capabilities
of intelligent behaviors and have been trying to understand the
nature of intelligence.
➢ It is clear that central to our intelligence is our ability to learn.
Thus a thorough understanding of human learning process is
crucial to understand human intelligence.
➢ ML will gain us the insight into the underlying principles of
human learning and that may lead to the discovery of more
effective education techniques. It will also contribute to the
design of machine learning systems.

Finally, it is desirable to explore alternative learning


mechanisms in the space of all possible learning methods.

We remark that Machine Learning has become feasible in many


important applications (and hence the popularity of the field) partly
because the recent progress in learning algorithms and theory, the
rapidly increase of computational power, the great availability of huge
amount of data, and interests in commercial ML application
development.

4
CHAPTER 2

HISTORY OF MACHINE LEARNING

2.1 Learning Means?

Learning is a phenomenon and process which has manifestations of


various aspects. Roughly speaking, learning process

Over the years, research in machine learning has been pursued with
varying degrees of intensity, using different approaches and placing
emphasis on different, aspects and goals. Within the relatively short
history of this discipline, one may distinguish three major periods,
each centered on a different concept:

➢ Neural modeling and decision-theoretic techniques.


➢ Symbolic concept-oriented learning.
➢ Knowledge-intensive approaches combining various learning
strategies

2.2 Neural Modelling (Self Organized System)?

➢ The distinguishing feature of the first concept was the interest


in building general purpose learning systems that start with little
or no initial structure or task-oriented knowledge.
➢ The major thrust of research based on this approach involved
constructing a variety of neural model-based machines, with
random or partially random initial structure.

5
➢ These systems were generally referred to as neural networks or
self-organizing systems. Learning in such systems consisted of
incremental changes in the probabilities that neuron-like
elements would transmit a signal.
➢ Due to the early computer technology, most of the research
under this neural network model was either theoretical or
involved the construction of special purpose experimental
hardware systems.

2.3 The Symbolic Concept Acquisition Paradigm

A second major paradigm started to emerge in the early sixties


stemming from the work of psychologist and early AI researchers on
models of human learning by Hunt. The paradigm utilized logic or
graph structure representations rather than numerical or statistical
methods Systems learned symbolic descriptions representing higher
level knowledge and made strong structural assumptions about the
concepts to be acquired.
Examples of work in this paradigm include research on human concept
acquisition and various applied pattern recognition systems.

2.4 The Modern Knowledge-Intensive Paradigm

The third paradigm represented the most recent period of research


starting in the mid seventies.

6
Researchers have broadened their interest beyond learning isolated
concepts from examples, and have begun investigating a wide
spectrum of learning methods, most based upon knowledge-rich
systems specifically, this paradigm can be characterizing by several
new trends, including:

1. Knowledge-Intensive Approaches: Researchers are strongly


emphasizing the use of task-oriented knowledge and the
constraints it provides in guiding the learning process One
lesson from the failures of earlier knowledge and poor learning
systems that is acquire and to acquire new knowledge a system
must already possess a great deal of initial knowledge.
2. Exploration of alternative methods of learning: In addition
to the earlier research emphasis on learning from examples,
researchers are now investigating a wider variety of learning
methods such as learning from instruction.

7
CHAPTER 3

MACHINE LEARNING OVERVIEW

Machine Learning can still be defined as learning the theory


automatically from the data, through a process of inference, model
fitting, or learning from examples:

➢ Automated extraction of useful information from a body of data


by building good probabilistic models.
➢ Ideally suited for areas with lots of data in the absence of a
general theory.

3.1 The Aim of Machine Learning

The field of machine learning can be organized around three primary


research Areas:

➢ Task-Oriented Studies: The development and analysis of


learning systems oriented toward solving a predetermined set,
of tasks (also known as the “engineering approach”).
➢ Cognitive Simulation: The investigation and computer
simulation of human learning processes (also known as the
“cognitive modeling approach”).
➢ Theoretical Analysis: The theoretical exploration of the space
of possible learning methods and algorithms independent
application domain.

8
Although many research efforts strive primarily towards one of
these objectives, progress in on objective often lends to progress in
another.

For example, in order to investigate the space of possible learning


methods, a reasonable starting point may be to consider the only
known example of robust learning behavior, namely humans (and
perhaps other biological systems) Similarly, psychological
investigations of human learning may held by theoretical analysis that
may suggest various possible learning models.

The need to acquire a particular form of knowledge in stone task-


oriented study may itself spawn new theoretical analysis or pose the
question: “how do humans acquire this specific skill (or knowledge)?”
The existence of these mutually supportive objectives reflects the
entire field of artificial intelligence where expert system research,
cognitive simulation, and theoretical studies provide some cross-
fertilization of problems and ideas.

3.2 Machine Learning as a Science

➢ The clear contender for a cognitive invariant in human is the


learning mechanism which is the ability facts, skills and more
abstractive concepts. Therefore understanding human learning
well enough to reproduce aspect of that learning behavior in a
computer system is, in itself, a worthy scientific goal.

9
➢ Moreover, the computer can render substantial assistance to
cognitive psychology, in that it may be used to test the
consistency and completeness of learning theories and enforce
a commitment to the fine-structure process level detail that
precludes meaningless tautological or untestable theories
(Bishop,2006).
➢ The study of human learning processes is also of considerable
practical significance. Gaining insights into the principles
underlying human learning abilities is likely to lead to more
effective educational techniques.
➢ Machine learning research is all about developing intelligent
computer assistant or a computer tutoring systems and many of
these goals are shared within the machine learning fields.
According to Jaime et al who stated computer tutoring are
starting to incorporate abilities to infer models of student
competence from observed performance.

10
CHAPTER 4

CLASSIFICATION OF MACHINE LEARNING

There are several areas of machine learning that could be exploited to


solve the problems of email management and our approach
implemented unsupervised machine learning method.

4.1 Unsupervised learning

It is a method of machine learning whereby the algorithm is presented


with examples from the input space only and a model is fit to these
observations. In unsupervised learning, a data set of input objects is
gathered. Unsupervised learning then typically treats input objects as
a set of random variables. A joint density model is then built for the
data set.

➢ The problem of unsupervised learning involved learning


patterns in the input when no specific output values are
supplied”.
➢ In the unsupervised learning problem, we observe only the
features and have no measurements of the outcome. Our task
is rather to describe how the data are organized or clustered”.
➢ Trevor Hastie explained that "In unsupervised learning or
clustering there is no explicit teacher, and the system forms
clusters or ‘natural groupings’ of the input patterns.

11
➢ “Natural” is always defined explicitly or implicitly in the
clustering system itself; and given a particular set of patterns
or cost function; different clustering algorithms lead to
different clusters.

There are various categories in the field of artificial intelligence. The


classifications of machine learning systems are:

1. Supervised Machine Learning: Supervised learning is a machine


learning technique for learning a function from training data. The
training data consist of pairs of input objects (typically vectors), and
desired outputs. The output of the function can be a continuous
value (called regression), or can predict a class label of the input
object (called classification).
➢ The task of the supervised learner is to predict the value of
the function for any valid input object after having seen a
number of training examples (i.e. pairs of input and target
output).
➢ To achieve this, the learner has to generalize from the
presented data to unseen situations in a "reasonable" way.
➢ Supervised learning is a machine learning technique
whereby the algorithm is first presented with training data
which consists of examples which include both the inputs
and the desired outputs; thus enabling it to learn a function.

12
2. Unsupervised Machine Learning: Unsupervised learning is a
type of machine learning where manual labels of inputs are not
used. It is distinguished from supervised learning approaches which
learn how to perform a task, such as classification or regression,
using a set of human prepared examples. Unsupervised learning
means we simply have a training set of vectors without function
values of them.

Fig. 4.1: Machine Learning Supervise Process

13
CHAPTER 5

MACHINE LEARNING ALGORITHMS

Machine learning algorithms are organized into taxonomy, based


on the desired outcome of the algorithm. Common algorithm types
include:

5.1 Supervised learning: Where the algorithm generates a


function that maps inputs to desired outputs. One standard
formulation of the supervised learning task is the classification
problem: the learner is required to learn (to approximate the
behavior of) a function which maps a vector into one of several
classes by looking at several input-output examples of the
function.
1. Unsupervised learning: Which models a set of inputs,
labeled examples are not available.
2. Semi-supervised learning Which combines both labeled
and unlabeled examples to generate an appropriate function
or classifier.
3. Reinforcement learning: Where the algorithm learns a
policy of how to act given an observation of the world.

4. Learning to learn: Where the algorithm learns its own


inductive bias based on previous experience.

14
5. Transduction: Similar to supervised learning, but does not
explicitly construct a function: instead, tries to predict new
outputs based on training inputs, training outputs, and new
inputs.

The performance and computational analysis of machine learning


algorithms is a branch of statistics known as computational learning
theory. Machine learning is about designing algorithms that allow a
computer to learn.

Learning is not necessarily involves consciousness but learning is


a matter of finding statistical regularities or other patterns in the data.
Thus, many machine learning algorithms will barely resemble how
human might approach a learning task.

However, learning algorithms can give insight into the relative


difficulty of learning in different environments.

5.2 Algorithm Types

In the area of supervised learning which deals much with


classification. These are the algorithms types:

a. Linear Classifiers

1. Fisher’s linear discriminant

2. Naïve Bayes Classifier

3. Perceptron

15
4. Support Vector Machine

b. Quadratic Classifiers

c. Boosting

d. Neural networks

e. Bayesian Networks

f. Decision Tree

5.2.a. Linear Classifiers:

In machine learning, the goal of classification is to group items that


have similar feature values, into groups. A linear classifier achieves
this by making a classification decision based on the value of the linear
combination of the features. If the input feature vector to the classifier
is a real vector x, then the output score is

where is a real vector of weights and f is a function that converts the


dot product of the two vectors into the desired output.

5.2.a.1 Fisher’s linear discriminant

The resulting combination may be used as a linear classifier or, more


commonly, for dimensionality reduction before later classification.

16
Linear discriminant analysis (LDA) and the related Fisher's linear
discriminant are methods used in machine learning to find a linear
combination of features which characterizes or separates two or more
classes of objects or events.

5.2.a.2 Naïve Bayes Classifier

A naive Bayes classifier is a simple probabilistic classifier based on


applying Bayes' theorem with strong
(naive) independence assumptions. A more descriptive term for the
underlying probability model would be "independent feature".
In simple terms, a naive Bayes classifier assumes that the presence or
absence of a particular feature is unrelated to the presence or absence
of any other feature, given the class variable. A naive Bayes classifier
considers each of these features to contribute independently to the
probability that this fruit is an apple, regardless of the presence or
absence of the other features.

5.2.a.3 Perceptron

The perceptron is an algorithm for supervised classification of an


input into one of several possible non-binary outputs.

The learning algorithm for perceptrons is an online algorithm, in that


it processes elements in the training set one at a time.

5.2.a.4 Support vector machines

In machine learning, support vector machines (SVMs) are supervised


learning models with associated learning algorithms that analyze

17
data and recognize patterns, used for classification and regression
analysis.

The basic SVM takes a set of input data and predicts, for each given
input, which of two possible classes forms the output, making it a non-
probabilistic binary linear classifier. Given a set of training examples,
each marked as belonging to one of two categories, an SVM training
algorithm builds a model that assigns new examples into one category
or the other.

An SVM model is a representation of the examples as points in space,


mapped so that the examples of the separate categories are divided by
a clear gap that is as wide as possible. New examples are then mapped
into that same space and predicted to belong to a category based on
which side of the gap they fall on.

5.2.b Quadratic classifier

A quadratic classifier is used in machine learning and statistical


classification to separate measurements of two or more classes of
objects or events by a quadric surface. It is a more general version of
the linear classifier.

5.2.c Boosting

Boosting is a machine learning meta-algorithm for


reducing bias in supervised learning. Boosting is based on the question
posed as “Can a set of weak learners create a single strong learner?”

18
A weak learner is defined to be a classifier which is only slightly
correlated with the true classification. In contrast, a strong learner is a
classifier that is arbitrarily well-correlated with the true classification.

5.2.d Neural networks

Neural networks are capable of machine learning and pattern


recognition. They are usually presented as systems of interconnected
"neurons" that can compute values from inputs by feeding information
through the network. Neural networking is the science of creating
computational solutions modeled after the brain.

Like the human brain, neural networks are trainable-once they are
taught to solve one complex problem, they can apply their skills to a
new set of problems without having to start the learning process from
scratch.

5.2.e Bayesian network

A Bayesian network, Bayes network, belief network, Bayes(ian)


model or probabilistic directed acyclic graphical model is
a probabilistic graphical model (a type ofstatistical model) that
represents a set of random variables and their conditional
dependencies via a directed acyclic graph (DAG).

For example, suppose that there are two events which could cause
grass to be wet: either the sprinkler is on or it's raining.

19
Also, suppose that the rain has a direct effect on the use of the sprinkler
(namely that when it rains, the sprinkler is usually not turned on). Then
the situation can be modeled with a Bayesian network (shown). All
three variables have two possible values, T (for true) and F (for false).

Fig. 5.1: Bayesian network

5.2.f Decision Trees

A decision tree is a hierarchical data structure implementing the


divide-and-conquer strategy. It is an efficient nonparametric method,
which can be used for both classification and regression.

A decision tree is a hierarchical model for supervised learning whereby


the local region is identified in a sequence of recursive splits in a
smaller number of steps. Given an input, at each node, a test is applied
and one of the branches is taken depending on the outcome. This
process starts at the root and is repeated recursively until a leaf node
is hit, at which point the value written in the leaf constitutes the output.

20
Fig.5.2: Decision Tree

5.3 Machine Learning Applications

The other aspect for classifying learning systems is the area of


application which gives a new dimension for machine learning. Below
are areas to which various existing learning systems have been applied.
They are:

➢ Computer Programming
➢ Game playing (chess, poker, and so on)
➢ Image recognition, Speech recognition
➢ Medical diagnosis
➢ griculture, Physics
➢ Email management, Robotics
➢ Music
➢ Mathematics
➢ Natural Language Processing and many more.

21
5.4 Examples of Machine Learning Problems

There are many examples of machine learning problems. Much of this


course will focus on classification problems in which the goal is to
categorize objects into a fixed set of categories. Here are several
examples:

➢ Optical character recognition: categorize images of


handwritten characters by the letters represented.
➢ Face detection: find faces in images (or indicate if a face is
present).
➢ Spam filtering: identify email messages as spam or non-spam.
➢ Topic spotting: categorize news articles (say) as to whether
they are about politics, sports, entertainment, etc.
➢ Spoken language understanding: within the context of a
limited domain, determine the meaning of something uttered by
a speaker to the extent that it can be classified into one of a fixed
set of categories.
➢ Medical diagnosis: diagnose a patient as a sufferer or non-
sufferer of some disease.
➢ Customer segmentation: predict, for instance, which
customers will respond to a particular promotion.
➢ Fraud detection: identify credit card transactions (for instance)
which may be fraudulent in nature.
➢ Weather prediction: predict, for instance, whether or not it will
rain tomorrow.

22
CHAPTER 6

FUTURE DIRECTIONS

Research in Machine Learning Theory is a combination of attacking


established fundamental questions, and developing new frameworks
for modeling the needs of new machine learning applications. While it
is impossible to know where the next breakthroughs will come, a few
topics one can expect the future to hold include:

➢ Better understanding how auxiliary information, such as


unlabeled data, hints from a user, or previously-learned tasks,
can best be used by a machine learning algorithm to improve its
ability to learn new things. Traditionally, Machine Learning
Theory has focused on problems of learning a task (say,
identifying spam) from labeled examples (email labeled as spam
or not).
➢ However, often there is additional information available. One
might have access to large quantities of unlabeled data (email
messages not labeled by their type, or discussion-group
transcripts on the web) that could potentially provide useful
information. One might have other hints from the user besides
just labels, e.g. highlighting relevant portions of the email
message.
➢ Further developing connections to economic theory. As
software agents based on machine learning are used in

23
competitive settings, “strategic” issues become increasingly
important.
➢ Most algorithms and models to date have focused on the case
of a single learning algorithm operating in an environment that,
while it may be changing, does not have its own motivations and
strategies.
➢ However, if learning algorithms are to operate in settings
dominated by other adaptive algorithms acting in their own
users’ interests, such as bidding on items or performing various
kinds of negotiations, then we have a true merging of computer
science and economic models.

24
REPORT SUMMARY

Machine Learning Theory is both a fundamental theory with many


basic and compelling foundational questions, and a topic of practical
importance that helps to advance the state of the art in software by
providing mathematical frameworks for designing new machine
learning algorithms. It is an exciting time for the field, as connections
to many other areas are being discovered and explored, and as new
machine learning applications bring new questions to be modeled and
studied.

25
REFERENCES

[1] Alpaydin, E. (2004). Introduction to Machine Learning.


Massachusetts, USA: MIT Press.

[2] Carling, A. (1992). Introducing Neural Networks. Wilmslow, UK:


Sigma Press

[3] Richard S. Sutton, A. G. (1998). Reinforcement Learning. MIT


Press.

[4] Tom, M. (1997). Machine Learning. Machine Learning, Tom


Mitchell, McGraw Hill, 1997: McGraw Hill.

[5] Rosenblatt, F. (1958) “The perceptron: a probabilistic model for


information storage and organization in the brain” Psychological
Review.

[6] Michalski, R S , Carhoncll, .J G , & Mitchcll, T. M. (1983) (Eds)


Machine Learning, an Artificial Intelligence Approach Palo Alto, CA:
Tioga Press.

26

You might also like