Explain How Files and Databases Are Used in Organizations

Download as docx, pdf, or txt
Download as docx, pdf, or txt
You are on page 1of 5

CAPE Information Technology

Unit 2
Module 1 Specific Objective 2

2. explain how files and Uses: including store, organize, search,


databases are used in retrieve; eliminate redundancies; data mining,
organizations; data marts and data warehouses.
Databases are used to hold information that is useful in an organization and it may be used to organize or
arrange data in such a way that will improve the efficiency in data response in an organization.

In organizing data, the DB is used to allow data to be stored in tables, and manipulated in queries. Several
database are used across organizations and within our society.

Databases allows organization to locate information quickly through use of criterias using database
querying facilities and supply the requested information upon request in a timely and organized manner.

With Database, retrieval of records is possible where users can specify key terms and these key terms in
criterias are powerful enough to extract the required data.

Databases are helpful to remove repetition of data which leads to increased storage caused from
redundancies which is an undesirable characteristic of database. Database are useful to help remove
redundancy patterns or data that can arise in database that can distort data as well as affect the results a
database produce.

Data Mining

Generally, data mining (sometimes called data or knowledge discovery) is the process of analyzing data
from different perspectives and summarizing it into useful information - information that can be used to
increase revenue, cuts costs, or both. Data mining software is one of a number of analytical tools for
analyzing data. It allows users to analyze data from many different dimensions or angles, categorize it, and
summarize the relationships identified. Technically, data mining is the process of finding correlations or
patterns among dozens of fields in large relational databases.

For example, one Midwest grocery chain used the data mining capacity of Oracle software to analyze
local buying patterns. They discovered that when men bought diapers on Thursdays and Saturdays, they
also tended to buy beer. Further analysis showed that these shoppers typically did their weekly grocery
shopping on Saturdays. On Thursdays, however, they only bought a few items. The retailer concluded that
they purchased the beer to have it available for the upcoming weekend. The grocery chain could use this
newly discovered information in various ways to increase revenue. For example, they could move the beer
display closer to the diaper display. And, they could make sure beer and diapers were sold at full price on
Thursdays.

Data
Data are any facts, numbers, or text that can be processed by a computer. Today, organizations are
accumulating vast and growing amounts of data in different formats and different databases. This
includes:

 operational or transactional data such as, sales, cost, inventory, payroll, and accounting
 nonoperational data, such as industry sales, forecast data, and macro economic data
CAPE Information Technology
Unit 2
Module 1 Specific Objective 2

 meta data - data about the data itself, such as logical database design or data dictionary definitions

Information
The patterns, associations, or relationships among all this data can provide information. For example,
analysis of retail point of sale transaction data can yield information on which products are selling and
when.

Knowledge
Information can be converted into knowledge about historical patterns and future trends. For example,
summary information on retail supermarket sales can be analyzed in light of promotional efforts to
provide knowledge of consumer buying behavior. Thus, a manufacturer or retailer could determine which
items are most susceptible to promotional efforts.

Data Warehouses
Dramatic advances in data capture, processing power, data transmission, and storage capabilities are
enabling organizations to integrate their various databases into data warehouses. Data warehousing is
defined as a process of centralized data management and retrieval. Data warehousing, like data mining, is
a relatively new term although the concept itself has been around for years. Data warehousing represents
an ideal vision of maintaining a central repository of all organizational data. Centralization of data is
needed to maximize user access and analysis. Dramatic technological advances are making this vision a
reality for many companies. And, equally dramatic advances in data analysis software are allowing users
to access this data freely. The data analysis software is what supports data mining.

What can data mining do?

Data mining is primarily used today by companies with a strong consumer focus - retail, financial,
communication, and marketing organizations. It enables these companies to determine relationships
among "internal" factors such as price, product positioning, or staff skills, and "external" factors such as
economic indicators, competition, and customer demographics. And, it enables them to determine the
impact on sales, customer satisfaction, and corporate profits. Finally, it enables them to "drill down" into
summary information to view detail transactional data.

With data mining, a retailer could use point-of-sale records of customer purchases to send targeted
promotions based on an individual's purchase history. By mining demographic data from comment or
warranty cards, the retailer could develop products and promotions to appeal to specific customer
segments.

For example, Blockbuster Entertainment mines its video rental history database to recommend rentals to
individual customers. American Express can suggest products to its cardholders based on analysis of their
monthly expenditures.

WalMart is pioneering massive data mining to transform its supplier relationships. WalMart captures
point-of-sale transactions from over 2,900 stores in 6 countries and continuously transmits this data to its
massive 7.5 terabyte Teradata data warehouse. WalMart allows more than 3,500 suppliers, to access data
on their products and perform data analyses. These suppliers use this data to identify customer buying
patterns at the store display level. They use this information to manage local store inventory and identify
new merchandising opportunities. In 1995, WalMart computers processed over 1 million complex data
CAPE Information Technology
Unit 2
Module 1 Specific Objective 2

queries.

The National Basketball Association (NBA) is exploring a data mining application that can be used in
conjunction with image recordings of basketball games. The Advanced Scout software analyzes the
movements of players to help coaches orchestrate plays and strategies. For example, an analysis of the
play-by-play sheet of the game played between the New York Knicks and the Cleveland Cavaliers on
January 6, 1995 reveals that when Mark Price played the Guard position, John Williams attempted four
jump shots and made each one! Advanced Scout not only finds this pattern, but explains that it is
interesting because it differs considerably from the average shooting percentage of 49.30% for the
Cavaliers during that game.

By using the NBA universal clock, a coach can automatically bring up the video clips showing each of the
jump shots attempted by Williams with Price on the floor, without needing to comb through hours of
video footage. Those clips show a very successful pick-and-roll play in which Price draws the Knick's
defense and then finds Williams for an open jump shot.

How does Data Mining work?

While large-scale information technology has been evolving separate transaction and analytical systems,
data mining provides the link between the two. Data mining software analyzes relationships and patterns
in stored transaction data based on open-ended user queries. Several types of analytical software are
available: statistical, machine learning, and neural networks. Generally, any of four types of relationships
are sought:

 Classes: Stored data is used to locate data in predetermined groups. For example, a restaurant
chain could mine customer purchase data to determine when customers visit and what they
typically order. This information could be used to increase traffic by having daily specials.

 Clusters: Data items are grouped according to logical relationships or consumer preferences. For
example, data can be mined to identify market segments or consumer affinities.

 Associations: Data can be mined to identify associations. The beer-diaper example is an example
of associative mining.

 Sequential patterns: Data is mined to anticipate behavior patterns and trends. For example, an
outdoor equipment retailer could predict the likelihood of a backpack being purchased based on a
consumer's purchase of sleeping bags and hiking shoes.

Data Mining consists of five major elements:

 Extract, transform, and load transaction data onto the data warehouse system.
Store and manage the data in a multidimensional database system.
 Provide data access to business analysts and information technology professionals.
 Analyze the data by application software.
 Present the data in a useful format, such as a graph or table.

Different levels of analysis are available:


CAPE Information Technology
Unit 2
Module 1 Specific Objective 2

 Artificial neural networks: Non-linear predictive models that learn through training and resemble
biological neural networks in structure.

 Genetic algorithms: Optimization techniques that use processes such as genetic combination,
mutation, and natural selection in a design based on the concepts of natural evolution.
 Decision trees: Tree-shaped structures that represent sets of decisions. These decisions generate
rules for the classification of a dataset. Specific decision tree methods include Classification and
Regression Trees (CART) and Chi Square Automatic Interaction Detection (CHAID) . CART and
CHAID are decision tree techniques used for classification of a dataset. They provide a set of rules
that you can apply to a new (unclassified) dataset to predict which records will have a given
outcome. CART segments a dataset by creating 2-way splits while CHAID segments using chi
square tests to create multi-way splits. CART typically requires less data preparation than CHAID.

 Nearest neighbor method: A technique that classifies each record in a dataset based on a
combination of the classes of the k record(s) most similar to it in a historical dataset (where k 1).
Sometimes called the k-nearest neighbor technique.

 Rule induction: The extraction of useful if-then rules from data based on statistical significance.

 Data visualization: The visual interpretation of complex relationships in multidimensional data.


Graphics tools are used to illustrate data relationships.

What technological infrastructure is required?

Today, Data Mining applications are available on all size systems for mainframe, client/server, and PC
platforms. System prices range from several thousand dollars for the smallest applications up to $1 million
a terabyte for the largest. Enterprise-wide applications generally range in size from 10 gigabytes to
over
11 terabytes. NCR has the capacity to deliver applications exceeding 100 terabytes. There are two critical
technological drivers:

 Size of the database: the more data being processed and maintained, the more powerful the system
required.
CAPE Information Technology
Unit 2
Module 1 Specific Objective 2

 Query complexity: the more complex the queries and the greater the number of
queries being processed, the more powerful the system required.

Relational database storage and management technology is adequate for many data mining
applications less than 50 gigabytes. However, this infrastructure needs to be significantly
enhanced to support larger applications. Some vendors have added extensive indexing
capabilities to improve query performance. Others use new hardware architectures such as
Massively Parallel Processors (MPP) to achieve order-of- magnitude improvements in query
time. For example, MPP systems from NCR link hundreds of high- speed Pentium
processors to achieve performance levels exceeding those of the largest supercomputers.

http://www.anderson.ucla.edu/faculty/jason.frand/teacher/technologies/palace/
datamining.htm

Issues raised by
Data Mining

One of the key issues raised by data mining technology is not a business or technological one,
but a social one. It is the issue of individual privacy. Data mining makes it possible to
analyze routine business transactions and glean a significant amount of information
about individuals buying habits and preferences.

Another issue is that of data integrity. Clearly, data analysis can only be as good as the data
that is being analyzed. A key implementation challenge is integrating conflicting or
redundant data from different sources. For example, a bank may maintain credit cards
accounts on several different databases. The addresses (or even the names) of a single
cardholder may be different in each. Software must translate data from one system to
another and select the address most recently entered.

A hotly debated technical issue is whether it is better to set up a relational database


structure or a multidimensional one. In a relational structure, data is stored in tables,
permitting ad hoc queries. In a multidimensional structure, on the other hand, sets of cubes
are arranged in arrays, with subsets created according to category. While
multidimensional structures facilitate multidimensional data mining, relational structures
thus far have performed better in client/server environments. And, with the explosion of the
Internet, the world is becoming one big client/server environment.

Finally, there is the issue of cost. While system hardware costs have dropped dramatically
within the past five years, data mining and data warehousing tend to be self-reinforcing.
The more powerful the data mining queries, the greater the utility of the information being
gleaned from the data, and the greater the pressure to increase the amount of data being
collected and maintained, which increases the pressure for faster, more powerful data mining
queries. This increases pressure for larger, faster systems, which are more expensive.

You might also like