You are on page 1of 6

Page.1Q1.

Explain the meaning of data cleaning and data


formating.
Ans-Data cleaning
This step complements the previous one. It is also the most time
consuming due to a lot of possible techniques that can be
implemented so as to optimize data quality for future modeling stage.
Possible techniques for data cleaning include:
Data normalization. For example decimal scaling into the range
(0,1), or standard deviation normalization.
Data smoothing. Discretization of numeric attributes is one
example, this is helpful or even necessary for logic based methods.
Treatment of missing values. There is not simple and safe solution
for the cases where some of the attributes have significant number of
missing values. Generally, it is good to experiment with and without
these attributes in the modelling phase, in order to find out the
importance of the missing values.
Data formatting
Final data preparation step which represents syntactic modifications to the data that do
not change its meaning, but are required by the particular modelling tool chosen for the
DM task. These include:
reordering of the attributes or records: some modelling tools require reordering of the
attributes (or records) in the dataset: putting target attribute at the beginning or at the
end, randomizing order of records (required by neural networks for example)
changes related to the constraints of modelling tools: removing commas or tabs,
special characters, trimming strings to maximum allowed number of characters,
replacing special characters with allowed set of special characters.
Q2.What is metadata ? Explain various purpose in which
metadata is used.
Ans-Meta data is data about data. Since data in a dataware house is both voluminous
and dynamic, it needs constant monitoring. This can be done only if we have a separate
set of data about data is stored. This is the purpose of meta data.
Meta data is useful for data transformation and load data management and query
generation. This chapter introduces a few of the commonly used meta data functions
for each of them. Meta data, by definition is data about data or data that describes
the data. In simple terms, the data warehouse contains data that describes different
situations. But there should also be some data that gives details about the data stored in
a data warehouse. This data is metadata. Metadata, apart form other things, will be
used for the following purposes.
1. data transformation and loading
2. data management
3. query generation
Q3.Write the steps in designing of fact tables.
Ans-DESIGNING OF FACT TABLES
The above listed methods, when iterated repeatedly will help to finally arrive at a set of
entities that go into a fact table. The next question is how big a fact table can be? An
answer could be that it should be big enough to store all the facts, still making the task
of collecting data from this table reasonably fast. Obviously, this depends on the
hardware architecture as well as the design of the database. A suitable
hardware architecture can ensure that the cost of collecting data is reduced by the
inherent capability of the hardware on the other hand the database designed should
ensure that whenever a data is asked for, the time needed to search for the same is
minimum. In other words, the designer should be able to balance the value of
information made available by the database and cost of making the same data
available to the user. A larger database obviously stores more details, so is definitely
useful, but the cost of storing a larger database as well as the cost of searching and
evaluating the same becomes higher. Technologically, there is perhaps no limit on the
size of the database. How does one optimize the cost- benefit ratio? There are no
standard formulae, but some of the following facts can be taken not of.
i.
Understand the significance of the data stored with respect to
time. Only those data that are still needed for processing need to
be stored. For example customer details after a period of time
may become irrelevant. Salary details paid in 1980s may be of
little use in analyzing the employee cost of 21st century etc. As
and when the data becomes obsolete, it can be removed.
ii. Find out whether maintaining of statistical samples of each of the subsets could be
resorted to instead of storing the entire data. For example, instead of storing the sales
details of all the 200 towns in the last 5 years, one can store details of 10 smaller
towns, five metros, 10 bigger cities and 20 villages. After all data warehousing most
often is resorted to get trends and not the actual figures. The subsets of these individual
details can always be extrapolated
Q3.List and explain the aspects to be looked into while designing the summary
tables.
Ans-ASPECTS TO BE LOOKED INTO WHILE DESIGNING THE
SUMMARY TABLES
The main purpose of using summary tables is to cut down the time taken to execute a
specific query. The main methodology involves minimizing the volume of data being
scanned each time the query is to be answered. In other words, partial answers to the
query are already made available. For example, in the above cited example of mobile
market, if one expects
i) the citizens above 18 years of age
ii) with salaries greater than 15,000 and
iii) with professions that involve traveling are the potential customers, then, every time
the query is to be processed (may be every month or every quarter), one will have to
look at the entire data base to compute these values and then combine them suitably to
get the relevant answers. The other method is to prepare summary tables, which have
the values pertaining toe ach of these sub-queries, before hand, and then combine them
as and when the query is raised .
Itcan be noted that the summaries can be prepared in the background (or when the
number of queries running are relatively sparse) and only the aggregation can be done
on the fly. Summary table are designed by following the steps given below
i) Decide the dimensions along which aggregation is to be done.
ii) Determine the aggregation of multiple facts.
iii) Aggregate multiple facts into the summary table.

Q4.Explain the role of access control issues in data mart


design.
Ans-ROLE OF ACCESS CONTROL ISSUES IN DATA MART
DESIGN
This is one of the major constraints in data mart designs. Any data warehouse, with its
huge volume of data is, more often than not, subject to various access controls as to
who could access which part of data. The easiest case is where the data is partitioned so
clearly that a user of each partition cannot access any other data. In such cases, each of
these can be put in a data mart and the user of each can
access only his data . In the data ware house, the data pertaining to all these marts are
stored, but the partitioning are retained. If a super user wants to get an overall view of
the data, suitable aggregations can be generated. However, in certain other cases the
demarcation may not be so clear. In such cases, a judicious analysis of the privacy
constraints so as to optimize the privacy of each data mart is maintained.
Data marts, as described in the previous sections can be designed, based on several
splits noticeable either in the data or the organization or in privacy laws. They may also
be designed to suit the user access tools. In the latter case, there is not much choice
available for design parameters. In the other cases, it is always desirable to design the
data mart to suit the design of the ware house itself. This helps to maintain maximum
control on the data base instances, by ensuring that the same design is replicated in
each of the data marts. Similarly the summary informations on each of the data mart
can be a smaller replica of the summary of the data ware house it self.
Q5.List the application and reasons for the growing popularity of data mining.
Ans-REASONS FOR THE GROWING POPULARITY OF DATA
MINING
a) Growing Data Volume
The main reason for necessity of automated computer systems for intelligent data
analysis is the enormous volume of existing and newly appearing data that require
processing. The amount of data accumulated each day by various business, scientific,
and governmental organizations around the world is daunting. It becomes impossible
for human analysts to cope with such overwhelming amounts of data.
b) Limitations of Human Analysis
Two other problems that surface when human analysts process data are the inadequacy
of the human brain when searching for complex multifactor dependencies in data, and
the lack of objectiveness in such an analysis. A human expert is always a hostage of the
previous experience of investigating other systems. Sometimes this helps, sometimes
this hurts, but it is almost impossible to get rid of this fact.
c) Low Cost of Machine Learning
One additional benefit of using automated data mining systems is that this process has a
much lower cost than hiring an many highly trained professional statisticians. While
data mining does not eliminate human participation in solving the task completely, it
significantly simplifies the job and allows an analyst who is not a professional in
statistics and programming to manage the process of extracting knowledge from data.
Q6What is data mining ? What kind of data can be mined ?
Ans-There are many definitions for Data mining. Few important definitions are given
below.
Data mining refers to extracting or mining knowledge from large amounts of data.
Data mining is the process of exploration and analysis, by automatic or semiautomatic
means, of large quantities of data in order to discover meaningful patterns and rules.
WHAT KIND OF DATA CAN BE MINED?
In principle, data mining is not specific to one type of media or data. Data mining
should be applicable to any kind of information repository. However, algorithms and
approaches may differ when applied to different types of data. Indeed, the challenges
presented by different types of data vary significantly.
Data mining is being put into use and studied for databases, including relational
databases, object-relational databases and object oriented databases, data warehouses,
transactional databases, unstructured and semi structured repositories such as the World
Wide Web, advanced databases such as spatial databases,
multimedia databases, time-series databases and textual databases, and even flat files.
Here are some examples in more detail
Flat files: Flat files are actually the most common data source for data mining
algorithms, especially at the research level. Flat files are simple data files in text or
binary format with a structure known by the data mining algorithm to be applied. The
data in these files can be transactions, time-series data, scientific measurements.
Relational Databases: A relational database consists of a set of tables containing either
values of entity attributes, or values of attributes from entity relationships. Tables have
columns and rows, where columns represent attributes and rows represent tuples.
Q7.Give the top level syntax of the data mining query languages DMQL.
Ans-A data mining language helps in effective knowledge discovery from the data
mining systems. Designing a comprehensive data mining language is challenging
because data mining covers a wide spectrum of tasks from data characterization to
mining association rules, data classification and evolution analysis.
Each task has different requirements. The design of an effective data mining query
language requires a deep understanding of the power, limitation and underlying
mechanism of the various kinds of data mining tasks.
Q8. Explain the meaning of data mining with apriori algorithm
Ans-APriori algorithm data mining discovers items that are frequently associated
together. Let us look at the example of a store that sells DVDs, Videos, CDs, Books
and Games. The store owner might want to discover which of these items customers
are likely to buy together. This can be used to increase the stores cross sell and upsell
ratios. Customers in this particular store may like buying a DVD and a Game in 10 out
of every 100 transactions or the sale of Videos may hardly ever be associated
with a sale of a DVD. With the information above, the store could strive for more
optimum placement of DVDs and Games as the sale of one of them may improve the
chances of the sale of the other frequently associated item. On the other hand, the
mailing campaigns may be fine tuned to reflect the fact that offering discount
coupons on Videos may even negatively impact the sales of DVDs offered in the same
campaign. A better decision could be not to offer both DVDs and Videos in a
campaign. To arrive at these decisions, the store may have had to analyze 10,000 past
transactions of customers using calculations that seperate frequent and consequently
important associations from weak and unimportant associations.

iv) Determine the level of aggregation and the extent of embedding.


v) Design time into the table.
vi) Index the summary table.
2.Q9.Explain the working principle of decision tree used for data mining.
Ans-DATA MINING WITH DECISION TREES
Decision trees are powerful and popular tools for classification and prediction. The
attractiveness of tree-based methods is due in large part to the fact that, it is simple and
decision trees represent rules. Rules can readily be expressed so that we humans can
understand them or in a database access language like SQL so that records falling into
a particular category may be retrieved. In some applications, the accuracy of a
classification or prediction is the only thing that matters; if a direct mail firm obtains a
model that can accurately predict which members of a prospect pool are most
likely to respond to a certain solicitation, they may not care how or why the model
works.
Decision tree working concept
Decision tree is a classifier in the form of a tree structure where each node is either:
a leaf node, indicating a class of instances, or
a decision node that specifies some test to be carried out on a single attribute0value,
with one branch and sub-tree for each possible outcome of the test.
A decision tree can be used to classify an instance by starting at the root of the tree and
moving through it until a leaf node, which provides the classification of the instance.
Example: Decision making in the Bombay stock market
Suppose that the major factors affecting the Bombay stock market are:
what it did yesterday;
what the New Delhi market is doing today;
bank interest rate;
unemployment rate;
Indias prospect at cricket.
Q10. What is Bayes theorem ? Explain the working procedure of Bayesisan
classifier.
Ans-Bayes Theorem
Let X be a data sample whose class label is unknown. Let H be some hypothesis, such
as that the data sample X belongs to a specified class C. For classification problems,
we want to determine P (H/X), the probability that the hypothesis H holds given the
observed data sample X. P(HX) is the posterior probability, or a posteriori probability,
of H conditioned on X. For example, suppose the world of data samples consists of
fruits, described by their color and shape. Suppose that X is red and round, and that H
is the hypothesis that X is an apple. Then P(HX) reflects our confidence that
X is an apple given that we have seen that X is red and round. In contrast, P(H) is the
prior probability, or a priori probability, of H. For our example, this is the probability
that any given data sample is an apple, regardless of how the data sample looks. The
posterior probability, P(HX), is based on more information (Such as background
knowledge) than the prior probability, P(H), which is independent of X.
Similarly, P(XH) is the posterior probability of X conditioned on H. That is, it is the
probability that X is red and round given that we know that it is true that X is an apple.
P(X) is the prior probability of X. . P(X), P(H), and P(XH) may be estimated from the
given data, as we shall see below. Bayes theorem is useful in that it provides a way of
calculating the posterior probability,
P(HX), from P(H), P(X), and
P(XH). Bayes theorem is
P(HX) = P(XH) P(H)
P(X)
Q11.Explain how neural network can be used for data mining ?
Ams-A neural processing element receives inputs from other connected processing
elements. These input signals or values pass through weighted connections, which
either amplify or diminish the signals. Inside the neural processing element, all of these
input signals are summed together to give the total input to the unit. This total input
value is then passed through a mathematical function to produce an output or decision
value ranging from 0 to 1. Notice that this is a real valued (analog) output, not a digital
0/1 output. If the input signal matches the connection weights exactly, then the output
is close to 1. If the input signal totally mismatches the connection weights then the
output is close to 0. Varying degrees of similarity are represented by the intermediate
values. Now, of course, we can force the neural processing element to make a binary
(1/0) decision, but by using analog values ranging between 0.0 and 1.0 as the outputs,
we are retaining more information to pass on to the next layer of neural processing
units. In a very real sense, neural networks are analog computers.
Each neural processing element acts as a simple pattern recognition machine. It checks
the input signals against its memory traces (connection weights) and produces an
output signal that corresponds to the degree of match between those patterns. In typical
neural networks, there are hundreds of neural processing elements whose pattern
recognition and decision making abilities are harnessed together to
solve problems.
Backpropagation
Backpropagation learns by iteratively processing a set of training samples, comparing
the networks
prediction for each sample with the actual known class label. For each sample with the
actual known
class label. For each training sample, the weights are modified so as to minimize the
means squared error
between the networks prediction and the actual class. These modifications are made in
the backwards
direction, that is, from the output layer, through each hidden layer down to the first
hidden layer (hence the
name backpropagation). Although it is not guaranteed, in general the weights will
eventually coverage,
and the learning process stops. The algorithm is summarized in Figure each step is
described below.
Initialize the weights: The weights in the network are initialized to small random
numbers (e.g.,
ranging from 1.0 to 1.0, or -0.5 to 0.5). Each unit has a bias associated with it, as
explained below. The

K-means algorithm
This algorithm has as an input a predefined number of clusters, that is the k from its
name. Means stands for an average, an average location of all the members of a
particular cluster. When dealing with clustering techniques, one has to adopt a notion
of a high dimensional space, or space in which orthogonal dimensions are all attributes
from the table of data we are analyzing. The value of each attribute of an
example represents a distance of the example from the origin along the attribute axes.
The coordinates of this point are averages of attribute values of all examples that
belong to the cluster. The steps of the K-means algorithm are given below.
1. Select randomly k points (it can be also examples) to be the
seeds for the centroids of k clusters.
2. Assign each example to the centroid closest to the example,
forming in this way k exclusive clusters of examples.
3. Calculate new centroids of the clusters. For that purpose average
all attribute values of the examples belonging to the same cluster (centroid).
Q12.Explain the STAR-FLAKE schema in detail.
Ans-STAR FLAKE SCHEMAS
One of the key factors for a data base designer is to ensure that a database should be
able to answer all types of queries, even those that are not initially visualized by the
developer. To do this, it is essential to understand how the data within the database is
used. In a decision support system, which is what a data ware is supposed to provide
basically, a large number of different questions are asked about the same set of facts.
For example, given a sales data question like
i) What is the average sales quantum of a particular item?
ii) Which are the most popular brands in the last week?
iii) Which item has the least tumaround item.
iv) How many customers returned to procure the same item within one month. Etc.,.
Can be asked. They are all based on the sales data, but the method of viewing the data
to answer the question is different. The answers need to be given by rearranging or
cross referencing different facts.
Q13.Explain the method for designing dimension tables.
Ans-DESIGNING DIMENSION TABLES
After the fact tables have been designed, it is essential to design the dimension tables.
However, the design of dimension tables need not be considered a critical activity,
though a good design helps in improving the performance. It is also desirable to keep
the volumes relatively small, so that restructuring cost will be
less. Now we see some of the commonly used dimensions.
Star dimension
They speed up the query performance by denormalising reference information into a
single table. They presume that the bulk of queries coming are such that they analyze
the facts by applying a number of constraints to a single dimensioned data.
For example, the details of sales from a stores can be stored in horizontal rows and
select one/few of the attributes. Suppose a cloth store stores details of the sales one
below the other and questions like how many while shirts of size 85" are sold in one
week are asked. All that the query has to do is to put the relevant constraints to get the
information. This technique works well in solutions where there are a number of
entitles, all related to the key dimension entity.
Q14.Explain the Horizontal partioning in briefly.
Ans-Needless to say, the dataware design process should try to maximize the
performance of the system. One of the ways to ensure this is to try to optimize by
designing the data base with respect to specific hardware architecture. Obviously, the
exact details of optimization depends on the hardware platforms.
Normally the following guidelines are useful:
i. maximize the processing, disk and I/O operations.
ii. Reduce bottlenecks at the CPU and I/O
The following mechanisms become handly
4.3.1 Maximising the processing and avoiding bottlenecks
One of the ways of ensuring faster processing is to split the data query into several
parallel queries, convert them into parallel threads and run them parallel. This method
will work only when there are sufficient number of processors or sufficient processing
power to ensure that they can actually run in parallel. (again not that to run five threads,
it is not always necessary that we should have five processors. But to ensure optimality,
even a lesser number of processors should be able to do the job, provided they are able
to do it fast enough to avoid bottlenecks at eh processor).
Normalisation
The usual approach in normalization in database applications is to ensure that the data
is divided into two or more tables, such that when the data in one of them is updated, it
does not lead to anamolies of data (The student is advised to refer any book on data
base management systems for details, if interested).
The idea is to ensure that when combined, the data available is consistent.
However, in data warehousing, one may even tend to break the large table into several
denormalized smaller tables. This may lead to lots of extra space being used. But it
helps in an indirect way It avoids the overheads of joining the data during queries.
To make things clear consider the following table
Q16.Explain the need of data mart in detail.
Ans-THE NEED FOR DATA MARTS
In a crude sense, if you consider a data ware house as a store house of data, a data mart
is a retail outlet of data. Searching for any data in a huge store house is difficult, but if
the data is available, you should be positively able to get it. On the other hand, in a
retail out let, since the volume to be searched from is small, you can be able to access
the data fast. But it is possible that the data you are searching for may not be available
there, in which case you have to go back to your main store house to search for the
data. Coming back to technical terminology, one can say the following are the reasons
for which data marts are created.
i) Since the volume of data scanned is small, they speed up the query processing.
ii) Data can be structured in a form suitable for a user access too
iii) Data can be segmented or partitioned so that they can be used on different platforms
and
also different control strategies become applicable.

biases are similarly initialized to small random numbers.


Each training sample, X, is processed by the following steps.

3.IDENTIFY THE ACCESS TOOL REQUIREMENTS


Data marts are required to support internal data structures that support the user access
tools. Data within those structures are not actually controlled by the ware house, but
the data is to be rearranged and up dated by the ware house. This arrangement (called
populating of data) is suitable for the existing requirements of data analysis. While the
requirements are few and less complicated, any populating
method may be suitable, but as the demands increase (as it happens over a period of
time) the populating methods should match the tools used.
As a rule, this rearrangement (or populating) is to be done by the ware house after
acquiring the data from the source. In other words, the data received from the source
should not directly be arranged in the form of structures as needed by the access tools.
This is because each piece of data is likely to be used by several access tools which
need different populating methods. Also, additional requirements may
come up later. Hence each data mart is to be populated from the ware house based on
the access tool requirements of the data ware house.
Q17.Explain the Data warehouse process manager in detail.
Ans-DATAWARE HOUSE PROCESS MANAGERS
These are responsible for the smooth flow, maintainance and upkeep of data into and
out of the database. The main types of process managers are
Load manager
Warehouse manager and
Query manager
We shall look into each of them briefly. Before that, we look at a schematic diagram
that defines the boundaries of the three types of managers.
Load manager
This is responsible for any data transformations and for loading of data into the
database. They should
effect the following
Data source interaction
Data transformation
Data load.
The actual complexity of each of these modules depend on the size of the database.
It should be able to interact with the source systems to verify the received data. This is
a very important aspect and any improper operations leads to invalid data affecting the
entire warehouse. The concept is normally achieved by making the source and data
ware house systems compatible.
Ware house Manager
The warehouse manager is responsible for maintaining data of the ware house. It
should also create and maintain a layer of meta data. Some of the responsibilities of the
ware house manager are
o Data movement
Meta data management
o Performance monitoring
o Archiving.
Data movement includes the transfer of data within the ware house, aggregation,
creation and maintenance of tables, indexes and other objects of importance. It should
be able to create new aggregations as well as remove the old ones. Creation of
additional rows / columns, keeping track of the aggregation
processes and creating meta data are also its functions.
Query Manager
We shall look at the last of manager, but not of any less importance, the query
manager. The main responsibilities include the control of the following.
o Users access to data
o Query scheduling
Query Monitoring
These jobs are varied in nature and have not been automated as yet.
The main job of the query manager is to control the users access to data and also to
present the data as a result of the query processing in a format suitable to the user. The
raw data, often from different sources, need to be compiled in a format suitable for
querying. The query manager will have to act as a mediator between the user on one
hand and the meta data on the other.
Q18.Explain the Data warehouse Delivery process in detail.
Ans-THE DATA WAREHOUSE DELIVERY PROCESS
This section deals with the dataware house from a different view point - how the
different components that go into it enable the building of a data ware house. The study
helps us in two ways:
i) to have a clear view of the data ware house building process.
ii) to understand the working of the data ware house in the context of the components.
Now we look at the concepts in details.
i. IT Strategy
The company must and should have an overall IT strategy and the data ware housing
has to be a part of the overall strategy. This would not only ensure that adequate
backup in terms of data and investments are available, but also will help in integrating
the ware house into the strategy. In other words, a data ware house can not be
visualized in isolation.
ii. Business case analysis
This looks at an obvious thing, but is most often misunderstood. The overall
understanding of the
business and the importance of various components there in is a must. This will ensure
that one can clearly justify the appropriate level of investment that goes into the data
ware house design and also the amount of returns accruing.
Unfortunately, in many cases, the returns out of the ware housing activity are not
quantifiable. At the end of the year, one cannot say - I have saved / generated 2.5 crore
Rs. because of data ware housing - sort of statements. Data ware house affects the
business and strategy plans indirectly - giving scope for undue expectations on one
hand and total neglect on the other. Hence, it is essential that the designer
must have a sound understanding of the overall business, the scope for his concept

Q19.Briefly explain the system management tools.


Ans- SYSTEM MANAGEMENT TOOLS
The most important jobs done by this class of managers includes the following
1. Configuration managers
2. schedule managers
3. event managers
4. database mangers
5. back up recovery managers 6. resource and performance a monitors.
We shall look into the working of the first five classes, since last type of managers are
less critical in nature.
Configuration manager
This tool is responsible for setting up and configuring the hardware. Since several types
of machines are being addressed, several concepts like machine configuration,
compatibility etc. are to be taken care of, as also the platform on which the system
operates.
Schedule manager
The scheduling is the key for successful warehouse management. Almost all operations
in the ware house need some type of scheduling. Every operating system will have its
own scheduler and batch control mechanism. But these schedulers may not be capable
of fully meeting the requirements of a data warehouse.
Event manager
An event is defined as a measurable, observable occurrence of a defined action. If this
definition is quite vague, it is because it encompasses a very large set of operations.
The event manager is a software that continuously monitors the system for the
occurrence of the event and then take any action that is suitable (Note that the event is a
measurable and observable occurrence). The action to be taken is also normally
specific to the event.
Database manager
The database manger normally will also have a separate (and often independent)
system manager module. The purpose of these managers is to automate certain
processes and simplify the execution of others. Some of operations are listed as
follows.
Ability to add/remove users
o User management
o Manipulate user quotas
o Assign and deassign the user profiles
Q20.What is schema? Distinguish between facts and dimensions.
Ans- schema- A schema, by definition, is a logical arrangements of facts that facilitate
ease of storage and retrieval, as described by the end users. The end user is not
bothered about the overall arrangements of the data or the fields in it. For example, a
sales executive, trying to project the sales of a particular item is only interested in the
sales details of that item where as a tax practitioner looking at the same data will be
interested only in the amounts received by the company and the profits made.
Distinguish between facts and dimensions
The star schema looks a good solution to the problem of ware housing. It simply states
that one should identify the facts and store it in the read-only area and the dimensions
surround the area. Whereas the dimensions are liable to change, the facts are not. But
given a set of raw data from the sources, how does one identify the facts and the
dimensions? It is not always easy, but the following steps can help in that
direction.
i) Look for the fundamental transactions in the entire business process. These basic
entities are the facts.
ii) Find out the important dimensions that apply to each of these facts. They are the
candidates for dimension tables.
iii) Ensure that facts do not include those candidates that are actually dimensions, with
a set of facts attached to it.
iv) Ensure that dimensions do not include these candidates that are actually facts.
Q21.Explain how to categorize data mining system.
Ans- CATEGORIZE DATA MINING SYSTEMS
There are many data mining systems available or being developed. Some are
specialized systems dedicated to a given data source or are confined to limited data
mining functionalities, other are more versatile and comprehensive. Data mining
systems can be categorized according to various criteria among
other classification are the following
a) Classification according to the type of data source mined: this classification
categorizes data mining systems according to the type of data handled such as spatial
data, multimedia data, time-series data, text data, World Wide Web, etc.
b) Classification according to the data model drawn on: this classification
categorizes data mining systems based on the data model involved such as relational
database, object-oriented database, data warehouse, transactional, etc.
Classification according to the king of knowledge discovered: this classification
categorizes data mining systems based on the kind of knowledge discovered or data
mining functionalities, such as characterization, discrimination, association,
classification, clustering, etc. Some systems tend to be comprehensive systems offering
several data mining functionalities together.
Q22 ..A DATA MINING QUERY LANGUAGE
A data mining query language provides necessary primitives that allow users to
communicate with data mining systems. But novice users may find data mining query
language difficult to use and the syntax difficult to remember. Instead , user may prefer
to communicate with data mining systems through a graphical user interface (GUI). In
relational database technology , SQL serves as a standard core language
for relational systems , on top of which GUIs can easily be designed. Similarly, a data
mining query language may serve as a core language for data mining system
implementations, providing a basis for the development of GUI for effective data
mining. A data mining GUI may consist of the following functional components
a)
Data collection and data mining query composition - This component
allows the user to specify task-relevant data sets and to compose data
mining queries. It is similar to GUIs used for the specification of
relational queries.
b) Presentation of discovered patterns This component allows the display of the
discovered patterns in various forms, including tables, graphs, charts, curves and other
visualization techniques.

(data ware house) in the project, so that he can answer the probing questions.

4.Q.Enlist the desirable schemes required for a good architecture


of data mining systems.
Ans- ARCHITECTURES OF DATA MINING SYSTEMS
A good system architecture will enable the system to make best use of the software
environment , accomplish data mining tasks in an efficient and timely manner,
interoperate and exchange information with other information systems, be adaptable to
users different requirements and evolve with time. To know what are the desired
architectures for data mining systems, we view data mining is integrated
with database/data warehousing and coupling with the following schemes
a) no-coupling
b) loose coupling
c) semitight coupling
d) tight-coupling
No-coupling It means that data mining system will not utilize any function of a
database or data warehousing system. Here in this system , it fetches data from a
particular source such as a file , processes data using some data mining algorithms and
then store the mining result in another file. This system has some disadvantages
1) Database system provides a great deal of flexibility and efficiency at storing ,
organizing, accessing and processing data. Without this in a file, Data mining system
may spend a more amount of time finding, collecting , cleaning and transforming data.
Qno--CLUSTERING IN DATA MINING
Requirements for clustering
Clustering is a division of data into groups of similar objects. Each group, called
cluster, consists of objects that are similar between themselves and dissimilar to objects
of other groups. Representing data by fewer clusters necessarily loses certain fine
details (akin to lossy data compression), but achieves
simplification. It represents many data objects by few clusters, and hence, it models
data by its clusters. Data modeling puts clustering in a historical perspective rooted in
mathematics, statistics, and numerical analysis. From a machine learning perspective
clusters correspond to hidden patterns, the search for clusters is unsupervised learning,
and the resulting system represents a data concept. Therefore, clustering
is unsupervised learning of a hidden data concept. Data mining deals with large
databases that impose on clustering analysis additional severe computational
requirements.
Requirements for clustering
Clustering is a challenging and interesting field potential applications pose their own
special requirements.
The following are typical requirements of clustering in data mining.
Scalability: Many clustering algorithms work well on small data sets containing
fewer than 200 data objects However, a large database may contain millions of objects.
Clustering on a sample of a given large data set may lead to biased results. Highly
scalable clustering algorithms are needed.
Ability to deal with different types of attributed: Many algorithms are designed to
cluster interval-based (numerical) data. However, applications may require clustering
other types of data, such as binary, categorical (nominal), and ordinal data, or mixtures
of these data types.
Nominal, Ordinal and Ratio-Scaled Variables
Nominal Variables
A nominal variable is a generalization of the binary variable in that it can take on more
than two states. For example, map_color is a nominal variable that may have, say, five
states: red, yellow, green, pink and blue.
Nominal variables can be encoded by asymmetric binary variables by creating a new
binary variable for each of the M nominal states. For an object with a given state value,
the binary variable representing that state is set to 1, while variable map_color, a binary
variable can be created for each of the five colors listed above. For an object having the
color yellow, the yellow variable is set to 1, while the remaining four variables are set
to 0..
Ordinal Variables
A discrete ordinal variable resembles a nominal variable, except that the M states of the
ordinal value are ordered in a meaningful sequence. Ordinal variables are very useful
for registering subjective assessments of qualities that cannot be measured objectively.
For example professional ranks are often enumerated in a sequential order, such as
assistant, associate, and full. A continuous ordinal variable
looks like a set of continuous data of an unknown scale; that is, the relative ordering of
the values is essential but their actual magnitude is not. For example, the relative
ranking in a particular sport (e.g., gold, silver, bronze) is often more essential than the
actual values of a particular measure.
Ratio-Scaled Variables
A ratio-scaled variable makes a positive measurement on a nonlinear scale, such as an
exponential scale, approximately following the formula AeBt or Ae-Bt, Where A and
B are positive constants. Typical examples include the growth of a bacteria population,
or the decay of a radioactive element. To compute the dissimilarity between objects
described by ratio-scaled variable There are three methods to handle ratio-scaled
variables for computing the dissimilarity between objects.
Neural Network Topologies
The arrangement of neural processing units and their interconnections can have a
profound impact on the processing capabilities of the neural networks. In general, all
neural networks have some set of processing units that receive inputs from the outside
world, which we refer to appropriately as the input units. Many neural networks also
have one or more layers of hidden processing units that receive inputs only from
other processing units. A layer or slab of processing units receives a vector of data or
the outputs of a previous layer of units and processes them in parallel. The set of
processing units that represents the final result of the neural network computation is
designated as the output units. There are three major connection topologies that
define how data flows between the input, hidden, and output processing units.
Feed-Forward Networks
Feed-forward networks are used in situations when we can bring all of the information

Q24.With the help of a block diagram explain the typical process flow in a data
Warehouse.
Ans- TYPICAL PROCESS FLOW IN A DATA WAREHOUSE
Any data ware house must support the following activities
i) Populating the ware house (i.e. inclusion of data)
ii) day-to-day management of the ware house.
iii) Ability to accommodate the changes.
The processes to populate the ware house have to be able to extract the data, clean it
up, and make it available to the analysis systems. This is done on a daily / weekly basis
depending on the quantum of the data population to be incorporated.
The day to day management of data ware house is not to be confused with maintenance
and management of hardware and software. When large amounts of data are stored and
new data are being continually added at regular intervals, maintaince of the quality of
data becomes an important element. Ability to accommodate changes implies the
system is structured in such a way as to be able to cope with future changes without the
entire system being remodeled. Based on these, we can view the processes that a
typical data ware house scheme should support as follows.
Q25.How the Naive Bayesian classification works.
Ans- Naive Bayesian Classification
The nave Bayesian classifier, or simple Bayesian classifier, works as follows:
1. Each data sample is represented by an n-dimensional feature vector, X = (x1,
x2, . . . . xn), depicting n measurements made on the sample from n attributes,
respectively, A1, A2, . .An.
2. Suppose that there are m classes, C1, C2, . Cm. Given an unknown data sample, X
(i.e., having no class label), the classifier will predict that X belongs to the class having
the highest posterior probability, conditioned on X. That is, the nave Bayesian
classifier assigns an unknown sample X to the class Ci if and only if
P(CiX) > P(CjX) for 1 j m, j I Thus we maximize P(Ci/X). The class Ci for
which P(Ci/X) is maximized is called the maximum posteriori hypothesis. By Bayes
theorem P(CiX) = P(XCi) P(Ci)
P(X) 3. As P(X) is constant for all classes, only P(XCi) P(Ci) need be maximized. If
the class prior probabilities are not known, then it is commonly assumed that the
classes are equally likely,
that is, P(C1) = P(C2) = = P(Cm), and we would therefore maximize P(XCi).
Otherwise, we maximize P(XCi) P(Ci). Note that the class prior probabilities may be
estimated by P(Ci) = si/s where si is the number of training samples of class Ci, and s is
the total number of training samples.
Training Bayesian Belief Networks
In the learning or training of a belief network, a number of scenarios are possible. The
network structure may be given in advance or inferred from the data. The network
variables may be observable or hidden in all or some of the training samples. The case
of hidden data is also referred to as missing values or incomplete data.
If the network structure is known and the variables are observable, then training the
network is straightforward. It consists of computing the CPT entries, as is similarly
done when computing the probabilities involved in native Bayesian classification.
Neural Network Topologies
The arrangement of neural processing units and their interconnections can have a
profound impact on the processing capabilities of the neural networks. In general, all
neural networks have some set of processing units that receive inputs from the outside
world, which we refer to appropriately as the input units. Many neural networks also
have one or more layers of hidden processing units that receive
inputs only from other processing units. A layer or slab of processing units receives a
vector of data or the outputs of a previous layer of units and processes them in parallel.
The set of processing units that represents the final result of the neural network
computation is designated as the output units. There are three major connection
topologies that define how data flows between the input, hidden, and output
processing units.
Backpropagation
Backpropagation learns by iteratively processing a set of training samples, comparing
the networks prediction for each sample with the actual known class label. For each
sample with the actual known class label. For each training sample, the weights are
modified so as to minimize the means squared error between the networks prediction
and the actual class. These modifications are made in the backwards
direction, that is, from the output layer, through each hidden layer down to the first
hidden layer (hence the name backpropagation). Although it is not guaranteed, in
general the weights will eventually coverage, and the learning process stops. The
algorithm is summarized in Figure each step is described below.
Initialize the weights: The weights in the network are initialized to small random
numbers (e.g., ranging from 1.0 to 1.0, or -0.5 to 0.5). Each unit has a bias associated
with it, as explained below. The biases are similarly initialized to small random
numbers.
Nonlinear Regression
Polynomial regression can be modeled by adding polynomial terms to the basic linear
model. By applying transformations to the variables, we can convert the nonlinear
model into a linear one that can then be solved by the method of least squares.
Transformation of a polynomial regression model to a linear regression model.
Consider a cubic polynomial relationship given by
Y = + 1X1 + 2X2 + 3X3
To convert this equation to linear form, we define new variables:
X 1 = X X2 = X2 X3 = X3
Using the above Equation can then be converted to linear form by applying the above
assignments, resulting in the equation Y = + 1X1 + 2X2 + 3X3 which is solvable
by the method of least squares.

to bear on a problem at once, and we can present it to the neural network. It is like a
pop quiz, where the teacher walks in, writes a set of facts on the board, and says, OK,
tell me the answer. You must take the data, process it, and jump to a conclusion. In
this type of neural network, the data flows through the network in one
direction, and the answer is based solely on the current set of inputs.
5. Explain the concept of data warehousing and data mining.
Ans. A data warehouse is a collection of a large amount of data and these data is the
pieces of information Which is use to suitable managerial decisions. (a storehouse of
data) eg:- student data to the details of the citizens of a city or the sales of previous
years or the number of patients that came to a hospital with different ailments. Such
data becomes a storehouse of information.
Data mining is the process of exploration and analysis, by automatic or semiautomatic
means, of large quantities of data in order to discover meaningful patterns and rules.
The main concept of datamining using a variety of techniques to identify nuggets of
information or decision making knowledge in bodies of data, and extracting these in
such a way that they can be put to use in the areas such as decision support, prediction,
forecasting and estimation.
Q15. Define data mining query in term of primitives.
Ans: a) Growing Data Volume: The main reason for necessity of automated
computer systems for intelligent data analysis is the enormous volume of existing and
newly appearing data that require processing. The amount of data accumulated each
day by various business, scientific, and governmental organizations around the world is
daunting.
b) Limitations of Human Analysis: Two other problems that surface when human
analysts processdata are the inadequacy of the human brain when searching for
complex multifactor dependencies in data, and the lack of objectiveness in such an
analysis.
c) Low Cost of Machine Learning: While data mining does not eliminate human
participation in solving the task completely, it significantly simplifies the job and
allows an analyst who is not a professional in statistics and programming to manage
the process of extracting knowledge from data.
Qno-List various applications of Data mining in various fields.

YaExplain in brief the data mining applications


Ans: Data mining has many varied field of application which is listed below:
Retail/Marketing:
Identify buying patterns from customers.Find associations among customer
demographic characteristics
Predict response to mailing campaigns
Market basket analysis
Banking:
Detect patterns of fraudulent credit card useIdentity loyal customers
Predict customers, determine credit card spending
Identify stock trading
Insurance and Health Care:
Claims analysis
Identify behavior pattern of risky customers.Identify fraudulent behavior
Transportation:
Determine the distribution schedules among outlets.Analyze loading
Medicine:
Characterize patient behavior to predict office visits.
Identify successful medical therapies for different illnesses.
Q20.What are the guidelines for KDD environment.
Ans: The following are the guidelines for KDD environment are:1. Support extremely large data sets: Data mining deals with extremely large data
sets consisting of billions of records and without proper platforms to store and handle
these volumes of data, no reliable
data mining is possible. Parallel servers with databases optimized for decision support
system oriented queries are useful. Fast and flexible access to large data sets is of very
important.
2. Support hybrid learning: Learning tasks can be divided into three areas:
a.classification tasks b. knowledge engineering tasks c. problem-solving tasks. All
algorithms can not perform well in all the
above areas as discussed in previous chapters. Depending on our requirement one has
to choose the appropriate one.
3. Establish a data warehouse: A data warehouse contains historic data and is subject
oriented and static, that is, users do not update the data but it is created on a regular
time-frame on the basis of the operational data of an organization.
4. Introduce data cleaning facilities: Even when a data warehouse is in operation, the
data is certain to contain all sorts of heterogeneous mixture. Special tools for cleaning
data are necessary and some advanced tools are available, especially in the field of deduplication of client files.
5. Facilitate working with dynamic coding: Creative coding is the heart of the
knowledge discovery process. The environment should enable the user to experiment
with different coding schemes, store partial results make
attributes discrete, create time series out of historic data, select random sub-samples,
separate test sets and so on.
Q21. Explain data mining for financial data analysis.
Ans: Financial data collected in the banking and financial industries are often relatively
complete, reliable and of high quality, which facilitates systematic data analysis and
data mining. The various issues are
a) Design and construction of data warehouses for multidimensional data analysis
and data mining:
Data warehouses need to be constructed for banking and financial data.
Multidimensional data analysis methods should be used to analyze the general
properties of such data. Data warehouses, data cubes, multifeature and discoverydriven data cubes, characteristic and comparative analyses and outlier analyses all play
important roles in financial data analysis and mining.
b) Loan payment prediction and customer credit policy analysis: Loan payment
prediction and customer credit analysis are critical to the business of a bank. Many
factors can strongly or weakly influence loan payment performance and customer

Q23. What is the importance of period of retention of data?


Ans: A businessman says he wants to the data to be retained for as long as possible 5,
10, 15 years the longer the better. The more data we have, the better the information
generated. But such a view thing is unnecessarily simplistic. If a company wants to
have an idea of the recorder levels, details of sales for last 6 months to one year may be
enough. Sales pattern of 5 years is unlikely to be relevant today. So, It is important to
determine the retention period for each function but once it is drawn, it becomes easy to
decide on the optimum value of data to be stored.
Q25. Give the advantages and disadvantages of equal segment partitioning.
Ans: The advantage is that the slots are reusable. Suppose we are sure that we will no
more need the data of 10 years back, then we can simply delete the data of that slot and
use it again. Of course there is a serious draw back in the scheme if the partitions tend
to differ too much in size. The number of visitors visiting a till station, say in summer
months, will be much larger than in winter months and hence Purchase
recommendations can e advertised on the web, in weekly flyers or on the sales receipts
to help improve customer service, aid customers in selecting items and increase sales
37. Define aggregation. Explain steps require designing summary table.
Ans: Association: - A collection of items and a set of records, which containsome
number of items from the given collection, an association function is anoperation
against this set of records which return affinities or patterns that existamong the
collection of items. Summary table are designed by following the steps
given as follows: a) decide the dimensions along which aggregation is to be done.
b) Determine the aggregation of multiple facts. c) Aggregate multiple facts into
the summary table. d) Determine the level of aggregation and the extent of
embedding. e) Design time into the table. f) Index the summary table.
Q30.Explain horizontal and vertical partitioning and differentiate them.
Ans: HORIZONTAL PARTITIONING-This is essentially means that the table is
partitioned after the first few thousand entries, and the next few thousand entries etc.
This is because in most cases, not all the information in the fact table needed all the
time. Thus horizontal partitioning helps to reduce the query access time, by directly
cutting down the amount of data to be scanned by the queries.
a) Partition by time into equal segments : This is the most straight forward method of
partitioning by months or years etc. This will help if the queries often come regarding
the fortnightly or monthly performance / sales etc.
b) Partitioning by time into different sized segments: This is very useful technique
to keep the physical table small and also the operating cost low.
VERTICAL PARTITIONING- A vertical partitioning schema divides the table
vertically. Each row is divided into 2 or more partitions. i) We may not need to access
all the data pertaining to a student all the time.
For example, we may need either only his personal details like age, address etc. or only
the examination details of marks scored etc. Then we may choose to split them into
separate tables, each containing data only about the relevant fields. This will speed up
accessing.
Q27. Explain data mining for retail industry application.
Ans: The retail industry is a major application area for data mining since it collects
huge amount of data on sales, customer shopping history, goods transportation, and
consumption and service records and so on. The quantity of data collected continues to
expand rapidly, due to web or e-commerce.
a) Design and construction of data warehouses on the benefits of data mining: The
first aspect is to design a warehouse. Here it involves deciding which dimensions and
levels to include and what preprocessing to perform in order to facilitate quality and
efficient data mining.
b) Multidimensional analysis of sales, customers, products, time and region: The
retail industry requires timely information regarding customer needs, product sales,
trends and fashions as well as the quality, cost, profit and service of commodities. It is
therefore important to provide powerful multidimensional analysis and visualization
tools, including the construction of sophisticated data cubes according to the needs of
data analysis.
36.Explain multi dimensional schemas.
Ans: This is a very
convenient method of
analyzing data, when it
goes beyond the normal
tabular relations.
For example, a store
maintains a table of
each item it sells over a
month as a table, in each
of its 10 outlets. This is
a 2 dimensional table. One the other hand, if the company wants a data of all items
sold by its outlets, it can be done by simply by
superimposing the 2 dimensional table for each of
these items one behind the other. Then it
becomes a 3 dimensional view. Then the query,
instead of looking for a 2 dimensional rectangle of
data, will look for a 3 dimensional cuboid of data.
There is no reason why the dimensioning should
stop at 3 dimensions. In fact almost all queries can
be thought of as approaching a multi-dimensioned
unit of data from a multidimensional volume of the
schema. A lot of designing effort goes into optimizing such searches.
Q26. Explain the Query generation.
Ans: Meta data is also required to generate queries. The query manger uses the
metadata to build a history of all queries run and generator a query profile for each
user, or group of uses. We simply list a few of
the commonly used meta data for the query. The names are self explanatory. o QueryTable accessed- Column accessed, Name, Reference identifier. o Restrictions appliedColumn name, Table name, Reference identifier ,Restrictions. o Join criteria appliedColumn name, Table name, Reference Identifier, Column name, Table name,
Reference identifier. o Aggregate function used-Column name, Reference identifier,
Aggregate function. o Syntax o Resources o Disk

credit rating. Data mining methods, such as feature selection and attribute relevance
ranking may help identify important factors and eliminate irrelevant ones.
c) Classification and clustering of customers for targeted marketing: Classification
and clustering methods can be used for customer group identification and targeted
marketing..
Upload By:Abhimnayu kumar singh

You might also like