You are on page 1of 19

 1.Are there prerequisites for installing DB2 Cube View version 8?

A:
Yes, you need to install fixpack 2+.
 Can I upgrade my AIX database to version 8, but keep Warehouse
Manager on version 7? A: Yes, you can upgrade the AIX database to DB2
version 8 and still use Warehouse Manager version 7 to populate it.
Can you explain what the connector for WebSphere Site Analyzer provides and how
it relates to Warehouse Manager? A: There are pre-built connectors that make it
easier to access specific information supporting analysis applications. The product name
is DB2 Warehouse Manager Connector for the Web.

DB2 Warehouse Manager makes it easy to access and analyze data from the WebSphere
Site Analyzer. It provides built-in support to read the Web logs, clickstream data (i.e.,
organize it according to user access), and map it into relational tables for further analysis
and mining.

The goal of the Web connector is to bring in the clickstream data, which is comprised of:

• Site Health: Structural and performance analysis; for example, broken links, page
sizes, page performance, response time, and server capacity.
• Traffic Measures: Hits, pageviews, visits, patterns, and server utilization.
• Visitor Behavior: Origin, site behavior, and profiling.
• Commerce data: Using the WebSphere Commerce Analyzer to incorporate the
business transactions data such as orders and marketing campaigns.

How can I connect to remote servers or databases by using the Preferences


notebook? A: You can specify the servers and the databases you want to connect to
in the Preferences notebook on the Server logon and the Database logon pages.

When you start the Intelligent Miner for Data client, the Preferences notebook is
displayed. You can also go to this notebook by selecting "Options => Preferences" from
the main window.

On the Server logon page, you can specify the following parameters to connect to a
particular server in the appropriate entry fields of the Selection box:

• Server name
• User ID
• Password

To connect to a server, specify these parameters in the Selection box and click the "OK"
button.

You can quickly switch between different servers if you create a list of remote servers in
the Matrix box. If you select a server from this list, the parameters in the Selection box
are changed accordingly. The server that is displayed in the Selection box is the server
you are connecting to.
To connect to the server displayed in the Selection box, click the "OK" button. If the
entry fields in the Selection box are empty and you click the "OK" button, the local server
is used.

To create a list of servers in the Matrix box, specify the required parameters in the
Selection box and click the "Add" button. Repeat these steps until the list of servers is
complete.

• To delete an entry from the list, select this entry from the list and click the
"Delete" button.
• To update an entry in the list, select this entry from the list; the entry will be
displayed in the Selection box. Make your changes in the Selection box and then
click the "Update" button.

On the Database logon page, you can add, update, or delete databases in the same way as
you can do on the Server logon page. However, if you want to connect to a database, the
list of databases in the Matrix box is searched for the user ID and the appropriate
password. The Selection box is only used to add, update, or delete an entry in the list of
database in the Matrix box.

The Default database represents all databases that do not require a password. Do not
update the Default database. If you add a user ID to this database entry, you must also
specify a password. You can only remove the user ID by following these steps:

• Close the Intelligent Miner.


• Make a copy of the file idmgprefs.properties in your Intelligent Miner or home
directory.
• Open the file idmgprefs.properties with a text editor.
• Search for the string Default.
• Remove the user ID that follows the slash. Be careful not to remove any blanks.
• Save the file and reopen the Intelligent Miner.
• If the Intelligent Miner does not start, replace the edited file with the copy of the
unedited file or remove the edited file. If you remove the edited file, the file
idmgprefs.properties is newly generated, but your personal settings are lost.
 Note: Any changes you do on the Server logon or the Database logon pages are
only effective after you click the "OK" or the "Apply" button.
How can I import a mining database that has computed fields defined in DB2 data
objects? A: Before importing a mining database, the description file (.des) must be
changed. Edit the .des file and change all DB2 related names such as the database server
name, schema, table name, and, if required, the table space and OS/390 database name to
the names that are to be used after the mining database has been imported. See Chapter 8,
"Exporting and importing mining bases," of Using the Intelligent Miner for Data V6.1 for
more information.

If you do not change the .des file before importing the mining database and later change
the DB2 data objects on the Intelligent Miner GUI, all the computed fields defined in the
changed DB2 data objects will be lost and the computed fields would have to be added
again manually.

How can Intelligent Miner for Data (IM4D) be used with decision support systems
like the Query Management Facility (QMF) and DB2 OLAP Server? A: IM4D
can be used by decision support systems:

1. With visualizations to help in variable selection.


2. For results interpretation.
3. To help filter and present final results.
4. To automate the data mining process.

How do I create notification message(s) for a process or for a step in a process?


A: Use the following to create notification message(s):

1. Right-click on the process or the step and select Notification . . . from the pop-up
menu to open the Notification Page.
2. Select a value in the Notify on group box.
o Success: click the Success radio button to create a notification instance for
the successful completion of the selected step or process.
o Failure: click the Failure radio button to create a notification instance for
the failure of the selected step or process.
o Completion: Click the Completion radio button to create a notification
instance for the completion of the selected step or process in cases other
than success or failure.
3. In the Type field, select the type of notification from the drop-down list (for
example, e-mail).
4. In the Users field, specify the users and groups that will be notified. Click [...]
button to open the Select Users window.
5. In the Mail server field, specify the TCP/IP address or the name of the SMTP
server.
6. In the Current messages field, click on the Edit . . . button to change the default
messages (Note: this is optional).
7. Click the Add button to add the notification to the notification list.
8. Click OK to create the specified notification message.
 Process complete.
How do I define a mail server when creating a notification for a process or process
step? A: There needs to be a setup of the Warehouse User ID for e-mail notification
at:
Administration -> Warehouse -> Users and Groups -> Warehouse Users.

When you define the user you also define the e-mail address where notifications will be
sent for that user. There is a field to enter the e-mail ID on the properties notebook for the
user.
For each step that needs a notification, open the schedule notebook to the
Notification Tab: Highlight the step --> Right Click -> select Notification.

The Schedule Notebook opens to the Notification Tab.

On the Notification Tab, specify the Warehouse User to send the Notification and your
SMTP mail Server. Make sure you press the Add button to add the notification to the
right side of the screen. You can add more than one notification so you can notify more
than one user if necessary.

How do I get around: DB210041E Request for DB2 Query Patroller license failed?
A: Check the following:

1. Nodelock file to show any license.


2. .../SQLLIB/CFG directory to show DQPLEVEL and DB2UDBEE.LVL.
3. Run the command DB2LICM -L Logon as the instance and connect to the
database.

These steps should correct the problem:

1. Bind the Query Patroller bind files with:

db2 bind @db2qp.lst grant public blocking all


db2 bind @db2qp_sp.lst

2. Run db2stop force.


3. Open another session and log on as iwm or su - iwm.
4. Enter ps -ef|grep iwm_ to see if there are processes with a name like iwm_*, and
kill them with kill -9 pid.
5. Enter rm /tmp/iwm*.
6. Run ipclean -a.
7. On the instance session, run db2start.

How do I set the log level higher for more detailed information within Data
Warehouse Center 7.2? A: Within DWC, log level capability can be set from 0 to
4. There is a log level 5, yet it cannot be turned on using the GUI, but must be turned on
manually. A command line trace can be used for any trace level, and this is the only way
to turn on a level 5 trace:

1. Go to start, programs, IBM DB2, command line processor.


2. Connect to the control database:

db2 => connect to Control_Database_name

3. Update the configuration table:


db2 => update iwh.configuration set value_int = 5 where name = 'TRACELVL'
and (component = '<component name>')

Valid components are:

a. Logger trace = log


b. Agent trace = agent
c. Server trace = RTK
d. DDD = DDD
e. ODBC = VWOdbc
4. For multiple traces the format is:

db2 => update iwh.configuration set value_int = 5 where name = 'TRACELVL'


and (component = '<component name>' or component = '<component name>')

5. Reset the connection:

db2 => connect reset

6. Stop and restart the Warehouse server and logger.


7. Perform the failing operation.
8. Be sure to reset the trace level to 0 using the command line when you are done:

db2 => update iwh.configuration set value_int = 0 where name = 'TRACELVL'


and (component = '<component name>')

9. When you run a trace, the Data Warehouse Center writes information to text files.
Data Warehouse Center programs that are called from steps also write any trace
information to this directory. These files are located in the directory specified by
the VWS_LOGGING environment variable.
10. The default value of VWS_LOGGING is:
a. Windows and OS/2 = x:\sqllib\logging
b. UNIX = /var/IWH
c. AS/400 = /QIBM/UserData/IWH

How does Intelligent Miner for Data (IM4D) use Data Warehouse? A: IM4D
can use Data Warehouse to:

1. Consolidate crucial business data, cleaned, consistent, and organized in ways to


facilitate analysis.
2. Collect data access and manipulation functions.
3. Reduce the cost and difficulty of data mining.
4. Provide higher quality information from data mining processes.
 I am cataloging data from a data warehouse to be used by end users. When
initializing the Information Catalog, why do I receive an error about the
need to purchase a license for the product? A: You are attempting to use
the Information Catalog Manager Initialization utility without a valid license.
You must purchase the DB2 Warehouse Manager or the IBM DB2 OLAP Server
and install the Information Catalog Manager component which includes the
Information Catalog Initialization utility.
 I am merging data which copies two flat files into a relational table. When
I try to promote to test, why do I receive the error: DWC07356E RC2 =
8565? A: Verify that the SQL transformation properties within the "step-
name" listed in the error message have the column mapping correctly linked.
 I am using DB2 7.2 with Data Warehouse Center (DWC) to define a flat file
within the warehouse sources. After defining the flat file from a drive, why
do I get the error DWC07356E RC=7356 RC2=8430 when trying to access
the fields tab, instead of showing the actual fields of the specified flat file?
A: The RC is a generic return code. The secondary return code
(RC2=8430) suggests checking the text file. When you create a text file like
action.txt, Windows adds a double extension such as action.txt.txt. You must
rename the file.
 I am using DB2 UDB version 7.2 on Windows NT. In the Data Warehouse
Center, I want to create a query the normal way rather than editing SQL
ASSIST. I need to put a restriction within a join instead of the where clause
and would like to be able to add columns and change the query using the
GUI. Is this possible? A: The GUI is supported by SQL ASSIST. SQL
ASSIST does not support the desired action at this time. It appears that the
product's inability to meet your request is a limitation with SQL ASSIST. In the
version of SQL ASSIST used by Data Warehouse Center, you can only create
'JOIN ... ON' join conditions of the form 'column operator column'.
I am using Data Warehouse Center 7.2 on a Windows 2000 platform. I define some
tables in Warehouse Target and use it in some SQL steps. Then I drop them from the
SQL steps and also from the DB2 database. When I try to drop them from
Warehouse Target, why do I receive a DWC03157E error? A: It seems you have
join and star schema information that was not deleted for these tables. There are existing
tables you are not able to delete. They cannot tell which are legitimate joins and which
need to be deleted. So if you have these problems again you can repeat these instructions
for any new tables.

Connect to the control database using the command line processor and run these
commands:

> select iwhid, name, irname from iwh.dataresource where largetablename = 'exact
tablename for the table to delete'
Make sure you use the IWHID for the right table in these next two statements in this
order.
> delete from iwh.attributelink where iwhid in (select source_iwhid from
iwh.relationship where relation_name = 'AttLink1_TO_Attr_Rel' and target_iwhid in
(select target_iwhid from iwh.relationship where relation_name = 'DR_TO_Attr_Rel '
and source_iwhid = 'IWHID'))
> delete from iwh.relationship where relation_name = 'AttLink1_TO_Attr_Rel' and
target_iwhid in (select target_iwhid from iwh.relationship where relation_name =
'DR_TO_Attr_Rel' and source_iwhid = 'IWHID')
For example:
> delete from iwh.attributelink where iwhid = '0000018467'
> delete from iwh.attributelink where iwhid = '0000006334'
> delete from iwh.relationship where relation_name = 'StarSchema_TO_AttLnk'
> delete from iwh.relationship where relation_name = 'StarSchema_TO_DR'
> delete from iwh.attributelink where iwhid in (select source_iwhid from
iwh.relationship where relation_name = 'AttLink1_TO_Attr_Rel' and target_iwhid in
(select target_iwhid from iwh.relationship where relation_name = 'DR_TO_Attr_Rel '
and source_iwhid = '0000017585'))
> delete from iwh.attributelink where iwhid in (select source_iwhid from
iwh.relationship where relation_name = 'AttLink1_TO_Attr_Rel' and target_iwhid in
(select target_iwhid from iwh.relationship where relation_name = 'DR_TO_Attr_Rel '
and source_iwhid = '0000017398'))
> delete from iwh.attributelink where iwhid in (select source_iwhid from
iwh.relationship where relation_name = 'AttLink1_TO_Attr_Rel' and target_iwhid in
(select target_iwhid from iwh.relationship where relation_name = 'DR_TO_Attr_Rel '
and source_iwhid = '0000012324'))
> delete from iwh.relationship where relation_name = 'AttLink1_TO_Attr_Rel' and
target_iwhid in (select target_iwhid from iwh.relationship where relation_name =
'DR_TO_Attr_Rel' and source_iwhid = '0000017585')
> delete from iwh.relationship where relation_name = 'AttLink1_TO_Attr_Rel' and
target_iwhid in (select target_iwhid from iwh.relationship where relation_name =
'DR_TO_Attr_Rel' and source_iwhid = '0000012324')
> delete from iwh.relationship where relation_name = 'AttLink1_TO_Attr_Rel' and
target_iwhid in (select target_iwhid from iwh.relationship where relation_name =
'DR_TO_Attr_Rel' and source_iwhid = '0000017398')
 I created multiple steps with on success task flow, then tried to use the work
in progress and run the first step continuous to each step in succession.
After the first step, why does the second step not start? A: Your steps
are most likely still in test mode. When a step is in test mode the "on success"
task flow does not take effect. You must promote to production mode.
 Informix XPS has an "external table" concept that would be very useful
for DB2 Data Warehouses. Does DB2 have anything similar? A: DB2
can support external objects (like flat files) and make them look like relational
tables through a feature called "user-defined table functions." You create a
function that maps the data from the external source into a set of columns and
then DB2 can manipulate this data (read only) just like an SQL table. You can
then apply any SQL logic against it.
 Once I am within Intelligent Miner for Data (IM4D), how do I get back to
the logon screen? A: Within the IM4D toolbar, select options, then
preferences.
 There are ODBC Merant drivers within DB2 Data Warehouse Center.
What versions of relational databases (Oracle or Informix) are supported
with the Merant drivers? A: You can find the supported relational
databases for Merant Drivers from the ReleaseNotes.txt file within the fix pack
download, e.g. DB2 version 7.1 fix pack 3 release notes.
What are different situations when dealing with DB2 Cube View? A:
There are normally four situations to look at:

• If you have no OLAP installed, then you can:


o Write OLAP applications directly to DB2 Cube Views or to DB2 OLAP
Server.
o Use existing front-end tools.
o Define enterprise-wide OLAP strategy.
• If DB2 OLAP is installed, then you can:
o Retain DB2 OLAP - use current applications/tools.
o Add DB2 Cube Views for OLAP acceleration.
o Write OLAP applications directly to DB2 Cube Views.
o Accelerate DB2 OLAP cube load cycle.
• If an application is installed with Essbase, then you can:
o Retain Essbase and application - use current applications/tools.
o Add Cube Views for OLAP acceleration.
o Simplify some cube management in Essbase.
o Accelerate Essbase cube load cycle.
• If a front-end tool installed that does OLAP, then you can:
o Add DB2 Cube Views for cube slice dispensing.
o Continue using local persistent or cached cubes.
o Ensure consistent cubes across tools and other applications.
o Add DB2 OLAP to expand advanced analysis capabilities.

What are some advantages of using Intelligent Miner for Data (IM4D) with Data
Warehouse? A: Advantages of using IM4D with Data Warehouse include:

 Clean, consistent, organized data.


 Reduced cost and difficulty of mining.
 Higher quality information from data mining.
 Help to filter and present final results.

 What are some common mistakes and misconceptions about data marts
and data warehouses? A: Some common mistakes and misconceptions
are:

A data mart is not a data warehouse: The scope of a data mart is a given decision
support application, whereas the data warehouse consolidates all data for corporate-
wide analysis.
A data mart is not just a small data warehouse: Although a data mart is usually
smaller, the difference in size is not a dominant differentiation aspect.

A collection of data marts is not a data warehouse: Corporate-wide consolidation


of informational data in a "unified source" is what is usually missing across
independently developed data marts.

A data warehouse is not a Decision Support System (DSS): A DSS comprises the
data warehouse as well as the application environment to query, manipulate, analyze,
and visualize data and information.

Note: As a rule of thumb, a data mart is used with a single entity and a data warehouse
is used corporate-wide.

What are suggestions for improving performance when using Intelligent Miner for
Data (IM4D) with DB2? A: To improve performance using IM4D with DB2, you
should:

1. Create indexes on DB2 tables.


2. Set the IDM_CS_TIMEOUT variable to avoid timeouts in slow networks.
3. Minimize input variables to neural classification (only contributors).
4. Presort association input data by transaction ID.
5. Presort sequential pattern input data by transaction group and transaction ID.
6. Use an appropriate memory size setting in the prefences window to avoid
swapping.
7. Shorten clustering run times by increasing accuracy rate and decreasing the
number of passes.

What are the basic elements of a Multi-Dimensional Data Model (MDDM) and an
example of each? A: To build an initial MDDM, the following base elements have
to be identified:

1. Measure -- Information analysts query data items to measure the performance or


behavior of a business process or a business object.
Example: End-users analyzing the "performance" of the movie theater ticket
selling process will basically be interested in the numbers of tickets sold. Also,
sample types of measures are quantities, sizes, amounts, durations, and delay.
2. Dimension -- Information analysts examine an entity or a collection of related
entities to identify the context of the measures they work with.
Example: Quantity of tickets sold are associated with a given combination of
movie, time, and geography. Also, dimensions determine the contextual
background for measures.
3. Fact -- Facts are a collection of related measures together with their associated
dimensions represented by the dimension keys.
Example: You have a fact table called "Ticket Sales." The schema contains a fact
identifier, dimension keys which link to the dimensions, measures, and various
kinds of supportive attributes; these are usually derived from other data items in
the model. Also, a fact can represent a business object, a business transaction, or
an event that is used by the information analyst.
4. Grain and granularity -- The grain of a dimension is the lowest level of detail
available within that dimension. The granularity of a measure is determined by the
combination of the grains of all its dimensions.
Example: The granularity of a measure determines the level of detail at which
information analysts will be able to analyze it. Also, the finer the granularity, the
more detailed the information analysis can be performed.
5. Dimension hierarchy and aggregation level -- Dimensions consist of one or
more dimension hierarchies. Each dimension hierarchy can include several
aggregation levels.
Example: The geography dimension could, for instance, consist of a hierarchy
corresponding to the relationship between cities and countries. Also, each
dimension hierarchy usually includes several so-called aggregation levels. City
and country are examples of such aggregation levels within the geography
dimension.

What are the characteristics of data warehousing? A: The characteristics of data


warehousing are:

• Data organization with a corporate-wide scope of use for decision support and
informational applications.
• Non-volatile store of business data, transactions, and events.
• Built on the basis of a historical(temporal) data model, possibly combined with
other information dimensions.
• Integrated, consolidated, secured, and cleansed data.
• The source of information for all decision support and informational
applications.

What are the concepts behind data warehousing? A: The concepts of data
warehousing are:

1. A common data warehouse populating subsystem.


2. A central data warehouse(CDW) within a managed data warehouse
environment.
3. Data marts holding information specifically designed for end-users.
4. Subsystems to populate the data marts and to provide groups of end-users
transparent access either to primitive data in the CDW or to derived information
in their data mart.
5. A well-designed decision support subsystem suitable for other informational
applications

What are the functional requirements for a data warehouse? A: A data


warehouse must be able to support various types of information applications.
• Decision support processing is the principle type of information application in a
data warehouse, but the use of a data warehouse is not restricted to a decision
support system.
• It is possible that each information application has its own set of requirements in
terms of data, the way that data is modeled, and the way it is used.
• The data warehouse is where these applications get their "consolidated data."

A data warehouse must consolidate primitive data and it must provide all facilities to
derive information from it, as required by the end-users.

• Detailed primitive data is of prime importance, but data volumes tend to be big
and users usually require information derived from the primitive data.

Data in a data warehouse must be organized such that it can be analyzed or explored from
different angles.

• Analysis of the historical context (the time dimension) is of prime importance.


• Examples of other important contextual dimensions are geography, organization,
products, suppliers, customers, and so on.

What are the methods of classification used by Intelligent Miner for Data (IM4D)?
A: IM4D classification methods include:

1. Linear discriminant analysis.


2. Logistic regression.
3. Multinomial logistic regression.
4. Tree induction (tree learning).
5. Neural networks.

The last two are most commonly used by IM4D.

What are the potential drawbacks of a data mart? A: The potential drawbacks of
a data mart are:

• Each data mart modeled from the data sources.


o No common, consolidated source for the various marts.
• No common data consolidation or cleansing.
o Vital but also very costly and complicated.
• Cross data mart issues (consistency, integration, etc.) are difficult to resolve.
o Especially when the number of marts increases.

What are the process steps for Intelligent Miner for Data (IM4D)? A: IM4D
process steps include:

1. Identify business problems.


2. Collect data.
3. Prepare data.
4. Mine data.
5. Analyse results.
6. Report results.

What are the types of Intelligent Miner for Data (IM4D) kernels? A: Types of
IM4D kernels include:

1. Associations.
2. Similar time sequences.
3. Sequential patterns.
4. Clustering (demographic, neural).
5. Predicting values (RBF, neural).
6. Classification (tree and neural induction).

What factors could cause performance concerns with Intelligent Miner for Data
(IM4D)? A: Factors that might cause performance concerns include:

1. Amount of data being mined.


2. Hardware features, capacity.
3. Intelligent Miner kernel used.
4. Data variables specified.
5. Database tables versus flat files.
6. Quality of the data.

What is Common Format Interface Files (CFIF) and its purposes? A: The CFIF
is part of the data warehouse design process. The files consolidate data in a consistent,
well-defined structure, irrespective of the source they came from. The main purposes of
the CFIFs are:

• Provide a mechanism to reconstruct records of consistent data out of the data in


the capture files.
• Resolve conflicts due to data being available and captured in different sources.
• Filter data out of the population process, the data which is irrelevant, unwanted,
inconsistent, and incomplete.

Note: The types of capture files include: extract files or tables, image copies, changed
data files or tables, DBMS logs or journal, message files, and event logs.

What is Operation Data Store and its characteristics? A: Operation Data Store
(ODS) is a Business Intelligence environment/solution component that supports time-
sensitive operational decision support (e.g., customer services). The ODS is narrowly
focused on a particular set of business processes.

The characteristics are:


• User updatable to the ODS and operational sources.
• Focus on current data (near real time access).
• High data volatility.
• Focus on integrated, detailed, and granular data.
• Complements or extends operational systems.

Examples:

• Call Center
• Internet
• Transportation Capacity Management
• Network Optimization
• Risk Approval
• Load Authorization
• Fraud Detection

What is a DB2 Cube View? A: A DB2 Cube View accelerates OLAP queries by
using more efficient DB2 materialized query tables. Please read an overview of DB2
Cube Views.

What is a Management Information System and its characteristics? A:


Management Information System (MIS) is a query and reporting system that performs
aggregations and summarizations against operational data directly, sometimes storing
summarized and aggregated data privately. An MIS provides query and reporting
capabilities for decision makers (managers, executives). It is also know as "a virtual
warehouse."

The characteristics are:

• Direct access to source data and data transformations with the data access process.
• Focused on the current state of the data or whatever "history" is available in the
data sources.
• Complex query transformations required when processing heterogeneous sources.
• Difficult to incorporate other data (external data, unstructured data) in the process.
• Tendency to "support" proliferation of source extracts to personal data stores

What is a data mart and the advantages of using it? A: A data mart is an
implementation of a data warehouse with a small and more tightly restricted scope of data
and data warehouse functions, serving a single department or part of an organization.

The advantages are:


• A data mart can match very closely with data requirements of specific end-user
workgroups, departments, or small organizations. As such, data marts can provide
highly appreciated solutions.
• Since data marts have a smaller scope than data warehouses, they can usually be
established quickly and can, therefore, generate a faster return on investment.
• Data marts can also be used proficiently for demonstration or prototyping
purposes, or for proving to end-users that data organizations coupled with good
tools for decision support can really be beneficial.
• Data marts usually integrate well with Decision Support System tools and
complementary products. However, beware of tool integration and corporate IS
strategies too heavily based on tool selection.

What is a warehouse source? A: Warehouse sources identify the tables and files
that will provide data to your warehouse. The Data Warehouse Center uses the
specifications in the warehouse sources to access the data. The sources can be nearly
any relational or nonrelational source (table, view, or file) that has connectivity to your
network.

What is an agent site? A: An agent site is a logical name for a workstation where
agent software is installed. The agent site name is not the same as the TCP/IP host name.
A single workstation can have only one TCP/IP host name. However, you can define
multiple agent sites on a single workstation. A logical name identifies each agent site.

The default agent site, named the Default DWC AgentSite, is a local agent on Windows
NT that the Data Warehouse Center defines during initialization of the warehouse control
database.

What is the difference between data mining and statistics? A: Some differences
between data mining and statistics are:

1. Data mining does not require statistical assumptions.


2. Data mining algorithms are scalable.
3. You can access databases directly with data mining.
4. Parallel mining is possible for large applications.
5. Use of data mining results is more straightforward.

What layers make up Intelligent Miner for Data (IM4D)? A: IM4D layers
include:

1. Data access - a source of relational data from DB2 and flat files. Non-DB2
databases can also be accessed via IBM DataJoiner.
2. Client/server - an area between jobs that run on the server, having to do with
storage and processing. On the client, it has to do with setup, control and results.
3. Environment - an interface to the external world. It consists of a set of classes
representing the basic objects of Intelligent Miner.
4. Results - the results of processing data by the different mining kernels.
5. Mining kernels - engines for doing the actual mining.

What processing functions restrict the amount of data to be processed by Intelligent


Miner for Data (IM4D)? A: IM4D processing functions to restrict data are:

1. Filter fields.
2. Filter records.
3. Get random sample.
4. Filter records using a value set.
5. Discard records with missing values.

What sources does Intelligent Miner for Data (IM4D) use? A: Sources used by
IM4D include:

1. Internal data consisting of production transaction details like name, address,


purchase history, product inventory, or billing history, i.e. normal production data.
2. Compiled/external data like business data from groups like Dun & Bradstreet, the
census bureau, Standard and Poor's, often purchased from other sources.
3. Behavioral data like book subscriptions, credit card usage details, responses to
mailings, travel patterns, and job history. This could be a by-product of the
production applications, or it might be externally purchased.
4. Modeled data that has already had some analysis, coding and transformations
applied to it, such as mosaic code and prism code. This can be obtained from
other sources, and has already been processed and coded in some way. An
example might be data about people with their salary ranges and geographical
location, yet without actual identification of individuals.

What technologies are used in Intelligent Miner for Data (IM4D)? A: Examples
of IM4D technologies include:

1. Clustering - a pensions supplier identifies a group of high earning, self-employed


people in their mid thirties, and creates a plan for them.
2. Analysis - a supermarket identifies people who buy product A.
3. Prediction - a finance company looks at existing customers and identifies new
applicants likely to default on payment.
4. Classification - a company selects people most likely to respond to a mail shot.

When installing Query Patroller Server 7.1, what database name is to be used
during installation configuration? A: It is suggested you create a new database.
On Windows NT, the database is created automatically by entering the database name.
On Windows 2000, the database has to be created before the installation.

Where are the Intelligent Miner for Data (IM4D) log files? A: Intelligent
Miner for Data (IM4D) has built-in diagnostic functions. One setting is to start tracing
activity. You can set levels of tracing from 1 to 10. Ten is the highest and gives the
most detail. Records will be written to a log file with an extension of TRC. You can
also get an error file for major problems with an extension of ERR. These will be put
in the idmres directory.

Where can I download IM4D version 6.1.1? A: From the DEMOpkg FTP site,
you can download IM4D version 6.1.1 either as a zip file for Windows or as a CD
image (for all supported platforms) so you can burn your own CD.

Where can I get Fixpack 3 (special build) for Data Warehouse Center version 7.1
and what fixes are included? A: The special build is available for Windows
platforms only. All languages are available for this replacement. Replace "english-us" in
the URL with the appropriate language code.

After applying the special build, execute command "DB2LEVEL" and you will see the
following information:

DB21085I Instance "DB2" uses DB2 code release "SQL07020" with level identifier
"03010105" and information tokens "DB2 v7.1.0.42," "special_3834," and "WR21254w."

This special build includes the following fixes:

• Default agent hostname may become invalid when updating userid and/or
password.
• Migration may generate SQL steps with improper SQL when source data is from
flat files and the customer originally was on Visual Warehouse.
• Trace directory may reset to invalid directory after updates to Warehouse
Properties.
• Performance issue when modifying large SQL steps within the DWC navigator.
• DWC Logger stops logging messages if message contains variables > 512 bytes.
• Command Center output of "WITH" queries not output to Query Result Page
when user has chosen to display results on Query Result Page.

Where do I find documentation for Intelligent Miner for Data version 6.1? A:
Follow the information in the users guide, which can be found as soft copy (HTML and
PDF) in English on the clients CD1 in the doc directory.

Depending on the language, use clients CD1 for Latin1 or clients CD2 for DBCS.

Why would I get a DWC06200E error when logging into Data Warehouse Center 7.2
using a remote control database? A: DWC06200E says an unexpected
communications error has occurred. To eliminate the error:

1. Verify connection and logon of server/client databases.


2. DWC agent, server, and logger are stopped and started.
3. Initialize DWC control database (which is the remote database).
4. Log onto DWC, click the advanced tab, and verify:
a. Control database: remote database.
b. Server host name: local server name.
c. Server service name: vwkernel.

Within Data Warehouse Center (DWC) version V7.2, I am having trouble demoting
a step that is related to the special character "#" in the step name. When demoting
the step from test to development I receive RC = 7356 and RC2 = 8535. How do I
demote a step? A: To demote the step, use the external trigger function noted with
the Data Warehouse Center.

You will need JDK1.1.7 or later installed and do the following:

1. Start a step from outside the DWC


2. Run the following command: java XTServer 11004
3. From another command window, run: java XTClient hostname 11004 user
password "step_name" 5
o For example: java XTClient test1 11004 test1 test1 "Load Demographics
Data" 5
o 5 is to demote to development mode

For additional information look at the external trigger.

Within Data Warehouse Center 7.2, how do I copy a table from one process into
another process? A: There is no way to do this directly, yet there is a workaround:

1. Define the warehouse source.


2. Open the original process and copy the table into the warehouse source specified
above.
3. Create a new process.
4. Open the new process and specify the warehouse source and target.
5. Select the warehouse source specified above.
I am currently using DB2 version 8 and trying to FAQ help
find the 'Manage Information Catalog Wizard' Search tips
option but am not able to do so. How can I get that
component?

A: You will
need to
install
Warehouse
Manager
version 8
and select
'Information
Catalog
Manager
Tools' under
the
installation
options.
Once this is
successfully
installed,
you should
be able to
start the
wizard by
clicking
Start -->
Programs --
> IBM DB2
--> Set-up
tools-->
Manage
Information
Catalog
wizard.
Reference:
81258
Date: 03 Jul
2007 [
Keyword:
Data
Warehouse
]
When a sequence seq1 is defined for multiple users (schemas), sometimes a user
(say A) when accessing seq1 via implicit schema, receives SQL0551 about not having
privilege to access B.seq1 (where B is another user/schema) when B made a prior
access or is currently accessing seq1 in the same session or via a parallel session.
How do I resolve this? A: This error occurs when accessing the sequence (seq1)
using either of the expressions:

VALUES (PREVVAL FOR seq1)


VALUES (NEXTVAL FOR seq1)

The error is resolved by varying the format of the above expressions as in:

1. VALUES PREVVAL FOR seq1


2. VALUES(PREVVAL FOR seq1)
3. VALUES (PREVVAL FOR seq1)

However, the particular format that resolves the error is not consistent for all users -- so
while format 1 resolves the error for user A, it may not resolve the error for user B.

This behavior is a defect that will be fixed in fixpack 7 for DB2 UDB version 7.2. The
associated APAR is IY29232.

You might also like