You are on page 1of 19

http://dwbi.

org/

1. Question1. What Are Business Objects?


Answer :
Business Objects are basically called domain objects. A domain model shows a set of
domain objects and explains how they related to each other. Hence a business model
also will show a set of domain objects and will show the relationship between them.

2. Question2. What Is A Universe?


Answer :
Universe is called as the semantic layer between the Database and the one who designs
inorder to create the objects and the classes. Universe hence will map to the data in the
Database.

3. Question3. What Is An Object?


Answer :
Object is just an entity. The object is called as the instance of a class.

4. Question4. What Is Business Objects Repository?


Answer :
The term Business objects Repository is the metadata only. Repository will be like a
database. Business Object Repository will store business objects like user, data, access
styles, access permissions. Business objects Repository is a centralized concept and a
set of data structures can be stored.

5. Question5. What Is Domain?


Answer :
Domain is called as a logical grouping of system tables.

6. Question6. How Many Types Of Domain Are There In A Basic Setup?


Answer :
There are normally 3 types of domain in a basic setup. They are
1. Secure
2. Universe
3. Document
7. Question7. What Is Called A Category?
Answer :
category is called as the grouping of various entities.

8. Question8. What Is A Bomain.key?


Answer :
BOMain. Key is a file which is containing all the information about the repository site
.Therefore it will contain the whole address of the repository security domain.

9. Question9. Can We Have Multiple Domains?


Answer :
Yes. We can have multiple domains

10. Question10. Which Is The Domain Can't Be Have Multiple?


Answer :
The domain which cannot be having multiple is the security Domain.

11. Question11. When Is The Repository Created?


Answer :
Repository Creation will depend on the version. Repository creation will happen after
installing the software in 5i/6i versions after installing the software, whereas in the Xi
version the repository will be created at the time of installation.

12. Question12. How The Access To The Rows Of Database Is

Restricted?
Answer :
The restrictions also will depend on the Version. In XI version we can be done this by
using the row-level security in the designer module. In 5i/6i version the restriction of the
access to the database is done by the supervisor.

13. Question13. What Is The Term Object Qualific ation Referring To?
Answer :
Object qualification is just representing what kind of object is that. The three types of
object qualifiers are:
1. Measure
2. Dimension
3. Detailed

14. Question14. Which Will The Size Of A Database?


Answer :
The size of a database will be between 4 and 8 Tera Bytes.

15. Question15. Which Is A Loop?


Answer :
Loop is circular flow.

16. Question16. How Can We Overcome Loop?


Answer :
Loop can be overcome by Alias and Context.

17. Question17. How Is Called As A Join?


Answer :
Join is normally used for linking tables depending upon how the data is requested by the
user.

18. Question18. Explain The Different Types Of Joins?


Answer :
Joins are of different types. They are
1. Inner join
2. Outer Join
3. Left Join
4. Right Join
5. Full Outer join

19. Question19. What Are Called Linked Universes?


Answer :
Sometimes the data provided may be from two different universes and the data in these
two universes can be linked someway. These types of universes are called Linked
Universes.

20. Question20. What Is Called Alerter?


Answer :
Alerter is used to give alert or attention to a block of data by highlighting it.

21. Question21. What Is Called Filters?

Answer :

Filters are used to get necessary data from a group of data.

22. Question22. What Is Called Breaks?

Answer :

Breaks are used to group the data without changing anything in the format.

23. Question23. What Are Called Conditions?

Answer :

conditions are used to get some data based on certain criteria’s or conditions.

24. Question24. What Is The Difference Between Master-detail And

Breaks?

Answer :

In Breaks common entities will be deleted. In master-detail we will declare a certain


entity as a master to get the detailed information or it will report in this case that the table
format is changed.

25. Question25. What Are Metrics?

Answer :

Metrics are a system of parameters .Metrics is a way of quantitative and periodic


assessment of a process that is to be measured. Metrics are normally used to track
trends, productivity.
26. Question26. What Is A Set?

Answer :

Set is called as the grouping of a number of users.

27. Question27. What Is The Advantage Of Afd?

Answer :

AFD is used to create Dash Boards.

28. Question28. Where Normally Afd's Are Stored?

Answer :

AFDs are normally stored in repository, corporate or personal.

29. Question29. What Is The Source Of Metrics?

Answer :

The source of Metrics is by measure objects.

30. Question30. How Many Analysis Are Used In Business Objects?

Answer :

2 types of analysis are used in Business Objects. They are


1. Slice-Dice Analysis
2. Drill Analysis

31. Question31. What Are The General Issues In Migration Process?

Answer :
The general issues in migration process are
1. Alignment
2. Performance

32. Question32. What Is Solution For Improving Performance?

Answer :

The main solution for improving performance is to make use of the Aggregate tables.

33. Question33. Types Of Sap Businessobjects Universe In Sap

Businessobjects?

Answer :

Simple SAP Businessobjects universe and complex SAP Businessobjects


universe.

34. Question34. What Is The Main Use Of Bca?

Answer :

The main use of BCA is that BCA can be used for refreshing and scheduling and
exporting and saving in different format like .html, .rtf, .xls and .pdf

35. Question35. What Are Called Universal Parameters?

Answer :

RDBMS connection, size rights, Name of the universe, and description are called as the
universal Parameters.

36. Question36. What Is The Difference Between User And Versatile?

Answer :
SUPERVISOR offers various types of users of Business Objects products. The user
profile determines by default what products a user can use. User profiles can be
customized to reflect the needs of the users.

? User (all products but DESIGNER and SUPERVISOR)

? Versatile (configurable)

USERs use the Business Objects end-user products to query, report, and analyze data.
They may also use the optional REPORTER and EXPLORER modules for
multidimensional analysis. End users can produce documents containing data from one
or more data sources.

A versatile user is a customized user who may be given access by a supervisor to any
combination of Business Objects products.

37. Question37. What Are The User Requirements In This Universe?

Answer :

The user’s requirements in this universe are the Database connections, key column, join
and check for loop if you need measures, metrics. The steps to be followed should be
1. The connection should be defined.
2. Create the classes and objects
3. Give the joins and then resolve the loops.
4. Generate the universe.

38. Question38. What Is The Use Of Business Objects Data Services?

Answer :

Business Objects Data Services provides a graphical interface that allows you to easily
create jobs that extract data from heterogeneous sources, transform that data to meet
the business requirements of your organization, and load the data into a single location.

39. Question39. Define Data Services Components.

Answer :

Data Services includes the following standard components:

o Designer
o Repository
o Job Server
o Engines
o Access Server
o Adapters
o Real-time Services
o Address Server
o Cleansing Packages, Dictionaries, and Directories
o Management Console

40. Question40. What Are The Steps Included In Data Integration

Process?

Answer :

o Stage data in an operational datastore, data warehouse, or data mart.


o Update staged data in batch or real-time modes.
o Create a single environment for developing, testing, and deploying the entire data
integration platform.
o Manage a single metadata repository to capture the relationships between different
extraction and access methods and provide integrated lineage and impact analysis.

41. Question41. Define The Terms Job, Workflow, And Dataflow

Answer :

o A job is the smallest unit of work that you can schedule independently for execution.
o A work flow defines the decision-making process for executing data flows.
o Data flows extract, transform, and load data. Everything having to do with data,
including reading sources, transforming data, and loading targets, occurs inside a data
flow.

42. Question42. Arrange These Objects In Order By Their Hierarchy:

Dataflow, Job, Project, And Workflow.

Answer :

Project, Job, Workflow, Dataflow.

43. Question43. What Are Reusable Objects In Dataservices?


Answer :

Job, Workflow, Dataflow.

44. Question44. What Is A Transform?

Answer :

A transform enables you to control how datasets change in a dataflow.

45. Question45. What Is A Script?

Answer :

A script is a single-use object that is used to call functions and assign values in a
workflow.

46. Question46. What Is A Real Time Job?

Answer :

Real-time jobs "extract" data from the body of the real time message received and from
any secondary sources used in the job.

47. Question47. What Is An Embedded Dataflow?

Answer :

An Embedded Dataflow is a dataflow that is called from inside another dataflow.

48. Question48. What Is The Difference Between A Data Store And A

Database?

Answer :

A datastore is a connection to a database.


49. Question49. How Many Types Of Datastores Are Present In Data

Services?

Answer :

Three.

o Database Datastores: provide a simple way to import metadata directly from RDBMS.
o Application Datastores: let users easily import metadata from most Enterprise Resource
Planning (ERP) systems.
o Adapter Datastores: can provide access to an application’s data and metadata or just
metadata.

50. Question50. What Is The Use Of Compact Repository?

Answer :

Remove redundant and obsolete objects from the repository tables.

51. Question51. What Are Memory Datastores?

Answer :

Data Services also allows you to create a database datastore using Memory as the
Database type. Memory Datastores are designed to enhance processing performance of
data flows executing in real-time jobs.

52. Question52. What Are File Formats?

Answer :

A file format is a set of properties describing the structure of a flat file (ASCII). File
formats describe the metadata structure. File format objects can describe files in:

o Delimited format — Characters such as commas or tabs separate each field.


o Fixed width format — The column width is specified by the user.
o SAP ERP and R/3 format.

53. Question53. Which Is Not A Datastore Type?


Answer :

File Format

54. Question54. What Is Repository? List The Types Of Repositories.

Answer :

The Data Services repository is a set of tables that holds user-created and predefined
system objects, source and target metadata, and transformation rules. There are 3 types
of repositories.

o A local repository
o A central repository
o A profiler repository

55. Question55. What Is The Difference Between A Repository And A

Datastore?

Answer :

A Repository is a set of tables that hold system objects, source and target metadata, and
transformation rules. A Datastore is an actual connection to a database that holds data.

56. Question56. What Is The Difference Between A Parameter And A

Variable?

Answer :

A Parameter is an expression that passes a piece of information to a work flow, data flow
or custom function when it is called in a job. A Variable is a symbolic placeholder for
values.

57. Question57. When Would You Use A Global Variable Instead Of A

Local Variable?
Answer :

o When the variable will need to be used multiple times within a job.
o When you want to reduce the development time required for passing values between
job components.
o When you need to create a dependency between job level global variable name and job
components.

58. Question58. What Is Substitution Parameter?

Answer :

The Value that is constant in one environment, but may change when a job is migrated
to another environment.

59. Question59. List Some Reasons Why A Job Might Fail To

Execute?

Answer :

Incorrect syntax, Job Server not running, port numbers for Designer and Job Server not
matching.

60. Question60. List Factors You Consider When Determining

Whether To Run Work Flows Or Data Flows Serially Or In Parallel?

Answer :

Consider the following:

o Whether or not the flows are independent of each other


o Whether or not the server can handle the processing requirements of flows running at
the same time (in parallel)

61. Question61. What Does A Lookup Function Do? How Do The

Different Variations Of The Lookup Function Differ?


Answer :

All lookup functions return one row for each row in the source. They differ in how they
choose which of several matching rows to return.

62. Question62. List The Three Types Of Input Formats Accepted By

The Address Cleanse Transform.

Answer :

Discrete, multiline, and hybrid.

63. Question63. Name The Transform That You Would Use To

Combine Incoming Data Sets To Produce A Single Output Data Set

With The Same Schema As The Input Data Sets.

Answer :

The Merge transform.

64. Question64. What Are Adapters?

Answer :

Adapters are additional Java-based programs that can be installed on the job server to
provide connectivity to other systems such as Salesforce.com or the
JavaMessagingQueue. There is also a SoftwareDevelopment Kit (SDK) to allow
customers to create adapters for custom applications.

65. Question65. List The Data Integrator Transforms

Answer :

o Data_Transfer
o Date_Generation
o Effective_Date
o Hierarchy_Flattening
o History_Preserving
o Key_Generation
o Map_CDC_Operation
o Pivot Reverse Pivot
o Table_Comparison
o XML_Pipeline

66. Question66. List The Data Quality Transforms

Answer :

o Global_Address_Cleanse
o Data_Cleanse
o Match
o Associate
o Country_id
o USA_Regulatory_Address_Cleanse

67. Question67. What Are Cleansing Packages?

Answer :

These are packages that enhance the ability of Data Cleanse to accurately process
various forms of global data by including language-specific reference data and parsing
rules.

68. Question68. What Is Data Cleanse?

Answer :

The Data Cleanse transform identifies and isolates specific parts of mixed data, and
standardizes your data based on information stored in the parsing dictionary, business
rules defined in the rule file, and expressions defined in the pattern file.

69. Question69. What Is The Difference Between Dictionary And

Directory?

Answer :
Directories provide information on addresses from postal authorities. Dictionary files are
used to identify, parse, and standardize data such as names, titles, and firm data.

70. Question70. Give Some Examples Of How Data Can Be Enhanced

Through The Data Cleanse Transform, And Describe The Benefit Of

Those Enhancements.

Answer :

o Enhancement Benefit
o Determine gender distributions and target
o Gender Codes marketing campaigns
o Provide fields for improving matching
o Match Standards results

71. Question71. A Project Requires The Parsing Of Names Into Given

And Family, Validating Address Information, And Finding Duplicates

Across Several Systems. Name The Transforms Needed And The Task

They Will Perform.

Answer :

o Data Cleanse: Parse names into given and family.


o Address Cleanse: Validate address information.
o Match: Find duplicates.

72. Question72. Describe When To Use The Usa Regulatory And

Global Address Cleanse Transforms.

Answer :

Use the USA Regulatory transform if USPS certification and/or additional options such
as DPV and Geocode are required. Global Address Cleanse should be utilized when
processing multi-country data.
73. Question73. Give Two Examples Of How The Data Cleanse

Transform Can Enhance (append) Data.

Answer :

The Data Cleanse transform can generate name match standards and greetings. It can
also assign gender codes and prenames such as Mr. and Mrs.

74. Question74. What Are Name Match Standards And How Are They

Used?

Answer :

Name match standards illustrate the multiple ways a name can be represented. They are
used in the match process to greatly increase match results.

75. Question75. What Are The Different Strategies You Can Use To

Avoid Duplicate Rows Of Data When Re -loading A Job.

Answer :

o Using the auto-correct load option in the target table.


o Including the Table Comparison transform in the data flow.
o Designing the data flow to completely replace the target table during each execution.
o Including a preload SQL statement to execute before the table loads.

76. Question76. What Is The Use Of Auto Correct Load?

Answer :

It does not allow duplicated data entering into the target table.It works like Type 1 Insert
else Update the rows based on Non-matching and matching data respectively.

77. Question77. What Is The Use Of Array Fetch Size?


Answer :

Array fetch size indicates the number of rows retrieved in a single request to a source
database. The default value is 1000. Higher numbers reduce requests, lowering network
traffic, and possibly improve performance. The maximum value is 5000

78. Question78. What Are The Differences Between Row -by-row

Select And Cached Comparison Table And Sorted Input In Table

Comparison Transform?

Answer :

o Row-by-row select - look up the target table using SQL every time it receives an input
row. This option is best if the target table is large.
o Cached comparison table — To load the comparison table into memory. This option is
best when the table fits into memory and you are comparing the entire target table
o Sorted input — To read the comparison table in the order of the primary key column(s)
using sequential read.This option improves performance because Data Integrator reads
the comparison table only once.Add a query between the source and the
Table_Comparison transform. Then, from the query’s input schema, drag the primary
key columns into the Order By box of the query.

79. Question79. What Is The Use Of Using Number Of Loaders In

Target Table?

Answer :

Number of loaders loading with one loader is known as Single loader Loading. Loading
when the number of loaders is greater than one is known as Parallel Loading. The
default number of loaders is 1. The maximum number of loaders is 5.

80. Question80. What Is The Use Of Rows Per Commit?

Answer :

Specifies the transaction size in number of rows. If set to 1000, Data Integrator sends a
commit to the underlying database every 1000 rows.
81. Question81. What Is The Difference Between Lookup (),

Lookup_ext () And Lookup_seq ()?

Answer :

o lookup() : Briefly, It returns single value based on single condition


o lookup_ext(): It returns multiple values based on single/multiple condition(s)
o lookup_seq(): It returns multiple values based on sequence number

82. Question82. What Is The Use Of History Preserving Transform?

Answer :

The History Preserving transform allows you to produce a new row in your target rather
than updating an existing row. You can indicate in which columns the transform identifies
changes to be preserved. If the value of certain columns change, this transform creates
a new row for each row flagged as UPDATE in the input data set.

83. Question83. What Is The Use Of Map -operation Transform?

Answer :

The Map-Operation transform allows you to change operation codes on data sets to
produce the desired output. Operation codes: INSERT, UPDATE, DELETE, NORMAL or
DISCARD.

84. Question84. What Is Heirarchy Flatenning?

Answer :

Constructs a complete hierarchy from parent/child relationships, and then produces a


description of the hierarchy in vertically or horizontally flattened format.

o Parent Column, Child Column


o Parent Attributes, Child Attributes.

85. Question85. What Is The Use Of Case Transform?

Answer :
Use of the Case transform is to simplify branch logic in data flows by consolidating case
or decision-making logic into one transform. The transform allows you to split a data set
into smaller sets based on logical branches.

86. Question86. What Must You Define In Order To Audit A Data

Flow?

Answer :

You must define audit points and audit rules when you want to audit a data flow.

87. Question87. List Some Factors For Performance Tuning In Data

Services?

Answer :

The following sections describe ways you can adjust Data Integrator performance

o Source-based performance options


o Using array fetch size
o Caching data
o Join ordering
o Minimizing extracted data
o Target-based performance options
o Loading method and rows per commit
o Staging tables to speed up auto-correct loads
o Job design performance options
o Improving throughput
o Maximizing the number of pushed-down operations
o Minimizing data type conversion
o Minimizing local conversion
o Improving Informix repository performance

You might also like