51.
What
is I_T_FIELDS?
List
of the transfer structure fields. Only these fields are actually
filled in the data table and can be sensibly addressed in the
program.
52.
What
is C_T_DATA?
Table
with the data received from the API in the format of source structure
entered in table ROIS (field ROIS-STRUCTURE).
53.
What
is I_UPDMODE?
Transfer
mode as requested in the Scheduler of the BW. Not normally required.
54.
What
is I_T_SELECT?
Table
with the selection criteria stored in the Scheduler of the SAP BW.
This is not normally required.
55.
What
are the different Update Modes?
Direct
Delta:
In this method, extraction data from document postings is transferred
directly to BW delta queue.
Queued
Delta:
In this method, extraction data from document postings is collected
in an extraction queue, from which a periodic collective run is used
to transfer the data to BW delta queue.
o
The
transfer sequence and the order in which the data was created are the
same in both Direct and Queued Delta.
Unserialized
V3 Update:
In
this method, the extraction data is written to the update tables and
then is transferred to the BW delta queues without taking the
sequence into account.
56.
What
are the different ways Data Transfer?
Full
Update:
All the data from the InfoStructure is transferred according to the
selection criteria defined in the scheduler in the SAP BW.
Delta
Update:
Only the data that has been changed or is new since the last update
is transferred.
57.
Which
Object connects Aggregates and InfoCube?
ReadPointer
connects Aggregates and InfoCube. We can view the ReadPointer in
table RSDDAGGRDIR, the field name is RN_SID, whenever we are rolling
up the data, it contains the request number, it will check with the
next request for second roll up. Just follow the table for a
particular InfoCube and roll up the data.
58.
What
is switching ON and OFF of aggregates? How do we do that?
When
we switch off an aggregate, it is not available to supply data to
queries, but the data remains in the aggregate, so if required, we
can turn it on and update the data, instead of re-aggregating all the
data. However if we deactivate an aggregate, it is not available for
reporting and also we lose the aggregated data. So when you activate
it, it starts the aggregation anew. To do this select the relevant
aggregate and choose the Switch On/Off (red and green button). An
aggregate that is switched off is marked in column Filled/Switched
off with Grey Button.
59.
While
creating aggregates system gives manual or automatic option. What are
these?
If
we select the automatic option, system will propose aggregates based
on the BW statistics. i.e., how many times the InfoCube is used to
fetch data, etc. Else we can manually select the dataset which should
form the aggregate.
60.
What
are the options when defining aggregates?
Manual
Automatic
61.
What
are Aggregates and when are they used?
An
aggregate is a materialized, aggregated view of the data in an
InfoCube. In an aggregate, the dataset of an InfoCube is saved
redundantly and persistently in a consolidated form. Aggregates make
it possible to access InfoCube data quickly in Reporting. Aggregates
can be used in following cases:
o
To
speed up the execution and navigation of a specific query.
o
Use
attributes often in queries.
o
To
speed up reporting with characteristic hierarchies by aggregating
specific hierarchy levels.
62.
When
I run Initial load it failed then what should I do?
Deletion
of an initial load can be done in the InfoPackage. First set the QM
status of the request to red if not yet done, then delete it from all
data targets. After that we go to the InfoPackage and choose from
menu scheduler
“initialization options for the source system”. There you should
see your red request. Mark it and delete it. Accept deletion question
and accept post information message. Now the request should be
deleted from the initialization options. Now you can run a new
init.
You can also run a repair request. That's a full request. With this you correct your data in the data target because of failed deltas or wrong inits. You do this in the InfoPackage too. Choose menu scheduler repair full request. But if you want to use the init/delta load you have to make a successful init first.
You can also run a repair request. That's a full request. With this you correct your data in the data target because of failed deltas or wrong inits. You do this in the InfoPackage too. Choose menu scheduler repair full request. But if you want to use the init/delta load you have to make a successful init first.
63.
What
are the inverted fields in DataSource?
They
allow to do reverse posting. It would actually multiply the field by
-1.
64.
What
are setup tables and why should we delete the setup tables first
before extraction?
a.
Setup
tables are filled with data from application tables. They are the
OLTP tables storing transaction record. They interface between the
application tables and extractor. LO extractor takes data from Setup
table while initialization and full upload. It is not needed to
access the application table for data selection. As setup tables are
required only for full and init load we can delete the data after
loading in order to avoid duplicate data.
65.
What
are the setup tables? Why use setup tables?
In
LO Extraction Mechanism when we fill the setup tables the extract
structure is filled with the data. When we schedule InfoPackage using
FULL / INITI DELTA from BW, the data is picked from the setup tables
66.
When
filling the set tables, is there any need of delete the setup tables?
a.
Yes. By
deleting the setup tables we are deleting the data that is in the
setup tables from the previous update. This avoids updating the
records twice into the BW.
67.
Why
we need to delete the setup table first then filling?
During
the Setup run, these setup tables are filled. Normally it's a good
practice to delete the existing setup tables before executing the
setup runs so as to avoid duplicate records for the same selections.
68.
With
what data the setup table is filling (Is it R3 data)?
The
init loads in BW pull data from the Setup tables. The setup tables
are only used in case of first init/full loads.
69.
Will
there be any data in the application tables after sending data to
Setup tables?
There
will be data in application tables even after fill up of setup
tables. Setup tables are just temp tables that fill up from
application tables for setting up Init/Full loads for BW.
70.
How
to work master data delta?
We
always do full load for Master Data. It would always overwrite the
previous entries.
71.
Master
data is stored in Master Data tables. Then what is the importance of
dimensions?
Dimension
tables link Master Data tables with the fact table through SID's.
72.
I
replicated the DataSource to BW system. I want to add one more field
to DataSource. How do I do it?
Add
the field to extract structure and replicate the DataSource again
into BW and this field will appear in BW also.
73.
Suppose
one million records are uploaded to InfoCube. Now I want to delete 20
records in InfoCube. How can we delete 20 records?
This
you could do with selective deletion
74.
What
is the InfoCube for inventory?
InfoCube:
0IC_C03
75.
What
is the maintenance of DataSource?
It
is the maintenance of required fields in a particular DataSource for
which there are reporting requirements in BW and data for the same
needs to be extracted.
76.
What
is the maintenance of Extract structure
Extract
structures are maintained in case of LO DataSources. There are
multiple extract structures for each DataSource in the LO for
different applications. Any enhancements to DataSource in case of LO
are done using maintenance of extract structures.
77.
What
are MC EKKO, MC EKPO in the maintenance of DataSource?
These
are purchasing related communication structures.
78.
How
is the Delta Load different for an InfoCube and ODS?
An
InfoCube will have additive Delta, but you will still be able to see
all individual records in the InfoCube contents. This is because if
you choose to delete the current request - then the records have to
be rolled back to the prior status. You build a query on the InfoCube
and on the query you will find that the data is actually summed up.
The ODS records will not have duplicate records. You will have only
one record.
79.
What
is the difference between the transactions LBWF and RSA7?
RSA7
is to view BW delta queue. This gets overwritten each time.
LBWF
is the Log for LO Extract Structures. This is populated only when the
User parameter MCL is set, and is recommended only for testing
purposes.
80.
What
exactly happens (background) when we are inactivating/activating the
extract structure for LO Cockpit?
If
the extract structure is activated then any online transaction or on
the compilation of setup tables, the data is posted to the extract
structures depending on the update method selected. Activation marks
the DataSource with green else it is yellow. The
activation/deactivation makes entries to the TMCEXACT table.
81.
What
is content extraction?
These
are extractors supplied by SAP for specific business modules. Eg.
2FI_AR_4: Customers: Line Items with Delta Extraction / 2FI_GL_6:
General Ledger Sales Figures via Delta Extraction.
82.
What
is direct Update of InfoObject?
This
is updating of InfoObject without using Update Rules but only the
Transfer Rules.
83.
You
get New status or Additive Delta. If I set here (on R/3) what is the
need of setting in BW.
In
R/3 the record mode determines this as seen in the RODELTAM table
i.e., whether it will be a new status or additive delta for the
respective DataSource. Based on this you need to select the
appropriate update type for the data target in BW. For e.g., ODS
supports additive as well as Overwrite function. Depending on which
DataSource is updating the ODS, and the record mode supported by this
DataSource, you need to do the right selection in BW.
84.
Where
does BW extract data from during Generic Extraction and LO
Extraction?
All
deltas are taken from the delta queue. The way of populating the
delta queue differs for LO and other DataSources.
85.
What
is the importance of ODS Object?
ODS
is mainly used as a staging area.
86.
Differences
between star and extended star schema?
Star
schema:
Only characteristics of the dimension tables can be used to access
facts. No structured drill downs can be created. Support for many
languages is difficult.
Extended
star schema:
Master data tables and their associated fields (attributes), External
hierarchy tables for structured access to data, Text tables with
extensive multilingual descriptions are supported using SIDs.
87.
What
are the major errors in BW and R3 pertaining to BW?
Errors
in loading data (ODS loading, InfoCube loading, delta loading etc)
Errors
in activating BW or other objects.
88.
When
are tables created in BW?
When
the objects are activated, the tables are created. The location
depends on the Basis installation.
89.
What
is M table?
Master
Data table.
90.
What
is F table?
Fact
table
91.
What
is data warehousing?
Data
Warehousing is a concept in which the data is stored and analysis is
performed over it.
92.
What
is a RemoteCube and how is it accessed and used?
A
RemoteCube is an InfoCube whose data is not managed in the BW but
externally. Only the structure of the RemoteCube is defined in BW.
The data is read for reporting using a BAPI from another system.
With
a RemoteCube, we can report using data in external systems without
having to physically store transaction data in BW. We can, for
example, include an external system from market data providers using
a RemoteCube.
This
is best used only for small volume of data and when less users access
the query.
93.
Tell
about a situation when you implemented a RemoteCube.
RemoteCube
is used when we like to report on transactional data. In a RemoteCube
data is not stored on BW side. Ideally used when detailed data is
required and we want to bypass loading of data into BW.
94.
Differences
between MultiCube and RemoteCube.
A
MultiCube is a type of InfoProvider that combines data from a number
of InfoCubes and makes them available as a whole to reporting.
A
RemoteCube is an InfoCube whose transaction data is not managed in
the BW but externally. Only the structure of the RemoteCube is
defined in BW. The data is read for reporting using a BAPI from
another system.
95.
How
you did Data modeling in your project? Explain
We
had collected data from the user and created HLD (High level Design
document) and we analyzed to find the source for the data. Then data
models were done indicating dataflow, lookups. While designing the
data model considerations were given to use existing objects (like
ODS and InfoCube) not storing redundant data, volume of data, Batch
dependency.
96.
There
is an InfoObject called 0PLANT I activated and using it after some
days one more person came and activated it again. What will happen,
whether there will be any effect merge or no effect.
Reactivating
the InfoObject shouldn't affect unless he has made some changes to
that and then reactivated it.
97.
I
have two processes one process contains ABAP program. After the
successful completion of the first process it should trigger second
one how to know whether the first is successful or not?
For
process chains go through this
linkhttp://help.sap.com/saphelp_nw2004s/helpdata/en/8f/c08b3baaa59649e10000000a11402f/frameset.htm
98.
I
want to create an InfoObject that is a dependent InfoObject. How to
do it?
Go
to the first InfoObject screen in administration work bench go to
compounding tab, create the InfoObject that is dependent on the
former InfoObject and activate.
99.
Delta
has been done successfully in LO. Later some fields were added to
that particular DataSources then there will be any effect to the
previous data records.
No.
If there is data in the DataSource we can only append the fields. No
data will be lost. But you need to have separate mechanism to fill in
the historical data for the newly added fields.
100.
There
are 5 characteristics in an InfoCube. We have to assign these
characteristics to a dimension based on what we assign
characteristics to dimension?
Depends
on the characteristic and cardinality.
The
characteristics that logically belong together can be grouped
together in a Dimension.
First
we will decide the dimensions of the InfoCube. After that we will
assign necessary InfoObjects to the corresponding dimensions