Saturday, July 9, 2011

Some SAP-BW T-CODEs

RSA1 Administrator Workbench
RSA2 Data sources Extractor
RSA3 Data sources Extractor Checker
RSA5 Installation of Datasource from Business Content
RSA6 Data Sources & Hierarchy Maintenence
RSA7 Delta Queue in BW side Using Data Marts
RSA8 Post process Data Sources & Hierarchy Maintenence
RRMX Bex Analyzer
PFCG Roles Creation and Maintenance
RSSM Authorization of Business Information Warehouse
RSRT Query Monitor
RSMO Monitor Administrator Workbench
RSMON Administration View of Administrator Workbench
RSDMWB Data Mining Workbench
RSANWB Analysis Process Designer
RSDMPROM Edit Multiprovider
RSISET Edit Infoset
RSDCUBE Edit Infocube
RSDODS Edit DSO
RSH1 Hierarchy Maintenance
RSPLAN BI Integrated Planning
RSRV Analyse & Repair BI Objects
FILE Definition of Log.Data file / Change/ Overview
AL11 SAP Directories " View & Configure"
WE05 IDOC List
SM51 Server - Process Overview
SM59 Cofiguration of RFC-Connections
SM37 Job Selection & Job Overview
SE37 Function Module
SE11 ABAP Dictionary
SE16 DATA Browser (Tables)
SM01 Locking / Unlocking Transaction Code
SMOBILEBW Mange & Monitor BW Requests for Mobile Clients
WE05 IDOC List
RSPC Process Chain Create and Maintain
RSPCM Monitor all Process Chains
SE38 ABAP Editor
SE80 Object Navigator: ABAP Development Workbench
SMOEAC Object Navigator CRM: Administration Console
RSPC Monitor the Process Chains
SU53 Authorization details of the USER itself
RS12 Select Lock Entries (Can unlock queries )
SM12 Select Lock Entries (Can unlock Objects )
RSECADMIN Authorization administration
SCOT SAP Connect administration
SOST SMTP Queue Status Messages
RSKC Maintenence of the permitted Extra Characcters in BW
RSO2 Generic Datasource
RSODADMIN Metadata Administration
CTW_META Generating Metadata
RSCUSTV1 Settings for Flat file Uploads
RSORBCT Business Content
SUIM User Informationsystem
SM02 System Information
SMICM ICM Monitor
LISTCUBE Call List Viewer for Data Targets
BPS0 Business Planning and Simulation
OS07 CPU Monitor

Routines

There are 5 types of routines:

1. At Info Package :- To give the dynamic file name to info package (we need to assign the file name to P_Filename variable).

2. START ROUTINE :- Available at Transformations. It will trigger before transformations. generally used to filtering the records or Fetching global data to be used in characteristic routine.

SOURCE_PACKAGE will having the total content information and structure will be same as the source transformation.

3. END ROUTINE :- Available at Transformation and trigger after transformations. Generally end user is used for updating data based on existing data. RESULT_PACKAGE will have process information and structure will be same as the target.

4. EXPERT ROUTINE :- In Transfer rules Edit -> Expert routine will be available. SOURCE_PACKAGE and RESULT_PACKAGE will be available here. Transformations will be deleted once click on Expert Routine and Start and End Routine will be deactivated once you click on Expert Routine. Generally we used this to split single record into multiple records.

5. CHARACTERISTIC ROUTINE :- Inside the transformations on Char/Key Figure. It will trigger after start routine and before end routine. All global objects declared in Start routine will be available here.

Question:  I have field routine, start routine and end routine in my transformation for an object. which one executes first?

Ans:  Start Routine executes first.
          Field Routine executes second.
          End Routine will be executed after these two.

Monday, July 4, 2011

SAP BI Interview Questions- Part 1

1. What is the t-code to see log of transport connection?in RSA1 -> Transport Connection you can collect the Queries and the Role and after this you can transport them (enabling the transport in SE10, import it in STMS
1. RSA1
2. Transport connection (button on the left bar menu)
3. Sap transport -> Object Types (button on the left bar menu)
4. Find Query Elements -> Query
5. Find your query
6. Group necessery object
7. Transport Object (car icon)
8. Release transport (SE10 T-code)
9. load transport (STMS T-code)

2.Lo; mm inventory data source with marker significance?
Marker is as like check point when u upload the data from inventory data source
2lis_03_bx data source for current stock and BF for movement type
after uploading data from BX u should rlise the request in cube or i menn to say compress it then load data from another data source BF and set this updated data to no marker update so marker is use as a check point if u dont do this u getting data missmatch at bex level bcz system get confuse .
(2LIS_03_BF Goods Movement From Inventory Management-- -----Unckeck the no marker update tab)
(2LIS_03_BX Stock Initialization for Inventory Management-- ---select the no marker update check box)
2LIS_03_UM Revaluations ----Uncheck the no marker update tab) in the infopackege of "collaps"

3. How can you navigate to see the error idocs ?
If it is fine check the IDOCs in source system go to BD87->give Ur user ID and date->execute->you can find Red status Idocs select the erroneous Idoc->Rt.click and select Manual process.



You need to Reprocess this IDOC which are RED. For this you can take help of Any of your Team (ALE IDOC Team or BAsis Team)Or Else
youcan push it manually. Just search it in bd87 screen only to Reprocess.
Also, Try to find why this IDocs are stuck there.

4)Difference between v1, v2, v3 jobs in extraction?

V1 Update: when ever we create a transaction in R/3(e.g.,Sales Order) then the entries get into the R/3 Tables(VBAK, VBAP..) and this takes place in V1 Update.
V2 Update: V2 Update starts a few seconds after V1 Update and in this update the values get into Statistical Tables, from where we do the extraction into BW.
V3 Update: Its purely for BW extraction.
But in the Document below, V1, V2 and V3 are defined in a different way. Can You please explain me in detial what exactly V1, V2 and V3 updates means?

5.What are statistical update and document update?
Synchronous Updating (V1 Update)
The statistics update is made synchronously with the document update.
While updating, if problems that result in the termination of the
statistics update occur, the original documents are NOT saved. The cause
of the termination should be investigated and the problem solved.
Subsequently, the documents can be entered again.
Radio button: V2 updating

6.Do you have any idea how to improve the performance of the BW..?
Asynchronous Updating (V2 Update)
With this update type, the document update is made separately from the statistics update. A termination of the statistics update has NO influence on the document update (see V1 Update).
Radio button: Updating in U3 update program
Asynchronous Updating (V3 Update)
With this update type, updating is made separately from the document update. The difference between this update type and the V2 Update lies,however, with the time schedule. If the V3 update is active, then the update can be executed at a later time.

In contrast to V1 and V2 Updates, no single documents are updated. The V3 update is, therefore, also described as a collective update.

7)How can you decide the query performance is slow or fast ?
You can check that in RSRT tcode.
execute the query in RSRT and after that follow the below steps
Goto SE16 and in the resulting screen give table name as RSDDSTAT for BW 3.x and RSDDSTAT_DM for BI 7.0 and press enteryou can view all the details about the query like time taken to execute the query and the timestmaps

8)What is statistical setup and what is the need and why?
Follow these steps to filling the set up table.

1. Go to transaction code RSA3 and see if any data is available related to your DataSource. If data is there in RSA3 then go to transaction code LBWG (Delete Setup data) and delete the data by entering the application name.
2. Go to transaction SBIW --> Settings for Application Specific Datasource --> Logistics --> Managing extract structures --> Initialization --> Filling the Setup table --> Application specific setup of statistical data --> perform setup (relevant application)
3. In OLI*** (for example OLI7BW for Statistical setup for old documents : Orders) give the name of the run and execute. Now all the available records from R/3 will be loaded to setup tables.
4. Go to transaction RSA3 and check the data.
5. Go to transaction LBWE and make sure the update mode for the corresponding DataSource is serialized V3 update.
6. Go to BW system and create infopackage and under the update tab select the initialize delta process. And schedule the package. Now all the data available in the setup tables are now loaded into the data target.
7. Now for the delta records go to LBWE in R/3 and change the update mode for the corresponding DataSource to Direct/Queue delta. By doing this record will bypass SM13 and directly go to RSA7. Go to transaction code RSA7 there you can see green light # Once the new records are added immediately you can see the record in RSA7.
8. Go to BW system and create a new infopackage for delta loads. Double click on new infopackage. Under update tab you can see the delta update radio button..
9. Now you can go to your data target and see the delta record.

9.Why we have construct setup tables?
The R/3 database structure for accounting is much more easier than the Logistical structure.
Once you post in a ledger that is done. You can correct, but that give just another posting.
BI can get information direct out of this (relatively) simple database structure.
In LO, you can have an order with multiple deliveries to more than one delivery addresses. And the payer can also be different.
When 1 item (orderline) changes, this can have its reflection on order, supply, delivery, invoice, etc.
Therefore a special record structure is build for Logistical reports.and this structure now is used for BI.
In order to have this special structre filled with your starting position, you must run a set-up. from that moment on R/3 will keep filling this LO-database.
If you wouldn't run the setup. BI would start with data from the moment you start the filling of LO (with the logistica cocpit)

10.How can you eliminate the duplicate records in TD, MD?Try to check the system logs through SM21 for the same.


11.What use marker in MM?
Marker update is just like check point.
ie it will give the snapshot of the stock on a particular date ie when was the marker updated.
Because we are using Noncumulative keyfigure it will lot of time to calculate the current stock for example at report time. to overcome this we use marker update
Marker updates do not summarize the data.. In inventory management scenarios, we have to calculate opening stock and closing stock on a daily basis. In order to facilitate this, we set a marker which will add and subtract the values for each record.
In the absence of marker update, the data will be added up and will not provide the correct values.

12.Tell me web template?
You get information on where the web template details are stored from the following tables :
RSZWOBJ Storage of the Web Objects
RSZWOBJTXT Texts for Templates/Items/ Views
RSZWOBJXREF Structure of the BW Objects in a Template

RSZWTEMPLATE Header Table for BW HTML Templates
You can check these tables and search for your web template entry . However, If I understand your question correctly , you will have to open the template in the WAD and then make the corrections in the same to correct it.

13.What is dashboard?
A dash board can be created using the web application Designer (WAD) or the visual composer (VC). A dashboard is just a collection of reports, views and links etc in a single view. For e.g. igoogle is a dashboard.

A dashboard is a graphical reporting interface, which displays KPIs (Key Performance Indicators) as charts and graphs. A dashboard is a performance management system

When we look at the all organization measures how they are performing with helicopter view, we need a report that teaches and shows the trend in a graphical display quickly. These reports are called as Dashboard Reports, still we can report these measures individually, but by keeping all measures in a single page, we are creating single access point to the users to view all information available to them. Absolutely this will save lot of precious time, gives clarity on decision that needs to be taken, helps the users to understand the measure(s) trend with business flow creating dashboard
Dashboards : Could be built with Visual Composer & WAD
create your dashboard in BW,

(1) Create all BEx Queries with required variants,tune them perfectly.
(2) Differentiate table queries and graph queries.
(3) Choose the graph type required that meet your requirement.
(4) Draw the layout how the Dashboard page looks like.
(5) Create a web template that has navigational block / selection information.
(6) Keep navigational block fields are common across the measures.
(7) Include the relevant web items into web template.
(8) Deploy the URL/Iview to users through portal/intranet

The steps to be followed in the creation of Dashboard using WAD are summarized as below:

1) Open a New Web template in WAD.
2) Define the tabular layout as per the requirements so as to embed the necessary web items.
3) Place the appropriate web items in the appropriate tabular grids
4) Assign queries to the web items (A Query assigned to a web item is called as a data provider)
5) Care should be taken to ensure that the navigation block’s selection parameters are common across all the BEx queries of the affected dataproviders.
6) Properties of the individual web items are to be set as per the requirements. They can be modified in Properties window or in the HTML code.
7) The URL when this web template is executed should be used in the portal/intranet




14.How can you solve the data mismatch tickets between r/3 and bw?
Check the mapping at BW side for 0STREET in transfer rules.Check the data in PSA for the same field.If the PSA is also doesn't have complete data then check the field in RSA3 in source system.

15. What is thumb rule?


16)What is replacement path tell me one scenario?
http://www.sd- solutions. com/documents/ SDS_BW_Replaceme nt%20Path% 20Variables. html

17.What is difference between PSA & IDOC?

BI7 is PSA used only for Data load from Source System into BW


18). what we do in Business Blue Print Stage?
SAP has defined a business blueprint phase to help extract pertinent information about your company that is necessary for implementation. These blueprints are in the form of questionnaires that are designed to probe for information that uncovers how your company does business. As such, they also serve to document the implementation. Each business blueprint document essentially outlines your future business processes and business requirements. The kinds of questions asked are germane to the particular business function, as seen inthe following sample questions:1) What information do you capture on a purchase order?2) What information is required to complete a purchase order?Accelerated SAP question and answer database:The question and answer database (QADB) is a simple although aging tool designed to facilitate the creation and maintenance of your business blueprint.This database stores the questions and the answers and serves as the heart of your blue print. Customers are provided with a customer input template for each application that collects the data. The question and answer format is standard across applications to facilitate easier use by the project team.Issues database: Another tool used in the blueprinting phase is the issues database. Thisdatabase stores any open concerns and pending issues that relate to the implementation. Centrally storing this information assists in gathering and then managing issues to resolution, so that important matters do not fall through the cracks. You can then track the issues in database, assign them to teammembers, and update the database accordingly.

19). How do we gather the requirements for an Implementation Project?One of the biggest and most important challenges in any implementation is gathering and understanding the end user and process team functional requirements. These functional requirements represent the scope of analysis needs and expectations (both now and in the future) of the end user. These typically involve all of the following:- Business reasons for the project and business questions answered by the implementation- Critical success factors for the implementation- Source systems that are involved and the scope of information needed from each- Intended audience and stakeholders and their analysis needs- Any major transformation that is needed in order to provide the information- Security requirements to prevent unauthorized useThis process involves one seemingly simple task: Find out exactly what theend users' analysis requirements are, both now and in the future, and buildthe BW system to these requirements. Although simple in concept, in practicegathering and reaching a clear understanding and agreement on a complete setof BW functional requirements is not always so simple.

20) How do we decide what cubes has to be created?
Its depends on your project requirement. Customized cubes are not mandatory for all the projects. If your bussines requirement is differs from given scenario ( BI content cubes ) then only we will opt for customized cubes.Normally your BW customization or creation of new info providers all are depending on your source system.If your source system other that R3 then you should go with customization of your all objects.If your source system is R3 and your users are using only R3 standard business scenarios like SD,MM or FI... etc., then you dont want to create any info providers or you dont want to enhance any thing in the existing BW Business Content. But 99% this is not possible. Because surely they should have included their new business scenario or new enhancements.For example, In my first project we implemented for Solution Manager BW implemention. There we have activated all the business content in CRM. But the source system have new scenarios for message escalation, ageing calculation etc., According their business scenrio we could't use standard business content. For that we have taken only existing info objects and created new info objects which are not there in the business content. After that we have created custom data source to info providers as well asreports.

21) Who used to make the Technical and Functional Specifications?
Technical Specification:Here we will mention all the BW objects (info objects, data sources, info sources and info providers). Then we are going to say the data flow and behaviour of the data load (either delta or full) also we can tell the duration of the cube activation or creation. Pure BW technical things are available in this document. This is not for End users document.Functional Specification:Here we will describe the business requirements. That means here we are going to say which are all business we are implementing like SD, MM and FI etc., then we are going to tell the KPI and deliverable reports detail to the users. This document is going to mingle with both Function Consultants and Business Users. This document is applicable for end users also.
22) Give me one example of a Functional Specification and explain what information we will get from that?Functional Specs are requirements of the business user.Technical Specs translate these requirements in a technical fashion.Let's say Functional Spec says,1. the user should be able to enter the Key date, Fiscal Year, Fiscal Version.2. The Company variable should be defaulted to USA but then if the user wants to change it, they can check the drop down list and choose other countries.3. The calculations or formulas for the report will be displayed in precision of one decimal point.4. The report should return values for 12 months of data depending on the fiscal year that the user enters Or it should display in quarterly values. Functional specs are also called as Software requirements.Now from this Techinal Spec follows, to resolve each of the line items listed above.1. To give the option of key date, Fiscal year and Fiscal Version – certain Info Obejcts should be availble in the system. If available, then should we create any variables for them - so that they are used as user entry variable. To create any varaibles, what is the approch, where do you do it, what is the technical of the objects you'll use, what'll be the technical name of the objects you'll crete as a result of this report.2. Same explanation goes for the rest. How do you set up the varaible,
3. What changes in properties willu do to get the precision.4. How will you get the 12 months of data.What will be the technical and display name of the report, who'll be authorized to run this report, etc are clearly specified in the technical specs.


23) What is Customization? How do we do in LO?

How to do basic LO extraction for SAP-R3-BW1. Go to transaction code RSA3 and see if any data is available related to your DataSource. If data is there in RSA3 then go to transaction code LBWG (Delete Setup data) and delete the data by entering the application name.2. Go to transaction SBIW --> Settings for Application Specific Datasource --> Logistics --> Managing extract structures --> Initialization --> Filling the Setup table --> Application specific setup of statistical data --> perform setup (relevant application)3. In OLI*** (for example OLI7BW for Statistical setup for old documents : Orders) give the name of the run and execute. Now all the available records from R/3 will be loaded to setup tables.4. Go to transaction RSA3 and check the data.5. Go to transaction LBWE and make sure the update mode for the corresponding DataSource is serialized V3 update.6. Go to BW system and create infopackage and under the update tab select the initialize delta process. And schedule the package. Now all the data available in the setup tables are now loaded into the data target.7.Now for the delta records go to LBWE in R/3 and change the update mode for the corresponding DataSource to Direct/Queue delta. By doing this record will bypass SM13 and directly go to RSA7. Go to transaction code RSA7 there you can see green light # Once the new records are added immediately you can see the record in RSA7.


24) When we use Maintain Data Source, What we do? What we will maintain?
Go to BW system and create a new infopackage for delta loads. Double click on new infopackage. Under update tab you can see the delta update radio button.

25) Tickets and Authorization in SAP Business Warehouse What is tickets? And example?Tickets are the tracking tool by which the user will track the work which we do. It can be a change requests or data loads or what ever. They will of types critical or moderate. Critical can be (Need to solve in 1 day or half a day) depends on the client. After solving the ticket will be closed by informing the client that the issue is solved. Tickets are raised at the time of support project these may be any issues, problems.... .etc. If the support person faces any issues then he will ask/request to operator to raise a ticket.
Operator will raise a ticket and assign it to the respective person. Critical means it is most complicated issues ....depends how you measure this...hope it helps. The concept of Ticket varies from contract to contract in between companies. Generally Ticket raised by the client can be considered based on the priority. Like High Priority, Low priority and so on. If a ticket is of high priority it has to be resolved ASAP. If the ticket is of low> priority it must be considered only after attending to high priority tickets. The typical tickets in a production Support work could be: 1. Loading any of the missing master data attributes/texts. 2. Create ADHOC hierarchies. 3. Validating the data in Cubes/ODS. 4. If any of the loads runs into errors then resolve it. 5. Add/remove fields in any of the master data/ODS/Cube. 6. Data source Enhancement. 7. Create ADHOC reports.
1. Loading any of the missing master data attributes/texts - This would be done by scheduling the infopackages for the attributes/texts mentioned by the client. 2. Create ADHOC hierarchies. - Create hierarchies in RSA1 for the info-object. 3. Validating the data in Cubes/ODS. - By using the Validation reports or by comparing BW data with R/3. 4. If any of the loads runs into errors then resolve it. - Analyze the error and take suitable action. 5. Add/remove fields in any of the master data/ODS/Cube. - Depends upon the requirement 6. Data source Enhancement. 7. Create ADHOC reports. - Create some new reports based on the requirement of client.

26) Change attribute run.
Generally attribute change run is used when there is any change in the master data..it is used for realingment of the master data..Attribute change run is nothing but adjusting the master data after its been loaded from time to time so that it can change or generate or adjust the sid's so that u may not have any problem when loading the trasaction data in to data targets.the detail explanation about Attribute change run.The hierarchy/attribute change run which activates hierarchy and attribute changes and adjusts the corresponding aggregates is devided, into 4 phases:1. Finding all affected aggregates2. set up all affected aggregates again and write the result in the new aggregate table.3. Activating attributes and hierarchies4. rename the new aggregate table. When renaming, it is not possible to execute queries. In some databases, which cannot rename the indexes, the indexes are also created in this phase.
27) Different types of Delta updates?
Delta loads will bring any new or changed records after the last upload.This method is used for better loading in less time. Most of the std SAP data sources come as delta enabled, but some are not. In this case you can do a full load to the ODS and then do a delta from the ODS to the cube. If you create generic datasources, then you have the option of creating a delta onCalday, timestamp or numeric pointer fields (this can be doc number, etc).You'll be able to see the delta changes coming in the delta queue through RSA7 on the R3 side.To do a delta, you first have to initialize the delta on the BW side and then set up the delta.The delta mechanism is the same for both Master data and Transaction data loads.============ ========= ==There are three deltasDirect Delta: With this update mode, the extraction data is transferred with each document posting directly into the BW delta queue. In doing so, each document posting with delta extraction is posted for exactly one LUW in the respective BW delta queues.Queued Delta: With this update mode, the extraction data is collected for the affected application instead of being collected in an extraction queue, and can be transferred as usual with the V3 update by means of an updating collective run into the BW delta queue. In doing so, up to 10000 deltaextractions of documents for an LUW are compressed for each Data Source into the BW delta queue, depending on the application.Non-serialized V3 Update: With this update mode, the extraction data for the application considered is written as before into the update tables with the help of a V3 update module. They are kept there as long as the data is selected through an updating collective run and are processed. However, in contrast to the current default settings (serialized V3 update), the data in the updating collective run are thereby read without regard to sequence from the update tables and are transferred to the BW delta queue.
28) Production supportIn production support there will be two kind jobs which you will be doing mostly 1, looking into the data load errors. 2, solving the tickets raised by the user. Data loading involves monitoring process chains, solving the errors related to data load, other than this you will also be doing some enhancements to the present cubes and master data but that done on requirement. User will raise a ticket when they face any problem with the query, like report showing wrong values incorrect data etc.if the system response is slow or if the query run time is high. Normally the production support activities include * Scheduling * R/3 Job Monitoring * B/W Job Monitoring * Taking corrective action for failed data loads. * Working on some tickets with small changes in reports or in AWB objects. The activities in a typical Production Support would be as follows: 1.Data Loading - could be using process chains or manual loads. 2. Resolving urgent user issues - helpline activities 3. Modifying BW reports as per the need of the user. 4. Creating aggregates in Prod system 5. Regression testing when version/patch upgrade is done. 6. Creating adhoc hierarchies. We can perform the daily activities in Production 1. monitoring Dataload failures thru RSMO 2. Monitoring Process Chains Daily/weekly/ monthly 3. Perform Change run Hirerachy 4. Check Aggr's Rollup.

29) How to convert a BeX query Global structure to local structure (Steps involved)
BeX query Global structure to local structureSteps; ***a local structure when you want to add structure elements that are unique to the specific query. Changing the global structure changes the structure for all the queries that use the global structure. That is reason you go for a local structure.Coming to the navigation part--In the BEx Analyzer, from the SAP Business Explorer toolbar, choose the open query icon (icon tht looks like a folder) On the SAP BEx Open dialog box:Choose Queries.Select the desired InfoCubeChoose New.On the Define the query screen:In the left frame, expand the Structure node.Drag and drop the desired structure into either the Rows or Columnsframe.Select the global structure.Right-click and choose Remove reference.A local structure is created.Remember that you cannot revert back the changes made to global structure inthis regard. You will have to delete the local structure and then drag ndrop global structure into query definition.*When you try to save a global structure, a dialogue box prompts you tocomfirm changes to all queries. that is how you identify a global structure*

30) What is the use of Define cell in BeX & where it is useful?
Cell in BEX:::Use*When you define selection criteria and formulas for structural components and there are two structural components of a query, generic cell definitions are created at the intersection of the structural components that determine the values to be presented in the cell.Cell-specific definitions allow you to define explicit formulas, along with implicit cell definition, and selection conditions for cells and in this way, to override implicitly created cell values. This function allows you to design much more detailed queries.In addition, you can define cells that have no direct relationship to the structural components. These cells are not displayed and serve as containers for help selections or help formulas.you need two structures to enable cell editor in bex. In every query you have one structure for key figures, then you have to do another structure with selections or formulas inside.Then having two structures, the cross among them results in a fix reporting area of n rows * m columns. The cross of any row with any column can be defined as formula in cell editor.This is useful when you want to any cell had a diferent behaviour that the general one described in your query defininion.For example imagine you have the following where % is a formula kfB/KfA *100.kfA kfB %chA 6 4 66%chB 10 2 20%chC 8 4 50%Then you want that % for row chC was the sum of % for chA and % chB. Then in cell editor you are enable to write a formula specifically for that cell as sum of the two cell before. chC/% = chA/% + chB/% then:kfA kfB %chA 6 4 66%chB 10 2 20%chC 8 4 86%
31) What is 0Recordmode?A. it is an info object , 0Record mode is used to identify the delta images in BW which is used in DSO .it is automatically activated when u activate DSO in BW. Like that in R/3 also have field 0cancel. It holds delta images in R/3. When ever u extracting data from R/3 using LO or Generic.. Etc. this field 0Cancel is mapping with 0Record mode in BW. Like this BW identify the Delta images.
32)What is the difference between filter & Restricted Key Figures? Examples & Steps in BI?
Filter restriction applies to entire query. RKF is restriction applied on a keyfigure.Suppose for example, you want to analyse data only after 2006...showing sales in 2007,2008 against Materials..You have got a keyfigure called Sales in your cube
Now you will put global restriction at query level by putting Fiscyear > 2006 in the Filter.This will make only data which have fiscyear >2006 available for query to process or show.
Now to meet your requirement. ..like belowMaterial Sales in 2007 Sales in 2008M1 200 300M2 400 700You need to create two RKF's.Sales in 2007 is one RKF which is defined on keyfigure Sales restricted by Fiscyear = 2007Similarly,Sales in 2008 is one RKF which is defined on Keyfigure Sales restricted by Fiscyear = 2008Now i think u understood the differenceFilter will make the restriction on query level..Like in above case putting filter Fiscyear>2006 willmake data from cube for yeaers 2001,2002,2003, 2004,2005 ,2006 unavailable to the query for showing up.So query is only left with data to be shown from 2007 and 2008.Within that data.....you can design your RKF to show only 2007 or something like that...
33)How to create condition and exceptions in Bi.7.0? But I know in Bw3.5 version.?
From a query name or description, you would not be able to judge whether the query is having any exception.There are two ways of finding exception against a query:1. Execute queries one by one, the one which is having background colour as exception reporting are with exceptions.2. Open queries in the BEX Query Designer. If you are finding exception tab at the right side of filter and rows/column tab, the query is having exception.
34)The FI Business Flow related to BW. case studies or scenarios
FI FlowBasically there are 5 major topics/areas in FI,1. GL Accounting -related tables are SKA1, SKB1 Master dataBSIS and BSAS are the Transaction Data2. Account Receivables- related to CustomerAll the SD related data when transfered to FI these are created.Related Tables BSID and BSAD3. Account Payables - related VendorAll the MM related documents data when transfered to FI these are createdRelated Tables BSIK and BSAKAll the above six tables data is present in BKPF and BSEG tablesYou can link these tables with the hlp of BELNR and GJAHR and with Dates also.4. Special Purpose Ledger.. which is rarely used.5. Asset ManagmentIn CO there are Profit center AccountingCost center Accounting will be there.

Difference between ODS and INFOCUBE


 Infocubes have a multidimensional structure with dimension tables(max 16, 13 custom) and one fact table. they are meant for summarised records.
 ODS store data at a more granular level. they have flat structures like a table in R/3. They have a unique feature "overwrite" which is absent in case of cubes.
 You can use ODS to load to cube further.
 Anyway, one major difference is the manner of data storage. In ODS, data is stored in flat tables. By flat we mean to say ordinary transparent table whereas in a CUBE, it composed of multiple tables arranged in a STAR SCHEMA joined by SIDs. The purpose is to do MULTI-DIMENSIONAL Reporting
 Another difference is : In ODS, you can update an existing record given the KEY. In CUBES, theres no such thing. It will accept duplicate records and during reporting, SUM the keyfigures up. Theres no EDIT previous record contents just ADD. With ODS, the procedure is UPDATE IF EXISTING (base from the Table Key) otherwise ADD RECORD.
 ODS Stores line item level detail, more granular Can't create aggregates on ODS ODS are based on flat tables Only two dimensional reporting possible on ODS. Overwrite feature available while loading records
 Infocube
- Stores summarized data, less granular.
- Aggregates can be created on top of Infocubes for better performance of Queries.
- Multi-dimensional reporting possible on Infocube.
- Theres no overwrite feature while loading records.
 Infocubes are MDM objects that fact table and dimension table are available whereas ODS is not a MDM object there are no fact tables and dimension tables. It consists of flat transparent tables.
 In infocubes there are characteristics and keyfigures but in ods key fields and data fields. we can keep non key characteristics in data fields.
 Some times we need detailed reports we can get through ODS. ODS are used to store data in a granular form i.e level of detail is more. The data in the infocube is in aggregated form.
 From reporting point of view ods is used for operational reporting where as infocubes for multidimensional reporting.
 ODS are used to merge data from one or more infosources but infocubes does not have that facility.
 The default update type for an ODS object is overwrite for infocube it is addition. ODS are used to implement delta in BW. Data is loaded into the ODS object as new records or updating existing records in change log or overwrite existing records in active data table using 0record mode.
 You cannot load data using Idoc transfer method in ODS but u can do in infocube.
 You cannot create aggregate on ODS. You cannot create infosets on infocube.
 ODS objects can be used.
 When you want to use the facility of overwrite. If you want to overwrite nonkey characteristics and key figures. If you want detailed reports you can use ODS.
If you want to merge data from two or more infosources you can use ODS. It allows you to drill down from infocube to ODS through RRI interface.
 ODS objects can be used in the following scenarios. ODS is not a mandatory but depending on the requirements we have to use it.
 When you want to use the facility of overwrite. If you want to overwrite nonkey characteristics and key figures in the data fields column.
 If you want detailed reports, you can use ODS.
 If you want to merge data from two or more infosources you can use ODS.
 It allows you to drill down from infocube to ODS through RRI interface if u want detailed data from ODS.
 If you want to create an external file.
 The most important difference between ODS and BW is the existence of key fields in the ODS. In the ODS you can have up to 16 info objects as key fields. Any other info objects will either be added or overwritten! So if you have flat files and want to be able to upload them multiple times you should not load them directly into the info cube, otherwise you need to delete the old request before uploading a new one. There is the disadvantage that if you delete rows in the flat file the rows are not deleted in the ODS.
 I also use ODS-Objects to upload control data for update or transfer routines. You can simply do a select on the ODS-Table /BIC/A00 to get the data.
 ODS is used as an intermediate storage area of operational data for the data ware house ODS contains high granular data . ODS are based on flat tables, resulting in simple modeling of ODS . We can cleanse transform merge sort data to build staging tables that can later be used to populate INOFCUBE .
 An infocube is a multidimentionsl dat acontainer used as a basis for analysis and reporting processing. The infocube is a fact table and their associated dimension tables in a star schema. It looks like a fact table appears in the middle of the graphic, along with several surrounding dimension tables. The central fact is usually very large, measured in gigabytes. it is the table from which you retrieve the interesting data. the size of the dimension tables amounts to only 1 to 5 percent of hte size of the fact table. Common dimensions are unit & time etc.
 There are different type of infocubes in BW, such as basic infocubes, remote infocubes etc.
 An ODS is a flat data container used for reporting and data cleansing/quality assurance purpose. They are not based on star schema and are used primaily for detail reporting rather than for dimensional analyais.
 An infocube has a fact table, which contains his facts (key figures) and a relation to dimension tables. This means that an infocube exists of more than one table. These tables all relate to each other. This is also called the star scheme, because the dimension tables all relate to the fact table, which is the central point. A dimension is for example the customer dimension, which contains all data that is important for the customer.
 An ODS is a flat structure. It is just one table that contains all data. Most of the time you use an ODS for line item data. Then you aggregate this data to an infocube.
 ODS holds transactional level data..Its just as a flat table. Its not based on multidimensional model. ODS have three tables 1. Active table 2. change log 3. New table
 Cube holds aggregated data which is not as detailed as ODS. Cube is based on multidimensional model. Cube have 2 tables 1. E table 2. F table.