GRC BW Instalation

Embed Size (px)

DESCRIPTION

GRC BW

Citation preview

  • BI Content How-to and FAQs for

    GRC RM-PC-FN 3.0

    Table of Contents:

    1. Introduction

    2. Technical Requirements

    3. Instructions and Information:

    I. Installation Steps:

    A) PC/RM Installation (perform installation and listed steps)

    B) BW Installation (perform installation and listed steps)

    C) Customizing Extraction in PC/RM (perform steps 1-5)

    D) Replicate DataSources in BW (perform listed steps)

    E) Activate Business Content in BW (perform listed steps)

    II. Post-Installation Steps:

    F) Confirm Business Content activation in BW (perform listed steps)

    G) Follow-up to activation in BW (perform steps 1-3)

    H) Load Supporting Data in BW (prior to loading PC/RM data into BW)

    III. BW Maintenance & Operation:

    I) Loading PC/RM Data into BW (perform steps 1-3)

    J) Executing BW Queries and Web Templates (mostly informational)

    IV. Helpful Information:

    K) Known Problems and Limitations (apply Notes if issues arise)

    L) Frequently Asked Questions

  • Introduction:

    This document pertains to BI Content for PC 3.0 and RM 3.0. The content for PC 2.5

    and RM 2.0 are completely different from a model perspective than PC/RM 3.0.

    Therefore, an upgrade from the older content versions to 3.0 is not possible.

    The BI Content model as delivered can be installed and connected to either PC or RM, or

    both applications. You do not need both installed/configured in the backend. The

    content will work if it is only connected to PC or RM even if some or all of the BI objects

    pertaining to the non-connected application are activated. However, the delivered

    content is designed for an integrated scenario which means that if both PC and RM are to

    be connected to BI then both should be installed on the same client. If they reside on

    different systems or clients then workarounds are possible but not supported.

    BI will extract the PC/RM data from the datamart in PC/RM so this needs to be loaded

    prior to any BI extracts. The BI Content contains extractors for the PC and RM data

    based on timeframes. The desired timeframes and selected frequency need to be

    specified in the PC/RM system prior to BI extractions.

    Delivered process chains in the BI system are used to execute the proper flow of the

    extracts. A timeframe can be updated using subsequent extracts. There is no delta

    mechanism so all records for the timeframe will be extract each time but the receiving

    objects in BI are set to overwrite to avoid duplicating the data. So, in effect, an update

    to the existing data (not an append) will occur. However, a full timeframe refresh option

    is also available from the process chains since subsequent timeframe extracts will not

    remove any data from BI which had been deleted in PC/RM.

    Numerous queries and web templates have been delivered with the content. Analysis

    authorization has also been delivered based on objects Organizational Unit and

    Organization Unit in Regulation but some setup is involved. Several jumps from BI

    queries to reports in the PC/RM system have also been delivered but some configuration

    is required. These steps are all described in this document.

  • Technical Requirements:

    Required PC/RM system configuration Minimum GRC PC/RM 3.0 SP02 (however,

    SP03 is strongly recommended)

    Recommended BI system configuration:

    SAP_BW 7.0 SP22

    BI_CONT 7.04 SP04 or higher

  • Instructions and Information:

    Installation Steps:

    A) PC/RM System

    Install minimum GRC PC/RM 3.0 SP02 in the GRC system (however, SP03 is strongly

    recommended)

    B) BW System

    Install SAP_BW 7.0 SP22, BI_CONT 7.04 SP04 or higher in the BW system

    Create a BW source system including the necessary RFC connection to the PC/RM

    backend system. Steps and prerequisites involved on the BW and PC/RM system can be

    found in the online help under Creating SAP Source Systems in:

    http://help.sap.com/saphelp_nw70/helpdata/en/ac/4a4e38493e4774e10000009b38f889/co

    ntent.htm

    C) Customizing Extraction

    The following configuration needs to occur in the PC/RM source system to enable

    integration with BW:

    1) Maintain Timeframe Frequencies:

    In this Customizing activity, you specify the frequencies to be used in the Process

    Control and Risk Management applications. Beyond the standard frequencies defined in

    section Standard Settings below, other reasonable user-defined frequencies might be

    "quarterly", "semi-annually" or other non-calendar based cycles.

    Path: SPRO SAP Reference IMG GRC Process Control or Risk Management

    General Settings Key Attributes Maintain Timeframe Frequencies

    http://help.sap.com/saphelp_nw70/helpdata/en/ac/4a4e38493e4774e10000009b38f889/content.htmhttp://help.sap.com/saphelp_nw70/helpdata/en/ac/4a4e38493e4774e10000009b38f889/content.htm
  • 2) Maintain Timeframes:

    In this Customizing activity, you define the timeframes to be used in the Process Control

    and Risk Management applications. Timeframes can be considered as specific named

    time buckets within one year (such as "January" or "Week 24") and need to be defined to

    cover the entire year.

    Each timeframe is assigned to one previously defined frequency, and then combined with

    defined offset settings to create a user-defined timeframe.

    Path: SPRO SAP Reference IMG GRC Process Control or Risk Management

    General Settings Key Attributes Maintain Timeframes

  • 3) Transfer (activate) Business Content DataSources in the PC/RM source system

    (transaction SBIW):

    Activities

    Note that these steps need to be performed in both systems if PC and RM are run as

    separate applications. This will be necessary during the business content activation in

    BW as described below in step E.

    Execute the function Transfer Application Component Hierarchy under Business

    Content DataSources (alternatively transaction RSA9 can be used)

    Execute the function Transfer Business Content DataSources under Business Content

    DataSources (alternatively transaction RSA5 can be used). Transfer the delivered

    PC/RM DataSources into the active version. The DataSources are located under

    application components: GRC-PCRM30 and optionally also GRC-DP (DataSources for

    Process Control Events)

  • 4) Maintain BI Extraction Settings (transaction GRFN_BI_TF_CUST)

    In this activity you can customize the timeframes and timeframe frequencies that will be

    used to extract data into BI system. These settings will serve as input to the data

    selection ABAP routines in the BI InfoPackages during extraction. However, the data to

    be extracted also must be filled in the datamart (also referred to as the report buffer).

    Activities:

    1. Go to transaction GRFN_BI_TF_CUST (GRC BI Extraction Customizing) on the PC or RM system and make one and only one entry.

    2. Specify appropriate values for the fields

  • a. Frequency: Enter the frequency of the extracted data. The frequencies should have already been defined using transaction SPRO. (SAP

    Customizing Implementation Guide GRC Process Control General

    Settings Key Attributes Maintain Timeframe Frequencies)

    b. Timeframe: Enter the starting extraction timeframe. The timeframes should have already been defined using transaction SPRO. (SAP

    Customizing Implementation Guide GRC Process Control General

    Settings Key Attributes Maintain Timeframes)

    c. From Year: Enter the beginning of the time interval to be used to extract data. This field, in conjunction with the Timeframe field, is used to

    determine the start of the extraction time interval. Ex. If the Timeframe

    has been set to M05 and the From Year is set to 2008, then during

    extraction into BI, only data that is valid from 01 May, 2008 onwards will

    be extracted

    d. To Year: Enter the end year of the extraction time interval. The end date of the extraction is determined by the frequency. If the end date based on

    the frequency runs into the year following the To Year value, then that

    date will also be considered (even though it does not belong to the to

    year year).

    e. Mixed frequencies: This feature is reserved for the future and is not currently available. This setting will enable prefixing the timeframe ID

    with the frequency.

    f. Inconsistencies: It is recommended to keep this as Inconsistencies logged as Error (default setting). Any extraction-time inconsistencies will

    be logged as either warnings or errors depending on this flag.

    g. Do not extract special timeframes: Aggregation of Deficiencies (AoD) and Signoff data may belong to a different frequency and timeframe than

    specified in above settings. Hence it is recommended to keep this

    unchecked to extract AoD and Signoff relevant data as well.

    h. Click the Save button to save your changes

    5) Maintain DataMart calculation (transaction GRFN_DM_MAINTAIN Maintain

    Datamart)

  • Use this activity to create and maintain data that will be used to load data into BI system.

    The datamart is also referred to as the report buffer. If you need to extract data to BW,

    the corresponding snapshot has to be created in the datamart first, otherwise the

    extraction will issue warning about missing data for requested timeframe.

    In case you need to reload the given timeframe into BW with the most recent data,

    you must update the corresponding snapshot in the datamart first.

    Activities

    1. Creating and Filling Datamart Timeframes:

    Use transaction GRFN_DM_MAINTAIN for maintaining datamart timeframes.

    2. Steps to create Datamart entries

    1. On your PC source system: a. Ensure that the required timeframe and timeframe frequencies have been

    created using transaction SPRO

    b. Ensure timeframe customizing settings have been appropriately configured using transaction GRFN_BI_TF_CUST

    2. Go to transaction GRFN_DM_MAINTAIN and select Maintain Datamart then click Execute to enter the Datamart Log

    3. Select Goto Languages from the menu options of the Datamart Log. Here you should maintain the languages that will be considered when extracting text

    elements into the datamart, and eventually, into the BI system. Once maintained,

    go back to the main screen of the DataMart Log.

  • 4. Click on the Create button and enter the required information. 5. Depending on your reporting needs, specify the appropriate timeframe and year in

    the appropriate text boxes. Only one timeframe can be specified for each create

    datamart entry created. *** The App. Component field should always be FN

    as in the screenshot below. Entering PC or RM will not work, but entering

    FN (for Foundation) will process PC, RM and FN data. Enter FN even if

    you have only PC or RM installed, as well as when you have both of them

    installed.

    6. Click the Save button to create the new datamart. 7. In order to fill the datamart, select the datamart entry (or multiple entries) with

    status Created (yellow) and click on the Upload Data mart icon. Data will not

    be filled in entries with status Completed (Green) or Error (Red). Depending

    on the frequency setting defined in GRPC_TF_CUST transaction, the data will be

    organized and loaded into this datamart entry. The status of the schedule upload

    will be set to Completed upon successful load of the data.

    8. If you rerun the upload for an already loaded datamart entry (the Created entry), the existing data in the datamart entry (the Completed entry) will be available

  • for extraction to BI as long as the new upload is collecting the data. Once the new

    upload has finished collecting the data to be loaded into the datamart entry, the

    old data will be REPLACED with the new one. The status of the Created entry

    will be temporarily changed to In Process while the datamart is recalculated.

    Another way to fill or refresh the datamart once the entries have been created is to

    go to transaction GRFN_DM_MAINTAIN and select either of the two

    highlighted entries below: Fill Data Mart (Online) or Fill Data Mart (Batch).

    This will fill or refresh all of the datamart entries in one step.

    D) Replicate DataSources

    After activating the delivered DataSources in the PC/RM source system (step 3 under

    Customizing Extraction) you must go to BW and replicate these DataSources. In BW,

    call transaction RSA1 and under the Modeling tab select the Source System window.

    Find the PC/RM source system (do this for both source systems if these are run as

    separate applications) by expanding the SAP node under Source Systems, perform a

    right-click on it, and choose Replicate DataSources.

  • Optionally, you can verify the replication of the DataSources by double-clicking on the

    PC/RM source system which will take you to the DataSources window. Navigate the

    nodes to find the GRC PC/RM 3.0 (GRC-PCRM30) application component and check

    for the existence of the DataSources in the subnodes.

  • E) Activate Business Content in BW

    SAP BI Content is configured to work with the delivered, integrated PC/RM scenario.

    SAP does not support any enhancements to, or special configuration of, the PC/RM or BI

    delivered product. Any suggestions given in this document are intended for customers

    using the products in ways not intended by delivered design but SAP will not support

    such customizations.

    PC and RM are separate applications and they have distinct business content that can be

    activated and run independently of one another. However, they share the same

    foundational objects so you must activate all of the 0GRC_FN objects regardless of

    whether you are using the PC and/or RM content along with the process chains

    0GRC_PCRM30*.

    The business content was designed for the situation where PC and RM are truly

    integrated and are both installed on the same client in one backend system and extracted

    using one BW source system. If you have a landscape where they are stand alone

    applications and reside in different clients or even different systems then the values of the

    shared FN objects such as Risk ID could potentially collide. In this situation

    customization (similar to MDM) must occur. For example, it may be necessary to

    append the system to the key fields (entities) in order to keep the values from different

    systems distinct. However, such modifications are not supported by SAP.

    Additionally, during business content activation in these situations it is also necessary to

    select both source systems during the InfoArea activation as described below (the

    DataSources must be transferred in step C-3 and replicated in step D for BOTH source

    systems). Activating the process chains takes special care and one option to be integrated

    in the steps below is presented here:

    1. Load ALL chains with the PC source system selected 2. Delete chain GRC PC/RM 3.0: Load RM Targets 3. Reactivate the following chains with the RM source system selected:

    GRC PC/RM 3.0: Load Attribute Text

    GRC PC/RM 3.0: Load Master Data Text

    GRC PC/RM 3.0: Load Master Data Attributes

    GRC PC/RM 3.0: Load RM Targets

    GRC PC/RM 3.0: Reload

    GRC PC/RM 3.0: Start Load Authorizations

    GRC PC/RM 3.0: Start Load Hierarchies

    Activating content:

    In order to activate the delivered business content execute transaction RSA1 and select

    the Business Content menu and proceed as follows:

  • Select the Self-defd source system as well as the PC/RM source system(s)

    Choose Collection Mode Collect Automatically

    Choose Grouping In Data Flow Before

  • Collect the following InfoObject Catalogs: 0GRC_FN30*, 0GRC_PC30*,

    0GRCRM30* and select Install

  • Collect the following InfoAreas: 0GRC_FN, 0GRC_PC30, 0GRC_RM30 and

    select Install

    Choose Grouping Only Necessary

    If separate source systems are used for PC and RM then follow the instructions at

    the top of this section (E). Otherwise, simply collect all 0GRC_PCRM30* Process

    Chains and select Install

  • Collect the following Roles: SAP_BW_GRC_FN30_ROLE,

    SAP_BW_GRC_PC30_ROLE, SAP_BW_GRC_RM30_ROLE and select Install (this

    should include all associated queries and web templates)

    Post-Installation Steps:

    F) Confirm Business Content object activation

  • Optionally, you can call transaction RSA1 in BW and under Modeling ensure that the

    objects indicated in help.sap.com and their corresponding sub-objects are all present and

    active as follows:

    Under InfoObjects check for the existence and active status of InfoObjects under the

    following InfoObject Catalogs: 0GRC_FN30_CHA01, 0GRC_FN30_KFY01,

    0GRC_PC30_CHA01, 0GRC_PC30_KFY01, 0GRC_RM30_CHA01,

    0GRC_RM30_KFY01:

  • Under InfoProvider check for the existence and active status of InfoProviders,

    Transformations, DTPs and InfoPackages under the following InfoAreas: 0GRC_FN,

    0GRC_PC30, 0GRC_RM30:

  • Call transaction RSPC and ensure that all Process Chains are present and active under

    GRC PC/RM 3.0 (Process Control and Risk Management 3.0):

  • In the Business Explorer (BEx) use the Query Designer to ensure that all queries are

    active. You can drill down by roles to see queries by InfoProvider:

    The delivered roles are:

    SAP_BW_GRC_FN30_ROLE

    SAP_BW_GRC_PC30_ROLE

    SAP_BW_GRC_RM30_ROLE

    These roles should be assigned to users so they can readily access the delivered queries

    from either the User Menu is BI or from BEx.

    If any objects exist but are inactive, repeat the activation steps from above, Activate

    Business Content in BW. If some objects are still inactive after repeated attempts,

    reactivate them manually in the workbench. If any objects are missing then go to the BI

    Content tab and reactivate them from content according to the directions in the section

    above.

  • G) Follow-up to activation (to be done in BW)

    1) According to Note 871132, ensure that InfoObject 0RTYPE (Exchange Rate Type)

    does NOT have the conversion routine ALPHA specified in tab General. This could

    cause errors when executing queries containing currency conversion routines. Please

    follow the directions in the note carefully.

    2) In order to initiate BW 7.0 Authorizations, the following InfoObjects need to be active

    in RSA1: 0TCAACTVT, 0TCAIPROV, OTCAKYFNM, OTCAVALID. Additionally,

    ensure that the delivered objects 0GFN_OU and 0GPC_OURE are active as these objects

    play a significant role in PC/RM authorizations in BI Content. All of these InfoObjects

    also need to be made authorization relevant (if they are not already) by selecting this

    option in the Business Explorer tab. Additional instructions pertaining to enabling

    authorizations for PC/RM in BW are covered under the subsequent section on

    authorizations in this document.

    3) Query sender/receiver assignments (for jumps to the backend PC/RM system) have

    been delivered with BI Content but the Receiver Object URLs associated with the

    jumps to the backend system need to be maintained to reflect the customers landscape

    (for technical details see the FAQ section in this document under How do I retrieve

    comments / long text from the PC/RM system?). The delivered RSBBS entries contain

  • template URLs and specific parameters used to call the backend report so be careful to

    only make the necessary server and client adjustments to the URLs.

    Call transaction RSBBS and select the following queries as the Sender then select

    Continue (Enter):

    Sender Queries Receiver Reporting Objects

    Long Description Group:

    0GRM3MP01_Q0007 Show Risk Long Text Description

    0GRM3MP02_Q0001 Show Response Long Text Description

    0GPC3MP01_Q0003 Detailed Control Description

    0GPC3MP13_Q0002 CAPA Root Cause

    Case Details Group:

    0GPC3MP01_Q0003 Evaluation Case Details

    0GPC3MP02_Q0002 Evaluation Case Details

    0GPC3MP02_Q0003 Evaluation Case Details

    0GPC3MP02_Q0004 Evaluation Case Details

    0GPC3MP02_Q0006 Evaluation Case Details

    0GPC3MP02_Q0007 Evaluation Case Details

    0GPC3MP02_Q0008 Evaluation Case Details

    0GPC3MP02_Q0010 Evaluation Case Details

    0GPC3MP02_Q0011 Evaluation Case Details

    0GPC3MP02_Q0012 Evaluation Case Details

    0GPC3MP02_Q0013 Evaluation Case Details

    0GPC3MP02_Q0014 Evaluation Case Details

    0GPC3MP02_Q0015 Evaluation Case Details

    0GPC3MP03_Q0002 Issue Details

    0GPC3MP10_Q0002 Evaluation Case Details

    0GPC3MP11_Q0002 Evaluation Case Details

    0GPC3MP13_Q0002 CAPA Details

    0GPC3MP14_Q0002 Remediation Details

    For each of the above queries, select the listed Receiver and click on Change then

    perform a dropdown under Report where the URL exists at the bottom of the pop-up

    Change screen. Here you can adjust the URLs as indicated in steps 1-3 below the

    following two groups:

    Long Description Group

  • Case Details Group

  • For each of the above cases leave the majority of the URL in tact but just make the

    changes indicated below (rm below will be substituted with pc for the PC related

    jumps in the delivered URLs):

    1. Replace : or : with the address associated with your PC/RM server (such as rmserver.mycompany.corp:50050,

    for example)

    2. Replace the rm_client in sap-client= with the PC/RM source system client number (such as 200, for example)

    3. Then select Transfer, Transfer, then Save to preserve your changes

    4) The Permitted Extra Characters (RSKC) needs to be set to solely the value

    ALL_CAPITAL to allow expected character values from the PC/RM source system

    plus it will handle most unexpected customer values as well. Otherwise, some data loads

    may fail particularly with errors during the execution of the DTP related to invalid or

    non permitted values or characters.

    5) Apply Note 1459101 to correct an issue with special timeframe extracts for objects

    Signoff and AOD.

  • H) Load Supporting Data

    Some supporting data such as Exchange Rates, Currencies and Units of Measure need to

    be loaded into BW in order to enhance the PC/RM reporting capabilities. These first

    need to be established in the PC/RM system then loaded into BW by performing a right-

    click on the PC/RM source system in RSA1 and selecting Transfer Exchange Rates and

    Transfer Global Settings:

  • BEx maps have also been included in a number of delivered PC web templates based on

    countries and regions. First, the following master data text and attributes need to be

    loaded via delivered DataSources for the following InfoObjects: 0LANGU, 0REGION,

    0COUNTRY. Then, in order to display the graphics the BEx Map settings need to be

    maintained in InfoObjects 0COUNTRY and 0REGION and the shape files and

    abbreviations must be loaded by following the instructions in:

    http://help.sap.com/saphelp_bw/helpdata/en/1a/f405387bcc513be10000009b38f8cf/conte

    nt.htm

    Be sure that the BW master data, the shape files and the PC/RM application values for

    the country and region are all in sync. Be sure to maintain this data at the organization

    unit level in the PC/RM application. To do this, execute the transaction

    GRFN_STR_CHANGE in the PC/RM system. Then navigate by Organizational Unit on

    the left and double-click to make your selection. Maintain the values in the Country and

    Region tab on the right.

    http://help.sap.com/saphelp_bw/helpdata/en/1a/f405387bcc513be10000009b38f8cf/content.htmhttp://help.sap.com/saphelp_bw/helpdata/en/1a/f405387bcc513be10000009b38f8cf/content.htm
  • I) Loading PC/RM Data:

    Loading data from the PC/RM source system is driven by BW process chains. It is

    guided by timeframe customizing on the source system, InfoPackage selection criteria,

    and sometimes the timeframe override in BW.

    The timeframe(s) to be extracted into BI (for timeframe dependent loads) are specified in

    the source system using transaction GRFN_BI_TF_CUST GRC BI Extraction

    Customizing. The settings here indicate which timeframe(s) will be relevant for

    extraction from the datamart. Therefore, the datamart first needs to be filled with the

    desired timeframes for extraction.

    Only full extracts (limited by timeframe, where applicable) are delivered as there is no

    delta extraction of data from the source system enabled in the DataSources.

    Also, only one timeframe selection can be specified in customizing at a time and PC/RM

    is designed to handle only one timeframe frequency/granularity - so choose the lowest

    level that is desired to report on.

    If selection criteria are entered in the InfoPackage then the timeframe customizing entries

    are overridden. However, this is contrary to the delivered design and is not

    recommended unless necessary. But if you decide to stray from this design then you must

    ensure that each InfoPackage is in sync with the frequency/granularity which you have

    been loading (GRFN_BI_TF_CUST).

    If transaction RS_BCT_GRC_FN_TFLOAD is used for timeframe reload (covered

    below) then the timeframe customizing entries are overridden.

    Hierarchies are the exception to the InfoPackage selection rule (see step 3 below).

    Also, the information associated with the DataSources 0GPC_SIGNOFF (Org. Units

    Sign-Off Info) and 0GPC_AOD (Aggregation of Deficiencies) behave differently and are

    scheduled for particular timeframe frequencies in PC. In order to extract these special

    timeframes you need to leave blank the checkbox Do not extract special timeframes in

    timeframe customizing (GRFN_BI_TF_CUST). Otherwise, if you mark this selection

    then the customizing settings will be respected and any special timeframes not included

    in GRFN_BI_TF_CUST will be ignored.

  • Basically, the PC/RM data is loaded by executing the following three process chains in

    order, one after the other one completes (details below):

    1. GRC PC/RM 3.0: Start Load Main Data

    2. GRC PC/RM 3.0: Start Load Authorizations

    3. GRC PC/RM 3.0: Start Load Hierarchies

    1. Standard Load of Master Data, Text and InfoProviders (Main Data):

    Typically, the standard load of data will be performed by scheduling one of the delivered

    process chains. This will need to happen in conjunction with the manual maintenance of

    the custom timeframe in the source system to ensure that the desired timeframe(s) is

    being extracted.

    0GRC_PCRM30_LOAD_ALL) using

    transaction RSPC in BW. This chain will delete the PSAs prior to loading to avoid

    duplication of data since the DTP Extraction Modes are defined as Full. The key

    figures of the DSOs are set to overwrite so the existing contents are configured to be

    updated, not appended. However, the contents of the target InfoCubes are deleted prior

    to each load in order to avoid duplication of data. Subsequent timeframe extracts will not

    remove any data from BI which had been deleted in PC/RM so a timefame refresh option

    is also available from the process chains (covered below).

    2. Load Authorizations:

    In order to update the authorizations data you need to

    (0GRC_PCRM30_AUTHS). This chain will

    delete the PSAs prior to loading to avoid duplication of data since the DTP Extraction

    Mode is defined as Full. This is also a full extract of data without any timeframe

    limitation as only current data is stored in the source system. The contents of the DSO

    are also deleted prior to loading since this is a complete refresh of the data with the

    results being simply the current authorizations being contained in the DSO. The

    processing of the authorization data is covered in section J under Authorizations Setup

    and Generation.

  • 3. Load Hierarchies:

    In order to update the hierarchy data you need to

    (0GRC_PCRM30_HIER). These extracts are timeframe

    dependent so the timeframe is determined by a combination of the value in

    GRFN_BI_TF_CUST in the source system and the selection process in the InfoPackage.

    First you must select a hierarchy version from the available list in each InfoPackage prior

    to performing the extract. From the Hierarchy Selection tab click on Available

    Hierarchies from OLTP to refresh the list. The list will represent the hierarchies which

    can be built from the timeframes specified in timeframe customizing. Select one of the

    listed hierarchies and save the InfoPackage. Note that prior to each hierarchy extract it

    may be necessary to update the hierarchy selection in the InfoPackage in order to refresh

    the list and select the latest hierarchy available in the PC/RM system.

    Repeat this process for all delivered InfoPackages for hierarchies belonging to

    InfoObjects 0GPC_CEC, 0GFN_TF, 0GRM_RG, 0GPC_CSP, 0GFN_OU, 0GRM_OG,

    0GPC_AG, 0GRM_CA and 0GPC_OURE

    Finally, you can execute the process chain GRC PC/RM 3.0: Start Load Hierarchies

    (0GRC_PCRM30_HIER).

    If the delivered Organizational Unit hierarchy setting is changed from Entire hierarchy is

    time-dependent to Time-Dependent Hierarchy Structure then this is not supported and

    the results cannot be guaranteed.

    Since hierarchy extraction uses 3.x InfoSources there are no DTPs involved and no

    deletion of the PSAs is necessary to avoid duplicate data loading.

    Timeframe Reload (situational):

    Since subsequent timeframe extracts will not remove any data from BI which had been

    deleted in PC/RM, a full timeframe refresh option is also available from the process

    chains. If a timeframe needs to be reloaded then the following procedure must be

    followed. Run transaction RS_BCT_GRC_FN_TFLOAD in BW and enter the

    timeframe to reload. Make sure the entry is in sync with the frequency/granularity

    which you have been loading specified in timefame customizing of the PC/RM system

    (transaction GRFN_BI_TF_CUST). The InfoPackage selection routines will allow this

    value to override the value entered in GRC BI Extraction Customizing in the source

    system. Initiate the extraction/load by executing process chain GRC PC/RM 3.0: Reload

    - Start Reload Timeframe (0GRC_PCRM30_RELOAD_TF). This chain will delete the

    data and PSAs for the timeframe from the master data, text and InfoProviders and then

    reload the data for this timeframe. Remember to remove the value from

  • RS_BCT_GRC_FN_TFLOAD or subsequent timeframe dependent loads (the Standard

    load and the Hierarchies load) will fail.

    Complete Data Deletion (situational):

    If you ever need to completely remove data from the system execute process chain GRC

    PC/RM 3.0: Delete All Data (Master/Trans/Hierarchy)

    (0GRC_PCRM30_DELETE_ALL). This would only be required in special cases and is

    not part of routine, scheduled loads.

    Description of Process Chains:

    Two streams of process chains exist in the automation of data loads GRC PC/RM

    3.0: Start Load Main Data and GRC PC/RM 3.0: Reload - Start Reload Timeframe.

    The process chains GRC PC/RM 3.0: Start Load Hierarchies, GRC PC/RM 3.0: Start

    Load Authorizations, and GRC PC/RM 3.0: Delete All Data (Master/Trans/Hierarchy)

    are stand alone but the two aforementioned call subsequent chains so their flow will be

    defined in detail below.

    Flow of GRC PC/RM 3.0: Start Load Main Data (0GRC_PCRM30_LOAD_ALL)

    process chain:

    GRC PC/RM 3.0: Start Load Main Data GRC PC/RM 3.0: Load Preliminary Data

    GRC PC/RM 3.0: Load Attribute Text GRC PC/RM 3.0: Load Master Data Text

    GRC PC/RM 3.0: Load Master Data Attributes GRC PC/RM 3.0: Load PC Targets

    GRC PC/RM 3.0: Load RM Targets

    Flow of GRC PC/RM 3.0: Reload - Start Reload Timeframe

    (0GRC_PCRM30_RELOAD_TF) process chain:

    GRC PC/RM 3.0: Reload - Start Reload Timeframe GRC PC/RM 3.0: Reload -

    Delete Timeframe MD & Targets GRC PC/RM 3.0: Load Preliminary Data GRC

    PC/RM 3.0: Reload - Load Timeframe Dep. Attribute Text GRC PC/RM 3.0: Load

    Master Data Text GRC PC/RM 3.0: Load Master Data Attributes GRC PC/RM 3.0:

    Load PC Targets GRC PC/RM 3.0: Load RM Targets

  • J) Executing BW Queries and Web Templates:

    All delivered queries and web templates are found under the following roles in BW

    (transaction PFCG): SAP_BW_GRC_FN30_ROLE, SAP_BW_GRC_PC30_ROLE,

    SAP_BW_GRC_RM30_ROLE. These roles do not provide secured access to these

    reporting objects but simply provide a method of convenient access to the objects once

    the role is assigned to a user. Optionally, access to the queries is secured either by

    InfoProvider or at the individual query level. To secure by InfoProvider users must have

    sufficient authorization both in analysis authorizations (via the special characteristics)

    and within PFCG roles/profiles. To secure at the query level authorization is granted by

    PFCG roles/profiles. Analysis authorization and PFCG maintenance is outside the scope

    of this document.

    If queries fail to execute first try to regenerate them using program

    RSR_GEN_DIRECT_ALL_QUERIES (transaction SE38) in BW. Select InfoCubes

    0GFN3*, 0GPC3* and 0GRM3*.

    Query analysis authorizations (field level authorizations) are based on BW InfoObjects

    Organizational Units (0GFN_OU) and Org. Unit in Regulation (0GPC_OURE). Query

    authorization variables are used to restrict access during query execution. A full

    explanation of this procedure is under the subsequent section on authorizations in this

    document.

    The following sender/receiver assignments (jumps) from one query to another reporting

    object are provided with the PC/ RM Business Content. They are defined in transaction

    RSBBS and are executed by performing a right-click Go To from the query output

    screen. The column/row from which the jump is executed (as well as the assignment

    details in RSBBS) determines the selection criteria for the receiving reporting object and

    thus the output.

    There are three types of receiving reporting objects below: Query, Web Template (WT)

    and URL. The URLs represent jumps to the backend PC/RM system in order to report on

    data that is not loaded into BW. Below are all of the delivered queries representing these

    three types of jumps.

    RM Sender Queries RM Receiver Objects

    Risk by Objective Org. Unit - Objective assignment

    Risk Overview Show Risk Long Text Description

    Response Overview Show Response Long Text Desc

    Opportunity by Objective Org. Unit - Objective assignment

    KRI Evaluations by Org Unit KRI Instance History Chart

    PC Sender Queries PC Receiver Objects

    Heat Map for Control Evaluations Control Evaluations Summary

    Control Evaluations Summary Control Evaluations Details

  • Control Evaluations Details Evaluation Case Details

    Control Evaluations Details Detailed Control Description

    Control Testing Evaluation History Overall Ctrl Testing Evals Hist for Org. Unit

    Failed Controls for Account Group - Risks Failed Testings

    Failed Controls for Account Group - Risks Failed Design Assessments

    Failed Controls for Account Group - Risks Failed Self Assessments

    Failed Controls for Account Group - Risks Failed Monitorings

    Failed Testings for Account Group - Risks Evaluation Case Details

    Failed Design Assessments for Acct Group - Risks Evaluation Case Details

    Failed Self Assessments for Acct Group - Risks Evaluation Case Details

    Failed Controls for CO-Risks Failed Testings

    Failed Controls for CO-Risks Failed Design Assessments

    Failed Controls for CO-Risks Failed Self Assessments

    Failed Controls for CO-Risks Failed Monitorings

    Failed Testings for CO-Risks Evaluation Case Details

    Failed Design Assessments for CO-Risks Evaluation Case Details

    Failed Self Assessments for CO-Risks Evaluation Case Details

    Failed Controls for Subprocess - Risks Failed Testings

    Failed Controls for Subprocess - Risks Failed Design Assessments

    Failed Controls for Subprocess - Risks Failed Self Assessments

    Failed Controls for Subprocess - Risks Failed Monitorings

    Failed Testings for Subprocess - Risks Evaluation Case Details

    Failed Design Assessments for Subprocess - Risks Evaluation Case Details

    Failed Self Assessments for Subprocess - Risks Evaluation Case Details

    Failed Monitorings for Account Group - Risks Evaluation Case Details

    Failed Monitorings for CO-Risks Evaluation Case Details

    Failed Monitorings for Subprocess - Risks Evaluation Case Details

    Issue Status by Organization Issue Status Details by Organization

    Issues Status Details by Organization Issue Details

    Indirect ELC Evaluation Summary Indirect ELC Evaluation Details

    Indirect ELC Evaluation Details Evaluation Case Details

    Subprocess Evaluations Summary Subprocess Evaluations Details

    Subprocess Evaluations Details Evaluation Case Details

    Issues Trend Analysis Issues Trend Analysis Details

    Issues Trend Analysis Details Issue Details

    Issue Trends By Org. Unit Issues Trend Analysis Details

    Issue Trends By Subprocess Issues Trend Analysis Details

    Issue Trends By Risk Level Issues Trend Analysis Details

    Issue Trends By Issue Priority Issues Trend Analysis Details

    CAPA Trend Analysis CAPA Trend Analysis Details

    CAPA Trend Analysis Details CAPA Details

    CAPA Trend Analysis Details CAPA Root Cause

    CAPA Trends by Org. Unit CAPA Trend Analysis Details

    CAPA Trends by Subprocess CAPA Trend Analysis Details

    CAPA Trends by Risk Level CAPA Trend Analysis Details

    Trends by CAPA Statuses CAPA Trend Analysis Details

  • Remediation Trend Analysis Remed Plans Trend Analysis Details

    Remediation Trend Analysis Dtls - Global Scope Remediation Details

    Trend by Org. Unit Remed Plans Trend Analysis Details

    Trend by Subprocess Remed Plans Trend Analysis Details

    Trend by Risk Levels Remed Plans Trend Analysis Details

    Trend by Remediation Statuses Remed Plans Trend Analysis Details

    Currency conversion can be performed on all delivered queries that contain currency

    value key figures. The currency translation type 0GRC_TFEND (transaction (RSCUR) is

    tied to all PC/RM currency type key figures in the BW queries. It evokes the user entry

    variables 0PGRCXRT (Exchange Rate Type) and 0PGRCCUR (Target Currency) and all

    conversion values used are derived from table TCURR and evaluated in reference to the

    key date of the query.

    The following Timeframe related variables are used in the PC/RM queries:

    End Date of Timeframe (0P_GFN_TF_PE) is an SAP Exit variable referencing the

    InfoObject 0DATE. It determines the last day of the user entered timeframe from user

    entry, single value variable 0P_GFN_TF in order to supply the key date of the query.

    Function Module RSVAREXIT_0P_GFN_TF_PE Timeframe End Date from

    0P_GFN_TF is used to determine the proper date.

    End Date of Timeframe Interval (0P_GFN_TF_IE) is an SAP Exit variable referencing

    the InfoObject 0DATE. It determines the last day of the user entered timeframe range

    from user entry, range value variable 0I_GFN_TF in order to supply the key date of the

    query. Function Module RSVAREXIT_0P_GFN_TF_IE Timefame Interval End Date

    from 0I_GFN_TF is used to determine the proper date.

    The following nine hierarchies are provided with BI Content for PC/RM: 0GPC_CEC,

    0GFN_TF, 0GRM_RG, 0GPC_CSP, 0GFN_OU, 0GRM_OG, 0GPC_AG, 0GRM_CA,

    0GPC_OURE. These hierarchies are loaded as well as activated using the process chain

    0GRC_PCRM30_HIER. The details of this process are covered in section I-3 Loading

    PC/RM Data. Some of the delivered queries display their output by these hierarchies.

    Authorizations Setup and Generation:

    Field level query security (analysis authorizations) is delivered in PC/RM 3.0 content

    using BI 7.x Analysis Authorizations technology based on Organizational Unit

    (0GFN_OU) and Org. Unit in Regulation (0GPC_OURE). Authorizations based on, for

    example, Org Unit simply means that you can only report on records from an

    InfoProvider which contain the Org Unit values you are authorized to see. However, you

    can see the values for all other fields associated with these records (such as subprocess

    level data) with no restrictions on those particular fields. You are free to turn off the

  • authorization on these objects and/or adding new authorization relevant objects but such

    changes are customizations and are not supported by SAP.

    The values determining the BI analysis authorizations for a user are derived from the Org

    Unit and Org Unit in Regulation authorizations in the PC/RM backend system and are

    based strictly on current (not time dependent) authorizations. These values are loaded

    into the BW system then authorizations are generated using a standard BW tool

    (RSECADMIN). In order to leverage these generated authorizations the InfoObjects

    which the authorization is based on were made authorization relevant. Additionally,

    query authorization variables are created on these InfoObjects and used within the queries

    to facilitate the security. Previously, field level query output security in PC/RM 2.5 was

    supported by exit variables in the queries and provides a form of time dependent

    reporting authorization.

    The DataSource 0GFN_AUTHORIZATIONS has been delivered to extract the

    authorizations data from the PC/RM backend system. This DataSource feeds the DSO

    0GFN_DS01 which is a copy of the template DSO 0TCA_DS01 used for processing the

    supporting authorization data. Once loaded, the data is processed in BW using

    transaction RSECADMIN.

    Transaction RSECADMIN is used to generate authorizations based on the authorization

    data stored in DSO 0GFN_DS01. The authorization data consists of individual values or

    ranges of values for each authorization relevant InfoObject by user. The generated

    authorizations contain these current authorization values for each user assigned to their

    user IDs (see the User tab of RSECADMIN).

    Authorizations can be assigned to a user via RSECADMIN or in PFCG. But the

    recommended method is to use RSECADMIN for all analysis authorization objects while

    PFCG should be limited to just role based authorization. Beware that some authorization

    assignments (such as * in the S_RS_COMP object) can lead to the authorization

    0BI_ALL being assigned to a user which effectively negates all analysis authorization

    functionality. This can be seen in RSECADMIN -> User -> Assignment -> Display ->

    Role-Based tab.

    The InfoObjects 0GFN_OU and 0GPC_OURE have been delivered authorization

    relevant by checking this option in the InfoObject definition tab BEx Explorer.

    0GPC_OURE is used solely in the PC content while 0GFN_OU is used in both the PC

    and RM content. Users running queries defined against an InfoCube containing either of

    these InfoObjects must have analysis authorization based on these objects even if it is

    simply the value *.

    Use of authorization variables is suggested in all queries belonging to an InfoProvider

    containing any of the authorization relevant InfoObjects. The variables will use the

    values in the generated authorizations to create the valid scope of users access. If the

    users selection criteria are a subset of the users authorization scope, then the result of

    the query will be the intersection of the users authorizations and the query filter

  • selection. If the selection criteria are outside of the users authorization then an

    authorization error will occur when the query is executed.

    Authorization variable 0S_GFN_OU_AUTH has been delivered to provide the users

    authorizations while allowing the user to enter query selection criteria (ready for input).

    In other words, the output of the query will be based on the intersection of the entered

    criteria and the defined authorizations.

    Authorization variable 0S_GFN_OU_AUTH_H has been delivered to provide the users

    authorizations without allowing the user to enter query selection criteria (not ready for

    input). In other words, the output of the query will be based solely on the authorized Org

    Unit values for this user.

    Special high level authorization characteristics exist in the system to facilitate reporting

    security. The InfoObjects 0TCAACTVT, 0TCAIPROV, 0TCAKYFNM and

    0TCAVALID must be activated from content and also made authorization relevant (in

    the Business Explorer tab in the InfoObject definition). An authorization must also be

    created in RSECADMIN containing these InfoObjects with the value * assigned to

    them. This authorization needs to be assigned to all BEx end users in order to run

    queries. This is in addition to the required PFCG role assignments required to run

    queries: S_RS_COMP, S_RS_COMP1, S_RFC, S_TCODE, etc).

    For more information on the BI 7.0 Analysis Authorizations concept please see: An

    Expert Guide to new SAP BI Security Features -

    http://www.sdn.sap.com/irj/scn/elearn?rid=/library/uuid/659fa0a2-0a01-0010-b39c-

    8f92b19fbfea

    http://www.sdn.sap.com/irj/scn/elearn?rid=/library/uuid/659fa0a2-0a01-0010-b39c-8f92b19fbfeahttp://www.sdn.sap.com/irj/scn/elearn?rid=/library/uuid/659fa0a2-0a01-0010-b39c-8f92b19fbfea
  • K) Known Problems and Limitations:

    1) There is a limitation when using the F4 help function (from the variable screen

    dropdowns during query execution) when a user exit variable is used for the key date

    of the query. This is a known BEx limitation in the variable screen only and does not

    affect the query results. In this situation the key date is not derived by the exit variable

    until the query is run so the time dependent attributes and text associated with a

    characteristic reflect the current system date (the default) in a dropdown in the variable

    screen.

    However, once the query result is rendered, if the attributes or text are displayed they will

    reflect the derived key date of the query so this issue does not affect the output. Since

    time dependent reporting delivered with the content is based on the timeframe end date,

    all objects displayed in the query output will be in sync with this derived date.

    In the PC/RM 3.0 content the situation may arise when user entry timeframe variables

    0P_GFN_TF or 0I_GFN_TF are used to determine the timeframe end date in SAP exit

    variables 0P_GFN_TF_PE or 0P_GFN_TF_IE, respectively.

    If users have the tendency to filter query selections in the variable screen based on

    characteristics with time dependency, then a workaround can be created if this is

    bothersome to them. Here is a suggested solution:

    Have the user enter the Timeframe End Date using a variable associated with the key date

    of the query. Since more than one timeframe may be associated with the end date (in

    situations where multiple frequencies or special timeframes have been used) the desired

    timeframe needs to be either entered manually or determined by a user exit. If an exit is

    used then the primary timeframe frequency could be store in a table in order to determine

    the proper timeframe associated with the end date. Then the selection of the query will

    filter on the timeframe and the key date of the query will be equal to the timeframe end

    date.

    Steps to perform in the BI system:

    - Create a table ZGFN_RPT_TF_FREQ with field FREQUENCY CHAR(10) (from data element /BI0/OIGFN_TFFREQ) and store the primary reporting frequency to be

    loaded into BI. Set the Data Class to APPL2 (Organization and customizing) under

    Technical Settings and the Enhancement Category for Structure to Can Be

    Enhanced (DEEP) under Extras (unless there are different customer standards).

    - Create a single value, mandatory, user entry input variable ZP_TF_END_DATE (Timeframe End Date) based on InfoObject 0DATE (this will allow you to use the

    variable in the key date section and when filtering on TF end date as well)

  • - Ensure that the setting in InfoObject 0GFN_TFEND for filter value selection (in the BEx tab in RSD1) is set to Only Values in InfoProvider

    - Set the query key date (and all references in the query) to use the new date variable ZP_TF_END_DATE

    - Create single value, mandatory (initial value not allowed), not ready for input, customer exit variable ZP_TF_FROM_END_DATE (Timeframe from End Date)

    based on InfoObject 0GFN_TF. Replace the delivered timeframe variable

    (0P_GFN_TF) with this new variable in all query occurrences. *Note that under this

    solution if you wish to use a range of timeframes in the query then instead an interval

    variable would need to be created defined as user entry input.

    - Function Module EXIT_SAPLRRS0_001 is used to include code for customer exit BEx variables. The code is written in Include ZXRSRU01 but the exit first needs to

    be activated under a project in CMOD. The code for looking up the proper

    timeframe (from the user input timeframe end date and the primary reporting

    frequency) by following all steps above should look something like the following

    but additionally customer coding standards should also be adhered to.

    DATA: L_S_RANGE TYPE RRRANGESID. DATA: L_S_VAR_RANGE TYPE RRRANGEEXIT. CASE I_VNAM. WHEN 'ZP_TF_FROM_END_DATE'. DATA: TF_FREQ LIKE ZGFN_RPT_TF_FREQ, FREQ LIKE ZGFN_RPT_TF_FREQ- FREQUENCY, TF_MD LIKE /BI0/PGFN_TF, TF LIKE /BI0/PGFN_TF - GFN_TF. IF I_STEP = 2. SELECT * from ZGFN_RPT_TF_FREQ into TF_FREQ. IF SY - SUBRC = 0. FREQ = TF_FREQ- FREQUENCY. READ TABLE I_T_VAR_RANGE INTO L_S_VAR_RANGE WITH KEY VNAM = 'Z P_TF_END_DATE'. IF SY - SUBRC = 0. SELECT * from /BI0/PGFN_TF into TF_MD WHERE GFN_TFFREQ = FREQ and GFN_TFEND = L_S_VAR_RANGE- LOW. IF SY - SUBRC = 0. TF = TF_MD - GFN_TF. CLEAR L_S_RANGE. L_S_RANGE- LOW = TF. L_S_RANGE- SIGN = 'I'. L_S_RANGE- OPT = 'E Q'. APPEND L_S_RANGE to E_T_RANGE. ENDIF. ENDSELECT.

  • ENDIF. ENDIF. ENDSELECT. ENDIF.

    2) Errors can occur when attempting to perform a currency conversion during query

    execution in either the variable screen or when using F4 Help in the currency translation

    type field. Please apply Note 1348575 to resolve this issue.

    3) If errors occur when running or editing process chains with regard to the PSA deletion

    steps, please refer to Note 1344779 Correction: Multiple PSA references in variant

    from content. Possible error messages which could be related this problem are:

    TSV_TNEW_BLOCKS_NO_ROLL_MEMORY and No roll storage space of length

    ####### available for internal storage. Please read the note before analyzing memory

    and space related issues.

    4) According to Note 871132, ensure that InfoObject 0RTYPE (Exchange Rate Type)

    does NOT have the conversion routine ALPHA specified in tab General. This could

    cause errors when executing queries containing currency conversion routines. Please

    follow the directions in the note carefully.

    5) If errors occur when executing queries containing currency conversion routines,

    please apply Note 871132 InfoObject: 0RTYPE has ALPHA conversion exit.

    6) Apply Note 1459101 to correct an issue with special timeframe extracts for objects

    Signoff and AOD. Without this correction the flag Do not extract special timeframes in

    timeframe customizing (GRFN_BI_TF_CUST) will be ignored and the special

    timeframes will be extracted along with all timeframes included in GRFN_BI_TF_CUST

    (instead of skipping the special timeframes).

  • L) Frequently Asked Questions

    How do I prepare for loading data into BI?

    Getting ready to integrate BI with PC/RM data involves preparation in both the PC/RM

    and BI systems. The steps are covered in detail under Installation Steps (sections A

    H) but here they are summarized:

    1. Install GRC PC/RM 3.0 SP03 in the GRC system

    2. Install SAP_BW 7.0 SP22, BI_CONT 7.04 SP04 or higher in the BW system

    3. Create a BW source system and RFC connection to the PC/RM backend system

    4. Maintain Timeframe Frequencies in the PC/RM system

    5. Maintain Timeframes in the PC/RM system

    6. Transfer Application Component Hierarchy for BI in the PC/RM system

    7. Transfer (activate) Business Content DataSources in the PC/RM system

    8. Maintain BI Extraction Customizing Settings (transaction GRFN_BI_TF_CUST in PC/RM)

    a. Set the frequency, starting timeframe and from/to year to extract into BI

    9. Maintain DataMart calculation (transaction GRFN_DM_MAINTAIN in PC/RM)

    a. Use this activity to create and maintain data that will be used to load data into BI system. The datamart is also referred to as the report buffer.

    10. Replicate DataSources in the BI system

    11. Activate Business Content in the BI system

    12. Follow-up steps to activation (to be done in BW):

    a. According to Note 871132, ensure that InfoObject 0RTYPE (Exchange Rate Type) does NOT have the conversion routine ALPHA specified in

    tab General.

    b. In order to initiate BW 7.0 Authorizations, the following InfoObjects need to be active in RSA1: 0TCAACTVT, 0TCAIPROV, OTCAKYFNM,

    OTCAVALID.

    c. Query sender/receiver assignments (for jumps) have been delivered with BI Content but the URLs associated with the jumps to the backend

    system need to be maintained to reflect the customers landscape.

    d. The Permitted Extra Characters (RSKC) needs to be set to solely the value ALL_CAPITAL

    13. Load Supporting Data:

    a. Supporting data such as Exchange Rates, Currencies and Units of Measure need to be loaded into BW in order to enhance the PC/RM reporting

    capabilities.

    b. Master data text and attributes to support the BEx maps need to be loaded for the following InfoObjects: 0LANGU, 0REGION, 0COUNTRY.

  • c. BEx Map settings need to be maintained in InfoObjects 0COUNTRY and 0REGION and the shape files and abbreviations must be loaded

    What is the flow of data in BI and how do I update, refresh and delete it?

    Loading data from the PC/RM source system is driven by BW process chains. It is

    guided by timeframe customizing on the source system, InfoPackage selection criteria,

    and sometimes the timeframe override in BW. The details are covered completely in

    section I Loading PC/RM Data but here they are summarized:

    The timeframe(s) to be extracted from the PC/RM datamart into BI are specified in the source system using transaction GRFN_BI_TF_CUST GRC BI Extraction

    Customizing.

    Only full extracts (limited by timeframe, where applicable) are delivered as there is no delta extraction of data from the source system enabled in the

    DataSources.

    Only one timeframe frequency/granularity is allowed - so choose the lowest level that is desired to report on.

    The PC/RM data is loaded by executing the following three process chains:

    :

    Chain is used to load the master data, text and InfoProviders from the PC/RM system

    Deletes the PSAs prior to loading to avoid duplication of data since the DTP Extraction Modes are defined as Full

    Key figures of the DSOs are set to overwrite so the existing contents are configured to be updated, not appended

    Contents of the target InfoCubes are deleted prior to each load in order to avoid duplication of data

    Subsequent timeframe extracts will not remove any data from BI which had been deleted in PC/RM so a timefame refresh option is also available

    from the process chains (covered below)

    Information associated with the DataSources 0GPC_SIGNOFF (Org. Units Sign-Off Info) and 0GPC_AOD (Aggregation of Deficiencies)

    behave differently and are scheduled for particular timeframe frequencies

    in PC (see detailed instructions on this)

    :

    Chain is used to load the authorizations from the PC/RM system

  • Deletes the PSAs prior to loading to avoid duplication of data since the DTP Extraction Mode is defined as Full

    This is a full extract of data without any timeframe limitation

    Contents of the DSO are deleted prior to loading since this is a complete refresh of the data

    Processing of the authorization data is covered in section J under Authorizations Setup and Generation

    :

    Chain is used to load the hierarchies from the PC/RM system

    The timeframe of the extract is determined by a combination of the value in GRFN_BI_TF_CUST in the source system and the selection process in

    the InfoPackage

    Prior to performing the extract select a hierarchy version from the available list in each InfoPackage

    Prior to each hierarchy extract it may be necessary to update the hierarchy selection in the InfoPackage

    InfoPackages are delivered for hierarchies belonging to InfoObjects 0GPC_CEC, 0GFN_TF, 0GRM_RG, 0GPC_CSP, 0GFN_OU,

    0GRM_OG, 0GPC_AG, 0GRM_CA, 0GPC_OURE

    Timeframe reloads can be performed if necessary:

    Since subsequent timeframe extracts will not remove any data from BI which had

    been deleted in PC/RM, a full timeframe refresh option is also available from the

    process chains. If a timeframe needs to be reloaded then the following procedure

    must be followed:

    Run transaction RS_BCT_GRC_FN_TFLOAD in BW and enter timeframe to reload

    Execute process chain 0GRC_PCRM30_RELOAD_TF

    Chain will delete data and PSAs for the timeframe from the master data, text and InfoProviders and then reload the data for this timeframe

    Remove the value from RS_BCT_GRC_FN_TFLOAD or subsequent timeframe dependent loads will fail

    Complete deletion of data can be performed if necessary:

    If you need to completely remove data from the system execute process chain

    0GRC_PCRM30_DELETE_ALL.

  • How do I extend the delivered DataSources in PC/RM 3.0 SP3?

    Please see Note 1314368 regarding customer defined fields which includes two white

    paper attachments. Use the white papers in conjunction with the table below while

    following these steps:

    1. Prepare the Customer Include (CI) in the PC/RM application. The naming convention is CI_GRxx_xxxxxx and the lists of tables/objects that can be

    enhanced are in the attachments of Note 1314368.

    2. Create the CI in reporting with naming convention CI_GRxx_xxxxxx_BI. It will be created using structure GRFN_S_BI_xx_ATTR_I as indicated in the

    table below. The names of the fields need to correspond to the names of the fields

    created in the first step (data is copied using move corresponding) but must not

    be of type string.

    3. Edit the corresponding DataSource in RSA6 and unhide the new fields then resave the DataSource.

    4. Replicate the modified DataSources from PC/RM to BI.

    Datamart Datamart Include Datamart CI Reporting/Extractor

    Structure Reporting/Extractor CI DataSource

    Orgunit GRFN_S_OU_ATTR_I CI_GRFN_ORGUNIT GRFN_S_BI_OU_ATTR_I CI_GRFN_ORGUNIT_BI 0GFN_OU_ATTR

    Account Group GRFN_S_AG_ATTR_I CI_GRPC_ACC_GROUP GRFN_S_BI_AG_ATTR_I CI_GRPC_ACC_GROUP_BI 0GPC_AG_ATTR

    Assessment GRFN_S_AS_ATTR_I CI_GRPC_CASEAS GRFN_S_BI_AS_ATTR_I CI_GRPC_CASEAS_BI 0GPC_AS_ATTR

    CAPA plan GRFN_S_CP_ATTR_I CI_GRPC_CASECP GRFN_S_BI_CP_ATTR_I CI_GRPC_CASECP_BI 0GPC_CP_ATTR

    Issue GRFN_S_IS_ATTR_I CI_GRPC_CASEIS GRFN_S_BI_IS_ATTR_I CI_GRPC_CASEIS_BI 0GPC_IS_ATTR

    Remediation Plan GRFN_S_PL_ATTR_I CI_GRPC_CASEPL GRFN_S_BI_PL_ATTR_I CI_GRPC_CASEPL_BI 0GPC_PL_ATTR

    Test Log GRFN_S_TL_ATTR_I CI_GRPC_CASETL GRFN_S_BI_TL_ATTR_I CI_GRPC_CASETL_BI 0GPC_TL_ATTR

    Control Objective GRFN_S_CO_ATTR_I CI_GRPC_COBJECTIVE GRFN_S_BI_CO_ATTR_I CI_GRPC_COBJECTIVE_BI 0GPC_COBJ_ATTR

    Control GRFN_S_CN_ATTR_I CI_GRPC_CONTROL GRFN_S_BI_CN_ATTR_I CI_GRPC_CONTROL_BI 0GPC_CN_ATTR

    Entity Level Control GRFN_S_EC_ATTR_I CI_GRPC_ECONTROL GRFN_S_BI_EC_ATTR_I CI_GRPC_ECONTROL_BI 0GPC_EC_ATTR

    Process GRFN_S_PR_ATTR_I CI_GRPC_PROCESS GRFN_S_BI_PR_ATTR_I CI_GRPC_PROCESS_BI 0GPC_PR_ATTR

    Signoff GRFN_S_SO_ATTR_I CI_GRPC_SIGNOFF GRFN_S_BI_SO_ATTR_I CI_GRPC_SIGNOFF_BI 0GPC_SIGNOFF

    Sub Process GRFN_S_SP_ATTR_I CI_GRPC_SUBPROCESS GRFN_S_BI_SP_ATTR_I CI_GRPC_SUBPROCESS_BI 0GPC_SP_ATTR

    Test step GRFN_S_V0_ATTR_I CI_GRPC_V0TS GRFN_S_BI_V0_ATTR_I CI_GRPC_V0TS_BI 0GPC_V0_ATTR

    Activity GRFN_S_AC_ATTR_I CI_GRRM_ACTIVITY GRFN_S_BI_AC_ATTR_I CI_GRRM_ACTIVITY_BI 0GRM_AC_ATTR

    Opportunity Risk GRFN_S_OR_ATTR_I CI_GRRM_OPPORTUNITY GRFN_S_BI_OR_ATTR_I CI_GRRM_OPPORTUNITY_BI 0GRM_OR_ATTR

    Risk GRFN_S_RS_ATTR_I CI_GRRM_RISK GRFN_S_BI_RS_ATTR_I CI_GRRM_RISK_BI 0GFN_RS_ATTR

    *Note that if you are upgrading from PC/RM 2.x to 3.0, the Business Content is not

    upgradeable so you need to first upgrade the PC/RM system then install the content and

  • add any previously existing customizations/extensions to the 3.0 content. Extensions to

    equivalent DataSources (since there are all new DataSources in the 3.0 content) would be

    done via the CIs mentioned above (if available) and all other customizations (such as

    transformations and InfoProvider modifications) would need to be mapped to their 3.0

    equivalent and manually implemented.

    How do I retrieve comments / long text from the PC/RM system?

    There is a limitation in BI of a maximum length of 60 characters per field. Therefore, descriptions or comment fields from the source system longer than 60 characters cannot be seamlessly stored in BI. In the PC/RM 3.0 business content some basic support has been provided to enable access to some of the fields exceeding 60 characters in the PC/RM system. At the technical level this data from the PC/RM source system is provided using the following objects:

    o Data Source 0GFN_LONG_TEXT_ID Informs BI if language dependent long text for a particular entity exists or not in the source system

    o Data Source 0GFN_LONG_TEXT_NO_LANGU - Informs BI if language independent

    long text for a particular entity exists or not in the source system (typically for comments)

    o RFC FM GRFN_BI_GET_LONG_TEXT RFC to retrieve the actual source system

    long texts (for the given entity, id, timeframe, language)

    o Web Dynpro application GRFN_BI_LONG_TEXT The application which displays the actual source system long texts to the query user. The parameters to this application are ENTITY_ID, FIELDNAME, KEY_DATE, LANGU, OBJECT_KEY, TF_FREQ.

    o Web Dynpro application GRPC_ASSESSMENT The application which displays the case details (including long texts) of the entity. The parameters to this application are ACTIVE_TAB, GUID, OBJECT_ID, READ_ONLY, REGULATION_ID, SHARED_CONTROL, WDACCESSIBILITY, WDTABLENAVIGATION, WORKITEM, WORKITEM_NAVIGATE.

    the report in the PC/RM system via the specified Web Dynpro applications and parameters:

    Sender Queries Receiver Rptng Objects WD App (parameters)

    GRFN_BI_LONG_TEXT 0GRM3MP01_Q0007 Risk Long Text Desc ENTITY_ID=RISK, FIELDNAME=RS_D, OBJECT_KEY=Risk,

    KEY_DATE=Timeframe End Date,

  • TF_FREQ=Timeframe Frequency

    0GRM3MP02_Q0001 Response Long Text Desc ENTITY_ID=RESPONSE, FIELDNAME=RP_DESC OBJECT_KEY=Response,

    KEY_DATE=Timeframe End Date,

    TF_FREQ=Timeframe Frequency 0GPC3MP01_Q0003 Detailed Control Desc ENTITY_ID=CONTROL, FIELDNAME=CN_D,

    KEY_DATE=Timeframe End Date,

    TF_FREQ=Timeframe Frequency, OBJECT_KEY=Control

    0GPC3MP13_Q0002 CAPA Root Cause ENTITY_ID=G_CP,

    FIELDNAME=CP_ROOT_CAUSE KEY_DATE=Timeframe End Date,

    TF_FREQ=Timeframe Frequency,

    OBJECT_KEY=CAPA Plan

    GRPC_ASSESSMENT

    0GPC3MP01_Q0003 Evaluation Case Details GUID=GRC CASE Id, READONLY=X 0GPC3MP02_Q0002 Evaluation Case Details GUID=GRC CASE Id, READONLY=X

    0GPC3MP02_Q0003 Evaluation Case Details GUID=GRC CASE Id, READONLY=X

    0GPC3MP02_Q0004 Evaluation Case Details GUID=GRC CASE Id, READONLY=X 0GPC3MP02_Q0006 Evaluation Case Details GUID=GRC CASE Id, READONLY=X

    0GPC3MP02_Q0007 Evaluation Case Details GUID=GRC CASE Id, READONLY=X

    0GPC3MP02_Q0008 Evaluation Case Details GUID=GRC CASE Id, READONLY=X 0GPC3MP02_Q0010 Evaluation Case Details GUID=GRC CASE Id, READONLY=X

    0GPC3MP02_Q0011 Evaluation Case Details GUID=GRC CASE Id, READONLY=X

    0GPC3MP02_Q0012 Evaluation Case Details GUID=GRC CASE Id, READONLY=X 0GPC3MP02_Q0013 Evaluation Case Details GUID=GRC CASE Id, READONLY=X

    0GPC3MP02_Q0014 Evaluation Case Details GUID=GRC CASE Id, READONLY=X 0GPC3MP02_Q0015 Evaluation Case Details GUID=GRC CASE Id, READONLY=X

    0GPC3MP03_Q0002 Issue Details GUID=Issue

    0GPC3MP10_Q0002 Evaluation Case Details GUID=GRC CASE Id, READONLY=X 0GPC3MP11_Q0002 Evaluation Case Details GUID=GRC CASE Id, READONLY=X

    0GPC3MP13_Q0002 CAPA Details GUID=CAPA Plan

    0GPC3MP14_Q0002 Remediation Details GUID=Remediation Plan

    Section G, step 3 covers the details of maintaining the RSBBS. The only required modifications are to the server and client specified in the delivered Receiver Object URLs. Some of the parameters mentioned above are sent to the Web Dynpro application using the URL and others are sent using the Assignment Details of the RSBBS entry. But these settings are delivered and should not be modified.

    How do I synchronize the InfoProvider timeframe selection with the key date for

    time dependent master data attributes in a query?

    Several of the business content queries have been delivered with variables created to

    handle this situation (0GRM3MP03_Q0001, for example). The challenge arises from the

    Timeframe being defined at a higher level of granularity than 0CALDAY (which is used

    to determine the key date of the query) so the Timeframe end date needs to be derived

    and used as the query key date for the purpose of time dependent attribute determination.

    The delivered variables, when properly used, will automatically calculate the proper key

    date based on the timeframe entered by the user for the query.

    The solution requires using 0GFN_TF variable 0P_GFN_TF (single value required

    entry) or 0I_GFN_TF (interval required entry) in conjunction with SAP Exit variable

    0P_GFN_TF_PE. The user enters the desired timeframe (or timeframe interval) in order

  • to select the proper records from the InfoProvider. Then the exit variable uses this value

    to determine the proper timeframe end date. The variable 0P_GFN_TF_PE is entered in

    the Key Date field of the query properties screen and is thus used to determine time

    dependent attribute values in the query output.

    Custom queries need to use these variables if time dependent attributes are included,

    otherwise the default value (the current system date) will be used as the querys key date

    and this could result in inconsistent results in the output.