Case Study: Answers
ANSWERS
(a) (i) Dashboard for vendors: (One Module in a data visualization tool for investigations for drilling down information about specific vendors and implementing partners without having to log in to SAP.
ANSWER: As the initial phase for the analytics component for the audit section will focus on ‘Procurement of goods and services’ and vendor analysis being a part of it, assumes a very important role. Understanding of various vendors that an organization outsources its products and services and the amount each vendor charges for the same products and services is very important as it helps in better selection of vendors based on different categories like prices, discounts, commodity etc.
When a visual has a hierarchy, we can drill down to reveal additional details. A visual for vendors and IP's that looks at count by a hierarchy made up of commodities/services supplied/discounts/location.
- Drill through filters: With drill through filters in Power BI Desktop, we can create a page in our reports that focuses on specific entities such as suppliers, vendors, IP’s
- Page-level filters: These are used to filter charts that are present in individual page.
- Report-level filters: They are used to simultaneously filter charts that are present in all pages of a report.
Connecting
SAP DW with Power BI: Implementation of 2.0 SAP
Connector requires the SAP .NET Connector 3.0. Download the SAP .NET Connector
3.0 from SAP. While installation, in optional setup steps, make sure we select
Install assemblies to GAC. In case of other versions respective tech notes can
be referred
To
create a new connection: Select Get Data. Select either
SAP Business Warehouse Application Server or SAP Business Warehouse Message
Server, and then connect. In the new connection dialog, select the
implementation. Selecting 2.0 for Implementation, enables Execution mode, Batch
size and Enable characteristic structures. Finally select OK.
Accessing
dashboards without log in: The option “Publish to Web” is
available under File Menu when we open the report in Power BI Service. An embed
code is generated, which can be directly inserted into your application. If
dashboard is no more required, then we can delete existing embed codes through
the Power BI Service (Settings -> Manage Embed Codes).
SQL trigger can be used for
the same. Triggers always fire after the update or insert. If the criteria don’t
match, then it rolls back.
CREATE OR REPLACE TRIGGER trg_dup
BEFORE INSERT ON Vendors
FOR EACH ROW
DECLARE counter integer;
BEGIN
SELECT * INTO counter FROM
(SELECT COUNT(rownum) FROM Vendors a
WHERE a.VENDOR_ID = :new.VENDOR_ID);
IF counter = 1 THEN
RAISE_APPLICATION_ERROR( -20003, 'Error: duplicate data.');
END IF;
END;
/
Best solution is to ensure data integrity. We ensure database integrity using constraint. In this case, a unique constraint. That is defined using a unique constraint or index:
ALTER TABLE VENDORS ADD CONSTRAINT UNQ_VENDORS_V_P_A ON (VENDOR_ID, PAYMENT_ID, AMOUNT);
(a) (iii) Office strategy for medium to long
term (Develop the overall Office
strategy for 2021 and beyond, based on lessons learned from the first phase)
1. Requirement gathering and defining of objectives: First and foremost, objective/s needs to be defined based on discussion with all the stakeholders. It will help in planning required data sources, human resources with specific skill sets, any required automations, high risk areas, required reports and their consumers, standard/non-standard operations, auditing tasks and how often they should run etc.
2.
Technology environment understanding: Early and regular interactions with IT
team will help in establishing a robust technical architecture to suit our
analytical requirements. Discussion of our established objectives will help in
clarifying our data requirements and will reduce delays due to any established
protocols and security concerns.
3. Audit analytics architecture: Setting up of centralized and secure audit environment, planning hardware and storage requirements, setting up of audit data repository and data dictionary
4. Data analytics strategy:
a. Data accessibility: Creation of
server profiles and database profiles for the audit team to gain secure and
authorized access to the data and the hardware platform(s) where the processing
will take place.
b. Data analysis protocols: Maintain the
security and confidentiality of source data and analytical results, avoid data
duplication, no local data storage read data directly from databases, data
extraction on dedicated audit analytics server
c. Business process understanding: Creation
of business process flow chart to identify data sources, possible area of
concerns (processes involving different applications or systems)
d. Resource alignment: Resources with
specific skills like AI and ML, statistical programmers, data analysts, IT
experts, audit or functional experts would be identified and assigned to
specific task. Multi skilled resources can act as bridge between different
functions.
5. Software update and rollouts: Proper rollout strategy need to be set up for any software/ hardware updates and rollout to avoid lengthy delay and sudden break downs. Such strategy will involve coordination of resources from across IT and audit departments.
6.Training and documentation: Proper documentation of defined processes, continuous standardization of reporting formats, regular training programs for audit team about analytics advantages, awareness about recent developments in audit analytics
The
work of audit and investigation requires coordinating with multiple
departments. Finance – for getting the actual numbers, procurement – for
getting the actual amounts, program performance and reporting – for data
related to IP’s and IT – for overall analytics environment functioning. Contacting
country and region offices would also be required to get on ground reports
about specific programs run by the offices.
An ideal case would have been a data warehouse seamlessly integrated with different departments based on laid down security policies and business processes. Data from each of the offices would be constantly pushed to the data warehouse using ETL process and accessible based on user roles and responsibilities. This would have greatly reduced the chances of physical meetings and coordination, thereby saving lot of time.
(c) What key challenges do you
foresee in implementing the plans?
(d) How will you respond to these
challenges?
As
the above questions are inter-related, I am answering them both in detail in
below section:
All the departments in an organization work to support a common goal. While departments operate separately, they are also interdependent. Below are some of the expected challenges and ways to respond:
2. Data silos: Due to lack of seamless integration of inter-departmental data, data silos have been created, limiting the view of data, threatening data integrity, wasting resources and discouraging collaborative work. That’s the reason the existing SAP DW has not yet matured. This has led to issues in collecting even basic holistic data, creation of ‘blind spots’ leaving red flags and eventually leading to potential leads being missed. Response/Solution: Reassessing the underlying data warehouse architecture would be the starting point. Centralized/component-based architecture, Star /snowflake schema, role of data marts (dependent and independent) etc. needs to be adjusted keeping in mind current status and future prospects.
3. Data infrastructure performance: Legacy data warehouses (whether on-premises or cloud based) haven’t kept up with all the data requests users have. Managing and preventing these issues can take up a lot of IT time, and the problems often compound over time. Hitting capacity limits slows down users and ties up database administrators too. Response/Solution: From a data infrastructure perspective, separating the compute and storage layers is essential to achieve business agility. When a data warehouse can handle your scalability needs and self-manage performance, that’s when you can really start being proactive.
4. Data preparation: As mentioned, different systems are at different maturity levels which has led to inconsistency in data and bound to encounters errors. Inconsistent data, duplicates, and missing data all result in data quality challenges which can lead to faulty reporting and analytics necessary for optimal decision-making. Data preparation challenges can be further divided into below sub-challenges:
- Data identification: Identifying
most appropriate sources from which data element can be acquired and analyzing
the formats, potential duplication, potential incremental value and source
reliability
- Data acquisition: Physical collection
of data from heterogenous sources
- Data cleansing: Elimination of
duplicates, erroneous values correction, replacing missing values and any other
transformation
- Data transformation: Enrichment of
data into more business descriptive content
Response/Solution: Auditing of data sources is required to see how many different databases/applications are present. Existing ETL design needs to be reviewed. Need to check ETL operations based on schema type, SCD’s (slowly changing dimensions), type of transformations required.
5.
Understanding business data: When building a data warehouse,
analytics and reporting will have to be taken into design considerations. In
order to do this, the business user will need to know exactly what analysis
will be performed. Envisioning these reports will be difficult for someone who
has no experience in business-intelligence data analysis, and who is unaware of
its capabilities. Response/Solution: Better understanding of business
processes can be of great help. Regular interaction with subject matter experts
during the planning stage can help in pin-pointing the exact analytics and
reporting required.
6. Lack of analytics tool: There is lack of dedicated analytics tool. An analytics tool becomes a necessity keeping in mind the goals of revealing outliers, risk profiling, pattern analysis, fraud risk assessment, testing of red flags and continuous auditing of key controls. Above mentioned in-depth analysis involve large-scale and resource intensive computations leading to new insights being discovered. Artificial intelligence (AI) and machine learning (ML) are already changing the face of every industry, where forecasting, segmentation, outlier analysis, natural language processing (NLP) are helping the businesses making better decisions. Response/Solution: For analytics related tasks option that are easy to integrate with SAP DW can be checked. Power BI already present suffices the need for visualizations. For predictive modelling python (with its packages – scikit-learn, numpy, scipy) and R can be of great help. SAS E miner, IBM SPSS modeler can also be looked as predictive modelling software.
Comments
Post a Comment