01_Intro_SAP BO DATA Integrator (1)
01_Intro_SAP BO DATA Integrator (1)
Data services is integrated with SAP BI/SAP R3/SAP Applications and non SAP Ware house.
Purpose:- It do ETL via batch Job and online method thru bulk and delta load processing of both
structured and unstructured data to generate a Ware House (sap and Non-sap)
Data Services is the combination of Data Integrator and Data Quality. Previously these are separate tools
like Data Integrator which is used to do the ETL part and Data Quality to do the data profiling and Data
Cleansing. Now with Data Services both DI and DQ are combined in to once interface so that it provides
the complete solution (data integration and Quality) under one platform.
This even combines the separate job servers & Repositories of DI and DI in to one.
Data Federator: - The output of the data federator is the virtual data. Federator provides the data as
input to the data services and using federator we can project data from multiple sources as a single
source.
SQL -- DS -- SQL
Flat File -- DS -- BI
R/3 -- DS -- BI
R/3 -- DS -- SQL
SQL -- DS -- BI
We can move the data from any source to any target DB using Data Services.
Data Services is a utility to do ETL process, it is not a ware house, so it doesn’t stage any amount of data
in it.
Data Services can create ETL process and can create a ware house (SAP / Non-Sap) .
1) Migration
2) Ware house or DB building
3) Data Quality
Data Profiling: - Pre processing of data before the ETL to check the health of the data. By profiling we
check the health of the data if it’s good or bad.
It supports Export Execution Command to load the data in to the ware house via batch mode process
It can do real time data load/ Full data load/ Incremental Data load
Management Console
Job Server
Repository Manager
Profiler
Meta Data Integrator
Local
Central
Profiler
No concept of Process chains/ DTP/ Info packages if you use the data services to load the data.
Designer
It includes all the artifacts of the project (Work Flow, Data Flow, Data Store, Tables)
It allows us to create users and user groups and assign the users to the user groups with privileges
We can execute the jobs from any geographic location as this is a web based tool
XML inputs can be loaded to the Ware house using the Access server
Repository Manager
Job Server
This is the server which is responsible to execute the jobs. Without assigning the local / central
repository we cannot execute the job.
Projects:-
Project is a folder where you store all the related jobs at once place. We can call it as a Folder to
organize jobs.
Jobs:-
Jobs are the executable part of the Data Services. This job is present under the project.
Batch Job
Online jobs
Work Flows:-
This work flow acts a folder to contain the related Data Flows. This Work Flows are re-usable
Conditionals:-
Conditional contains Work Flows or data flows and these are controlled by script whether to trigger or
not.
Scripts:-
Scripts are set of codes used to define or initialize the global variables, Control the flow of conditionals
or control the flow of execution, to print some statements at the runtime and also to assign specific
default values to the variables.
Data Flow:-
It is the place held to import the data from the data base/ sap to data services local repository
It is the collection of dimensions and fact tables to create the data ware house.
Transformations:-
These are the query transformations that are used to carry out the ETL process.
File Format :-
Variables:-
We can create and use the local and global variables and use them in the project. The variables starts
with “$” Symbol.
Functions:-
We have numerous inbuilt functions like (String, math, lookup , enrich and so on)
Template Table:-
These are the temporary tables that are used to hold the intermediate data or the final data.
Data Store:-
These data stores act a port from which you can define the connections to the source or the target
systems. You can create multiple configurations in one data store to connect this to the different
systems
ATL :-
ATL files are like the BIAR files. This is named after a company. ATL doesn’t hold any full form like BIAR.
The Project/ Job/ Work Flow/ Data Flow/ Tables can be exported to ATL so that they can be moved
between Dev Qual and from Qual Prod.
Similarly you can also import the Project/ Job/ Work Flow/ Data Flow/ Tables which are exported to ATL,
back in to the data services