High-Level Design architecture (HLD)
Link.
Low-Level Design architecture (LLD)
Link.
Architecture Data Flow
Here is a suggested template for Data Model + Data Mapping :
DA&DA - Domain Mapping INDUSTRIAL
DataPrep Flow
Schema showing the different STEPS of the application flow - with the data involved at each step
According to Data Ocean Blueprint and macro architecture
BW --> GCP Data Ocean INDUSTRIAL (prj-data-dm-industrial-[env]) --> GCP Product MECANO (prj-data-pmo-maintenance-dash-[env]) --> Output Tableau
A detailed description of the Talend flows can be found here.
Conceptual Data Model
The data model is accessible in GenMyModel.
Steps descriptions
Data lineage can be found in the Google Sheet here below
Note: Table on project prj-data-dm-industrial-[env]
DataSource
Description
ECC (PF1/WP1)
BW is the application where from where we collect information. The data is extracted using specific BW queries with detail in the link
Tools
Talend to collect the data from the source and store on GCP/Google Big Query.
Access rights
You can use Xtract server (ACEW1DTLNDENG02) to manually extract and monitor.
Only DA&AI DataEng's and Data Architects can access data on GCP/Google Big Query.
Source
Location
Xtract server (ACEW1DTLNDENG02)
Format
BW queries extracted via CSV file
Destination
Location
Extracted data will be stored on GCP/Cloud storage and Google Big Query.
Format
The format of the data saved in the databank
Sizing
Expected data volume for :
- incremental process from source to staging (as of Jan 8th 2024)
SELECT * FROM STG.log_files WHERE meta_run_id = '874d0cf9-94c4-4739-90ef-2d4d7f416d4c'
15 tables are updated and roughly 1.8mln rows are inserted globally (this number may slightly change). No table is supposed to be empty.
Assessment
Check the log tables in GCP on table log_tables and run_jobs to check that there is no error loading from source to staging/ods
Check the surrogate key must be unique in the data mart layer
Scheduling
Is there an automatic schedule ? Yes
At what frequency ? We have two Talend plans in the TMC:
- PL_MECANO_DASH : Every monday at 4am (Paris time)
- PL_MECANO_DAILY_LOAD : Every tuesday, wednesday, thursday, friday at 4 am (Paris time)
What is the trigger ? TMC
Timing
The average time expected for :
- PL_MECANO_DASH : around 40 minutes
- PL_MECANO_DAILY_LOAD : less than 5 minutes
Criticality
Medium
Logging
Table table log_tables, run_jobs, log_files, and reject_files in `prj-data-dm-industrial-[environment].STG.[table]`
Link to the Mapping document :
Mecano KT Call - Questions to Matteo/Sebastien - 15Jan2024 :
Requesting more info/documentation on below:
Step 0 :
Source Details/BW ServerName:
Source repository/Connection Details:
Step1 :
1. List of total Xtract jobs and their frequency? - Please update the wiki.
2. Naming convention to identify 'Xtract jobs' related to Mecano?
3. Xtract jobs and (CSV) files location?
4. R001_MECANO 0.1 Job and Date Calculation script .
5. Google Cloud Storage Bucket Name and location?
cs-ew1-prj-data-dm-industrial-prod-staging
Understand more about BW Query: - (Sai/Prabakar)
TALEND_QVMECANO_BW_QRY_MVPMNO04_0001
TALEND_QVMECANO_BW_QRY_MVPMNO04_0002
TALEND_QVMECANO_BW_QRY_MVPMOP04_0001
TALEND_QVMECANO_BW_QRY_MVPMOP04_0002
prj-data-dm-industrial :
Source → GCP Buckets --> prj-data-dm-industrial.STG. --> prj-data-dm-industrial.ODS --> prj-data-dm-industrial.DM . --> prj-data-dm-industrial. DS_MaintenanceDashboard |
||
\/
Reporting will be done from PRJ-DATA-MAINTENANCE-DASH-DEV project on DPL dataset (prj-data-pmo-dash-dev.DPL).
Tableau --> prj-data-maintenance-dash.DPL --> prj-data-maintenance-dash.DataOcean --> prj-data-maintenance-dash.DM → prj-data-dm-industrial.DS_MaintenanceDashboard


