You are viewing an old version of this page. View the current version.

Compare with Current View Page History

« Previous Version 20 Next »

Status

  Update in progress

Owner
Stakeholders

Purpose

The purpose of this document is to define the conversion approach to create Business Partners - General in S/4 HANA.

In SAP ECC, customer and vendor master data are maintained separately as distinct entities. Customers are managed through Customer Master Data, while vendors are handled via Vendor Master Data. These records store essential details such as company name, address, payment terms, and tax information.
In SAP S/4HANA, the Business Partner (BP) concept replaces the traditional customer and vendor master data approach. The BP model integrates both customer and vendor roles into a single entity, simplifying data management and ensuring consistency across different business functions


Conversion Scope

The scope of this document covers the approach for converting active Customer Master Data General and Vendor Master General from Legacy Source Systems into S/4HANA Business Partner (BP) General (Role 000000) Master Data Design Standard. 


Customer Master Data - General Information

The data from legacy system includes:

  1. Customer with AR Balance under the company codes within S4 Hana implementation scope.
  2. or (Customer doesn't have central deletion indicator AND Customer has sales transaction within the sales organizations in scope), e.g., the customer is used as any partner function in the sales document, such as sales order, delivery or billing.
  3. or (Customer doesn't have central deletion indicator and customer has financial posting under the company codes within S4 Hana implementation scope)
  4. or Customer is part of Customer Hierarchy Higher Node, and the sales area data is part of the sales areas within migration scope
  5. or there is customer consignment stock and the plant for the consignment stock is within S4 Hana implementation scope. 
  6. or (Customer is created within 2 year and there is no central block) and it has sales view within sales organization in scope
  7. or (Customer is created within 2 year and there is no central block) and it has company code view within company code in scope
  8. or 

The data from legacy system excludes:

  1. Customer has central deletion indicator and without AR/AP Balance under the company codes within S4 Hana implementation scope.
  2. Customer is used exclusively by entities not in scope, such as Oil & Gas and Aroma.



List of source systems and approximate number of records 
SourceScope

Source Approx No. of Records

Target SystemTarget Approx

No. of Records

WP2Customer Master Data General Information
S4 Hana ROW
PF2Customer Master Data General Information
S4 Hana ROW
WP2Customer Master Data General Information
S4 Hana China
PF2Customer Master Data General Information
S4 Hana China
WP2Customer Master Data General Information
S4 Hana CUI
PF2Customer Master Data General Information
S4 Hana CUI

Vendor Master Data - General Information

The data from legacy system includes:

The data from legacy system excludes:

  1.  

 List of source systems and approximate number of records 

SourceScope

Source Approx No. of Records

Target SystemTarget Approx

No. of Records
































Additional Information

Multi-language Requirement

The customer and vendor general data may contain international address. Therefore, the conversion will also need to support the multi-language address.  Below languages (International versions) are supported. 


International VersionDescription
CSimplified Chinese
RCyrillic
KKanji (Japanese)
AArabic
3Korean
TThai
HHangul

Document Management

It is possible the customer has attachment in the legacy system. The migration of attachment will be captured in conversion spec CNV-3004 - Attachment for customer master data. 

Legal Requirement

CMMC 2.0 is a mandatory DoD cybersecurity certification for contractors handling Controlled Unclassified Information (CUI) and Federal Contract Information (FCI). CUI includes sensitive technical data (e.g., design specs, system info) related to U.S. military and space applications. The Composites Business handles CUI and is therefore within CMMC scope. Without certification, the business risks disqualification from existing and future DoD programs.
It is mandatory to implement CMMC-compliant systems and processes to for all the organizations that are dealing with CUI.

Therefore, there will be one SAP instance specifically for CUI related entities. As Synithi is not CUI certified partner, the migration for CUI related entities will be covered by US based data consultant using separate tools.

Special Requirements

If the data conversion involves third-party systems or external data sources, such as Icertis, describe any additional requirements related to data mapping, transformation logic, validation rules or security measures that must be followed.

Due to compliance requirement, there will be one SAP instance for Rest of the World, one for China and one for CUI.

  1. For entities in China, the data will be loaded into SAP China instance while the entire migration process will remain the same as rest of the world.
  2. For entities which will reside in CUI, the migration will be handled by US based data consultant. 


Customer Master Data - General Information

 - To identify the record is for SAP China Instance, it will use below logic. 

  1. The customer has sales area data or company code data in below entities. 
  2. If the customer is used in both China entities and ROW, then the general data needs to be created in both SAP China and ROW instances.
SAP China Instance Specific Company CodesSAP China Instance Specific Sales Organization





 - To identify the record is for SAP CUI Instance, it will use below logic. 

  1. The customer has sales area data or company code data in below entities. 
  2. If the customer is used in both CUI entities and ROW, then the general data needs to be created in both SAP CUI and ROW instances.
SAP CUI Instance Specific Company CodesSAP CUI Instance Specific Sales Organization




  

In the meantime, for WP2/PF2 customer master general data, it is possible they are both coming from the same MDM PRS system, therefore, a de-duplication or reconciliation needs to be performed based on below logic. 

  1. In PF2, it will have the relationship that KNA1-ZZR_KUNNR_RCS (RCS Customer code) = WP2, KNA1-KUNNR, then it refers to the same customer. 
  2. In WP2, it will have the relationship that KNA1-ZZR_KUNNR_PRS (PRS Customer code)= PF2, KNA1-KUNNR, then it refers to the same customer. 


 Vendor Master Data - General Information

*please indicate how S2P will identify the SAP China Instances.



Target Design

The technical design of the target for this conversion approach. 

TableFieldData ElementField DescriptionData TypeLengthRequirement






















Data Cleansing

All data cleansing should take place in the data source system as defined in this document, unless system limitations prevent it.

Customer Master Data - General Information

IDCriticalityError Message/Report DescriptionRuleOutputSource System
3007-001
Missing Postal code in the general data

PF2/WP2
3007-002
Missing Street in the general data

PF2/WP2
3007-003
Missing region in the general data

PF2/WP2
3007-004
Review the international version address maintained for the customer

PF2/WP2
3007-005
Review the customer with obsolete region code

PF2/WP2
3007-006
Identify duplicate BP
1. customer vs customer


PF2/WP2
3007-007
Identify duplicate BP

2. vendor vs customer if applicable


PF2/WP2
3007-008
Block customer general data without any usage for more than 2 years

For customer without general block and customer is not used in any sales or finance transaction for more than 2 years under company company code


PF2/WP2


Vendor Master Data - General Information


IDCriticalityError Message/Report DescriptionRuleOutputSource System








Conversion Process 

The high-level process is represented by the diagram below:

The ETL (Extract, Transform, Load) process is a structured approach to data migration and management, ensuring high-quality data is seamlessly transferred across systems. Here’s a breakdown of its key components:
1. Extraction
The process begins with extracting metadata and raw data from source systems, such as Syensqo ECC system (i.e., WP2/PF2) periodically. The extracted data is then staged for transformation.


2. Transformation
Once extracted, the data undergoes cleansing, consolidation, and governance. This step ensures data integrity, consistency, and compliance with business rules. The transformation process includes:
- Data validation to remove inconsistencies.
- Standardization to align formats across datasets.
- Business rule application to refine data for operational use.


3. Loading
The transformed data is then loaded into the target S4 Hana system. 

Data Privacy and Sensitivity

N/A


Extraction

Extract data from a source into Syniti Migrate. There are 2 possibilities:

  1. The data exists. Syniti Migrate connects to the source and loads the data into Syniti Migrate. There are 3 methods:
    1. Perform full data extraction from relevant tables in the source system(s).
    2. Perform extraction through the application layer.
    3. Only if Syniti Migrate cannot connect to the source, data is loaded to the repository from the provided source system extract/report.
  2. The data does not exist (or cannot be converted from its current state).  The data is manually collected by the business directly in Syniti Migrate. This is to be conducted using DCT (Data Collection Template) in Syniti Migrate

The agreed Relevancy criteria is applied to the extracted records to identify the records that are applicable for the Target loads

Extraction Run Sheet

Req #Requirement DescriptionTeam Responsible
Extraction Scope Definition- Identify the source systems and databases involved.
- Define the data objects (tables, fields, records) to be extracted.
- Establish business rules for data selection.
Synithi
Extraction Methodology- Specify the extraction approach (full, incremental, or delta extraction).
- Determine the tools and technologies used.
- Define data filtering criteria to exclude irrelevant records.
Synithi
Extraction Execution Plan- Establish execution timelines and batch processing schedules.
- Assign responsibilities for extraction monitoring.
- Document dependencies on other migration tasks.
Synithi
Data Quality and Validation- Define error handling mechanisms for extraction failures.Synithi




Selection Screen

Selection Ref ScreenParameter NameSelection TypeRequirementValue to be entered/set
N/A



















Data Collection Template (DCT)

Target Ready Data Collection Template will be created for data with exception of some fields which require transformation as mentioned in the transformation rule.

DCT Rules


Customer Master Data - General Information

Field NameField DescriptionRule












 Vendor Master Data - General Information

Field NameField DescriptionRule













Extraction Dependencies

Item #Step DescriptionTeam Responsible
1

Source System Availability

  • Ensure that the source database or application is accessible.
  • Confirm that necessary credentials and permissions are granted
Syensqo IT
2

Data Structure

  • Identify relationships between tables, views, and stored procedures.
Synithi
3

Referential Integrity

  • Ensure dependent records are extracted together.
Synithi
4

Extraction Methodology

  • Define whether extraction is full, incremental, or delta-based.
  • Establish batch processing schedules for large datasets.
Synithi
5

Performance and Scalability Considerations

  • Optimize extraction queries to prevent system overload.
  • Ensure network bandwidth supports data transfer volumes.
Synithi
6

Security and Compliance

  • Adhere to regulatory standards for sensitive information if applicable
Synithi


Transformation

The Target fields are mapped to the applicable Legacy field that will be its source, this is a 3-way activity involving the Business, Functional team and Data team.  This identifies the transformation activity required to allow Syniti Migrate to make the data Target ready:

  1. Perform value mapping and data transformation rules.
    1. Legacy values are mapped to the to-be values (this could include a default value)
    2. Values are transformed according to the rules defined in Syniti Migrate
  2. Prepare target-ready data in the structure and format that is required for loading via prescribed Load Tool.  This step also produces the load data ready for business to perform Pre-load Data Validation

Transformation Run Sheet

Item #Step DescriptionTeam Responsible
1Transformation Scope Definition
- Identify the source and target data structures.
- Define business rules for data standardization.
- Establish data cleansing requirements to remove inconsistencies.
Data Team
2Data Mapping and Standardization
- Align source fields with target fields.
- Ensure unit consistency (e.g., currency, measurement units)
Data Team
3Business Rule Application
- Implement data enrichment/collection if applicable
- Apply conditional transformations based on predefined logic/business rules
Data Team
4Transformation Execution Plan
- Define batch processing schedules.
- Assign responsibilities for monitoring execution.
- Establish error-handling mechanisms
Synithi


Transformation Rules

Rule #Source systemSource TableSource FieldSource DescriptionTarget SystemTarget TableTarget FieldTarget DescriptionTransformation Logic









































Transformation Mapping

Mapping Table NameMapping Table Description 
MAP_BU_GROUPBP Grouping Mapping Table
MAP_REGIONCountry/Region Code Mapping Table
MAP_BPKINDBP Type Mapping Table


Transformation Dependencies

List the steps that need to occur before transformation can commence
Item #Step DescriptionTeam Responsible
1Source Data Integrity
- Ensure extracted data is complete, accurate, and consistent.
- Validate that data types and formats align with transformation requirements.
Synithi
2Referential Integrity
- Ensure dependent records are transformed together or in advance
Synithi
3Transformation Logic and Mapping
- Define data mapping rules between source and target schemas.
Data Team
4Performance and Scalability Considerations
- Optimize transformation processes for large datasets.
- Ensure system resources can handle transformation workloads
Synithi
5Logging and Error Handling
- Maintain detailed logs of transformation activities.
- Define error-handling procedures for failed transformations
Synithi


Pre-Load Validation

Project Team

The following pre-load validations will be performed by the Project Team.

Completeness

TaskAction
Compare Data Counts
  1. Verify row counts between source and target databases.
  2. Identify missing or duplicated records.


Validate the mandatory fieldsValidate there is value for all the mandatory fields
Validate Primary Keys and Unique Constraints
  1. Check for duplicate or missing primary key values, i.e., if there is same BP number.
  2. Ensure unique constraints are maintained.


Test Referential IntegrityConfirm dependent records exist in related tables

Accuracy

TaskAction
Validate the transformationValidate the fields which require transformation have the value after transformation instead of the original field value
Check Data Consistency
  1. Compare field values across systems
  2. Validate data formats and structures





Business

 The following pre-load validations will be performed by the business. 

Completeness

TaskAction





Accuracy

TaskAction





Load

The load process includes:

  1. Execute the automated data load into target system using load tool or product the load file if the load must be done manually
  2. Once the data is loaded to the target system, it will be extracted and prepared for Post Load Data Validation

Load Run Sheet

Item #Step DescriptionTeam Responsible
1Load Scope Definition
- Identify the target system and database structure.
- Define data objects (tables, fields, records) to be loaded.
- Establish business rules for data validation.
Data team
2Load Methodology
- Specify the loading tools and technologies (Migration Cockpit, LSMW, custom loading program).
Synithi
3Data Quality and Validation
- Ensure data integrity checks (null values, duplicates, format validation).
- Perform pre-load validations to verify completeness.
- Define error handling mechanisms for load failures
Synithi
4Load Execution Plan
- Establish execution timelines and batch processing schedules.
- Assign responsibilities for monitoring execution.
- Document dependencies on other migration tasks
Synithi
5Logging and Reporting
- Maintain detailed logs of loading activities.
- Generate summary reports on loaded data volume and quality.
- Define escalation procedures for errors
Synithi


Load Phase and Dependencies

The Business Partner General will be loaded in the pre-cutover period.

Before loading, it will have dependency on the configuration. The configuration needs to be transported into the respective system first, including the manual configuration such as the BP number range set up.

Configuration

Item #Configuration Item
1BP Grouping
2Customer/Vendor Account Group
3International Version
4Tax Category
5BP Type
6BP Number Range/Customer/Vendor Number range

 Conversion Objects

Object #Preceding Object Conversion Approach
1083Bank Master




Error Handling

Error TypeError DescriptionAction Taken
Configuration / Data TransformationThe value XXX for field XXX doesn't exist
  1. Check the mapping/conversion is done properly in the loading file
  2. Validate the target value is configured/transported in the target system
  3. Reach out to function team to validate the configuration
ConfigurationThere is mandatory field XXX missing
  1. Validate MDS if the fields are set as mandatory
  2. Validate if there is value in the pre-loading file
  3. Validate if the configuration for the mandatory fields are done properly
ConfigurationThe BP grouping is External or Internal Number range
  1. Validate the number range set up if this is External or Internal number range


Post-Load Validation

Project Team

The following post-load validations will be performed by the Project Team.

Completeness

TaskAction
Perform Source-to-Target Comparisons
  1. Validate that migrated data matches source records.
  2. Check for discrepancies in numerical values, text fields, and timestamps

Accuracy

TaskAction
Execute Sample Queries and Reports
  1. Run queries to validate business logic.
  2. Generate reports to compare expected vs. actual results
Conduct Post-Migration ReconciliationGenerate reports comparing pre- and post-migration data.



Business

Post-load validation is a critical step in data migration, ensuring that transferred data is accurate, complete, and functional within the target system.

1. Ensuring Data Integrity
After migration, data must be consistent with its original structure. Post-load validation checks for missing records, incorrect mappings, and formatting errors to prevent discrepancies.
2. Business Continuity
Faulty data can disrupt operations, leading to financial losses and inefficiencies. Validating post-load data ensures that applications function as expected, preventing downtime.
3. Error Detection and Resolution
By validating data post-migration, businesses can detect anomalies early, reducing the cost and effort required for corrections

Completeness

TaskAction





Accuracy

TaskAction
Perform Manual TestingConduct manual spot-checks for additional assurance.





Key Assumptions

  • BP Master Data Standard is up to date as on the date of documenting this conversion approach and data load. 
  • BP General (Role 000000) is in scope
  • There will only be SAP instance, one for ROW, and one for China only



Change log

Version Published Changed By Comment
CURRENT (v. 20) Apr 27, 2026 08:28 CELEDONIO-ext, Arnold
v. 219 Apr 24, 2026 15:27 RUAN-ext, Eric *20260424 update remove the redundant relevancy rule
v. 218 Apr 21, 2026 13:36 CELEDONIO-ext, Arnold Update Vendor mapping rules
v. 217 Apr 20, 2026 14:01 RUAN-ext, Eric 20260420 update for but100 to remove source and add ADR table without selecting the person records
v. 216 Apr 06, 2026 10:14 CELEDONIO-ext, Arnold
v. 215 Apr 02, 2026 14:51 RUAN-ext, Eric
v. 214 Mar 31, 2026 16:16 RUAN-ext, Eric
v. 213 Mar 30, 2026 10:35 CELEDONIO-ext, Arnold
v. 212 Mar 30, 2026 10:23 CELEDONIO-ext, Arnold
v. 211 Mar 26, 2026 14:24 RUAN-ext, Eric

Go to Page History

Workflow history

Title Last Updated By Updated Status  
There are no pages at the moment.

  • No labels