SupplyChain / Client / update_data_integration_flow
update_data_integration_flow#
- SupplyChain.Client.update_data_integration_flow(**kwargs)#
- Enables you to programmatically update an existing data pipeline to ingest data from the source systems such as, Amazon S3 buckets, to a predefined Amazon Web Services Supply Chain dataset (product, inbound_order) or a temporary dataset along with the data transformation query provided with the API. - See also: AWS API Documentation - Request Syntax- response = client.update_data_integration_flow( instanceId='string', name='string', sources=[ { 'sourceType': 'S3'|'DATASET', 'sourceName': 'string', 's3Source': { 'bucketName': 'string', 'prefix': 'string', 'options': { 'fileType': 'CSV'|'PARQUET'|'JSON' } }, 'datasetSource': { 'datasetIdentifier': 'string', 'options': { 'loadType': 'INCREMENTAL'|'REPLACE', 'dedupeRecords': True|False } } }, ], transformation={ 'transformationType': 'SQL'|'NONE', 'sqlTransformation': { 'query': 'string' } }, target={ 'targetType': 'S3'|'DATASET', 's3Target': { 'bucketName': 'string', 'prefix': 'string', 'options': { 'fileType': 'CSV'|'PARQUET'|'JSON' } }, 'datasetTarget': { 'datasetIdentifier': 'string', 'options': { 'loadType': 'INCREMENTAL'|'REPLACE', 'dedupeRecords': True|False } } } ) - Parameters:
- instanceId (string) – - [REQUIRED] - The Amazon Web Services Supply Chain instance identifier. 
- name (string) – - [REQUIRED] - The name of the DataIntegrationFlow to be updated. 
- sources (list) – - The new source configurations for the DataIntegrationFlow. - (dict) – - The DataIntegrationFlow source parameters. - sourceType (string) – [REQUIRED] - The DataIntegrationFlow source type. 
- sourceName (string) – [REQUIRED] - The DataIntegrationFlow source name that can be used as table alias in SQL transformation query. 
- s3Source (dict) – - The S3 DataIntegrationFlow source. - bucketName (string) – [REQUIRED] - The bucketName of the S3 source objects. 
- prefix (string) – [REQUIRED] - The prefix of the S3 source objects. 
- options (dict) – - The other options of the S3 DataIntegrationFlow source. - fileType (string) – - The Amazon S3 file type in S3 options. 
 
 
- datasetSource (dict) – - The dataset DataIntegrationFlow source. - datasetIdentifier (string) – [REQUIRED] - The ARN of the dataset. 
- options (dict) – - The dataset DataIntegrationFlow source options. - loadType (string) – - The dataset data load type in dataset options. 
- dedupeRecords (boolean) – - The dataset load option to remove duplicates. 
 
 
 
 
- transformation (dict) – - The new transformation configurations for the DataIntegrationFlow. - transformationType (string) – [REQUIRED] - The DataIntegrationFlow transformation type. 
- sqlTransformation (dict) – - The SQL DataIntegrationFlow transformation configuration. - query (string) – [REQUIRED] - The transformation SQL query body based on SparkSQL. 
 
 
- target (dict) – - The new target configurations for the DataIntegrationFlow. - targetType (string) – [REQUIRED] - The DataIntegrationFlow target type. 
- s3Target (dict) – - The S3 DataIntegrationFlow target. - bucketName (string) – [REQUIRED] - The bucketName of the S3 target objects. 
- prefix (string) – [REQUIRED] - The prefix of the S3 target objects. 
- options (dict) – - The S3 DataIntegrationFlow target options. - fileType (string) – - The Amazon S3 file type in S3 options. 
 
 
- datasetTarget (dict) – - The dataset DataIntegrationFlow target. - datasetIdentifier (string) – [REQUIRED] - The dataset ARN. 
- options (dict) – - The dataset DataIntegrationFlow target options. - loadType (string) – - The dataset data load type in dataset options. 
- dedupeRecords (boolean) – - The dataset load option to remove duplicates. 
 
 
 
 
- Return type:
- dict 
- Returns:
- Response Syntax- { 'flow': { 'instanceId': 'string', 'name': 'string', 'sources': [ { 'sourceType': 'S3'|'DATASET', 'sourceName': 'string', 's3Source': { 'bucketName': 'string', 'prefix': 'string', 'options': { 'fileType': 'CSV'|'PARQUET'|'JSON' } }, 'datasetSource': { 'datasetIdentifier': 'string', 'options': { 'loadType': 'INCREMENTAL'|'REPLACE', 'dedupeRecords': True|False } } }, ], 'transformation': { 'transformationType': 'SQL'|'NONE', 'sqlTransformation': { 'query': 'string' } }, 'target': { 'targetType': 'S3'|'DATASET', 's3Target': { 'bucketName': 'string', 'prefix': 'string', 'options': { 'fileType': 'CSV'|'PARQUET'|'JSON' } }, 'datasetTarget': { 'datasetIdentifier': 'string', 'options': { 'loadType': 'INCREMENTAL'|'REPLACE', 'dedupeRecords': True|False } } }, 'createdTime': datetime(2015, 1, 1), 'lastModifiedTime': datetime(2015, 1, 1) } } - Response Structure- (dict) – - The response parameters for UpdateDataIntegrationFlow. - flow (dict) – - The details of the updated DataIntegrationFlow. - instanceId (string) – - The DataIntegrationFlow instance ID. 
- name (string) – - The DataIntegrationFlow name. 
- sources (list) – - The DataIntegrationFlow source configurations. - (dict) – - The DataIntegrationFlow source parameters. - sourceType (string) – - The DataIntegrationFlow source type. 
- sourceName (string) – - The DataIntegrationFlow source name that can be used as table alias in SQL transformation query. 
- s3Source (dict) – - The S3 DataIntegrationFlow source. - bucketName (string) – - The bucketName of the S3 source objects. 
- prefix (string) – - The prefix of the S3 source objects. 
- options (dict) – - The other options of the S3 DataIntegrationFlow source. - fileType (string) – - The Amazon S3 file type in S3 options. 
 
 
- datasetSource (dict) – - The dataset DataIntegrationFlow source. - datasetIdentifier (string) – - The ARN of the dataset. 
- options (dict) – - The dataset DataIntegrationFlow source options. - loadType (string) – - The dataset data load type in dataset options. 
- dedupeRecords (boolean) – - The dataset load option to remove duplicates. 
 
 
 
 
- transformation (dict) – - The DataIntegrationFlow transformation configurations. - transformationType (string) – - The DataIntegrationFlow transformation type. 
- sqlTransformation (dict) – - The SQL DataIntegrationFlow transformation configuration. - query (string) – - The transformation SQL query body based on SparkSQL. 
 
 
- target (dict) – - The DataIntegrationFlow target configuration. - targetType (string) – - The DataIntegrationFlow target type. 
- s3Target (dict) – - The S3 DataIntegrationFlow target. - bucketName (string) – - The bucketName of the S3 target objects. 
- prefix (string) – - The prefix of the S3 target objects. 
- options (dict) – - The S3 DataIntegrationFlow target options. - fileType (string) – - The Amazon S3 file type in S3 options. 
 
 
- datasetTarget (dict) – - The dataset DataIntegrationFlow target. - datasetIdentifier (string) – - The dataset ARN. 
- options (dict) – - The dataset DataIntegrationFlow target options. - loadType (string) – - The dataset data load type in dataset options. 
- dedupeRecords (boolean) – - The dataset load option to remove duplicates. 
 
 
 
- createdTime (datetime) – - The DataIntegrationFlow creation timestamp. 
- lastModifiedTime (datetime) – - The DataIntegrationFlow last modified timestamp. 
 
 
 
 - Exceptions- SupplyChain.Client.exceptions.ServiceQuotaExceededException
- SupplyChain.Client.exceptions.ResourceNotFoundException
- SupplyChain.Client.exceptions.ThrottlingException
- SupplyChain.Client.exceptions.AccessDeniedException
- SupplyChain.Client.exceptions.ValidationException
- SupplyChain.Client.exceptions.InternalServerException
- SupplyChain.Client.exceptions.ConflictException