The factories/pipelines resource type can be deployed with operations that target:
For a list of changed properties in each API version, see change log.
To create a Microsoft.DataFactory/factories/pipelines resource, add the following JSON to your template.
{
"type": "Microsoft.DataFactory/factories/pipelines",
"apiVersion": "2018-06-01",
"name": "string",
"properties": {
"activities": [
{
"dependsOn": [
{
"activity": "string",
"dependencyConditions": [ "string" ],
"{customized property}": {}
}
],
"description": "string",
"name": "string",
"onInactiveMarkAs": "string",
"state": "string",
"userProperties": [
{
"name": "string",
"value": {}
}
],
"type": "string"
// For remaining properties, see Activity objects
}
],
"annotations": [ object ],
"concurrency": "int",
"description": "string",
"folder": {
"name": "string"
},
"parameters": {
"{customized property}": {
"defaultValue": {},
"type": "string"
}
},
"policy": {
"elapsedTimeMetric": {
"duration": {}
}
},
"runDimensions": {
"{customized property}": {}
},
"variables": {
"{customized property}": {
"defaultValue": {},
"type": "string"
}
}
}
}
Set the type property to specify the type of object.
For AppendVariable, use:
"type": "AppendVariable",
"typeProperties": {
"value": {},
"variableName": "string"
}
For AzureDataExplorerCommand, use:
"type": "AzureDataExplorerCommand",
"linkedServiceName": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
},
"policy": {
"retry": {},
"retryIntervalInSeconds": "int",
"secureInput": "bool",
"secureOutput": "bool",
"timeout": {},
"{customized property}": {}
},
"typeProperties": {
"command": {},
"commandTimeout": {}
}
For AzureFunctionActivity, use:
"type": "AzureFunctionActivity",
"linkedServiceName": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
},
"policy": {
"retry": {},
"retryIntervalInSeconds": "int",
"secureInput": "bool",
"secureOutput": "bool",
"timeout": {},
"{customized property}": {}
},
"typeProperties": {
"body": {},
"functionName": {},
"headers": {
"{customized property}": "string"
},
"method": "string"
}
For AzureMLBatchExecution, use:
"type": "AzureMLBatchExecution",
"linkedServiceName": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
},
"policy": {
"retry": {},
"retryIntervalInSeconds": "int",
"secureInput": "bool",
"secureOutput": "bool",
"timeout": {},
"{customized property}": {}
},
"typeProperties": {
"globalParameters": {
"{customized property}": {}
},
"webServiceInputs": {
"{customized property}": {
"filePath": {},
"linkedServiceName": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
}
}
},
"webServiceOutputs": {
"{customized property}": {
"filePath": {},
"linkedServiceName": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
}
}
}
}
For AzureMLExecutePipeline, use:
"type": "AzureMLExecutePipeline",
"linkedServiceName": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
},
"policy": {
"retry": {},
"retryIntervalInSeconds": "int",
"secureInput": "bool",
"secureOutput": "bool",
"timeout": {},
"{customized property}": {}
},
"typeProperties": {
"continueOnStepFailure": {},
"dataPathAssignments": {},
"experimentName": {},
"mlParentRunId": {},
"mlPipelineEndpointId": {},
"mlPipelineId": {},
"mlPipelineParameters": {},
"version": {}
}
For AzureMLUpdateResource, use:
"type": "AzureMLUpdateResource",
"linkedServiceName": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
},
"policy": {
"retry": {},
"retryIntervalInSeconds": "int",
"secureInput": "bool",
"secureOutput": "bool",
"timeout": {},
"{customized property}": {}
},
"typeProperties": {
"trainedModelFilePath": {},
"trainedModelLinkedServiceName": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
},
"trainedModelName": {}
}
For Copy, use:
"type": "Copy",
"inputs": [
{
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "DatasetReference"
}
],
"linkedServiceName": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
},
"outputs": [
{
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "DatasetReference"
}
],
"policy": {
"retry": {},
"retryIntervalInSeconds": "int",
"secureInput": "bool",
"secureOutput": "bool",
"timeout": {},
"{customized property}": {}
},
"typeProperties": {
"dataIntegrationUnits": {},
"enableSkipIncompatibleRow": {},
"enableStaging": {},
"logSettings": {
"copyActivityLogSettings": {
"enableReliableLogging": {},
"logLevel": {}
},
"enableCopyActivityLog": {},
"logLocationSettings": {
"linkedServiceName": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
},
"path": {}
}
},
"logStorageSettings": {
"enableReliableLogging": {},
"linkedServiceName": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
},
"logLevel": {},
"path": {},
"{customized property}": {}
},
"parallelCopies": {},
"preserve": [ object ],
"preserveRules": [ object ],
"redirectIncompatibleRowSettings": {
"linkedServiceName": {},
"path": {},
"{customized property}": {}
},
"sink": {
"disableMetricsCollection": {},
"maxConcurrentConnections": {},
"sinkRetryCount": {},
"sinkRetryWait": {},
"writeBatchSize": {},
"writeBatchTimeout": {},
"type": "string"
// For remaining properties, see CopySink objects
},
"skipErrorFile": {
"dataInconsistency": {},
"fileMissing": {}
},
"source": {
"disableMetricsCollection": {},
"maxConcurrentConnections": {},
"sourceRetryCount": {},
"sourceRetryWait": {},
"type": "string"
// For remaining properties, see CopySource objects
},
"stagingSettings": {
"enableCompression": {},
"linkedServiceName": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
},
"path": {},
"{customized property}": {}
},
"translator": {},
"validateDataConsistency": {}
}
For Custom, use:
"type": "Custom",
"linkedServiceName": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
},
"policy": {
"retry": {},
"retryIntervalInSeconds": "int",
"secureInput": "bool",
"secureOutput": "bool",
"timeout": {},
"{customized property}": {}
},
"typeProperties": {
"autoUserSpecification": {},
"command": {},
"extendedProperties": {
"{customized property}": {}
},
"folderPath": {},
"referenceObjects": {
"datasets": [
{
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "DatasetReference"
}
],
"linkedServices": [
{
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
}
]
},
"resourceLinkedService": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
},
"retentionTimeInDays": {}
}
For DatabricksNotebook, use:
"type": "DatabricksNotebook",
"linkedServiceName": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
},
"policy": {
"retry": {},
"retryIntervalInSeconds": "int",
"secureInput": "bool",
"secureOutput": "bool",
"timeout": {},
"{customized property}": {}
},
"typeProperties": {
"baseParameters": {
"{customized property}": {}
},
"libraries": [
{
"{customized property}": {}
}
],
"notebookPath": {}
}
For DatabricksSparkJar, use:
"type": "DatabricksSparkJar",
"linkedServiceName": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
},
"policy": {
"retry": {},
"retryIntervalInSeconds": "int",
"secureInput": "bool",
"secureOutput": "bool",
"timeout": {},
"{customized property}": {}
},
"typeProperties": {
"libraries": [
{
"{customized property}": {}
}
],
"mainClassName": {},
"parameters": [ object ]
}
For DatabricksSparkPython, use:
"type": "DatabricksSparkPython",
"linkedServiceName": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
},
"policy": {
"retry": {},
"retryIntervalInSeconds": "int",
"secureInput": "bool",
"secureOutput": "bool",
"timeout": {},
"{customized property}": {}
},
"typeProperties": {
"libraries": [
{
"{customized property}": {}
}
],
"parameters": [ object ],
"pythonFile": {}
}
For DataLakeAnalyticsU-SQL, use:
"type": "DataLakeAnalyticsU-SQL",
"linkedServiceName": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
},
"policy": {
"retry": {},
"retryIntervalInSeconds": "int",
"secureInput": "bool",
"secureOutput": "bool",
"timeout": {},
"{customized property}": {}
},
"typeProperties": {
"compilationMode": {},
"degreeOfParallelism": {},
"parameters": {
"{customized property}": {}
},
"priority": {},
"runtimeVersion": {},
"scriptLinkedService": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
},
"scriptPath": {}
}
For Delete, use:
"type": "Delete",
"linkedServiceName": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
},
"policy": {
"retry": {},
"retryIntervalInSeconds": "int",
"secureInput": "bool",
"secureOutput": "bool",
"timeout": {},
"{customized property}": {}
},
"typeProperties": {
"dataset": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "DatasetReference"
},
"enableLogging": {},
"logStorageSettings": {
"enableReliableLogging": {},
"linkedServiceName": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
},
"logLevel": {},
"path": {},
"{customized property}": {}
},
"maxConcurrentConnections": "int",
"recursive": {},
"storeSettings": {
"disableMetricsCollection": {},
"maxConcurrentConnections": {},
"type": "string"
// For remaining properties, see StoreReadSettings objects
}
}
For ExecuteDataFlow, use:
"type": "ExecuteDataFlow",
"linkedServiceName": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
},
"policy": {
"retry": {},
"retryIntervalInSeconds": "int",
"secureInput": "bool",
"secureOutput": "bool",
"timeout": {},
"{customized property}": {}
},
"typeProperties": {
"compute": {
"computeType": {},
"coreCount": {}
},
"continueOnError": {},
"dataFlow": {
"datasetParameters": {},
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "DataFlowReference",
"{customized property}": {}
},
"integrationRuntime": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "IntegrationRuntimeReference"
},
"runConcurrently": {},
"sourceStagingConcurrency": {},
"staging": {
"folderPath": {},
"linkedService": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
}
},
"traceLevel": {}
}
For ExecutePipeline, use:
"type": "ExecutePipeline",
"policy": {
"secureInput": "bool",
"{customized property}": {}
},
"typeProperties": {
"parameters": {
"{customized property}": {}
},
"pipeline": {
"name": "string",
"referenceName": "string",
"type": "PipelineReference"
},
"waitOnCompletion": "bool"
}
For ExecuteSSISPackage, use:
"type": "ExecuteSSISPackage",
"linkedServiceName": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
},
"policy": {
"retry": {},
"retryIntervalInSeconds": "int",
"secureInput": "bool",
"secureOutput": "bool",
"timeout": {},
"{customized property}": {}
},
"typeProperties": {
"connectVia": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "IntegrationRuntimeReference"
},
"environmentPath": {},
"executionCredential": {
"domain": {},
"password": {
"type": "string",
"value": "string"
},
"userName": {}
},
"loggingLevel": {},
"logLocation": {
"logPath": {},
"type": "File",
"typeProperties": {
"accessCredential": {
"domain": {},
"password": {
"type": "string"
// For remaining properties, see SecretBase objects
},
"userName": {}
},
"logRefreshInterval": {}
}
},
"packageConnectionManagers": {
"{customized property}": {
"{customized property}": {
"value": {}
}
}
},
"packageLocation": {
"packagePath": {},
"type": "string",
"typeProperties": {
"accessCredential": {
"domain": {},
"password": {
"type": "string"
// For remaining properties, see SecretBase objects
},
"userName": {}
},
"childPackages": [
{
"packageContent": {},
"packageLastModifiedDate": "string",
"packageName": "string",
"packagePath": {}
}
],
"configurationAccessCredential": {
"domain": {},
"password": {
"type": "string"
// For remaining properties, see SecretBase objects
},
"userName": {}
},
"configurationPath": {},
"packageContent": {},
"packageLastModifiedDate": "string",
"packageName": "string",
"packagePassword": {
"type": "string"
// For remaining properties, see SecretBase objects
}
}
},
"packageParameters": {
"{customized property}": {
"value": {}
}
},
"projectConnectionManagers": {
"{customized property}": {
"{customized property}": {
"value": {}
}
}
},
"projectParameters": {
"{customized property}": {
"value": {}
}
},
"propertyOverrides": {
"{customized property}": {
"isSensitive": "bool",
"value": {}
}
},
"runtime": {}
}
For ExecuteWranglingDataflow, use:
"type": "ExecuteWranglingDataflow",
"policy": {
"retry": {},
"retryIntervalInSeconds": "int",
"secureInput": "bool",
"secureOutput": "bool",
"timeout": {},
"{customized property}": {}
},
"typeProperties": {
"compute": {
"computeType": {},
"coreCount": {}
},
"continueOnError": {},
"dataFlow": {
"datasetParameters": {},
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "DataFlowReference",
"{customized property}": {}
},
"integrationRuntime": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "IntegrationRuntimeReference"
},
"queries": [
{
"dataflowSinks": [
{
"dataset": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "DatasetReference"
},
"description": "string",
"flowlet": {
"datasetParameters": {},
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "DataFlowReference",
"{customized property}": {}
},
"linkedService": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
},
"name": "string",
"rejectedDataLinkedService": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
},
"schemaLinkedService": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
},
"script": "string"
}
],
"queryName": "string"
}
],
"runConcurrently": {},
"sinks": {
"{customized property}": {
"dataset": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "DatasetReference"
},
"description": "string",
"flowlet": {
"datasetParameters": {},
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "DataFlowReference",
"{customized property}": {}
},
"linkedService": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
},
"name": "string",
"rejectedDataLinkedService": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
},
"schemaLinkedService": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
},
"script": "string"
}
},
"sourceStagingConcurrency": {},
"staging": {
"folderPath": {},
"linkedService": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
}
},
"traceLevel": {}
}
For Fail, use:
"type": "Fail",
"typeProperties": {
"errorCode": {},
"message": {}
}
For Filter, use:
"type": "Filter",
"typeProperties": {
"condition": {
"type": "Expression",
"value": "string"
},
"items": {
"type": "Expression",
"value": "string"
}
}
For ForEach, use:
"type": "ForEach",
"typeProperties": {
"activities": [
{
"dependsOn": [
{
"activity": "string",
"dependencyConditions": [ "string" ],
"{customized property}": {}
}
],
"description": "string",
"name": "string",
"onInactiveMarkAs": "string",
"state": "string",
"userProperties": [
{
"name": "string",
"value": {}
}
],
"type": "string"
// For remaining properties, see Activity objects
}
],
"batchCount": "int",
"isSequential": "bool",
"items": {
"type": "Expression",
"value": "string"
}
}
For GetMetadata, use:
"type": "GetMetadata",
"linkedServiceName": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
},
"policy": {
"retry": {},
"retryIntervalInSeconds": "int",
"secureInput": "bool",
"secureOutput": "bool",
"timeout": {},
"{customized property}": {}
},
"typeProperties": {
"dataset": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "DatasetReference"
},
"fieldList": [ object ],
"formatSettings": {
"type": "string"
// For remaining properties, see FormatReadSettings objects
},
"storeSettings": {
"disableMetricsCollection": {},
"maxConcurrentConnections": {},
"type": "string"
// For remaining properties, see StoreReadSettings objects
}
}
For HDInsightHive, use:
"type": "HDInsightHive",
"linkedServiceName": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
},
"policy": {
"retry": {},
"retryIntervalInSeconds": "int",
"secureInput": "bool",
"secureOutput": "bool",
"timeout": {},
"{customized property}": {}
},
"typeProperties": {
"arguments": [ object ],
"defines": {
"{customized property}": {}
},
"getDebugInfo": "string",
"queryTimeout": "int",
"scriptLinkedService": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
},
"scriptPath": {},
"storageLinkedServices": [
{
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
}
],
"variables": {
"{customized property}": {}
}
}
For HDInsightMapReduce, use:
"type": "HDInsightMapReduce",
"linkedServiceName": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
},
"policy": {
"retry": {},
"retryIntervalInSeconds": "int",
"secureInput": "bool",
"secureOutput": "bool",
"timeout": {},
"{customized property}": {}
},
"typeProperties": {
"arguments": [ object ],
"className": {},
"defines": {
"{customized property}": {}
},
"getDebugInfo": "string",
"jarFilePath": {},
"jarLibs": [ object ],
"jarLinkedService": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
},
"storageLinkedServices": [
{
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
}
]
}
For HDInsightPig, use:
"type": "HDInsightPig",
"linkedServiceName": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
},
"policy": {
"retry": {},
"retryIntervalInSeconds": "int",
"secureInput": "bool",
"secureOutput": "bool",
"timeout": {},
"{customized property}": {}
},
"typeProperties": {
"arguments": {},
"defines": {
"{customized property}": {}
},
"getDebugInfo": "string",
"scriptLinkedService": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
},
"scriptPath": {},
"storageLinkedServices": [
{
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
}
]
}
For HDInsightSpark, use:
"type": "HDInsightSpark",
"linkedServiceName": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
},
"policy": {
"retry": {},
"retryIntervalInSeconds": "int",
"secureInput": "bool",
"secureOutput": "bool",
"timeout": {},
"{customized property}": {}
},
"typeProperties": {
"arguments": [ object ],
"className": "string",
"entryFilePath": {},
"getDebugInfo": "string",
"proxyUser": {},
"rootPath": {},
"sparkConfig": {
"{customized property}": {}
},
"sparkJobLinkedService": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
}
}
For HDInsightStreaming, use:
"type": "HDInsightStreaming",
"linkedServiceName": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
},
"policy": {
"retry": {},
"retryIntervalInSeconds": "int",
"secureInput": "bool",
"secureOutput": "bool",
"timeout": {},
"{customized property}": {}
},
"typeProperties": {
"arguments": [ object ],
"combiner": {},
"commandEnvironment": [ object ],
"defines": {
"{customized property}": {}
},
"fileLinkedService": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
},
"filePaths": [ object ],
"getDebugInfo": "string",
"input": {},
"mapper": {},
"output": {},
"reducer": {},
"storageLinkedServices": [
{
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
}
]
}
For IfCondition, use:
"type": "IfCondition",
"typeProperties": {
"expression": {
"type": "Expression",
"value": "string"
},
"ifFalseActivities": [
{
"dependsOn": [
{
"activity": "string",
"dependencyConditions": [ "string" ],
"{customized property}": {}
}
],
"description": "string",
"name": "string",
"onInactiveMarkAs": "string",
"state": "string",
"userProperties": [
{
"name": "string",
"value": {}
}
],
"type": "string"
// For remaining properties, see Activity objects
}
],
"ifTrueActivities": [
{
"dependsOn": [
{
"activity": "string",
"dependencyConditions": [ "string" ],
"{customized property}": {}
}
],
"description": "string",
"name": "string",
"onInactiveMarkAs": "string",
"state": "string",
"userProperties": [
{
"name": "string",
"value": {}
}
],
"type": "string"
// For remaining properties, see Activity objects
}
]
}
For Lookup, use:
"type": "Lookup",
"linkedServiceName": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
},
"policy": {
"retry": {},
"retryIntervalInSeconds": "int",
"secureInput": "bool",
"secureOutput": "bool",
"timeout": {},
"{customized property}": {}
},
"typeProperties": {
"dataset": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "DatasetReference"
},
"firstRowOnly": {},
"source": {
"disableMetricsCollection": {},
"maxConcurrentConnections": {},
"sourceRetryCount": {},
"sourceRetryWait": {},
"type": "string"
// For remaining properties, see CopySource objects
}
}
For Script, use:
"type": "Script",
"linkedServiceName": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
},
"policy": {
"retry": {},
"retryIntervalInSeconds": "int",
"secureInput": "bool",
"secureOutput": "bool",
"timeout": {},
"{customized property}": {}
},
"typeProperties": {
"logSettings": {
"logDestination": "string",
"logLocationSettings": {
"linkedServiceName": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
},
"path": {}
}
},
"scriptBlockExecutionTimeout": {},
"scripts": [
{
"parameters": [
{
"direction": "string",
"name": {},
"size": "int",
"type": "string",
"value": {}
}
],
"text": {},
"type": "string"
}
]
}
For SetVariable, use:
"type": "SetVariable",
"policy": {
"secureInput": "bool",
"secureOutput": "bool"
},
"typeProperties": {
"setSystemVariable": "bool",
"value": {},
"variableName": "string"
}
For SparkJob, use:
"type": "SparkJob",
"linkedServiceName": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
},
"policy": {
"retry": {},
"retryIntervalInSeconds": "int",
"secureInput": "bool",
"secureOutput": "bool",
"timeout": {},
"{customized property}": {}
},
"typeProperties": {
"args": [ object ],
"className": {},
"conf": {},
"configurationType": "string",
"driverSize": {},
"executorSize": {},
"file": {},
"files": [ object ],
"filesV2": [ object ],
"numExecutors": {},
"pythonCodeReference": [ object ],
"scanFolder": {},
"sparkConfig": {
"{customized property}": {}
},
"sparkJob": {
"referenceName": {},
"type": "SparkJobDefinitionReference"
},
"targetBigDataPool": {
"referenceName": {},
"type": "BigDataPoolReference"
},
"targetSparkConfiguration": {
"referenceName": {},
"type": "SparkConfigurationReference"
}
}
For SqlServerStoredProcedure, use:
"type": "SqlServerStoredProcedure",
"linkedServiceName": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
},
"policy": {
"retry": {},
"retryIntervalInSeconds": "int",
"secureInput": "bool",
"secureOutput": "bool",
"timeout": {},
"{customized property}": {}
},
"typeProperties": {
"storedProcedureName": {},
"storedProcedureParameters": {}
}
For Switch, use:
"type": "Switch",
"typeProperties": {
"cases": [
{
"activities": [
{
"dependsOn": [
{
"activity": "string",
"dependencyConditions": [ "string" ],
"{customized property}": {}
}
],
"description": "string",
"name": "string",
"onInactiveMarkAs": "string",
"state": "string",
"userProperties": [
{
"name": "string",
"value": {}
}
],
"type": "string"
// For remaining properties, see Activity objects
}
],
"value": "string"
}
],
"defaultActivities": [
{
"dependsOn": [
{
"activity": "string",
"dependencyConditions": [ "string" ],
"{customized property}": {}
}
],
"description": "string",
"name": "string",
"onInactiveMarkAs": "string",
"state": "string",
"userProperties": [
{
"name": "string",
"value": {}
}
],
"type": "string"
// For remaining properties, see Activity objects
}
],
"on": {
"type": "Expression",
"value": "string"
}
}
For SynapseNotebook, use:
"type": "SynapseNotebook",
"linkedServiceName": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
},
"policy": {
"retry": {},
"retryIntervalInSeconds": "int",
"secureInput": "bool",
"secureOutput": "bool",
"timeout": {},
"{customized property}": {}
},
"typeProperties": {
"conf": {},
"configurationType": "string",
"driverSize": {},
"executorSize": {},
"notebook": {
"referenceName": {},
"type": "NotebookReference"
},
"numExecutors": {},
"parameters": {
"{customized property}": {
"type": "string",
"value": {}
}
},
"sparkConfig": {
"{customized property}": {}
},
"sparkPool": {
"referenceName": {},
"type": "BigDataPoolReference"
},
"targetSparkConfiguration": {
"referenceName": {},
"type": "SparkConfigurationReference"
}
}
For Until, use:
"type": "Until",
"typeProperties": {
"activities": [
{
"dependsOn": [
{
"activity": "string",
"dependencyConditions": [ "string" ],
"{customized property}": {}
}
],
"description": "string",
"name": "string",
"onInactiveMarkAs": "string",
"state": "string",
"userProperties": [
{
"name": "string",
"value": {}
}
],
"type": "string"
// For remaining properties, see Activity objects
}
],
"expression": {
"type": "Expression",
"value": "string"
},
"timeout": {}
}
For Validation, use:
"type": "Validation",
"typeProperties": {
"childItems": {},
"dataset": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "DatasetReference"
},
"minimumSize": {},
"sleep": {},
"timeout": {}
}
For Wait, use:
"type": "Wait",
"typeProperties": {
"waitTimeInSeconds": {}
}
For WebActivity, use:
"type": "WebActivity",
"linkedServiceName": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
},
"policy": {
"retry": {},
"retryIntervalInSeconds": "int",
"secureInput": "bool",
"secureOutput": "bool",
"timeout": {},
"{customized property}": {}
},
"typeProperties": {
"authentication": {
"credential": {
"referenceName": "string",
"type": "CredentialReference",
"{customized property}": {}
},
"password": {
"type": "string"
// For remaining properties, see SecretBase objects
},
"pfx": {
"type": "string"
// For remaining properties, see SecretBase objects
},
"resource": {},
"type": "string",
"username": {},
"userTenant": {}
},
"body": {},
"connectVia": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "IntegrationRuntimeReference"
},
"datasets": [
{
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "DatasetReference"
}
],
"disableCertValidation": "bool",
"headers": {
"{customized property}": "string"
},
"httpRequestTimeout": {},
"linkedServices": [
{
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
}
],
"method": "string",
"turnOffAsync": "bool",
"url": {}
}
For WebHook, use:
"type": "WebHook",
"policy": {
"secureInput": "bool",
"secureOutput": "bool"
},
"typeProperties": {
"authentication": {
"credential": {
"referenceName": "string",
"type": "CredentialReference",
"{customized property}": {}
},
"password": {
"type": "string"
// For remaining properties, see SecretBase objects
},
"pfx": {
"type": "string"
// For remaining properties, see SecretBase objects
},
"resource": {},
"type": "string",
"username": {},
"userTenant": {}
},
"body": {},
"headers": {
"{customized property}": "string"
},
"method": "POST",
"reportStatusOnCallBack": {},
"timeout": "string",
"url": {}
}
Set the type property to specify the type of object.
For AvroSink, use:
"type": "AvroSink",
"formatSettings": {
"fileNamePrefix": {},
"maxRowsPerFile": {},
"recordName": "string",
"recordNamespace": "string",
"type": "string",
"{customized property}": {}
},
"storeSettings": {
"copyBehavior": {},
"disableMetricsCollection": {},
"maxConcurrentConnections": {},
"metadata": [
{
"name": {},
"value": {}
}
],
"type": "string"
// For remaining properties, see StoreWriteSettings objects
}
For AzureBlobFSSink, use:
"type": "AzureBlobFSSink",
"copyBehavior": {},
"metadata": [
{
"name": {},
"value": {}
}
]
For AzureDatabricksDeltaLakeSink, use:
"type": "AzureDatabricksDeltaLakeSink",
"importSettings": {
"dateFormat": {},
"timestampFormat": {},
"type": "string",
"{customized property}": {}
},
"preCopyScript": {}
For AzureDataExplorerSink, use:
"type": "AzureDataExplorerSink",
"flushImmediately": {},
"ingestionMappingAsJson": {},
"ingestionMappingName": {}
For AzureDataLakeStoreSink, use:
"type": "AzureDataLakeStoreSink",
"copyBehavior": {},
"enableAdlsSingleFileParallel": {}
For AzureMySqlSink, use:
"type": "AzureMySqlSink",
"preCopyScript": {}
For AzurePostgreSqlSink, use:
"type": "AzurePostgreSqlSink",
"preCopyScript": {}
For AzureQueueSink, use:
"type": "AzureQueueSink"
For AzureSearchIndexSink, use:
"type": "AzureSearchIndexSink",
"writeBehavior": "string"
For AzureSqlSink, use:
"type": "AzureSqlSink",
"preCopyScript": {},
"sqlWriterStoredProcedureName": {},
"sqlWriterTableType": {},
"sqlWriterUseTableLock": {},
"storedProcedureParameters": {},
"storedProcedureTableTypeParameterName": {},
"tableOption": {},
"upsertSettings": {
"interimSchemaName": {},
"keys": {},
"useTempDB": {}
},
"writeBehavior": {}
For AzureTableSink, use:
"type": "AzureTableSink",
"azureTableDefaultPartitionKeyValue": {},
"azureTableInsertType": {},
"azureTablePartitionKeyName": {},
"azureTableRowKeyName": {}
For BinarySink, use:
"type": "BinarySink",
"storeSettings": {
"copyBehavior": {},
"disableMetricsCollection": {},
"maxConcurrentConnections": {},
"metadata": [
{
"name": {},
"value": {}
}
],
"type": "string"
// For remaining properties, see StoreWriteSettings objects
}
For BlobSink, use:
"type": "BlobSink",
"blobWriterAddHeader": {},
"blobWriterDateTimeFormat": {},
"blobWriterOverwriteFiles": {},
"copyBehavior": {},
"metadata": [
{
"name": {},
"value": {}
}
]
For CommonDataServiceForAppsSink, use:
"type": "CommonDataServiceForAppsSink",
"alternateKeyName": {},
"ignoreNullValues": {},
"writeBehavior": "Upsert"
For CosmosDbMongoDbApiSink, use:
"type": "CosmosDbMongoDbApiSink",
"writeBehavior": {}
For CosmosDbSqlApiSink, use:
"type": "CosmosDbSqlApiSink",
"writeBehavior": {}
For DelimitedTextSink, use:
"type": "DelimitedTextSink",
"formatSettings": {
"fileExtension": {},
"fileNamePrefix": {},
"maxRowsPerFile": {},
"quoteAllText": {},
"type": "string",
"{customized property}": {}
},
"storeSettings": {
"copyBehavior": {},
"disableMetricsCollection": {},
"maxConcurrentConnections": {},
"metadata": [
{
"name": {},
"value": {}
}
],
"type": "string"
// For remaining properties, see StoreWriteSettings objects
}
For DocumentDbCollectionSink, use:
"type": "DocumentDbCollectionSink",
"nestingSeparator": {},
"writeBehavior": {}
For DynamicsCrmSink, use:
"type": "DynamicsCrmSink",
"alternateKeyName": {},
"ignoreNullValues": {},
"writeBehavior": "Upsert"
For DynamicsSink, use:
"type": "DynamicsSink",
"alternateKeyName": {},
"ignoreNullValues": {},
"writeBehavior": "Upsert"
For FileSystemSink, use:
"type": "FileSystemSink",
"copyBehavior": {}
For InformixSink, use:
"type": "InformixSink",
"preCopyScript": {}
For JsonSink, use:
"type": "JsonSink",
"formatSettings": {
"filePattern": {},
"type": "string",
"{customized property}": {}
},
"storeSettings": {
"copyBehavior": {},
"disableMetricsCollection": {},
"maxConcurrentConnections": {},
"metadata": [
{
"name": {},
"value": {}
}
],
"type": "string"
// For remaining properties, see StoreWriteSettings objects
}
For LakeHouseTableSink, use:
"type": "LakeHouseTableSink",
"partitionNameList": {},
"partitionOption": {},
"tableActionOption": {}
For MicrosoftAccessSink, use:
"type": "MicrosoftAccessSink",
"preCopyScript": {}
For MongoDbAtlasSink, use:
"type": "MongoDbAtlasSink",
"writeBehavior": {}
For MongoDbV2Sink, use:
"type": "MongoDbV2Sink",
"writeBehavior": {}
For OdbcSink, use:
"type": "OdbcSink",
"preCopyScript": {}
For OracleSink, use:
"type": "OracleSink",
"preCopyScript": {}
For OrcSink, use:
"type": "OrcSink",
"formatSettings": {
"fileNamePrefix": {},
"maxRowsPerFile": {},
"type": "string",
"{customized property}": {}
},
"storeSettings": {
"copyBehavior": {},
"disableMetricsCollection": {},
"maxConcurrentConnections": {},
"metadata": [
{
"name": {},
"value": {}
}
],
"type": "string"
// For remaining properties, see StoreWriteSettings objects
}
For ParquetSink, use:
"type": "ParquetSink",
"formatSettings": {
"fileNamePrefix": {},
"maxRowsPerFile": {},
"type": "string",
"{customized property}": {}
},
"storeSettings": {
"copyBehavior": {},
"disableMetricsCollection": {},
"maxConcurrentConnections": {},
"metadata": [
{
"name": {},
"value": {}
}
],
"type": "string"
// For remaining properties, see StoreWriteSettings objects
}
For RestSink, use:
"type": "RestSink",
"additionalHeaders": {},
"httpCompressionType": {},
"httpRequestTimeout": {},
"requestInterval": {},
"requestMethod": {}
For SalesforceServiceCloudSink, use:
"type": "SalesforceServiceCloudSink",
"externalIdFieldName": {},
"ignoreNullValues": {},
"writeBehavior": "string"
For SalesforceServiceCloudV2Sink, use:
"type": "SalesforceServiceCloudV2Sink",
"externalIdFieldName": {},
"ignoreNullValues": {},
"writeBehavior": "string"
For SalesforceSink, use:
"type": "SalesforceSink",
"externalIdFieldName": {},
"ignoreNullValues": {},
"writeBehavior": "string"
For SalesforceV2Sink, use:
"type": "SalesforceV2Sink",
"externalIdFieldName": {},
"ignoreNullValues": {},
"writeBehavior": "string"
For SapCloudForCustomerSink, use:
"type": "SapCloudForCustomerSink",
"httpRequestTimeout": {},
"writeBehavior": "string"
For SnowflakeSink, use:
"type": "SnowflakeSink",
"importSettings": {
"additionalCopyOptions": {
"{customized property}": {}
},
"additionalFormatOptions": {
"{customized property}": {}
},
"type": "string",
"{customized property}": {}
},
"preCopyScript": {}
For SnowflakeV2Sink, use:
"type": "SnowflakeV2Sink",
"importSettings": {
"additionalCopyOptions": {
"{customized property}": {}
},
"additionalFormatOptions": {
"{customized property}": {}
},
"type": "string",
"{customized property}": {}
},
"preCopyScript": {}
For SqlDWSink, use:
"type": "SqlDWSink",
"allowCopyCommand": {},
"allowPolyBase": {},
"copyCommandSettings": {
"additionalOptions": {
"{customized property}": "string"
},
"defaultValues": [
{
"columnName": {},
"defaultValue": {}
}
]
},
"polyBaseSettings": {
"rejectSampleValue": {},
"rejectType": "string",
"rejectValue": {},
"useTypeDefault": {},
"{customized property}": {}
},
"preCopyScript": {},
"sqlWriterUseTableLock": {},
"tableOption": {},
"upsertSettings": {
"interimSchemaName": {},
"keys": {}
},
"writeBehavior": {}
For SqlMISink, use:
"type": "SqlMISink",
"preCopyScript": {},
"sqlWriterStoredProcedureName": {},
"sqlWriterTableType": {},
"sqlWriterUseTableLock": {},
"storedProcedureParameters": {},
"storedProcedureTableTypeParameterName": {},
"tableOption": {},
"upsertSettings": {
"interimSchemaName": {},
"keys": {},
"useTempDB": {}
},
"writeBehavior": {}
For SqlServerSink, use:
"type": "SqlServerSink",
"preCopyScript": {},
"sqlWriterStoredProcedureName": {},
"sqlWriterTableType": {},
"sqlWriterUseTableLock": {},
"storedProcedureParameters": {},
"storedProcedureTableTypeParameterName": {},
"tableOption": {},
"upsertSettings": {
"interimSchemaName": {},
"keys": {},
"useTempDB": {}
},
"writeBehavior": {}
For SqlSink, use:
"type": "SqlSink",
"preCopyScript": {},
"sqlWriterStoredProcedureName": {},
"sqlWriterTableType": {},
"sqlWriterUseTableLock": {},
"storedProcedureParameters": {},
"storedProcedureTableTypeParameterName": {},
"tableOption": {},
"upsertSettings": {
"interimSchemaName": {},
"keys": {},
"useTempDB": {}
},
"writeBehavior": {}
For WarehouseSink, use:
"type": "WarehouseSink",
"allowCopyCommand": {},
"copyCommandSettings": {
"additionalOptions": {
"{customized property}": "string"
},
"defaultValues": [
{
"columnName": {},
"defaultValue": {}
}
]
},
"preCopyScript": {},
"tableOption": {},
"writeBehavior": {}
Set the type property to specify the type of object.
For AzureBlobFSWriteSettings, use:
"type": "AzureBlobFSWriteSettings",
"blockSizeInMB": {}
For AzureBlobStorageWriteSettings, use:
"type": "AzureBlobStorageWriteSettings",
"blockSizeInMB": {}
For AzureDataLakeStoreWriteSettings, use:
"type": "AzureDataLakeStoreWriteSettings",
"expiryDateTime": {}
For AzureFileStorageWriteSettings, use:
"type": "AzureFileStorageWriteSettings"
For FileServerWriteSettings, use:
"type": "FileServerWriteSettings"
For LakeHouseWriteSettings, use:
"type": "LakeHouseWriteSettings"
For SftpWriteSettings, use:
"type": "SftpWriteSettings",
"operationTimeout": {},
"useTempFileRename": {}
Set the type property to specify the type of object.
For AmazonMWSSource, use:
"type": "AmazonMWSSource",
"additionalColumns": {},
"query": {},
"queryTimeout": {}
For AmazonRdsForOracleSource, use:
"type": "AmazonRdsForOracleSource",
"additionalColumns": {},
"oracleReaderQuery": {},
"partitionOption": {},
"partitionSettings": {
"partitionColumnName": {},
"partitionLowerBound": {},
"partitionNames": {},
"partitionUpperBound": {}
},
"queryTimeout": {}
For AmazonRdsForSqlServerSource, use:
"type": "AmazonRdsForSqlServerSource",
"additionalColumns": {},
"isolationLevel": {},
"partitionOption": {},
"partitionSettings": {
"partitionColumnName": {},
"partitionLowerBound": {},
"partitionUpperBound": {}
},
"produceAdditionalTypes": {},
"queryTimeout": {},
"sqlReaderQuery": {},
"sqlReaderStoredProcedureName": {},
"storedProcedureParameters": {}
For AmazonRedshiftSource, use:
"type": "AmazonRedshiftSource",
"additionalColumns": {},
"query": {},
"queryTimeout": {},
"redshiftUnloadSettings": {
"bucketName": {},
"s3LinkedServiceName": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
}
}
For AvroSource, use:
"type": "AvroSource",
"additionalColumns": {},
"storeSettings": {
"disableMetricsCollection": {},
"maxConcurrentConnections": {},
"type": "string"
// For remaining properties, see StoreReadSettings objects
}
For AzureBlobFSSource, use:
"type": "AzureBlobFSSource",
"recursive": {},
"skipHeaderLineCount": {},
"treatEmptyAsNull": {}
For AzureDatabricksDeltaLakeSource, use:
"type": "AzureDatabricksDeltaLakeSource",
"exportSettings": {
"dateFormat": {},
"timestampFormat": {},
"type": "string",
"{customized property}": {}
},
"query": {}
For AzureDataExplorerSource, use:
"type": "AzureDataExplorerSource",
"additionalColumns": {},
"noTruncation": {},
"query": {},
"queryTimeout": {}
For AzureDataLakeStoreSource, use:
"type": "AzureDataLakeStoreSource",
"recursive": {}
For AzureMariaDBSource, use:
"type": "AzureMariaDBSource",
"additionalColumns": {},
"query": {},
"queryTimeout": {}
For AzureMySqlSource, use:
"type": "AzureMySqlSource",
"additionalColumns": {},
"query": {},
"queryTimeout": {}
For AzurePostgreSqlSource, use:
"type": "AzurePostgreSqlSource",
"additionalColumns": {},
"query": {},
"queryTimeout": {}
For AzureSqlSource, use:
"type": "AzureSqlSource",
"additionalColumns": {},
"isolationLevel": {},
"partitionOption": {},
"partitionSettings": {
"partitionColumnName": {},
"partitionLowerBound": {},
"partitionUpperBound": {}
},
"produceAdditionalTypes": {},
"queryTimeout": {},
"sqlReaderQuery": {},
"sqlReaderStoredProcedureName": {},
"storedProcedureParameters": {}
For AzureTableSource, use:
"type": "AzureTableSource",
"additionalColumns": {},
"azureTableSourceIgnoreTableNotFound": {},
"azureTableSourceQuery": {},
"queryTimeout": {}
For BinarySource, use:
"type": "BinarySource",
"formatSettings": {
"compressionProperties": {
"type": "string"
// For remaining properties, see CompressionReadSettings objects
},
"type": "string",
"{customized property}": {}
},
"storeSettings": {
"disableMetricsCollection": {},
"maxConcurrentConnections": {},
"type": "string"
// For remaining properties, see StoreReadSettings objects
}
For BlobSource, use:
"type": "BlobSource",
"recursive": {},
"skipHeaderLineCount": {},
"treatEmptyAsNull": {}
For CassandraSource, use:
"type": "CassandraSource",
"additionalColumns": {},
"consistencyLevel": "string",
"query": {},
"queryTimeout": {}
For CommonDataServiceForAppsSource, use:
"type": "CommonDataServiceForAppsSource",
"additionalColumns": {},
"query": {}
For ConcurSource, use:
"type": "ConcurSource",
"additionalColumns": {},
"query": {},
"queryTimeout": {}
For CosmosDbMongoDbApiSource, use:
"type": "CosmosDbMongoDbApiSource",
"additionalColumns": {},
"batchSize": {},
"cursorMethods": {
"limit": {},
"project": {},
"skip": {},
"sort": {},
"{customized property}": {}
},
"filter": {},
"queryTimeout": {}
For CosmosDbSqlApiSource, use:
"type": "CosmosDbSqlApiSource",
"additionalColumns": {},
"detectDatetime": {},
"pageSize": {},
"preferredRegions": {},
"query": {}
For CouchbaseSource, use:
"type": "CouchbaseSource",
"additionalColumns": {},
"query": {},
"queryTimeout": {}
For Db2Source, use:
"type": "Db2Source",
"additionalColumns": {},
"query": {},
"queryTimeout": {}
For DelimitedTextSource, use:
"type": "DelimitedTextSource",
"additionalColumns": {},
"formatSettings": {
"compressionProperties": {
"type": "string"
// For remaining properties, see CompressionReadSettings objects
},
"skipLineCount": {},
"type": "string",
"{customized property}": {}
},
"storeSettings": {
"disableMetricsCollection": {},
"maxConcurrentConnections": {},
"type": "string"
// For remaining properties, see StoreReadSettings objects
}
For DocumentDbCollectionSource, use:
"type": "DocumentDbCollectionSource",
"additionalColumns": {},
"nestingSeparator": {},
"query": {},
"queryTimeout": {}
For DrillSource, use:
"type": "DrillSource",
"additionalColumns": {},
"query": {},
"queryTimeout": {}
For DynamicsAXSource, use:
"type": "DynamicsAXSource",
"additionalColumns": {},
"httpRequestTimeout": {},
"query": {},
"queryTimeout": {}
For DynamicsCrmSource, use:
"type": "DynamicsCrmSource",
"additionalColumns": {},
"query": {}
For DynamicsSource, use:
"type": "DynamicsSource",
"additionalColumns": {},
"query": {}
For EloquaSource, use:
"type": "EloquaSource",
"additionalColumns": {},
"query": {},
"queryTimeout": {}
For ExcelSource, use:
"type": "ExcelSource",
"additionalColumns": {},
"storeSettings": {
"disableMetricsCollection": {},
"maxConcurrentConnections": {},
"type": "string"
// For remaining properties, see StoreReadSettings objects
}
For FileSystemSource, use:
"type": "FileSystemSource",
"additionalColumns": {},
"recursive": {}
For GoogleAdWordsSource, use:
"type": "GoogleAdWordsSource",
"additionalColumns": {},
"query": {},
"queryTimeout": {}
For GoogleBigQuerySource, use:
"type": "GoogleBigQuerySource",
"additionalColumns": {},
"query": {},
"queryTimeout": {}
For GoogleBigQueryV2Source, use:
"type": "GoogleBigQueryV2Source",
"additionalColumns": {},
"query": {},
"queryTimeout": {}
For GreenplumSource, use:
"type": "GreenplumSource",
"additionalColumns": {},
"query": {},
"queryTimeout": {}
For HBaseSource, use:
"type": "HBaseSource",
"additionalColumns": {},
"query": {},
"queryTimeout": {}
For HdfsSource, use:
"type": "HdfsSource",
"distcpSettings": {
"distcpOptions": {},
"resourceManagerEndpoint": {},
"tempScriptPath": {}
},
"recursive": {}
For HiveSource, use:
"type": "HiveSource",
"additionalColumns": {},
"query": {},
"queryTimeout": {}
For HttpSource, use:
"type": "HttpSource",
"httpRequestTimeout": {}
For HubspotSource, use:
"type": "HubspotSource",
"additionalColumns": {},
"query": {},
"queryTimeout": {}
For ImpalaSource, use:
"type": "ImpalaSource",
"additionalColumns": {},
"query": {},
"queryTimeout": {}
For InformixSource, use:
"type": "InformixSource",
"additionalColumns": {},
"query": {},
"queryTimeout": {}
For JiraSource, use:
"type": "JiraSource",
"additionalColumns": {},
"query": {},
"queryTimeout": {}
For JsonSource, use:
"type": "JsonSource",
"additionalColumns": {},
"formatSettings": {
"compressionProperties": {
"type": "string"
// For remaining properties, see CompressionReadSettings objects
},
"type": "string",
"{customized property}": {}
},
"storeSettings": {
"disableMetricsCollection": {},
"maxConcurrentConnections": {},
"type": "string"
// For remaining properties, see StoreReadSettings objects
}
For LakeHouseTableSource, use:
"type": "LakeHouseTableSource",
"additionalColumns": {},
"timestampAsOf": {},
"versionAsOf": {}
For MagentoSource, use:
"type": "MagentoSource",
"additionalColumns": {},
"query": {},
"queryTimeout": {}
For MariaDBSource, use:
"type": "MariaDBSource",
"additionalColumns": {},
"query": {},
"queryTimeout": {}
For MarketoSource, use:
"type": "MarketoSource",
"additionalColumns": {},
"query": {},
"queryTimeout": {}
For MicrosoftAccessSource, use:
"type": "MicrosoftAccessSource",
"additionalColumns": {},
"query": {}
For MongoDbAtlasSource, use:
"type": "MongoDbAtlasSource",
"additionalColumns": {},
"batchSize": {},
"cursorMethods": {
"limit": {},
"project": {},
"skip": {},
"sort": {},
"{customized property}": {}
},
"filter": {},
"queryTimeout": {}
For MongoDbSource, use:
"type": "MongoDbSource",
"additionalColumns": {},
"query": {}
For MongoDbV2Source, use:
"type": "MongoDbV2Source",
"additionalColumns": {},
"batchSize": {},
"cursorMethods": {
"limit": {},
"project": {},
"skip": {},
"sort": {},
"{customized property}": {}
},
"filter": {},
"queryTimeout": {}
For MySqlSource, use:
"type": "MySqlSource",
"additionalColumns": {},
"query": {},
"queryTimeout": {}
For NetezzaSource, use:
"type": "NetezzaSource",
"additionalColumns": {},
"partitionOption": {},
"partitionSettings": {
"partitionColumnName": {},
"partitionLowerBound": {},
"partitionUpperBound": {}
},
"query": {},
"queryTimeout": {}
For ODataSource, use:
"type": "ODataSource",
"additionalColumns": {},
"httpRequestTimeout": {},
"query": {}
For OdbcSource, use:
"type": "OdbcSource",
"additionalColumns": {},
"query": {},
"queryTimeout": {}
For Office365Source, use:
"type": "Office365Source",
"allowedGroups": {},
"dateFilterColumn": {},
"endTime": {},
"outputColumns": {},
"startTime": {},
"userScopeFilterUri": {}
For OracleServiceCloudSource, use:
"type": "OracleServiceCloudSource",
"additionalColumns": {},
"query": {},
"queryTimeout": {}
For OracleSource, use:
"type": "OracleSource",
"additionalColumns": {},
"oracleReaderQuery": {},
"partitionOption": {},
"partitionSettings": {
"partitionColumnName": {},
"partitionLowerBound": {},
"partitionNames": {},
"partitionUpperBound": {}
},
"queryTimeout": {}
For OrcSource, use:
"type": "OrcSource",
"additionalColumns": {},
"storeSettings": {
"disableMetricsCollection": {},
"maxConcurrentConnections": {},
"type": "string"
// For remaining properties, see StoreReadSettings objects
}
For ParquetSource, use:
"type": "ParquetSource",
"additionalColumns": {},
"formatSettings": {
"compressionProperties": {
"type": "string"
// For remaining properties, see CompressionReadSettings objects
},
"type": "string",
"{customized property}": {}
},
"storeSettings": {
"disableMetricsCollection": {},
"maxConcurrentConnections": {},
"type": "string"
// For remaining properties, see StoreReadSettings objects
}
For PaypalSource, use:
"type": "PaypalSource",
"additionalColumns": {},
"query": {},
"queryTimeout": {}
For PhoenixSource, use:
"type": "PhoenixSource",
"additionalColumns": {},
"query": {},
"queryTimeout": {}
For PostgreSqlSource, use:
"type": "PostgreSqlSource",
"additionalColumns": {},
"query": {},
"queryTimeout": {}
For PostgreSqlV2Source, use:
"type": "PostgreSqlV2Source",
"additionalColumns": {},
"query": {},
"queryTimeout": {}
For PrestoSource, use:
"type": "PrestoSource",
"additionalColumns": {},
"query": {},
"queryTimeout": {}
For QuickBooksSource, use:
"type": "QuickBooksSource",
"additionalColumns": {},
"query": {},
"queryTimeout": {}
For RelationalSource, use:
"type": "RelationalSource",
"additionalColumns": {},
"query": {}
For ResponsysSource, use:
"type": "ResponsysSource",
"additionalColumns": {},
"query": {},
"queryTimeout": {}
For RestSource, use:
"type": "RestSource",
"additionalColumns": {},
"additionalHeaders": {},
"httpRequestTimeout": {},
"paginationRules": {},
"requestBody": {},
"requestInterval": {},
"requestMethod": {}
For SalesforceMarketingCloudSource, use:
"type": "SalesforceMarketingCloudSource",
"additionalColumns": {},
"query": {},
"queryTimeout": {}
For SalesforceServiceCloudSource, use:
"type": "SalesforceServiceCloudSource",
"additionalColumns": {},
"query": {},
"readBehavior": {}
For SalesforceServiceCloudV2Source, use:
"type": "SalesforceServiceCloudV2Source",
"additionalColumns": {},
"includeDeletedObjects": {},
"SOQLQuery": {}
For SalesforceSource, use:
"type": "SalesforceSource",
"additionalColumns": {},
"query": {},
"queryTimeout": {},
"readBehavior": {}
For SalesforceV2Source, use:
"type": "SalesforceV2Source",
"additionalColumns": {},
"includeDeletedObjects": {},
"queryTimeout": {},
"SOQLQuery": {}
For SapBwSource, use:
"type": "SapBwSource",
"additionalColumns": {},
"query": {},
"queryTimeout": {}
For SapCloudForCustomerSource, use:
"type": "SapCloudForCustomerSource",
"additionalColumns": {},
"httpRequestTimeout": {},
"query": {},
"queryTimeout": {}
For SapEccSource, use:
"type": "SapEccSource",
"additionalColumns": {},
"httpRequestTimeout": {},
"query": {},
"queryTimeout": {}
For SapHanaSource, use:
"type": "SapHanaSource",
"additionalColumns": {},
"packetSize": {},
"partitionOption": {},
"partitionSettings": {
"partitionColumnName": {}
},
"query": {},
"queryTimeout": {}
For SapOdpSource, use:
"type": "SapOdpSource",
"additionalColumns": {},
"extractionMode": {},
"projection": {},
"queryTimeout": {},
"selection": {},
"subscriberProcess": {}
For SapOpenHubSource, use:
"type": "SapOpenHubSource",
"additionalColumns": {},
"baseRequestId": {},
"customRfcReadTableFunctionModule": {},
"excludeLastRequest": {},
"queryTimeout": {},
"sapDataColumnDelimiter": {}
For SapTableSource, use:
"type": "SapTableSource",
"additionalColumns": {},
"batchSize": {},
"customRfcReadTableFunctionModule": {},
"partitionOption": {},
"partitionSettings": {
"maxPartitionsNumber": {},
"partitionColumnName": {},
"partitionLowerBound": {},
"partitionUpperBound": {}
},
"queryTimeout": {},
"rfcTableFields": {},
"rfcTableOptions": {},
"rowCount": {},
"rowSkips": {},
"sapDataColumnDelimiter": {}
For ServiceNowSource, use:
"type": "ServiceNowSource",
"additionalColumns": {},
"query": {},
"queryTimeout": {}
For ServiceNowV2Source, use:
"type": "ServiceNowV2Source",
"additionalColumns": {},
"expression": {
"operands": [
{
"operator": "string",
"type": "string",
"value": "string"
},
"queryTimeout": {}
For SharePointOnlineListSource, use:
"type": "SharePointOnlineListSource",
"httpRequestTimeout": {},
"query": {}
For ShopifySource, use:
"type": "ShopifySource",
"additionalColumns": {},
"query": {},
"queryTimeout": {}
For SnowflakeSource, use:
"type": "SnowflakeSource",
"exportSettings": {
"additionalCopyOptions": {
"{customized property}": {}
},
"additionalFormatOptions": {
"{customized property}": {}
},
"type": "string",
"{customized property}": {}
},
"query": {}
For SnowflakeV2Source, use:
"type": "SnowflakeV2Source",
"exportSettings": {
"additionalCopyOptions": {
"{customized property}": {}
},
"additionalFormatOptions": {
"{customized property}": {}
},
"type": "string",
"{customized property}": {}
},
"query": {}
For SparkSource, use:
"type": "SparkSource",
"additionalColumns": {},
"query": {},
"queryTimeout": {}
For SqlDWSource, use:
"type": "SqlDWSource",
"additionalColumns": {},
"isolationLevel": {},
"partitionOption": {},
"partitionSettings": {
"partitionColumnName": {},
"partitionLowerBound": {},
"partitionUpperBound": {}
},
"queryTimeout": {},
"sqlReaderQuery": {},
"sqlReaderStoredProcedureName": {},
"storedProcedureParameters": {}
For SqlMISource, use:
"type": "SqlMISource",
"additionalColumns": {},
"isolationLevel": {},
"partitionOption": {},
"partitionSettings": {
"partitionColumnName": {},
"partitionLowerBound": {},
"partitionUpperBound": {}
},
"produceAdditionalTypes": {},
"queryTimeout": {},
"sqlReaderQuery": {},
"sqlReaderStoredProcedureName": {},
"storedProcedureParameters": {}
For SqlServerSource, use:
"type": "SqlServerSource",
"additionalColumns": {},
"isolationLevel": {},
"partitionOption": {},
"partitionSettings": {
"partitionColumnName": {},
"partitionLowerBound": {},
"partitionUpperBound": {}
},
"produceAdditionalTypes": {},
"queryTimeout": {},
"sqlReaderQuery": {},
"sqlReaderStoredProcedureName": {},
"storedProcedureParameters": {}
For SqlSource, use:
"type": "SqlSource",
"additionalColumns": {},
"isolationLevel": {},
"partitionOption": {},
"partitionSettings": {
"partitionColumnName": {},
"partitionLowerBound": {},
"partitionUpperBound": {}
},
"queryTimeout": {},
"sqlReaderQuery": {},
"sqlReaderStoredProcedureName": {},
"storedProcedureParameters": {}
For SquareSource, use:
"type": "SquareSource",
"additionalColumns": {},
"query": {},
"queryTimeout": {}
For SybaseSource, use:
"type": "SybaseSource",
"additionalColumns": {},
"query": {},
"queryTimeout": {}
For TeradataSource, use:
"type": "TeradataSource",
"additionalColumns": {},
"partitionOption": {},
"partitionSettings": {
"partitionColumnName": {},
"partitionLowerBound": {},
"partitionUpperBound": {}
},
"query": {},
"queryTimeout": {}
For VerticaSource, use:
"type": "VerticaSource",
"additionalColumns": {},
"query": {},
"queryTimeout": {}
For WarehouseSource, use:
"type": "WarehouseSource",
"additionalColumns": {},
"isolationLevel": {},
"partitionOption": {},
"partitionSettings": {
"partitionColumnName": {},
"partitionLowerBound": {},
"partitionUpperBound": {}
},
"queryTimeout": {},
"sqlReaderQuery": {},
"sqlReaderStoredProcedureName": {},
"storedProcedureParameters": {}
For WebSource, use:
"type": "WebSource",
"additionalColumns": {}
For XeroSource, use:
"type": "XeroSource",
"additionalColumns": {},
"query": {},
"queryTimeout": {}
For XmlSource, use:
"type": "XmlSource",
"additionalColumns": {},
"formatSettings": {
"compressionProperties": {
"type": "string"
// For remaining properties, see CompressionReadSettings objects
},
"detectDataType": {},
"namespacePrefixes": {},
"namespaces": {},
"type": "string",
"validationMode": {},
"{customized property}": {}
},
"storeSettings": {
"disableMetricsCollection": {},
"maxConcurrentConnections": {},
"type": "string"
// For remaining properties, see StoreReadSettings objects
}
For ZohoSource, use:
"type": "ZohoSource",
"additionalColumns": {},
"query": {},
"queryTimeout": {}
Set the type property to specify the type of object.
For AmazonS3CompatibleReadSettings, use:
"type": "AmazonS3CompatibleReadSettings",
"deleteFilesAfterCompletion": {},
"enablePartitionDiscovery": {},
"fileListPath": {},
"modifiedDatetimeEnd": {},
"modifiedDatetimeStart": {},
"partitionRootPath": {},
"prefix": {},
"recursive": {},
"wildcardFileName": {},
"wildcardFolderPath": {}
For AmazonS3ReadSettings, use:
"type": "AmazonS3ReadSettings",
"deleteFilesAfterCompletion": {},
"enablePartitionDiscovery": {},
"fileListPath": {},
"modifiedDatetimeEnd": {},
"modifiedDatetimeStart": {},
"partitionRootPath": {},
"prefix": {},
"recursive": {},
"wildcardFileName": {},
"wildcardFolderPath": {}
For AzureBlobFSReadSettings, use:
"type": "AzureBlobFSReadSettings",
"deleteFilesAfterCompletion": {},
"enablePartitionDiscovery": {},
"fileListPath": {},
"modifiedDatetimeEnd": {},
"modifiedDatetimeStart": {},
"partitionRootPath": {},
"recursive": {},
"wildcardFileName": {},
"wildcardFolderPath": {}
For AzureBlobStorageReadSettings, use:
"type": "AzureBlobStorageReadSettings",
"deleteFilesAfterCompletion": {},
"enablePartitionDiscovery": {},
"fileListPath": {},
"modifiedDatetimeEnd": {},
"modifiedDatetimeStart": {},
"partitionRootPath": {},
"prefix": {},
"recursive": {},
"wildcardFileName": {},
"wildcardFolderPath": {}
For AzureDataLakeStoreReadSettings, use:
"type": "AzureDataLakeStoreReadSettings",
"deleteFilesAfterCompletion": {},
"enablePartitionDiscovery": {},
"fileListPath": {},
"listAfter": {},
"listBefore": {},
"modifiedDatetimeEnd": {},
"modifiedDatetimeStart": {},
"partitionRootPath": {},
"recursive": {},
"wildcardFileName": {},
"wildcardFolderPath": {}
For AzureFileStorageReadSettings, use:
"type": "AzureFileStorageReadSettings",
"deleteFilesAfterCompletion": {},
"enablePartitionDiscovery": {},
"fileListPath": {},
"modifiedDatetimeEnd": {},
"modifiedDatetimeStart": {},
"partitionRootPath": {},
"prefix": {},
"recursive": {},
"wildcardFileName": {},
"wildcardFolderPath": {}
For FileServerReadSettings, use:
"type": "FileServerReadSettings",
"deleteFilesAfterCompletion": {},
"enablePartitionDiscovery": {},
"fileFilter": {},
"fileListPath": {},
"modifiedDatetimeEnd": {},
"modifiedDatetimeStart": {},
"partitionRootPath": {},
"recursive": {},
"wildcardFileName": {},
"wildcardFolderPath": {}
For FtpReadSettings, use:
"type": "FtpReadSettings",
"deleteFilesAfterCompletion": {},
"disableChunking": {},
"enablePartitionDiscovery": {},
"fileListPath": {},
"partitionRootPath": {},
"recursive": {},
"useBinaryTransfer": {},
"wildcardFileName": {},
"wildcardFolderPath": {}
For GoogleCloudStorageReadSettings, use:
"type": "GoogleCloudStorageReadSettings",
"deleteFilesAfterCompletion": {},
"enablePartitionDiscovery": {},
"fileListPath": {},
"modifiedDatetimeEnd": {},
"modifiedDatetimeStart": {},
"partitionRootPath": {},
"prefix": {},
"recursive": {},
"wildcardFileName": {},
"wildcardFolderPath": {}
For HdfsReadSettings, use:
"type": "HdfsReadSettings",
"deleteFilesAfterCompletion": {},
"distcpSettings": {
"distcpOptions": {},
"resourceManagerEndpoint": {},
"tempScriptPath": {}
},
"enablePartitionDiscovery": {},
"fileListPath": {},
"modifiedDatetimeEnd": {},
"modifiedDatetimeStart": {},
"partitionRootPath": {},
"recursive": {},
"wildcardFileName": {},
"wildcardFolderPath": {}
For HttpReadSettings, use:
"type": "HttpReadSettings",
"additionalColumns": {},
"additionalHeaders": {},
"requestBody": {},
"requestMethod": {},
"requestTimeout": {}
For LakeHouseReadSettings, use:
"type": "LakeHouseReadSettings",
"deleteFilesAfterCompletion": {},
"enablePartitionDiscovery": {},
"fileListPath": {},
"modifiedDatetimeEnd": {},
"modifiedDatetimeStart": {},
"partitionRootPath": {},
"recursive": {},
"wildcardFileName": {},
"wildcardFolderPath": {}
For OracleCloudStorageReadSettings, use:
"type": "OracleCloudStorageReadSettings",
"deleteFilesAfterCompletion": {},
"enablePartitionDiscovery": {},
"fileListPath": {},
"modifiedDatetimeEnd": {},
"modifiedDatetimeStart": {},
"partitionRootPath": {},
"prefix": {},
"recursive": {},
"wildcardFileName": {},
"wildcardFolderPath": {}
For SftpReadSettings, use:
"type": "SftpReadSettings",
"deleteFilesAfterCompletion": {},
"disableChunking": {},
"enablePartitionDiscovery": {},
"fileListPath": {},
"modifiedDatetimeEnd": {},
"modifiedDatetimeStart": {},
"partitionRootPath": {},
"recursive": {},
"wildcardFileName": {},
"wildcardFolderPath": {}
Set the type property to specify the type of object.
For TarGZipReadSettings, use:
"type": "TarGZipReadSettings",
"preserveCompressionFileNameAsFolder": {}
For TarReadSettings, use:
"type": "TarReadSettings",
"preserveCompressionFileNameAsFolder": {}
For ZipDeflateReadSettings, use:
"type": "ZipDeflateReadSettings",
"preserveZipFileNameAsFolder": {}
Set the type property to specify the type of object.
For AzureKeyVaultSecret, use:
"type": "AzureKeyVaultSecret",
"secretName": {},
"secretVersion": {},
"store": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
}
For SecureString, use:
"type": "SecureString",
"value": "string"
Set the type property to specify the type of object.
For BinaryReadSettings, use:
"type": "BinaryReadSettings",
"compressionProperties": {
"type": "string"
// For remaining properties, see CompressionReadSettings objects
}
For DelimitedTextReadSettings, use:
"type": "DelimitedTextReadSettings",
"compressionProperties": {
"type": "string"
// For remaining properties, see CompressionReadSettings objects
},
"skipLineCount": {}
For JsonReadSettings, use:
"type": "JsonReadSettings",
"compressionProperties": {
"type": "string"
// For remaining properties, see CompressionReadSettings objects
}
For ParquetReadSettings, use:
"type": "ParquetReadSettings",
"compressionProperties": {
"type": "string"
// For remaining properties, see CompressionReadSettings objects
}
For XmlReadSettings, use:
"type": "XmlReadSettings",
"compressionProperties": {
"type": "string"
// For remaining properties, see CompressionReadSettings objects
},
"detectDataType": {},
"namespacePrefixes": {},
"namespaces": {},
"validationMode": {}
Name | Description | Value |
---|---|---|
type | The resource type | 'Microsoft.DataFactory/factories/pipelines' |
apiVersion | The resource api version | '2018-06-01' |
name | The resource name See how to set names and types for child resources in JSON ARM templates. |
string (required) Character limit: 1-260 Valid characters: Can't use: <>*#.%&:\\+?/ or control charactersStart with alphanumeric. |
properties | Properties of the pipeline. | Pipeline (required) |
Name | Description | Value |
---|---|---|
activities | List of activities in pipeline. | Activity[] |
annotations | List of tags that can be used for describing the Pipeline. | any[] |
concurrency | The max number of concurrent runs for the pipeline. | int Constraints: Min value = 1 |
description | The description of the pipeline. | string |
folder | The folder that this Pipeline is in. If not specified, Pipeline will appear at the root level. | PipelineFolder |
parameters | List of parameters for pipeline. | ParameterDefinitionSpecification |
policy | Pipeline Policy. | PipelinePolicy |
runDimensions | Dimensions emitted by Pipeline. | PipelineRunDimensions |
variables | List of variables for pipeline. | VariableDefinitionSpecification |
Name | Description | Value |
---|---|---|
dependsOn | Activity depends on condition. | ActivityDependency[] |
description | Activity description. | string |
name | Activity name. | string (required) |
onInactiveMarkAs | Status result of the activity when the state is set to Inactive. This is an optional property and if not provided when the activity is inactive, the status will be Succeeded by default. | 'Failed' 'Skipped' 'Succeeded' |
state | Activity state. This is an optional property and if not provided, the state will be Active by default. | 'Active' 'Inactive' |
userProperties | Activity user properties. | UserProperty[] |
type | Set the object type | AppendVariable AzureDataExplorerCommand AzureFunctionActivity AzureMLBatchExecution AzureMLExecutePipeline AzureMLUpdateResource Copy Custom DatabricksNotebook DatabricksSparkJar DatabricksSparkPython DataLakeAnalyticsU-SQL Delete ExecuteDataFlow ExecutePipeline ExecuteSSISPackage ExecuteWranglingDataflow Fail Filter ForEach GetMetadata HDInsightHive HDInsightMapReduce HDInsightPig HDInsightSpark HDInsightStreaming IfCondition Lookup Script SetVariable SparkJob SqlServerStoredProcedure Switch SynapseNotebook Until Validation Wait WebActivity WebHook (required) |
Name | Description | Value |
---|---|---|
activity | Activity name. | string (required) |
dependencyConditions | Match-Condition for the dependency. | String array containing any of: 'Completed' 'Failed' 'Skipped' 'Succeeded' (required) |
{customized property} |
Name | Description | Value |
---|---|---|
name | User property name. | string (required) |
value | User property value. Type: string (or Expression with resultType string). |
Name | Description | Value |
---|---|---|
type | Type of activity. | 'AppendVariable' (required) |
typeProperties | Append Variable activity properties. | AppendVariableActivityTypeProperties (required) |
Name | Description | Value |
---|---|---|
value | Value to be appended. Type: could be a static value matching type of the variable item or Expression with resultType matching type of the variable item | |
variableName | Name of the variable whose value needs to be appended to. | string |
Name | Description | Value |
---|---|---|
type | Type of activity. | 'AzureDataExplorerCommand' (required) |
linkedServiceName | Linked service reference. | LinkedServiceReference |
policy | Activity policy. | ActivityPolicy |
typeProperties | Azure Data Explorer command activity properties. | AzureDataExplorerCommandActivityTypeProperties (required) |
Name | Description | Value |
---|---|---|
parameters | Arguments for LinkedService. | ParameterValueSpecification |
referenceName | Reference LinkedService name. | string (required) |
type | Linked service reference type. | 'LinkedServiceReference' (required) |
Name | Description | Value |
---|---|---|
{customized property} |
Name | Description | Value |
---|---|---|
retry | Maximum ordinary retry attempts. Default is 0. Type: integer (or Expression with resultType integer), minimum: 0. | |
retryIntervalInSeconds | Interval between each retry attempt (in seconds). The default is 30 sec. | int Constraints: Min value = 30 Max value = 86400 |
secureInput | When set to true, Input from activity is considered as secure and will not be logged to monitoring. | bool |
secureOutput | When set to true, Output from activity is considered as secure and will not be logged to monitoring. | bool |
timeout | Specifies the timeout for the activity to run. The default timeout is 7 days. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | |
{customized property} |
Name | Description | Value |
---|---|---|
command | A control command, according to the Azure Data Explorer command syntax. Type: string (or Expression with resultType string). | |
commandTimeout | Control command timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9]))..) |
Name | Description | Value |
---|---|---|
type | Type of activity. | 'AzureFunctionActivity' (required) |
linkedServiceName | Linked service reference. | LinkedServiceReference |
policy | Activity policy. | ActivityPolicy |
typeProperties | Azure Function activity properties. | AzureFunctionActivityTypeProperties (required) |
Name | Description | Value |
---|---|---|
body | Represents the payload that will be sent to the endpoint. Required for POST/PUT method, not allowed for GET method Type: string (or Expression with resultType string). | |
functionName | Name of the Function that the Azure Function Activity will call. Type: string (or Expression with resultType string) | |
headers | Represents the headers that will be sent to the request. For example, to set the language and type on a request: "headers" : { "Accept-Language": "en-us", "Content-Type": "application/json" }. Type: string (or Expression with resultType string). | AzureFunctionActivityTypePropertiesHeaders |
method | Rest API method for target endpoint. | 'DELETE' 'GET' 'HEAD' 'OPTIONS' 'POST' 'PUT' 'TRACE' (required) |
Name | Description | Value |
---|---|---|
{customized property} | string |
Name | Description | Value |
---|---|---|
type | Type of activity. | 'AzureMLBatchExecution' (required) |
linkedServiceName | Linked service reference. | LinkedServiceReference |
policy | Activity policy. | ActivityPolicy |
typeProperties | Azure ML Batch Execution activity properties. | AzureMLBatchExecutionActivityTypeProperties (required) |
Name | Description | Value |
---|---|---|
globalParameters | Key,Value pairs to be passed to the Azure ML Batch Execution Service endpoint. Keys must match the names of web service parameters defined in the published Azure ML web service. Values will be passed in the GlobalParameters property of the Azure ML batch execution request. | AzureMLBatchExecutionActivityTypePropertiesGlobalPar... |
webServiceInputs | Key,Value pairs, mapping the names of Azure ML endpoint's Web Service Inputs to AzureMLWebServiceFile objects specifying the input Blob locations.. This information will be passed in the WebServiceInputs property of the Azure ML batch execution request. | AzureMLBatchExecutionActivityTypePropertiesWebServic... |
webServiceOutputs | Key,Value pairs, mapping the names of Azure ML endpoint's Web Service Outputs to AzureMLWebServiceFile objects specifying the output Blob locations. This information will be passed in the WebServiceOutputs property of the Azure ML batch execution request. | AzureMLBatchExecutionActivityTypePropertiesWebServic... |
Name | Description | Value |
---|---|---|
{customized property} |
Name | Description | Value |
---|---|---|
{customized property} | AzureMLWebServiceFile |
Name | Description | Value |
---|---|---|
filePath | The relative file path, including container name, in the Azure Blob Storage specified by the LinkedService. Type: string (or Expression with resultType string). | |
linkedServiceName | Reference to an Azure Storage LinkedService, where Azure ML WebService Input/Output file located. | LinkedServiceReference (required) |
Name | Description | Value |
---|---|---|
{customized property} | AzureMLWebServiceFile |
Name | Description | Value |
---|---|---|
type | Type of activity. | 'AzureMLExecutePipeline' (required) |
linkedServiceName | Linked service reference. | LinkedServiceReference |
policy | Activity policy. | ActivityPolicy |
typeProperties | Azure ML Execute Pipeline activity properties. | AzureMLExecutePipelineActivityTypeProperties (required) |
Name | Description | Value |
---|---|---|
continueOnStepFailure | Whether to continue execution of other steps in the PipelineRun if a step fails. This information will be passed in the continueOnStepFailure property of the published pipeline execution request. Type: boolean (or Expression with resultType boolean). | |
dataPathAssignments | Dictionary used for changing data path assignments without retraining. Values will be passed in the dataPathAssignments property of the published pipeline execution request. Type: object (or Expression with resultType object). | |
experimentName | Run history experiment name of the pipeline run. This information will be passed in the ExperimentName property of the published pipeline execution request. Type: string (or Expression with resultType string). | |
mlParentRunId | The parent Azure ML Service pipeline run id. This information will be passed in the ParentRunId property of the published pipeline execution request. Type: string (or Expression with resultType string). | |
mlPipelineEndpointId | ID of the published Azure ML pipeline endpoint. Type: string (or Expression with resultType string). | |
mlPipelineId | ID of the published Azure ML pipeline. Type: string (or Expression with resultType string). | |
mlPipelineParameters | Key,Value pairs to be passed to the published Azure ML pipeline endpoint. Keys must match the names of pipeline parameters defined in the published pipeline. Values will be passed in the ParameterAssignments property of the published pipeline execution request. Type: object with key value pairs (or Expression with resultType object). | |
version | Version of the published Azure ML pipeline endpoint. Type: string (or Expression with resultType string). |
Name | Description | Value |
---|---|---|
type | Type of activity. | 'AzureMLUpdateResource' (required) |
linkedServiceName | Linked service reference. | LinkedServiceReference |
policy | Activity policy. | ActivityPolicy |
typeProperties | Azure ML Update Resource management activity properties. | AzureMLUpdateResourceActivityTypeProperties (required) |
Name | Description | Value |
---|---|---|
trainedModelFilePath | The relative file path in trainedModelLinkedService to represent the .ilearner file that will be uploaded by the update operation. Type: string (or Expression with resultType string). | |
trainedModelLinkedServiceName | Name of Azure Storage linked service holding the .ilearner file that will be uploaded by the update operation. | LinkedServiceReference (required) |
trainedModelName | Name of the Trained Model module in the Web Service experiment to be updated. Type: string (or Expression with resultType string). |
Name | Description | Value |
---|---|---|
type | Type of activity. | 'Copy' (required) |
inputs | List of inputs for the activity. | DatasetReference[] |
linkedServiceName | Linked service reference. | LinkedServiceReference |
outputs | List of outputs for the activity. | DatasetReference[] |
policy | Activity policy. | ActivityPolicy |
typeProperties | Copy activity properties. | CopyActivityTypeProperties (required) |
Name | Description | Value |
---|---|---|
parameters | Arguments for dataset. | ParameterValueSpecification |
referenceName | Reference dataset name. | string (required) |
type | Dataset reference type. | 'DatasetReference' (required) |
Name | Description | Value |
---|---|---|
dataIntegrationUnits | Maximum number of data integration units that can be used to perform this data movement. Type: integer (or Expression with resultType integer), minimum: 0. | |
enableSkipIncompatibleRow | Whether to skip incompatible row. Default value is false. Type: boolean (or Expression with resultType boolean). | |
enableStaging | Specifies whether to copy data via an interim staging. Default value is false. Type: boolean (or Expression with resultType boolean). | |
logSettings | Log settings customer needs provide when enabling log. | LogSettings |
logStorageSettings | (Deprecated. Please use LogSettings) Log storage settings customer need to provide when enabling session log. | LogStorageSettings |
parallelCopies | Maximum number of concurrent sessions opened on the source or sink to avoid overloading the data store. Type: integer (or Expression with resultType integer), minimum: 0. | |
preserve | Preserve rules. | any[] |
preserveRules | Preserve Rules. | any[] |
redirectIncompatibleRowSettings | Redirect incompatible row settings when EnableSkipIncompatibleRow is true. | RedirectIncompatibleRowSettings |
sink | Copy activity sink. | CopySink (required) |
skipErrorFile | Specify the fault tolerance for data consistency. | SkipErrorFile |
source | Copy activity source. | CopySource (required) |
stagingSettings | Specifies interim staging settings when EnableStaging is true. | StagingSettings |
translator | Copy activity translator. If not specified, tabular translator is used. | |
validateDataConsistency | Whether to enable Data Consistency validation. Type: boolean (or Expression with resultType boolean). |
Name | Description | Value |
---|---|---|
copyActivityLogSettings | Specifies settings for copy activity log. | CopyActivityLogSettings |
enableCopyActivityLog | Specifies whether to enable copy activity log. Type: boolean (or Expression with resultType boolean). | |
logLocationSettings | Log location settings customer needs to provide when enabling log. | LogLocationSettings (required) |
Name | Description | Value |
---|---|---|
enableReliableLogging | Specifies whether to enable reliable logging. Type: boolean (or Expression with resultType boolean). | |
logLevel | Gets or sets the log level, support: Info, Warning. Type: string (or Expression with resultType string). |
Name | Description | Value |
---|---|---|
linkedServiceName | Log storage linked service reference. | LinkedServiceReference (required) |
path | The path to storage for storing detailed logs of activity execution. Type: string (or Expression with resultType string). |
Name | Description | Value |
---|---|---|
enableReliableLogging | Specifies whether to enable reliable logging. Type: boolean (or Expression with resultType boolean). | |
linkedServiceName | Log storage linked service reference. | LinkedServiceReference (required) |
logLevel | Gets or sets the log level, support: Info, Warning. Type: string (or Expression with resultType string). | |
path | The path to storage for storing detailed logs of activity execution. Type: string (or Expression with resultType string). | |
{customized property} |
Name | Description | Value |
---|---|---|
linkedServiceName | Name of the Azure Storage, Storage SAS, or Azure Data Lake Store linked service used for redirecting incompatible row. Must be specified if redirectIncompatibleRowSettings is specified. Type: string (or Expression with resultType string). | |
path | The path for storing the redirect incompatible row data. Type: string (or Expression with resultType string). | |
{customized property} |
Name | Description | Value |
---|---|---|
disableMetricsCollection | If true, disable data store metrics collection. Default is false. Type: boolean (or Expression with resultType boolean). | |
maxConcurrentConnections | The maximum concurrent connection count for the sink data store. Type: integer (or Expression with resultType integer). | |
sinkRetryCount | Sink retry count. Type: integer (or Expression with resultType integer). | |
sinkRetryWait | Sink retry wait. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | |
writeBatchSize | Write batch size. Type: integer (or Expression with resultType integer), minimum: 0. | |
writeBatchTimeout | Write batch timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | |
type | Set the object type | AvroSink AzureBlobFSSink AzureDatabricksDeltaLakeSink AzureDataExplorerSink AzureDataLakeStoreSink AzureMySqlSink AzurePostgreSqlSink AzureQueueSink AzureSearchIndexSink AzureSqlSink AzureTableSink BinarySink BlobSink CommonDataServiceForAppsSink CosmosDbMongoDbApiSink CosmosDbSqlApiSink DelimitedTextSink DocumentDbCollectionSink DynamicsCrmSink DynamicsSink FileSystemSink InformixSink JsonSink LakeHouseTableSink MicrosoftAccessSink MongoDbAtlasSink MongoDbV2Sink OdbcSink OracleSink OrcSink ParquetSink RestSink SalesforceServiceCloudSink SalesforceServiceCloudV2Sink SalesforceSink SalesforceV2Sink SapCloudForCustomerSink SnowflakeSink SnowflakeV2Sink SqlDWSink SqlMISink SqlServerSink SqlSink WarehouseSink (required) |
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'AvroSink' (required) |
formatSettings | Avro format settings. | AvroWriteSettings |
storeSettings | Avro store settings. | StoreWriteSettings |
Name | Description | Value |
---|---|---|
fileNamePrefix | Specifies the file name pattern {fileNamePrefix}_{fileIndex}.{fileExtension} when copy from non-file based store without partitionOptions. Type: string (or Expression with resultType string). | |
maxRowsPerFile | Limit the written file's row count to be smaller than or equal to the specified count. Type: integer (or Expression with resultType integer). | |
recordName | Top level record name in write result, which is required in AVRO spec. | string |
recordNamespace | Record namespace in the write result. | string |
type | The write setting type. | string (required) |
{customized property} |
Name | Description | Value |
---|---|---|
copyBehavior | The type of copy behavior for copy sink. | |
disableMetricsCollection | If true, disable data store metrics collection. Default is false. Type: boolean (or Expression with resultType boolean). | |
maxConcurrentConnections | The maximum concurrent connection count for the source data store. Type: integer (or Expression with resultType integer). | |
metadata | Specify the custom metadata to be added to sink data. Type: array of objects (or Expression with resultType array of objects). | MetadataItem[] |
type | Set the object type | AzureBlobFSWriteSettings AzureBlobStorageWriteSettings AzureDataLakeStoreWriteSettings AzureFileStorageWriteSettings FileServerWriteSettings LakeHouseWriteSettings SftpWriteSettings (required) |
Name | Description | Value |
---|---|---|
name | Metadata item key name. Type: string (or Expression with resultType string). | |
value | Metadata item value. Type: string (or Expression with resultType string). |
Name | Description | Value |
---|---|---|
type | The write setting type. | 'AzureBlobFSWriteSettings' (required) |
blockSizeInMB | Indicates the block size(MB) when writing data to blob. Type: integer (or Expression with resultType integer). |
Name | Description | Value |
---|---|---|
type | The write setting type. | 'AzureBlobStorageWriteSettings' (required) |
blockSizeInMB | Indicates the block size(MB) when writing data to blob. Type: integer (or Expression with resultType integer). |
Name | Description | Value |
---|---|---|
type | The write setting type. | 'AzureDataLakeStoreWriteSettings' (required) |
expiryDateTime | Specifies the expiry time of the written files. The time is applied to the UTC time zone in the format of "2018-12-01T05:00:00Z". Default value is NULL. Type: string (or Expression with resultType string). |
Name | Description | Value |
---|---|---|
type | The write setting type. | 'AzureFileStorageWriteSettings' (required) |
Name | Description | Value |
---|---|---|
type | The write setting type. | 'FileServerWriteSettings' (required) |
Name | Description | Value |
---|---|---|
type | The write setting type. | 'LakeHouseWriteSettings' (required) |
Name | Description | Value |
---|---|---|
type | The write setting type. | 'SftpWriteSettings' (required) |
operationTimeout | Specifies the timeout for writing each chunk to SFTP server. Default value: 01:00:00 (one hour). Type: string (or Expression with resultType string). | |
useTempFileRename | Upload to temporary file(s) and rename. Disable this option if your SFTP server doesn't support rename operation. Type: boolean (or Expression with resultType boolean). |
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'AzureBlobFSSink' (required) |
copyBehavior | The type of copy behavior for copy sink. Type: string (or Expression with resultType string). | |
metadata | Specify the custom metadata to be added to sink data. Type: array of objects (or Expression with resultType array of objects). | MetadataItem[] |
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'AzureDatabricksDeltaLakeSink' (required) |
importSettings | Azure Databricks Delta Lake import settings. | AzureDatabricksDeltaLakeImportCommand |
preCopyScript | SQL pre-copy script. Type: string (or Expression with resultType string). |
Name | Description | Value |
---|---|---|
dateFormat | Specify the date format for csv in Azure Databricks Delta Lake Copy. Type: string (or Expression with resultType string). | |
timestampFormat | Specify the timestamp format for csv in Azure Databricks Delta Lake Copy. Type: string (or Expression with resultType string). | |
type | The import setting type. | string (required) |
{customized property} |
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'AzureDataExplorerSink' (required) |
flushImmediately | If set to true, any aggregation will be skipped. Default is false. Type: boolean. | |
ingestionMappingAsJson | An explicit column mapping description provided in a json format. Type: string. | |
ingestionMappingName | A name of a pre-created csv mapping that was defined on the target Kusto table. Type: string. |
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'AzureDataLakeStoreSink' (required) |
copyBehavior | The type of copy behavior for copy sink. Type: string (or Expression with resultType string). | |
enableAdlsSingleFileParallel | Single File Parallel. |
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'AzureMySqlSink' (required) |
preCopyScript | A query to execute before starting the copy. Type: string (or Expression with resultType string). |
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'AzurePostgreSqlSink' (required) |
preCopyScript | A query to execute before starting the copy. Type: string (or Expression with resultType string). |
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'AzureQueueSink' (required) |
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'AzureSearchIndexSink' (required) |
writeBehavior | Specify the write behavior when upserting documents into Azure Search Index. | 'Merge' 'Upload' |
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'AzureSqlSink' (required) |
preCopyScript | SQL pre-copy script. Type: string (or Expression with resultType string). | |
sqlWriterStoredProcedureName | SQL writer stored procedure name. Type: string (or Expression with resultType string). | |
sqlWriterTableType | SQL writer table type. Type: string (or Expression with resultType string). | |
sqlWriterUseTableLock | Whether to use table lock during bulk copy. Type: boolean (or Expression with resultType boolean). | |
storedProcedureParameters | SQL stored procedure parameters. | |
storedProcedureTableTypeParameterName | The stored procedure parameter name of the table type. Type: string (or Expression with resultType string). | |
tableOption | The option to handle sink table, such as autoCreate. For now only 'autoCreate' value is supported. Type: string (or Expression with resultType string). | |
upsertSettings | SQL upsert settings. | SqlUpsertSettings |
writeBehavior | Write behavior when copying data into Azure SQL. Type: SqlWriteBehaviorEnum (or Expression with resultType SqlWriteBehaviorEnum) |
Name | Description | Value |
---|---|---|
interimSchemaName | Schema name for interim table. Type: string (or Expression with resultType string). | |
keys | Key column names for unique row identification. Type: array of strings (or Expression with resultType array of strings). | |
useTempDB | Specifies whether to use temp db for upsert interim table. Type: boolean (or Expression with resultType boolean). |
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'AzureTableSink' (required) |
azureTableDefaultPartitionKeyValue | Azure Table default partition key value. Type: string (or Expression with resultType string). | |
azureTableInsertType | Azure Table insert type. Type: string (or Expression with resultType string). | |
azureTablePartitionKeyName | Azure Table partition key name. Type: string (or Expression with resultType string). | |
azureTableRowKeyName | Azure Table row key name. Type: string (or Expression with resultType string). |
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'BinarySink' (required) |
storeSettings | Binary store settings. | StoreWriteSettings |
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'BlobSink' (required) |
blobWriterAddHeader | Blob writer add header. Type: boolean (or Expression with resultType boolean). | |
blobWriterDateTimeFormat | Blob writer date time format. Type: string (or Expression with resultType string). | |
blobWriterOverwriteFiles | Blob writer overwrite files. Type: boolean (or Expression with resultType boolean). | |
copyBehavior | The type of copy behavior for copy sink. | |
metadata | Specify the custom metadata to be added to sink data. Type: array of objects (or Expression with resultType array of objects). | MetadataItem[] |
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'CommonDataServiceForAppsSink' (required) |
alternateKeyName | The logical name of the alternate key which will be used when upserting records. Type: string (or Expression with resultType string). | |
ignoreNullValues | The flag indicating whether to ignore null values from input dataset (except key fields) during write operation. Default is false. Type: boolean (or Expression with resultType boolean). | |
writeBehavior | The write behavior for the operation. | 'Upsert' (required) |
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'CosmosDbMongoDbApiSink' (required) |
writeBehavior | Specifies whether the document with same key to be overwritten (upsert) rather than throw exception (insert). The default value is "insert". Type: string (or Expression with resultType string). Type: string (or Expression with resultType string). |
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'CosmosDbSqlApiSink' (required) |
writeBehavior | Describes how to write data to Azure Cosmos DB. Type: string (or Expression with resultType string). Allowed values: insert and upsert. |
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'DelimitedTextSink' (required) |
formatSettings | DelimitedText format settings. | DelimitedTextWriteSettings |
storeSettings | DelimitedText store settings. | StoreWriteSettings |
Name | Description | Value |
---|---|---|
fileExtension | The file extension used to create the files. Type: string (or Expression with resultType string). | |
fileNamePrefix | Specifies the file name pattern {fileNamePrefix}_{fileIndex}.{fileExtension} when copy from non-file based store without partitionOptions. Type: string (or Expression with resultType string). | |
maxRowsPerFile | Limit the written file's row count to be smaller than or equal to the specified count. Type: integer (or Expression with resultType integer). | |
quoteAllText | Indicates whether string values should always be enclosed with quotes. Type: boolean (or Expression with resultType boolean). | |
type | The write setting type. | string (required) |
{customized property} |
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'DocumentDbCollectionSink' (required) |
nestingSeparator | Nested properties separator. Default is . (dot). Type: string (or Expression with resultType string). | |
writeBehavior | Describes how to write data to Azure Cosmos DB. Type: string (or Expression with resultType string). Allowed values: insert and upsert. |
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'DynamicsCrmSink' (required) |
alternateKeyName | The logical name of the alternate key which will be used when upserting records. Type: string (or Expression with resultType string). | |
ignoreNullValues | The flag indicating whether to ignore null values from input dataset (except key fields) during write operation. Default is false. Type: boolean (or Expression with resultType boolean). | |
writeBehavior | The write behavior for the operation. | 'Upsert' (required) |
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'DynamicsSink' (required) |
alternateKeyName | The logical name of the alternate key which will be used when upserting records. Type: string (or Expression with resultType string). | |
ignoreNullValues | The flag indicating whether ignore null values from input dataset (except key fields) during write operation. Default is false. Type: boolean (or Expression with resultType boolean). | |
writeBehavior | The write behavior for the operation. | 'Upsert' (required) |
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'FileSystemSink' (required) |
copyBehavior | The type of copy behavior for copy sink. |
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'InformixSink' (required) |
preCopyScript | A query to execute before starting the copy. Type: string (or Expression with resultType string). |
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'JsonSink' (required) |
formatSettings | Json format settings. | JsonWriteSettings |
storeSettings | Json store settings. | StoreWriteSettings |
Name | Description | Value |
---|---|---|
filePattern | File pattern of JSON. This setting controls the way a collection of JSON objects will be treated. The default value is 'setOfObjects'. It is case-sensitive. | |
type | The write setting type. | string (required) |
{customized property} |
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'LakeHouseTableSink' (required) |
partitionNameList | Specify the partition column names from sink columns. Type: array of objects (or Expression with resultType array of objects). | |
partitionOption | Create partitions in folder structure based on one or multiple columns. Each distinct column value (pair) will be a new partition. Possible values include: "None", "PartitionByKey". | |
tableActionOption | The type of table action for LakeHouse Table sink. Possible values include: "None", "Append", "Overwrite". |
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'MicrosoftAccessSink' (required) |
preCopyScript | A query to execute before starting the copy. Type: string (or Expression with resultType string). |
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'MongoDbAtlasSink' (required) |
writeBehavior | Specifies whether the document with same key to be overwritten (upsert) rather than throw exception (insert). The default value is "insert". Type: string (or Expression with resultType string). Type: string (or Expression with resultType string). |
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'MongoDbV2Sink' (required) |
writeBehavior | Specifies whether the document with same key to be overwritten (upsert) rather than throw exception (insert). The default value is "insert". Type: string (or Expression with resultType string). Type: string (or Expression with resultType string). |
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'OdbcSink' (required) |
preCopyScript | A query to execute before starting the copy. Type: string (or Expression with resultType string). |
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'OracleSink' (required) |
preCopyScript | SQL pre-copy script. Type: string (or Expression with resultType string). |
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'OrcSink' (required) |
formatSettings | ORC format settings. | OrcWriteSettings |
storeSettings | ORC store settings. | StoreWriteSettings |
Name | Description | Value |
---|---|---|
fileNamePrefix | Specifies the file name pattern {fileNamePrefix}_{fileIndex}.{fileExtension} when copy from non-file based store without partitionOptions. Type: string (or Expression with resultType string). | |
maxRowsPerFile | Limit the written file's row count to be smaller than or equal to the specified count. Type: integer (or Expression with resultType integer). | |
type | The write setting type. | string (required) |
{customized property} |
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'ParquetSink' (required) |
formatSettings | Parquet format settings. | ParquetWriteSettings |
storeSettings | Parquet store settings. | StoreWriteSettings |
Name | Description | Value |
---|---|---|
fileNamePrefix | Specifies the file name pattern {fileNamePrefix}_{fileIndex}.{fileExtension} when copy from non-file based store without partitionOptions. Type: string (or Expression with resultType string). | |
maxRowsPerFile | Limit the written file's row count to be smaller than or equal to the specified count. Type: integer (or Expression with resultType integer). | |
type | The write setting type. | string (required) |
{customized property} |
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'RestSink' (required) |
additionalHeaders | The additional HTTP headers in the request to the RESTful API. Type: key value pairs (value should be string type). | |
httpCompressionType | Http Compression Type to Send data in compressed format with Optimal Compression Level, Default is None. And The Only Supported option is Gzip. Type: string (or Expression with resultType string). | |
httpRequestTimeout | The timeout (TimeSpan) to get an HTTP response. It is the timeout to get a response, not the timeout to read response data. Default value: 00:01:40. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | |
requestInterval | The time to await before sending next request, in milliseconds | |
requestMethod | The HTTP method used to call the RESTful API. The default is POST. Type: string (or Expression with resultType string). |
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'SalesforceServiceCloudSink' (required) |
externalIdFieldName | The name of the external ID field for upsert operation. Default value is 'Id' column. Type: string (or Expression with resultType string). | |
ignoreNullValues | The flag indicating whether or not to ignore null values from input dataset (except key fields) during write operation. Default value is false. If set it to true, it means ADF will leave the data in the destination object unchanged when doing upsert/update operation and insert defined default value when doing insert operation, versus ADF will update the data in the destination object to NULL when doing upsert/update operation and insert NULL value when doing insert operation. Type: boolean (or Expression with resultType boolean). | |
writeBehavior | The write behavior for the operation. Default is Insert. | 'Insert' 'Upsert' |
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'SalesforceServiceCloudV2Sink' (required) |
externalIdFieldName | The name of the external ID field for upsert operation. Default value is 'Id' column. Type: string (or Expression with resultType string). | |
ignoreNullValues | The flag indicating whether or not to ignore null values from input dataset (except key fields) during write operation. Default value is false. If set it to true, it means ADF will leave the data in the destination object unchanged when doing upsert/update operation and insert defined default value when doing insert operation, versus ADF will update the data in the destination object to NULL when doing upsert/update operation and insert NULL value when doing insert operation. Type: boolean (or Expression with resultType boolean). | |
writeBehavior | The write behavior for the operation. Default is Insert. | 'Insert' 'Upsert' |
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'SalesforceSink' (required) |
externalIdFieldName | The name of the external ID field for upsert operation. Default value is 'Id' column. Type: string (or Expression with resultType string). | |
ignoreNullValues | The flag indicating whether or not to ignore null values from input dataset (except key fields) during write operation. Default value is false. If set it to true, it means ADF will leave the data in the destination object unchanged when doing upsert/update operation and insert defined default value when doing insert operation, versus ADF will update the data in the destination object to NULL when doing upsert/update operation and insert NULL value when doing insert operation. Type: boolean (or Expression with resultType boolean). | |
writeBehavior | The write behavior for the operation. Default is Insert. | 'Insert' 'Upsert' |
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'SalesforceV2Sink' (required) |
externalIdFieldName | The name of the external ID field for upsert operation. Default value is 'Id' column. Type: string (or Expression with resultType string). | |
ignoreNullValues | The flag indicating whether or not to ignore null values from input dataset (except key fields) during write operation. Default value is false. If set it to true, it means ADF will leave the data in the destination object unchanged when doing upsert/update operation and insert defined default value when doing insert operation, versus ADF will update the data in the destination object to NULL when doing upsert/update operation and insert NULL value when doing insert operation. Type: boolean (or Expression with resultType boolean). | |
writeBehavior | The write behavior for the operation. Default is Insert. | 'Insert' 'Upsert' |
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'SapCloudForCustomerSink' (required) |
httpRequestTimeout | The timeout (TimeSpan) to get an HTTP response. It is the timeout to get a response, not the timeout to read response data. Default value: 00:05:00. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | |
writeBehavior | The write behavior for the operation. Default is 'Insert'. | 'Insert' 'Update' |
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'SnowflakeSink' (required) |
importSettings | Snowflake import settings. | SnowflakeImportCopyCommand |
preCopyScript | SQL pre-copy script. Type: string (or Expression with resultType string). |
Name | Description | Value |
---|---|---|
additionalCopyOptions | Additional copy options directly passed to snowflake Copy Command. Type: key value pairs (value should be string type) (or Expression with resultType object). Example: "additionalCopyOptions": { "DATE_FORMAT": "MM/DD/YYYY", "TIME_FORMAT": "'HH24:MI:SS.FF'" } | SnowflakeImportCopyCommandAdditionalCopyOptions |
additionalFormatOptions | Additional format options directly passed to snowflake Copy Command. Type: key value pairs (value should be string type) (or Expression with resultType object). Example: "additionalFormatOptions": { "FORCE": "TRUE", "LOAD_UNCERTAIN_FILES": "'FALSE'" } | SnowflakeImportCopyCommandAdditionalFormatOptions |
type | The import setting type. | string (required) |
{customized property} |
Name | Description | Value |
---|---|---|
{customized property} |
Name | Description | Value |
---|---|---|
{customized property} |
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'SnowflakeV2Sink' (required) |
importSettings | Snowflake import settings. | SnowflakeImportCopyCommand |
preCopyScript | SQL pre-copy script. Type: string (or Expression with resultType string). |
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'SqlDWSink' (required) |
allowCopyCommand | Indicates to use Copy Command to copy data into SQL Data Warehouse. Type: boolean (or Expression with resultType boolean). | |
allowPolyBase | Indicates to use PolyBase to copy data into SQL Data Warehouse when applicable. Type: boolean (or Expression with resultType boolean). | |
copyCommandSettings | Specifies Copy Command related settings when allowCopyCommand is true. | DWCopyCommandSettings |
polyBaseSettings | Specifies PolyBase-related settings when allowPolyBase is true. | PolybaseSettings |
preCopyScript | SQL pre-copy script. Type: string (or Expression with resultType string). | |
sqlWriterUseTableLock | Whether to use table lock during bulk copy. Type: boolean (or Expression with resultType boolean). | |
tableOption | The option to handle sink table, such as autoCreate. For now only 'autoCreate' value is supported. Type: string (or Expression with resultType string). | |
upsertSettings | SQL DW upsert settings. | SqlDWUpsertSettings |
writeBehavior | Write behavior when copying data into azure SQL DW. Type: SqlDWWriteBehaviorEnum (or Expression with resultType SqlDWWriteBehaviorEnum) |
Name | Description | Value |
---|---|---|
additionalOptions | Additional options directly passed to SQL DW in Copy Command. Type: key value pairs (value should be string type) (or Expression with resultType object). Example: "additionalOptions": { "MAXERRORS": "1000", "DATEFORMAT": "'ymd'" } | DWCopyCommandSettingsAdditionalOptions |
defaultValues | Specifies the default values for each target column in SQL DW. The default values in the property overwrite the DEFAULT constraint set in the DB, and identity column cannot have a default value. Type: array of objects (or Expression with resultType array of objects). | DWCopyCommandDefaultValue[] |
Name | Description | Value |
---|---|---|
{customized property} | string |
Name | Description | Value |
---|---|---|
columnName | Column name. Type: object (or Expression with resultType string). | |
defaultValue | The default value of the column. Type: object (or Expression with resultType string). |
Name | Description | Value |
---|---|---|
rejectSampleValue | Determines the number of rows to attempt to retrieve before the PolyBase recalculates the percentage of rejected rows. Type: integer (or Expression with resultType integer), minimum: 0. | |
rejectType | Reject type. | 'percentage' 'value' |
rejectValue | Specifies the value or the percentage of rows that can be rejected before the query fails. Type: number (or Expression with resultType number), minimum: 0. | |
useTypeDefault | Specifies how to handle missing values in delimited text files when PolyBase retrieves data from the text file. Type: boolean (or Expression with resultType boolean). | |
{customized property} |
Name | Description | Value |
---|---|---|
interimSchemaName | Schema name for interim table. Type: string (or Expression with resultType string). | |
keys | Key column names for unique row identification. Type: array of strings (or Expression with resultType array of strings). |
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'SqlMISink' (required) |
preCopyScript | SQL pre-copy script. Type: string (or Expression with resultType string). | |
sqlWriterStoredProcedureName | SQL writer stored procedure name. Type: string (or Expression with resultType string). | |
sqlWriterTableType | SQL writer table type. Type: string (or Expression with resultType string). | |
sqlWriterUseTableLock | Whether to use table lock during bulk copy. Type: boolean (or Expression with resultType boolean). | |
storedProcedureParameters | SQL stored procedure parameters. | |
storedProcedureTableTypeParameterName | The stored procedure parameter name of the table type. Type: string (or Expression with resultType string). | |
tableOption | The option to handle sink table, such as autoCreate. For now only 'autoCreate' value is supported. Type: string (or Expression with resultType string). | |
upsertSettings | SQL upsert settings. | SqlUpsertSettings |
writeBehavior | White behavior when copying data into azure SQL MI. Type: string (or Expression with resultType string) |
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'SqlServerSink' (required) |
preCopyScript | SQL pre-copy script. Type: string (or Expression with resultType string). | |
sqlWriterStoredProcedureName | SQL writer stored procedure name. Type: string (or Expression with resultType string). | |
sqlWriterTableType | SQL writer table type. Type: string (or Expression with resultType string). | |
sqlWriterUseTableLock | Whether to use table lock during bulk copy. Type: boolean (or Expression with resultType boolean). | |
storedProcedureParameters | SQL stored procedure parameters. | |
storedProcedureTableTypeParameterName | The stored procedure parameter name of the table type. Type: string (or Expression with resultType string). | |
tableOption | The option to handle sink table, such as autoCreate. For now only 'autoCreate' value is supported. Type: string (or Expression with resultType string). | |
upsertSettings | SQL upsert settings. | SqlUpsertSettings |
writeBehavior | Write behavior when copying data into sql server. Type: string (or Expression with resultType string). |
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'SqlSink' (required) |
preCopyScript | SQL pre-copy script. Type: string (or Expression with resultType string). | |
sqlWriterStoredProcedureName | SQL writer stored procedure name. Type: string (or Expression with resultType string). | |
sqlWriterTableType | SQL writer table type. Type: string (or Expression with resultType string). | |
sqlWriterUseTableLock | Whether to use table lock during bulk copy. Type: boolean (or Expression with resultType boolean). | |
storedProcedureParameters | SQL stored procedure parameters. | |
storedProcedureTableTypeParameterName | The stored procedure parameter name of the table type. Type: string (or Expression with resultType string). | |
tableOption | The option to handle sink table, such as autoCreate. For now only 'autoCreate' value is supported. Type: string (or Expression with resultType string). | |
upsertSettings | SQL upsert settings. | SqlUpsertSettings |
writeBehavior | Write behavior when copying data into sql. Type: string (or Expression with resultType string). |
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'WarehouseSink' (required) |
allowCopyCommand | Indicates to use Copy Command to copy data into SQL Data Warehouse. Type: boolean (or Expression with resultType boolean). | |
copyCommandSettings | Specifies Copy Command related settings when allowCopyCommand is true. | DWCopyCommandSettings |
preCopyScript | SQL pre-copy script. Type: string (or Expression with resultType string). | |
tableOption | The option to handle sink table, such as autoCreate. For now only 'autoCreate' value is supported. Type: string (or Expression with resultType string). | |
writeBehavior | Write behavior when copying data into azure Microsoft Fabric Data Warehouse. Type: DWWriteBehaviorEnum (or Expression with resultType DWWriteBehaviorEnum) |
Name | Description | Value |
---|---|---|
dataInconsistency | Skip if source/sink file changed by other concurrent write. Default is false. Type: boolean (or Expression with resultType boolean). | |
fileMissing | Skip if file is deleted by other client during copy. Default is true. Type: boolean (or Expression with resultType boolean). |
Name | Description | Value |
---|---|---|
type | Copy source type. | 'AmazonMWSSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
Name | Description | Value |
---|---|---|
type | Copy source type. | 'AmazonRdsForOracleSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
oracleReaderQuery | AmazonRdsForOracle reader query. Type: string (or Expression with resultType string). | |
partitionOption | The partition mechanism that will be used for AmazonRdsForOracle read in parallel. Type: string (or Expression with resultType string). | |
partitionSettings | The settings that will be leveraged for AmazonRdsForOracle source partitioning. | AmazonRdsForOraclePartitionSettings |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
Name | Description | Value |
---|---|---|
partitionColumnName | The name of the column in integer type that will be used for proceeding range partitioning. Type: string (or Expression with resultType string). | |
partitionLowerBound | The minimum value of column specified in partitionColumnName that will be used for proceeding range partitioning. Type: string (or Expression with resultType string). | |
partitionNames | Names of the physical partitions of AmazonRdsForOracle table. | |
partitionUpperBound | The maximum value of column specified in partitionColumnName that will be used for proceeding range partitioning. Type: string (or Expression with resultType string). |
Name | Description | Value |
---|---|---|
type | Copy source type. | 'AmazonRdsForSqlServerSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
isolationLevel | Specifies the transaction locking behavior for the SQL source. Allowed values: ReadCommitted/ReadUncommitted/RepeatableRead/Serializable/Snapshot. The default value is ReadCommitted. Type: string (or Expression with resultType string). | |
partitionOption | The partition mechanism that will be used for Sql read in parallel. Possible values include: "None", "PhysicalPartitionsOfTable", "DynamicRange". | |
partitionSettings | The settings that will be leveraged for Sql source partitioning. | SqlPartitionSettings |
produceAdditionalTypes | Which additional types to produce. | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | |
sqlReaderQuery | SQL reader query. Type: string (or Expression with resultType string). | |
sqlReaderStoredProcedureName | Name of the stored procedure for a SQL Database source. This cannot be used at the same time as SqlReaderQuery. Type: string (or Expression with resultType string). | |
storedProcedureParameters | Value and type setting for stored procedure parameters. Example: "{Parameter1: {value: "1", type: "int"}}". |
Name | Description | Value |
---|---|---|
partitionColumnName | The name of the column in integer or datetime type that will be used for proceeding partitioning. If not specified, the primary key of the table is auto-detected and used as the partition column. Type: string (or Expression with resultType string). | |
partitionLowerBound | The minimum value of the partition column for partition range splitting. This value is used to decide the partition stride, not for filtering the rows in table. All rows in the table or query result will be partitioned and copied. Type: string (or Expression with resultType string). | |
partitionUpperBound | The maximum value of the partition column for partition range splitting. This value is used to decide the partition stride, not for filtering the rows in table. All rows in the table or query result will be partitioned and copied. Type: string (or Expression with resultType string). |
Name | Description | Value |
---|---|---|
type | Copy source type. | 'AmazonRedshiftSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | Database query. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | |
redshiftUnloadSettings | The Amazon S3 settings needed for the interim Amazon S3 when copying from Amazon Redshift with unload. With this, data from Amazon Redshift source will be unloaded into S3 first and then copied into the targeted sink from the interim S3. | RedshiftUnloadSettings |
Name | Description | Value |
---|---|---|
bucketName | The bucket of the interim Amazon S3 which will be used to store the unloaded data from Amazon Redshift source. The bucket must be in the same region as the Amazon Redshift source. Type: string (or Expression with resultType string). | |
s3LinkedServiceName | The name of the Amazon S3 linked service which will be used for the unload operation when copying from the Amazon Redshift source. | LinkedServiceReference (required) |
Name | Description | Value |
---|---|---|
type | Copy source type. | 'AvroSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
storeSettings | Avro store settings. | StoreReadSettings |
Name | Description | Value |
---|---|---|
disableMetricsCollection | If true, disable data store metrics collection. Default is false. Type: boolean (or Expression with resultType boolean). | |
maxConcurrentConnections | The maximum concurrent connection count for the source data store. Type: integer (or Expression with resultType integer). | |
type | Set the object type | AmazonS3CompatibleReadSettings AmazonS3ReadSettings AzureBlobFSReadSettings AzureBlobStorageReadSettings AzureDataLakeStoreReadSettings AzureFileStorageReadSettings FileServerReadSettings FtpReadSettings GoogleCloudStorageReadSettings HdfsReadSettings HttpReadSettings LakeHouseReadSettings OracleCloudStorageReadSettings SftpReadSettings (required) |
Name | Description | Value |
---|---|---|
type | The read setting type. | 'AmazonS3CompatibleReadSettings' (required) |
deleteFilesAfterCompletion | Indicates whether the source files need to be deleted after copy completion. Default is false. Type: boolean (or Expression with resultType boolean). | |
enablePartitionDiscovery | Indicates whether to enable partition discovery. Type: boolean (or Expression with resultType boolean). | |
fileListPath | Point to a text file that lists each file (relative path to the path configured in the dataset) that you want to copy. Type: string (or Expression with resultType string). | |
modifiedDatetimeEnd | The end of file's modified datetime. Type: string (or Expression with resultType string). | |
modifiedDatetimeStart | The start of file's modified datetime. Type: string (or Expression with resultType string). | |
partitionRootPath | Specify the root path where partition discovery starts from. Type: string (or Expression with resultType string). | |
prefix | The prefix filter for the S3 Compatible object name. Type: string (or Expression with resultType string). | |
recursive | If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean). | |
wildcardFileName | Amazon S3 Compatible wildcardFileName. Type: string (or Expression with resultType string). | |
wildcardFolderPath | Amazon S3 Compatible wildcardFolderPath. Type: string (or Expression with resultType string). |
Name | Description | Value |
---|---|---|
type | The read setting type. | 'AmazonS3ReadSettings' (required) |
deleteFilesAfterCompletion | Indicates whether the source files need to be deleted after copy completion. Default is false. Type: boolean (or Expression with resultType boolean). | |
enablePartitionDiscovery | Indicates whether to enable partition discovery. Type: boolean (or Expression with resultType boolean). | |
fileListPath | Point to a text file that lists each file (relative path to the path configured in the dataset) that you want to copy. Type: string (or Expression with resultType string). | |
modifiedDatetimeEnd | The end of file's modified datetime. Type: string (or Expression with resultType string). | |
modifiedDatetimeStart | The start of file's modified datetime. Type: string (or Expression with resultType string). | |
partitionRootPath | Specify the root path where partition discovery starts from. Type: string (or Expression with resultType string). | |
prefix | The prefix filter for the S3 object name. Type: string (or Expression with resultType string). | |
recursive | If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean). | |
wildcardFileName | AmazonS3 wildcardFileName. Type: string (or Expression with resultType string). | |
wildcardFolderPath | AmazonS3 wildcardFolderPath. Type: string (or Expression with resultType string). |
Name | Description | Value |
---|---|---|
type | The read setting type. | 'AzureBlobFSReadSettings' (required) |
deleteFilesAfterCompletion | Indicates whether the source files need to be deleted after copy completion. Default is false. Type: boolean (or Expression with resultType boolean). | |
enablePartitionDiscovery | Indicates whether to enable partition discovery. Type: boolean (or Expression with resultType boolean). | |
fileListPath | Point to a text file that lists each file (relative path to the path configured in the dataset) that you want to copy. Type: string (or Expression with resultType string). | |
modifiedDatetimeEnd | The end of file's modified datetime. Type: string (or Expression with resultType string). | |
modifiedDatetimeStart | The start of file's modified datetime. Type: string (or Expression with resultType string). | |
partitionRootPath | Specify the root path where partition discovery starts from. Type: string (or Expression with resultType string). | |
recursive | If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean). | |
wildcardFileName | Azure blobFS wildcardFileName. Type: string (or Expression with resultType string). | |
wildcardFolderPath | Azure blobFS wildcardFolderPath. Type: string (or Expression with resultType string). |
Name | Description | Value |
---|---|---|
type | The read setting type. | 'AzureBlobStorageReadSettings' (required) |
deleteFilesAfterCompletion | Indicates whether the source files need to be deleted after copy completion. Default is false. Type: boolean (or Expression with resultType boolean). | |
enablePartitionDiscovery | Indicates whether to enable partition discovery. Type: boolean (or Expression with resultType boolean). | |
fileListPath | Point to a text file that lists each file (relative path to the path configured in the dataset) that you want to copy. Type: string (or Expression with resultType string). | |
modifiedDatetimeEnd | The end of file's modified datetime. Type: string (or Expression with resultType string). | |
modifiedDatetimeStart | The start of file's modified datetime. Type: string (or Expression with resultType string). | |
partitionRootPath | Specify the root path where partition discovery starts from. Type: string (or Expression with resultType string). | |
prefix | The prefix filter for the Azure Blob name. Type: string (or Expression with resultType string). | |
recursive | If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean). | |
wildcardFileName | Azure blob wildcardFileName. Type: string (or Expression with resultType string). | |
wildcardFolderPath | Azure blob wildcardFolderPath. Type: string (or Expression with resultType string). |
Name | Description | Value |
---|---|---|
type | The read setting type. | 'AzureDataLakeStoreReadSettings' (required) |
deleteFilesAfterCompletion | Indicates whether the source files need to be deleted after copy completion. Default is false. Type: boolean (or Expression with resultType boolean). | |
enablePartitionDiscovery | Indicates whether to enable partition discovery. Type: boolean (or Expression with resultType boolean). | |
fileListPath | Point to a text file that lists each file (relative path to the path configured in the dataset) that you want to copy. Type: string (or Expression with resultType string). | |
listAfter | Lists files after the value (exclusive) based on file/folder namesâ lexicographical order. Applies under the folderPath in data set, and filter files/sub-folders under the folderPath. Type: string (or Expression with resultType string). | |
listBefore | Lists files before the value (inclusive) based on file/folder namesâ lexicographical order. Applies under the folderPath in data set, and filter files/sub-folders under the folderPath. Type: string (or Expression with resultType string). | |
modifiedDatetimeEnd | The end of file's modified datetime. Type: string (or Expression with resultType string). | |
modifiedDatetimeStart | The start of file's modified datetime. Type: string (or Expression with resultType string). | |
partitionRootPath | Specify the root path where partition discovery starts from. Type: string (or Expression with resultType string). | |
recursive | If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean). | |
wildcardFileName | ADLS wildcardFileName. Type: string (or Expression with resultType string). | |
wildcardFolderPath | ADLS wildcardFolderPath. Type: string (or Expression with resultType string). |
Name | Description | Value |
---|---|---|
type | The read setting type. | 'AzureFileStorageReadSettings' (required) |
deleteFilesAfterCompletion | Indicates whether the source files need to be deleted after copy completion. Default is false. Type: boolean (or Expression with resultType boolean). | |
enablePartitionDiscovery | Indicates whether to enable partition discovery. Type: boolean (or Expression with resultType boolean). | |
fileListPath | Point to a text file that lists each file (relative path to the path configured in the dataset) that you want to copy. Type: string (or Expression with resultType string). | |
modifiedDatetimeEnd | The end of file's modified datetime. Type: string (or Expression with resultType string). | |
modifiedDatetimeStart | The start of file's modified datetime. Type: string (or Expression with resultType string). | |
partitionRootPath | Specify the root path where partition discovery starts from. Type: string (or Expression with resultType string). | |
prefix | The prefix filter for the Azure File name starting from root path. Type: string (or Expression with resultType string). | |
recursive | If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean). | |
wildcardFileName | Azure File Storage wildcardFileName. Type: string (or Expression with resultType string). | |
wildcardFolderPath | Azure File Storage wildcardFolderPath. Type: string (or Expression with resultType string). |
Name | Description | Value |
---|---|---|
type | The read setting type. | 'FileServerReadSettings' (required) |
deleteFilesAfterCompletion | Indicates whether the source files need to be deleted after copy completion. Default is false. Type: boolean (or Expression with resultType boolean). | |
enablePartitionDiscovery | Indicates whether to enable partition discovery. Type: boolean (or Expression with resultType boolean). | |
fileFilter | Specify a filter to be used to select a subset of files in the folderPath rather than all files. Type: string (or Expression with resultType string). | |
fileListPath | Point to a text file that lists each file (relative path to the path configured in the dataset) that you want to copy. Type: string (or Expression with resultType string). | |
modifiedDatetimeEnd | The end of file's modified datetime. Type: string (or Expression with resultType string). | |
modifiedDatetimeStart | The start of file's modified datetime. Type: string (or Expression with resultType string). | |
partitionRootPath | Specify the root path where partition discovery starts from. Type: string (or Expression with resultType string). | |
recursive | If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean). | |
wildcardFileName | FileServer wildcardFileName. Type: string (or Expression with resultType string). | |
wildcardFolderPath | FileServer wildcardFolderPath. Type: string (or Expression with resultType string). |
Name | Description | Value |
---|---|---|
type | The read setting type. | 'FtpReadSettings' (required) |
deleteFilesAfterCompletion | Indicates whether the source files need to be deleted after copy completion. Default is false. Type: boolean (or Expression with resultType boolean). | |
disableChunking | If true, disable parallel reading within each file. Default is false. Type: boolean (or Expression with resultType boolean). | |
enablePartitionDiscovery | Indicates whether to enable partition discovery. Type: boolean (or Expression with resultType boolean). | |
fileListPath | Point to a text file that lists each file (relative path to the path configured in the dataset) that you want to copy. Type: string (or Expression with resultType string). | |
partitionRootPath | Specify the root path where partition discovery starts from. Type: string (or Expression with resultType string). | |
recursive | If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean). | |
useBinaryTransfer | Specify whether to use binary transfer mode for FTP stores. Type: boolean (or Expression with resultType boolean). | |
wildcardFileName | Ftp wildcardFileName. Type: string (or Expression with resultType string). | |
wildcardFolderPath | Ftp wildcardFolderPath. Type: string (or Expression with resultType string). |
Name | Description | Value |
---|---|---|
type | The read setting type. | 'GoogleCloudStorageReadSettings' (required) |
deleteFilesAfterCompletion | Indicates whether the source files need to be deleted after copy completion. Default is false. Type: boolean (or Expression with resultType boolean). | |
enablePartitionDiscovery | Indicates whether to enable partition discovery. Type: boolean (or Expression with resultType boolean). | |
fileListPath | Point to a text file that lists each file (relative path to the path configured in the dataset) that you want to copy. Type: string (or Expression with resultType string). | |
modifiedDatetimeEnd | The end of file's modified datetime. Type: string (or Expression with resultType string). | |
modifiedDatetimeStart | The start of file's modified datetime. Type: string (or Expression with resultType string). | |
partitionRootPath | Specify the root path where partition discovery starts from. Type: string (or Expression with resultType string). | |
prefix | The prefix filter for the Google Cloud Storage object name. Type: string (or Expression with resultType string). | |
recursive | If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean). | |
wildcardFileName | Google Cloud Storage wildcardFileName. Type: string (or Expression with resultType string). | |
wildcardFolderPath | Google Cloud Storage wildcardFolderPath. Type: string (or Expression with resultType string). |
Name | Description | Value |
---|---|---|
type | The read setting type. | 'HdfsReadSettings' (required) |
deleteFilesAfterCompletion | Indicates whether the source files need to be deleted after copy completion. Default is false. Type: boolean (or Expression with resultType boolean). | |
distcpSettings | Specifies Distcp-related settings. | DistcpSettings |
enablePartitionDiscovery | Indicates whether to enable partition discovery. Type: boolean (or Expression with resultType boolean). | |
fileListPath | Point to a text file that lists each file (relative path to the path configured in the dataset) that you want to copy. Type: string (or Expression with resultType string). | |
modifiedDatetimeEnd | The end of file's modified datetime. Type: string (or Expression with resultType string). | |
modifiedDatetimeStart | The start of file's modified datetime. Type: string (or Expression with resultType string). | |
partitionRootPath | Specify the root path where partition discovery starts from. Type: string (or Expression with resultType string). | |
recursive | If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean). | |
wildcardFileName | HDFS wildcardFileName. Type: string (or Expression with resultType string). | |
wildcardFolderPath | HDFS wildcardFolderPath. Type: string (or Expression with resultType string). |
Name | Description | Value |
---|---|---|
distcpOptions | Specifies the Distcp options. Type: string (or Expression with resultType string). | |
resourceManagerEndpoint | Specifies the Yarn ResourceManager endpoint. Type: string (or Expression with resultType string). | |
tempScriptPath | Specifies an existing folder path which will be used to store temp Distcp command script. The script file is generated by ADF and will be removed after Copy job finished. Type: string (or Expression with resultType string). |
Name | Description | Value |
---|---|---|
type | The read setting type. | 'HttpReadSettings' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
additionalHeaders | The additional HTTP headers in the request to the RESTful API. Type: string (or Expression with resultType string). | |
requestBody | The HTTP request body to the RESTful API if requestMethod is POST. Type: string (or Expression with resultType string). | |
requestMethod | The HTTP method used to call the RESTful API. The default is GET. Type: string (or Expression with resultType string). | |
requestTimeout | Specifies the timeout for a HTTP client to get HTTP response from HTTP server. Type: string (or Expression with resultType string). |
Name | Description | Value |
---|---|---|
type | The read setting type. | 'LakeHouseReadSettings' (required) |
deleteFilesAfterCompletion | Indicates whether the source files need to be deleted after copy completion. Default is false. Type: boolean (or Expression with resultType boolean). | |
enablePartitionDiscovery | Indicates whether to enable partition discovery. Type: boolean (or Expression with resultType boolean). | |
fileListPath | Point to a text file that lists each file (relative path to the path configured in the dataset) that you want to copy. Type: string (or Expression with resultType string). | |
modifiedDatetimeEnd | The end of file's modified datetime. Type: string (or Expression with resultType string). | |
modifiedDatetimeStart | The start of file's modified datetime. Type: string (or Expression with resultType string). | |
partitionRootPath | Specify the root path where partition discovery starts from. Type: string (or Expression with resultType string). | |
recursive | If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean). | |
wildcardFileName | Microsoft Fabric LakeHouse Files wildcardFileName. Type: string (or Expression with resultType string). | |
wildcardFolderPath | Microsoft Fabric LakeHouse Files wildcardFolderPath. Type: string (or Expression with resultType string). |
Name | Description | Value |
---|---|---|
type | The read setting type. | 'OracleCloudStorageReadSettings' (required) |
deleteFilesAfterCompletion | Indicates whether the source files need to be deleted after copy completion. Default is false. Type: boolean (or Expression with resultType boolean). | |
enablePartitionDiscovery | Indicates whether to enable partition discovery. Type: boolean (or Expression with resultType boolean). | |
fileListPath | Point to a text file that lists each file (relative path to the path configured in the dataset) that you want to copy. Type: string (or Expression with resultType string). | |
modifiedDatetimeEnd | The end of file's modified datetime. Type: string (or Expression with resultType string). | |
modifiedDatetimeStart | The start of file's modified datetime. Type: string (or Expression with resultType string). | |
partitionRootPath | Specify the root path where partition discovery starts from. Type: string (or Expression with resultType string). | |
prefix | The prefix filter for the Oracle Cloud Storage object name. Type: string (or Expression with resultType string). | |
recursive | If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean). | |
wildcardFileName | Oracle Cloud Storage wildcardFileName. Type: string (or Expression with resultType string). | |
wildcardFolderPath | Oracle Cloud Storage wildcardFolderPath. Type: string (or Expression with resultType string). |
Name | Description | Value |
---|---|---|
type | The read setting type. | 'SftpReadSettings' (required) |
deleteFilesAfterCompletion | Indicates whether the source files need to be deleted after copy completion. Default is false. Type: boolean (or Expression with resultType boolean). | |
disableChunking | If true, disable parallel reading within each file. Default is false. Type: boolean (or Expression with resultType boolean). | |
enablePartitionDiscovery | Indicates whether to enable partition discovery. Type: boolean (or Expression with resultType boolean). | |
fileListPath | Point to a text file that lists each file (relative path to the path configured in the dataset) that you want to copy. Type: string (or Expression with resultType string). | |
modifiedDatetimeEnd | The end of file's modified datetime. Type: string (or Expression with resultType string). | |
modifiedDatetimeStart | The start of file's modified datetime. Type: string (or Expression with resultType string). | |
partitionRootPath | Specify the root path where partition discovery starts from. Type: string (or Expression with resultType string). | |
recursive | If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean). | |
wildcardFileName | Sftp wildcardFileName. Type: string (or Expression with resultType string). | |
wildcardFolderPath | Sftp wildcardFolderPath. Type: string (or Expression with resultType string). |
Name | Description | Value |
---|---|---|
type | Copy source type. | 'AzureBlobFSSource' (required) |
recursive | If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean). | |
skipHeaderLineCount | Number of header lines to skip from each blob. Type: integer (or Expression with resultType integer). | |
treatEmptyAsNull | Treat empty as null. Type: boolean (or Expression with resultType boolean). |
Name | Description | Value |
---|---|---|
type | Copy source type. | 'AzureDatabricksDeltaLakeSource' (required) |
exportSettings | Azure Databricks Delta Lake export settings. | AzureDatabricksDeltaLakeExportCommand |
query | Azure Databricks Delta Lake Sql query. Type: string (or Expression with resultType string). |
Name | Description | Value |
---|---|---|
dateFormat | Specify the date format for the csv in Azure Databricks Delta Lake Copy. Type: string (or Expression with resultType string). | |
timestampFormat | Specify the timestamp format for the csv in Azure Databricks Delta Lake Copy. Type: string (or Expression with resultType string). | |
type | The export setting type. | string (required) |
{customized property} |
Name | Description | Value |
---|---|---|
type | Copy source type. | 'AzureDataExplorerSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
noTruncation | The name of the Boolean option that controls whether truncation is applied to result-sets that go beyond a certain row-count limit. | |
query | Database query. Should be a Kusto Query Language (KQL) query. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).. |
Name | Description | Value |
---|---|---|
type | Copy source type. | 'AzureDataLakeStoreSource' (required) |
recursive | If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean). |
Name | Description | Value |
---|---|---|
type | Copy source type. | 'AzureMariaDBSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
Name | Description | Value |
---|---|---|
type | Copy source type. | 'AzureMySqlSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | Database query. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
Name | Description | Value |
---|---|---|
type | Copy source type. | 'AzurePostgreSqlSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
Name | Description | Value |
---|---|---|
type | Copy source type. | 'AzureSqlSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
isolationLevel | Specifies the transaction locking behavior for the SQL source. Allowed values: ReadCommitted/ReadUncommitted/RepeatableRead/Serializable/Snapshot. The default value is ReadCommitted. Type: string (or Expression with resultType string). | |
partitionOption | The partition mechanism that will be used for Sql read in parallel. Possible values include: "None", "PhysicalPartitionsOfTable", "DynamicRange". Type: string (or Expression with resultType string). | |
partitionSettings | The settings that will be leveraged for Sql source partitioning. | SqlPartitionSettings |
produceAdditionalTypes | Which additional types to produce. | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | |
sqlReaderQuery | SQL reader query. Type: string (or Expression with resultType string). | |
sqlReaderStoredProcedureName | Name of the stored procedure for a SQL Database source. This cannot be used at the same time as SqlReaderQuery. Type: string (or Expression with resultType string). | |
storedProcedureParameters | Value and type setting for stored procedure parameters. Example: "{Parameter1: {value: "1", type: "int"}}". |
Name | Description | Value |
---|---|---|
type | Copy source type. | 'AzureTableSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
azureTableSourceIgnoreTableNotFound | Azure Table source ignore table not found. Type: boolean (or Expression with resultType boolean). | |
azureTableSourceQuery | Azure Table source query. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
Name | Description | Value |
---|---|---|
type | Copy source type. | 'BinarySource' (required) |
formatSettings | Binary format settings. | BinaryReadSettings |
storeSettings | Binary store settings. | StoreReadSettings |
Name | Description | Value |
---|---|---|
compressionProperties | Compression settings. | CompressionReadSettings |
type | The read setting type. | string (required) |
{customized property} |
Name | Description | Value |
---|---|---|
type | Set the object type | TarGZipReadSettings TarReadSettings ZipDeflateReadSettings (required) |
Name | Description | Value |
---|---|---|
type | The Compression setting type. | 'TarGZipReadSettings' (required) |
preserveCompressionFileNameAsFolder | Preserve the compression file name as folder path. Type: boolean (or Expression with resultType boolean). |
Name | Description | Value |
---|---|---|
type | The Compression setting type. | 'TarReadSettings' (required) |
preserveCompressionFileNameAsFolder | Preserve the compression file name as folder path. Type: boolean (or Expression with resultType boolean). |
Name | Description | Value |
---|---|---|
type | The Compression setting type. | 'ZipDeflateReadSettings' (required) |
preserveZipFileNameAsFolder | Preserve the zip file name as folder path. Type: boolean (or Expression with resultType boolean). |
Name | Description | Value |
---|---|---|
type | Copy source type. | 'BlobSource' (required) |
recursive | If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean). | |
skipHeaderLineCount | Number of header lines to skip from each blob. Type: integer (or Expression with resultType integer). | |
treatEmptyAsNull | Treat empty as null. Type: boolean (or Expression with resultType boolean). |
Name | Description | Value |
---|---|---|
type | Copy source type. | 'CassandraSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
consistencyLevel | The consistency level specifies how many Cassandra servers must respond to a read request before returning data to the client application. Cassandra checks the specified number of Cassandra servers for data to satisfy the read request. Must be one of cassandraSourceReadConsistencyLevels. The default value is 'ONE'. It is case-insensitive. | 'ALL' 'EACH_QUORUM' 'LOCAL_ONE' 'LOCAL_QUORUM' 'LOCAL_SERIAL' 'ONE' 'QUORUM' 'SERIAL' 'THREE' 'TWO' |
query | Database query. Should be a SQL-92 query expression or Cassandra Query Language (CQL) command. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
Name | Description | Value |
---|---|---|
type | Copy source type. | 'CommonDataServiceForAppsSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | FetchXML is a proprietary query language that is used in Microsoft Common Data Service for Apps (online & on-premises). Type: string (or Expression with resultType string). |
Name | Description | Value |
---|---|---|
type | Copy source type. | 'ConcurSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
Name | Description | Value |
---|---|---|
type | Copy source type. | 'CosmosDbMongoDbApiSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
batchSize | Specifies the number of documents to return in each batch of the response from MongoDB instance. In most cases, modifying the batch size will not affect the user or the application. This property's main purpose is to avoid hit the limitation of response size. Type: integer (or Expression with resultType integer). | |
cursorMethods | Cursor methods for Mongodb query. | MongoDbCursorMethodsProperties |
filter | Specifies selection filter using query operators. To return all documents in a collection, omit this parameter or pass an empty document ({}). Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
Name | Description | Value |
---|---|---|
limit | Specifies the maximum number of documents the server returns. limit() is analogous to the LIMIT statement in a SQL database. Type: integer (or Expression with resultType integer). | |
project | Specifies the fields to return in the documents that match the query filter. To return all fields in the matching documents, omit this parameter. Type: string (or Expression with resultType string). | |
skip | Specifies the how many documents skipped and where MongoDB begins returning results. This approach may be useful in implementing paginated results. Type: integer (or Expression with resultType integer). | |
sort | Specifies the order in which the query returns matching documents. Type: string (or Expression with resultType string). Type: string (or Expression with resultType string). | |
{customized property} |
Name | Description | Value |
---|---|---|
type | Copy source type. | 'CosmosDbSqlApiSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
detectDatetime | Whether detect primitive values as datetime values. Type: boolean (or Expression with resultType boolean). | |
pageSize | Page size of the result. Type: integer (or Expression with resultType integer). | |
preferredRegions | Preferred regions. Type: array of strings (or Expression with resultType array of strings). | |
query | SQL API query. Type: string (or Expression with resultType string). |
Name | Description | Value |
---|---|---|
type | Copy source type. | 'CouchbaseSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
Name | Description | Value |
---|---|---|
type | Copy source type. | 'Db2Source' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | Database query. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
Name | Description | Value |
---|---|---|
type | Copy source type. | 'DelimitedTextSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
formatSettings | DelimitedText format settings. | DelimitedTextReadSettings |
storeSettings | DelimitedText store settings. | StoreReadSettings |
Name | Description | Value |
---|---|---|
compressionProperties | Compression settings. | CompressionReadSettings |
skipLineCount | Indicates the number of non-empty rows to skip when reading data from input files. Type: integer (or Expression with resultType integer). | |
type | The read setting type. | string (required) |
{customized property} |
Name | Description | Value |
---|---|---|
type | Copy source type. | 'DocumentDbCollectionSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
nestingSeparator | Nested properties separator. Type: string (or Expression with resultType string). | |
query | Documents query. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
Name | Description | Value |
---|---|---|
type | Copy source type. | 'DrillSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
Name | Description | Value |
---|---|---|
type | Copy source type. | 'DynamicsAXSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
httpRequestTimeout | The timeout (TimeSpan) to get an HTTP response. It is the timeout to get a response, not the timeout to read response data. Default value: 00:05:00. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
Name | Description | Value |
---|---|---|
type | Copy source type. | 'DynamicsCrmSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | FetchXML is a proprietary query language that is used in Microsoft Dynamics CRM (online & on-premises). Type: string (or Expression with resultType string). |
Name | Description | Value |
---|---|---|
type | Copy source type. | 'DynamicsSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | FetchXML is a proprietary query language that is used in Microsoft Dynamics (online & on-premises). Type: string (or Expression with resultType string). |
Name | Description | Value |
---|---|---|
type | Copy source type. | 'EloquaSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
Name | Description | Value |
---|---|---|
type | Copy source type. | 'ExcelSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
storeSettings | Excel store settings. | StoreReadSettings |
Name | Description | Value |
---|---|---|
type | Copy source type. | 'FileSystemSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
recursive | If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean). |
Name | Description | Value |
---|---|---|
type | Copy source type. | 'GoogleAdWordsSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
Name | Description | Value |
---|---|---|
type | Copy source type. | 'GoogleBigQuerySource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
Name | Description | Value |
---|---|---|
type | Copy source type. | 'GoogleBigQueryV2Source' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
Name | Description | Value |
---|---|---|
type | Copy source type. | 'GreenplumSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
Name | Description | Value |
---|---|---|
type | Copy source type. | 'HBaseSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
Name | Description | Value |
---|---|---|
type | Copy source type. | 'HdfsSource' (required) |
distcpSettings | Specifies Distcp-related settings. | DistcpSettings |
recursive | If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean). |
Name | Description | Value |
---|---|---|
type | Copy source type. | 'HiveSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
Name | Description | Value |
---|---|---|
type | Copy source type. | 'HttpSource' (required) |
httpRequestTimeout | Specifies the timeout for a HTTP client to get HTTP response from HTTP server. The default value is equivalent to System.Net.HttpWebRequest.Timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
Name | Description | Value |
---|---|---|
type | Copy source type. | 'HubspotSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
Name | Description | Value |
---|---|---|
type | Copy source type. | 'ImpalaSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
Name | Description | Value |
---|---|---|
type | Copy source type. | 'InformixSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | Database query. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
Name | Description | Value |
---|---|---|
type | Copy source type. | 'JiraSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
Name | Description | Value |
---|---|---|
type | Copy source type. | 'JsonSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
formatSettings | Json format settings. | JsonReadSettings |
storeSettings | Json store settings. | StoreReadSettings |
Name | Description | Value |
---|---|---|
compressionProperties | Compression settings. | CompressionReadSettings |
type | The read setting type. | string (required) |
{customized property} |
Name | Description | Value |
---|---|---|
type | Copy source type. | 'LakeHouseTableSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
timestampAsOf | Query an older snapshot by timestamp. Type: string (or Expression with resultType string). | |
versionAsOf | Query an older snapshot by version. Type: integer (or Expression with resultType integer). |
Name | Description | Value |
---|---|---|
type | Copy source type. | 'MagentoSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
Name | Description | Value |
---|---|---|
type | Copy source type. | 'MariaDBSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
Name | Description | Value |
---|---|---|
type | Copy source type. | 'MarketoSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
Name | Description | Value |
---|---|---|
type | Copy source type. | 'MicrosoftAccessSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | Database query. Type: string (or Expression with resultType string). |
Name | Description | Value |
---|---|---|
type | Copy source type. | 'MongoDbAtlasSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
batchSize | Specifies the number of documents to return in each batch of the response from MongoDB Atlas instance. In most cases, modifying the batch size will not affect the user or the application. This property's main purpose is to avoid hit the limitation of response size. Type: integer (or Expression with resultType integer). | |
cursorMethods | Cursor methods for Mongodb query | MongoDbCursorMethodsProperties |
filter | Specifies selection filter using query operators. To return all documents in a collection, omit this parameter or pass an empty document ({}). Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
Name | Description | Value |
---|---|---|
type | Copy source type. | 'MongoDbSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | Database query. Should be a SQL-92 query expression. Type: string (or Expression with resultType string). |
Name | Description | Value |
---|---|---|
type | Copy source type. | 'MongoDbV2Source' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
batchSize | Specifies the number of documents to return in each batch of the response from MongoDB instance. In most cases, modifying the batch size will not affect the user or the application. This property's main purpose is to avoid hit the limitation of response size. Type: integer (or Expression with resultType integer). | |
cursorMethods | Cursor methods for Mongodb query | MongoDbCursorMethodsProperties |
filter | Specifies selection filter using query operators. To return all documents in a collection, omit this parameter or pass an empty document ({}). Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
Name | Description | Value |
---|---|---|
type | Copy source type. | 'MySqlSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | Database query. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
Name | Description | Value |
---|---|---|
type | Copy source type. | 'NetezzaSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
partitionOption | The partition mechanism that will be used for Netezza read in parallel. Possible values include: "None", "DataSlice", "DynamicRange". | |
partitionSettings | The settings that will be leveraged for Netezza source partitioning. | NetezzaPartitionSettings |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
Name | Description | Value |
---|---|---|
partitionColumnName | The name of the column in integer type that will be used for proceeding range partitioning. Type: string (or Expression with resultType string). | |
partitionLowerBound | The minimum value of column specified in partitionColumnName that will be used for proceeding range partitioning. Type: string (or Expression with resultType string). | |
partitionUpperBound | The maximum value of column specified in partitionColumnName that will be used for proceeding range partitioning. Type: string (or Expression with resultType string). |
Name | Description | Value |
---|---|---|
type | Copy source type. | 'ODataSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
httpRequestTimeout | The timeout (TimeSpan) to get an HTTP response. It is the timeout to get a response, not the timeout to read response data. Default value: 00:05:00. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | |
query | OData query. For example, "$top=1". Type: string (or Expression with resultType string). |
Name | Description | Value |
---|---|---|
type | Copy source type. | 'OdbcSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | Database query. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
Name | Description | Value |
---|---|---|
type | Copy source type. | 'Office365Source' (required) |
allowedGroups | The groups containing all the users. Type: array of strings (or Expression with resultType array of strings). | |
dateFilterColumn | The Column to apply the {paramref name="StartTime"/} and {paramref name="EndTime"/}. Type: string (or Expression with resultType string). | |
endTime | End time of the requested range for this dataset. Type: string (or Expression with resultType string). | |
outputColumns | The columns to be read out from the Office 365 table. Type: array of objects (or Expression with resultType array of objects). itemType: OutputColumn. Example: [ { "name": "Id" }, { "name": "CreatedDateTime" } ] | |
startTime | Start time of the requested range for this dataset. Type: string (or Expression with resultType string). | |
userScopeFilterUri | The user scope uri. Type: string (or Expression with resultType string). |
Name | Description | Value |
---|---|---|
type | Copy source type. | 'OracleServiceCloudSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
Name | Description | Value |
---|---|---|
type | Copy source type. | 'OracleSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
oracleReaderQuery | Oracle reader query. Type: string (or Expression with resultType string). | |
partitionOption | The partition mechanism that will be used for Oracle read in parallel. Possible values include: "None", "PhysicalPartitionsOfTable", "DynamicRange". | |
partitionSettings | The settings that will be leveraged for Oracle source partitioning. | OraclePartitionSettings |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
Name | Description | Value |
---|---|---|
partitionColumnName | The name of the column in integer type that will be used for proceeding range partitioning. Type: string (or Expression with resultType string). | |
partitionLowerBound | The minimum value of column specified in partitionColumnName that will be used for proceeding range partitioning. Type: string (or Expression with resultType string). | |
partitionNames | Names of the physical partitions of Oracle table. | |
partitionUpperBound | The maximum value of column specified in partitionColumnName that will be used for proceeding range partitioning. Type: string (or Expression with resultType string). |
Name | Description | Value |
---|---|---|
type | Copy source type. | 'OrcSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
storeSettings | ORC store settings. | StoreReadSettings |
Name | Description | Value |
---|---|---|
type | Copy source type. | 'ParquetSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
formatSettings | Parquet format settings. | ParquetReadSettings |
storeSettings | Parquet store settings. | StoreReadSettings |
Name | Description | Value |
---|---|---|
compressionProperties | Compression settings. | CompressionReadSettings |
type | The read setting type. | string (required) |
{customized property} |
Name | Description | Value |
---|---|---|
type | Copy source type. | 'PaypalSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
Name | Description | Value |
---|---|---|
type | Copy source type. | 'PhoenixSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
Name | Description | Value |
---|---|---|
type | Copy source type. | 'PostgreSqlSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | Database query. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
Name | Description | Value |
---|---|---|
type | Copy source type. | 'PostgreSqlV2Source' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | Database query. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
Name | Description | Value |
---|---|---|
type | Copy source type. | 'PrestoSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
Name | Description | Value |
---|---|---|
type | Copy source type. | 'QuickBooksSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
Name | Description | Value |
---|---|---|
type | Copy source type. | 'RelationalSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | Database query. Type: string (or Expression with resultType string). |
Name | Description | Value |
---|---|---|
type | Copy source type. | 'ResponsysSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
Name | Description | Value |
---|---|---|
type | Copy source type. | 'RestSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: key value pairs (value should be string type). | |
additionalHeaders | The additional HTTP headers in the request to the RESTful API. Type: string (or Expression with resultType string). | |
httpRequestTimeout | The timeout (TimeSpan) to get an HTTP response. It is the timeout to get a response, not the timeout to read response data. Default value: 00:01:40. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | |
paginationRules | The pagination rules to compose next page requests. Type: string (or Expression with resultType string). | |
requestBody | The HTTP request body to the RESTful API if requestMethod is POST. Type: string (or Expression with resultType string). | |
requestInterval | The time to await before sending next page request. | |
requestMethod | The HTTP method used to call the RESTful API. The default is GET. Type: string (or Expression with resultType string). |
Name | Description | Value |
---|---|---|
type | Copy source type. | 'SalesforceMarketingCloudSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
Name | Description | Value |
---|---|---|
type | Copy source type. | 'SalesforceServiceCloudSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | Database query. Type: string (or Expression with resultType string). | |
readBehavior | The read behavior for the operation. Default is Query. Allowed values: Query/QueryAll. Type: string (or Expression with resultType string). |
Name | Description | Value |
---|---|---|
type | Copy source type. | 'SalesforceServiceCloudV2Source' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
includeDeletedObjects | This property control whether query result contains Deleted objects. Default is false. Type: boolean (or Expression with resultType boolean). | |
SOQLQuery | Database query. Type: string (or Expression with resultType string). |
Name | Description | Value |
---|---|---|
type | Copy source type. | 'SalesforceSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | Database query. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | |
readBehavior | The read behavior for the operation. Default is Query. Allowed values: Query/QueryAll. Type: string (or Expression with resultType string). |
Name | Description | Value |
---|---|---|
type | Copy source type. | 'SalesforceV2Source' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
includeDeletedObjects | This property control whether query result contains Deleted objects. Default is false. Type: boolean (or Expression with resultType boolean). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | |
SOQLQuery | Database query. Type: string (or Expression with resultType string). |
Name | Description | Value |
---|---|---|
type | Copy source type. | 'SapBwSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | MDX query. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
Name | Description | Value |
---|---|---|
type | Copy source type. | 'SapCloudForCustomerSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
httpRequestTimeout | The timeout (TimeSpan) to get an HTTP response. It is the timeout to get a response, not the timeout to read response data. Default value: 00:05:00. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | |
query | SAP Cloud for Customer OData query. For example, "$top=1". Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
Name | Description | Value |
---|---|---|
type | Copy source type. | 'SapEccSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
httpRequestTimeout | The timeout (TimeSpan) to get an HTTP response. It is the timeout to get a response, not the timeout to read response data. Default value: 00:05:00. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | |
query | SAP ECC OData query. For example, "$top=1". Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
Name | Description | Value |
---|---|---|
type | Copy source type. | 'SapHanaSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
packetSize | The packet size of data read from SAP HANA. Type: integer(or Expression with resultType integer). | |
partitionOption | The partition mechanism that will be used for SAP HANA read in parallel. Possible values include: "None", "PhysicalPartitionsOfTable", "SapHanaDynamicRange". | |
partitionSettings | The settings that will be leveraged for SAP HANA source partitioning. | SapHanaPartitionSettings |
query | SAP HANA Sql query. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
Name | Description | Value |
---|---|---|
partitionColumnName | The name of the column that will be used for proceeding range partitioning. Type: string (or Expression with resultType string). |
Name | Description | Value |
---|---|---|
type | Copy source type. | 'SapOdpSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
extractionMode | The extraction mode. Allowed value include: Full, Delta and Recovery. The default value is Full. Type: string (or Expression with resultType string). | |
projection | Specifies the columns to be selected from source data. Type: array of objects(projection) (or Expression with resultType array of objects). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | |
selection | Specifies the selection conditions from source data. Type: array of objects(selection) (or Expression with resultType array of objects). | |
subscriberProcess | The subscriber process to manage the delta process. Type: string (or Expression with resultType string). |
Name | Description | Value |
---|---|---|
type | Copy source type. | 'SapOpenHubSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
baseRequestId | The ID of request for delta loading. Once it is set, only data with requestId larger than the value of this property will be retrieved. The default value is 0. Type: integer (or Expression with resultType integer ). | |
customRfcReadTableFunctionModule | Specifies the custom RFC function module that will be used to read data from SAP Table. Type: string (or Expression with resultType string). | |
excludeLastRequest | Whether to exclude the records of the last request. The default value is true. Type: boolean (or Expression with resultType boolean). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | |
sapDataColumnDelimiter | The single character that will be used as delimiter passed to SAP RFC as well as splitting the output data retrieved. Type: string (or Expression with resultType string). |
Name | Description | Value |
---|---|---|
type | Copy source type. | 'SapTableSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
batchSize | Specifies the maximum number of rows that will be retrieved at a time when retrieving data from SAP Table. Type: integer (or Expression with resultType integer). | |
customRfcReadTableFunctionModule | Specifies the custom RFC function module that will be used to read data from SAP Table. Type: string (or Expression with resultType string). | |
partitionOption | The partition mechanism that will be used for SAP table read in parallel. Possible values include: "None", "PartitionOnInt", "PartitionOnCalendarYear", "PartitionOnCalendarMonth", "PartitionOnCalendarDate", "PartitionOnTime". | |
partitionSettings | The settings that will be leveraged for SAP table source partitioning. | SapTablePartitionSettings |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | |
rfcTableFields | The fields of the SAP table that will be retrieved. For example, column0, column1. Type: string (or Expression with resultType string). | |
rfcTableOptions | The options for the filtering of the SAP Table. For example, COLUMN0 EQ SOME VALUE. Type: string (or Expression with resultType string). | |
rowCount | The number of rows to be retrieved. Type: integer(or Expression with resultType integer). | |
rowSkips | The number of rows that will be skipped. Type: integer (or Expression with resultType integer). | |
sapDataColumnDelimiter | The single character that will be used as delimiter passed to SAP RFC as well as splitting the output data retrieved. Type: string (or Expression with resultType string). |
Name | Description | Value |
---|---|---|
maxPartitionsNumber | The maximum value of partitions the table will be split into. Type: integer (or Expression with resultType string). | |
partitionColumnName | The name of the column that will be used for proceeding range partitioning. Type: string (or Expression with resultType string). | |
partitionLowerBound | The minimum value of column specified in partitionColumnName that will be used for proceeding range partitioning. Type: string (or Expression with resultType string). | |
partitionUpperBound | The maximum value of column specified in partitionColumnName that will be used for proceeding range partitioning. Type: string (or Expression with resultType string). |
Name | Description | Value |
---|---|---|
type | Copy source type. | 'ServiceNowSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
Name | Description | Value |
---|---|---|
type | Copy source type. | 'ServiceNowV2Source' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
expression | Expression to filter data from source. | ExpressionV2 |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
Name | Description | Value |
---|---|---|
operands | List of nested expressions. | ExpressionV2[] |
operator | Expression operator value Type: string. | string |
type | Type of expressions supported by the system. Type: string. | 'Binary' 'Constant' 'Field' 'Unary' |
value | Value for Constant/Field Type: string. | string |
Name | Description | Value |
---|---|---|
type | Copy source type. | 'SharePointOnlineListSource' (required) |
httpRequestTimeout | The wait time to get a response from SharePoint Online. Default value is 5 minutes (00:05:00). Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | |
query | The OData query to filter the data in SharePoint Online list. For example, "$top=1". Type: string (or Expression with resultType string). |
Name | Description | Value |
---|---|---|
type | Copy source type. | 'ShopifySource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
Name | Description | Value |
---|---|---|
type | Copy source type. | 'SnowflakeSource' (required) |
exportSettings | Snowflake export settings. | SnowflakeExportCopyCommand (required) |
query | Snowflake Sql query. Type: string (or Expression with resultType string). |
Name | Description | Value |
---|---|---|
additionalCopyOptions | Additional copy options directly passed to snowflake Copy Command. Type: key value pairs (value should be string type) (or Expression with resultType object). Example: "additionalCopyOptions": { "DATE_FORMAT": "MM/DD/YYYY", "TIME_FORMAT": "'HH24:MI:SS.FF'" } | SnowflakeExportCopyCommandAdditionalCopyOptions |
additionalFormatOptions | Additional format options directly passed to snowflake Copy Command. Type: key value pairs (value should be string type) (or Expression with resultType object). Example: "additionalFormatOptions": { "OVERWRITE": "TRUE", "MAX_FILE_SIZE": "'FALSE'" } | SnowflakeExportCopyCommandAdditionalFormatOptions |
type | The export setting type. | string (required) |
{customized property} |
Name | Description | Value |
---|---|---|
{customized property} |
Name | Description | Value |
---|---|---|
{customized property} |
Name | Description | Value |
---|---|---|
type | Copy source type. | 'SnowflakeV2Source' (required) |
exportSettings | Snowflake export settings. | SnowflakeExportCopyCommand (required) |
query | Snowflake Sql query. Type: string (or Expression with resultType string). |
Name | Description | Value |
---|---|---|
type | Copy source type. | 'SparkSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
Name | Description | Value |
---|---|---|
type | Copy source type. | 'SqlDWSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
isolationLevel | Specifies the transaction locking behavior for the SQL source. Allowed values: ReadCommitted/ReadUncommitted/RepeatableRead/Serializable/Snapshot. The default value is ReadCommitted. Type: string (or Expression with resultType string). | |
partitionOption | The partition mechanism that will be used for Sql read in parallel. Possible values include: "None", "PhysicalPartitionsOfTable", "DynamicRange". Type: string (or Expression with resultType string). | |
partitionSettings | The settings that will be leveraged for Sql source partitioning. | SqlPartitionSettings |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | |
sqlReaderQuery | SQL Data Warehouse reader query. Type: string (or Expression with resultType string). | |
sqlReaderStoredProcedureName | Name of the stored procedure for a SQL Data Warehouse source. This cannot be used at the same time as SqlReaderQuery. Type: string (or Expression with resultType string). | |
storedProcedureParameters | Value and type setting for stored procedure parameters. Example: "{Parameter1: {value: "1", type: "int"}}". Type: object (or Expression with resultType object), itemType: StoredProcedureParameter. |
Name | Description | Value |
---|---|---|
type | Copy source type. | 'SqlMISource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
isolationLevel | Specifies the transaction locking behavior for the SQL source. Allowed values: ReadCommitted/ReadUncommitted/RepeatableRead/Serializable/Snapshot. The default value is ReadCommitted. Type: string (or Expression with resultType string). | |
partitionOption | The partition mechanism that will be used for Sql read in parallel. Possible values include: "None", "PhysicalPartitionsOfTable", "DynamicRange". Type: string (or Expression with resultType string). | |
partitionSettings | The settings that will be leveraged for Sql source partitioning. | SqlPartitionSettings |
produceAdditionalTypes | Which additional types to produce. | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | |
sqlReaderQuery | SQL reader query. Type: string (or Expression with resultType string). | |
sqlReaderStoredProcedureName | Name of the stored procedure for a Azure SQL Managed Instance source. This cannot be used at the same time as SqlReaderQuery. Type: string (or Expression with resultType string). | |
storedProcedureParameters | Value and type setting for stored procedure parameters. Example: "{Parameter1: {value: "1", type: "int"}}". |
Name | Description | Value |
---|---|---|
type | Copy source type. | 'SqlServerSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
isolationLevel | Specifies the transaction locking behavior for the SQL source. Allowed values: ReadCommitted/ReadUncommitted/RepeatableRead/Serializable/Snapshot. The default value is ReadCommitted. Type: string (or Expression with resultType string). | |
partitionOption | The partition mechanism that will be used for Sql read in parallel. Possible values include: "None", "PhysicalPartitionsOfTable", "DynamicRange". Type: string (or Expression with resultType string). | |
partitionSettings | The settings that will be leveraged for Sql source partitioning. | SqlPartitionSettings |
produceAdditionalTypes | Which additional types to produce. | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | |
sqlReaderQuery | SQL reader query. Type: string (or Expression with resultType string). | |
sqlReaderStoredProcedureName | Name of the stored procedure for a SQL Database source. This cannot be used at the same time as SqlReaderQuery. Type: string (or Expression with resultType string). | |
storedProcedureParameters | Value and type setting for stored procedure parameters. Example: "{Parameter1: {value: "1", type: "int"}}". |
Name | Description | Value |
---|---|---|
type | Copy source type. | 'SqlSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
isolationLevel | Specifies the transaction locking behavior for the SQL source. Allowed values: ReadCommitted/ReadUncommitted/RepeatableRead/Serializable/Snapshot. The default value is ReadCommitted. Type: string (or Expression with resultType string). | |
partitionOption | The partition mechanism that will be used for Sql read in parallel. Possible values include: "None", "PhysicalPartitionsOfTable", "DynamicRange". Type: string (or Expression with resultType string). | |
partitionSettings | The settings that will be leveraged for Sql source partitioning. | SqlPartitionSettings |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | |
sqlReaderQuery | SQL reader query. Type: string (or Expression with resultType string). | |
sqlReaderStoredProcedureName | Name of the stored procedure for a SQL Database source. This cannot be used at the same time as SqlReaderQuery. Type: string (or Expression with resultType string). | |
storedProcedureParameters | Value and type setting for stored procedure parameters. Example: "{Parameter1: {value: "1", type: "int"}}". |
Name | Description | Value |
---|---|---|
type | Copy source type. | 'SquareSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
Name | Description | Value |
---|---|---|
type | Copy source type. | 'SybaseSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | Database query. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
Name | Description | Value |
---|---|---|
type | Copy source type. | 'TeradataSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
partitionOption | The partition mechanism that will be used for teradata read in parallel. Possible values include: "None", "Hash", "DynamicRange". | |
partitionSettings | The settings that will be leveraged for teradata source partitioning. | TeradataPartitionSettings |
query | Teradata query. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
Name | Description | Value |
---|---|---|
partitionColumnName | The name of the column that will be used for proceeding range or hash partitioning. Type: string (or Expression with resultType string). | |
partitionLowerBound | The minimum value of column specified in partitionColumnName that will be used for proceeding range partitioning. Type: string (or Expression with resultType string). | |
partitionUpperBound | The maximum value of column specified in partitionColumnName that will be used for proceeding range partitioning. Type: string (or Expression with resultType string). |
Name | Description | Value |
---|---|---|
type | Copy source type. | 'VerticaSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
Name | Description | Value |
---|---|---|
type | Copy source type. | 'WarehouseSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
isolationLevel | Specifies the transaction locking behavior for the Microsoft Fabric Warehouse source. Allowed values: ReadCommitted/ReadUncommitted/RepeatableRead/Serializable/Snapshot. The default value is ReadCommitted. Type: string (or Expression with resultType string). | |
partitionOption | The partition mechanism that will be used for Sql read in parallel. Possible values include: "None", "PhysicalPartitionsOfTable", "DynamicRange". | |
partitionSettings | The settings that will be leveraged for Sql source partitioning. | SqlPartitionSettings |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | |
sqlReaderQuery | Microsoft Fabric Warehouse reader query. Type: string (or Expression with resultType string). | |
sqlReaderStoredProcedureName | Name of the stored procedure for a Microsoft Fabric Warehouse source. This cannot be used at the same time as SqlReaderQuery. Type: string (or Expression with resultType string). | |
storedProcedureParameters | Value and type setting for stored procedure parameters. Example: "{Parameter1: {value: "1", type: "int"}}". Type: object (or Expression with resultType object), itemType: StoredProcedureParameter. |
Name | Description | Value |
---|---|---|
type | Copy source type. | 'WebSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). |
Name | Description | Value |
---|---|---|
type | Copy source type. | 'XeroSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
Name | Description | Value |
---|---|---|
type | Copy source type. | 'XmlSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
formatSettings | Xml format settings. | XmlReadSettings |
storeSettings | Xml store settings. | StoreReadSettings |
Name | Description | Value |
---|---|---|
compressionProperties | Compression settings. | CompressionReadSettings |
detectDataType | Indicates whether type detection is enabled when reading the xml files. Type: boolean (or Expression with resultType boolean). | |
namespacePrefixes | Namespace uri to prefix mappings to override the prefixes in column names when namespace is enabled, if no prefix is defined for a namespace uri, the prefix of xml element/attribute name in the xml data file will be used. Example: "{"http://www.example.com/xml":"prefix"}" Type: object (or Expression with resultType object). | |
namespaces | Indicates whether namespace is enabled when reading the xml files. Type: boolean (or Expression with resultType boolean). | |
type | The read setting type. | string (required) |
validationMode | Indicates what validation method is used when reading the xml files. Allowed values: 'none', 'xsd', or 'dtd'. Type: string (or Expression with resultType string). | |
{customized property} |
Name | Description | Value |
---|---|---|
type | Copy source type. | 'ZohoSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
Name | Description | Value |
---|---|---|
enableCompression | Specifies whether to use compression when copying data via an interim staging. Default value is false. Type: boolean (or Expression with resultType boolean). | |
linkedServiceName | Staging linked service reference. | LinkedServiceReference (required) |
path | The path to storage for storing the interim data. Type: string (or Expression with resultType string). | |
{customized property} |
Name | Description | Value |
---|---|---|
type | Type of activity. | 'Custom' (required) |
linkedServiceName | Linked service reference. | LinkedServiceReference |
policy | Activity policy. | ActivityPolicy |
typeProperties | Custom activity properties. | CustomActivityTypeProperties (required) |
Name | Description | Value |
---|---|---|
autoUserSpecification | Elevation level and scope for the user, default is nonadmin task. Type: string (or Expression with resultType double). | |
command | Command for custom activity Type: string (or Expression with resultType string). | |
extendedProperties | User defined property bag. There is no restriction on the keys or values that can be used. The user specified custom activity has the full responsibility to consume and interpret the content defined. | CustomActivityTypePropertiesExtendedProperties |
folderPath | Folder path for resource files Type: string (or Expression with resultType string). | |
referenceObjects | Reference objects | CustomActivityReferenceObject |
resourceLinkedService | Resource linked service reference. | LinkedServiceReference |
retentionTimeInDays | The retention time for the files submitted for custom activity. Type: double (or Expression with resultType double). |
Name | Description | Value |
---|---|---|
{customized property} |
Name | Description | Value |
---|---|---|
datasets | Dataset references. | DatasetReference[] |
linkedServices | Linked service references. | LinkedServiceReference[] |
Name | Description | Value |
---|---|---|
type | Type of activity. | 'DatabricksNotebook' (required) |
linkedServiceName | Linked service reference. | LinkedServiceReference |
policy | Activity policy. | ActivityPolicy |
typeProperties | Databricks Notebook activity properties. | DatabricksNotebookActivityTypeProperties (required) |
Name | Description | Value |
---|---|---|
baseParameters | Base parameters to be used for each run of this job.If the notebook takes a parameter that is not specified, the default value from the notebook will be used. | DatabricksNotebookActivityTypePropertiesBaseParamete... |
libraries | A list of libraries to be installed on the cluster that will execute the job. | DatabricksNotebookActivityTypePropertiesLibrariesIte...[] |
notebookPath | The absolute path of the notebook to be run in the Databricks Workspace. This path must begin with a slash. Type: string (or Expression with resultType string). |
Name | Description | Value |
---|---|---|
{customized property} |
Name | Description | Value |
---|---|---|
{customized property} |
Name | Description | Value |
---|---|---|
type | Type of activity. | 'DatabricksSparkJar' (required) |
linkedServiceName | Linked service reference. | LinkedServiceReference |
policy | Activity policy. | ActivityPolicy |
typeProperties | Databricks SparkJar activity properties. | DatabricksSparkJarActivityTypeProperties (required) |
Name | Description | Value |
---|---|---|
libraries | A list of libraries to be installed on the cluster that will execute the job. | DatabricksSparkJarActivityTypePropertiesLibrariesIte...[] |
mainClassName | The full name of the class containing the main method to be executed. This class must be contained in a JAR provided as a library. Type: string (or Expression with resultType string). | |
parameters | Parameters that will be passed to the main method. | any[] |
Name | Description | Value |
---|---|---|
{customized property} |
Name | Description | Value |
---|---|---|
type | Type of activity. | 'DatabricksSparkPython' (required) |
linkedServiceName | Linked service reference. | LinkedServiceReference |
policy | Activity policy. | ActivityPolicy |
typeProperties | Databricks SparkPython activity properties. | DatabricksSparkPythonActivityTypeProperties (required) |
Name | Description | Value |
---|---|---|
libraries | A list of libraries to be installed on the cluster that will execute the job. | DatabricksSparkPythonActivityTypePropertiesLibraries...[] |
parameters | Command line parameters that will be passed to the Python file. | any[] |
pythonFile | The URI of the Python file to be executed. DBFS paths are supported. Type: string (or Expression with resultType string). |
Name | Description | Value |
---|---|---|
{customized property} |
Name | Description | Value |
---|---|---|
type | Type of activity. | 'DataLakeAnalyticsU-SQL' (required) |
linkedServiceName | Linked service reference. | LinkedServiceReference |
policy | Activity policy. | ActivityPolicy |
typeProperties | Data Lake Analytics U-SQL activity properties. | DataLakeAnalyticsUsqlActivityTypeProperties (required) |
Name | Description | Value |
---|---|---|
compilationMode | Compilation mode of U-SQL. Must be one of these values : Semantic, Full and SingleBox. Type: string (or Expression with resultType string). | |
degreeOfParallelism | The maximum number of nodes simultaneously used to run the job. Default value is 1. Type: integer (or Expression with resultType integer), minimum: 1. | |
parameters | Parameters for U-SQL job request. | DataLakeAnalyticsUsqlActivityTypePropertiesParameter... |
priority | Determines which jobs out of all that are queued should be selected to run first. The lower the number, the higher the priority. Default value is 1000. Type: integer (or Expression with resultType integer), minimum: 1. | |
runtimeVersion | Runtime version of the U-SQL engine to use. Type: string (or Expression with resultType string). | |
scriptLinkedService | Script linked service reference. | LinkedServiceReference (required) |
scriptPath | Case-sensitive path to folder that contains the U-SQL script. Type: string (or Expression with resultType string). |
Name | Description | Value |
---|---|---|
{customized property} |
Name | Description | Value |
---|---|---|
type | Type of activity. | 'Delete' (required) |
linkedServiceName | Linked service reference. | LinkedServiceReference |
policy | Activity policy. | ActivityPolicy |
typeProperties | Delete activity properties. | DeleteActivityTypeProperties (required) |
Name | Description | Value |
---|---|---|
dataset | Delete activity dataset reference. | DatasetReference (required) |
enableLogging | Whether to record detailed logs of delete-activity execution. Default value is false. Type: boolean (or Expression with resultType boolean). | |
logStorageSettings | Log storage settings customer need to provide when enableLogging is true. | LogStorageSettings |
maxConcurrentConnections | The max concurrent connections to connect data source at the same time. | int Constraints: Min value = 1 |
recursive | If true, files or sub-folders under current folder path will be deleted recursively. Default is false. Type: boolean (or Expression with resultType boolean). | |
storeSettings | Delete activity store settings. | StoreReadSettings |
Name | Description | Value |
---|---|---|
type | Type of activity. | 'ExecuteDataFlow' (required) |
linkedServiceName | Linked service reference. | LinkedServiceReference |
policy | Activity policy. | ActivityPolicy |
typeProperties | Execute data flow activity properties. | ExecuteDataFlowActivityTypeProperties (required) |
Name | Description | Value |
---|---|---|
compute | Compute properties for data flow activity. | ExecuteDataFlowActivityTypePropertiesCompute |
continueOnError | Continue on error setting used for data flow execution. Enables processing to continue if a sink fails. Type: boolean (or Expression with resultType boolean) | |
dataFlow | Data flow reference. | DataFlowReference (required) |
integrationRuntime | The integration runtime reference. | IntegrationRuntimeReference |
runConcurrently | Concurrent run setting used for data flow execution. Allows sinks with the same save order to be processed concurrently. Type: boolean (or Expression with resultType boolean) | |
sourceStagingConcurrency | Specify number of parallel staging for sources applicable to the sink. Type: integer (or Expression with resultType integer) | |
staging | Staging info for execute data flow activity. | DataFlowStagingInfo |
traceLevel | Trace level setting used for data flow monitoring output. Supported values are: 'coarse', 'fine', and 'none'. Type: string (or Expression with resultType string) |
Name | Description | Value |
---|---|---|
computeType | Compute type of the cluster which will execute data flow job. Possible values include: 'General', 'MemoryOptimized', 'ComputeOptimized'. Type: string (or Expression with resultType string) | |
coreCount | Core count of the cluster which will execute data flow job. Supported values are: 8, 16, 32, 48, 80, 144 and 272. Type: integer (or Expression with resultType integer) |
Name | Description | Value |
---|---|---|
datasetParameters | Reference data flow parameters from dataset. | |
parameters | Data flow parameters | ParameterValueSpecification |
referenceName | Reference data flow name. | string (required) |
type | Data flow reference type. | 'DataFlowReference' (required) |
{customized property} |
Name | Description | Value |
---|---|---|
parameters | Arguments for integration runtime. | ParameterValueSpecification |
referenceName | Reference integration runtime name. | string (required) |
type | Type of integration runtime. | 'IntegrationRuntimeReference' (required) |
Name | Description | Value |
---|---|---|
folderPath | Folder path for staging blob. Type: string (or Expression with resultType string) | |
linkedService | Staging linked service reference. | LinkedServiceReference |
Name | Description | Value |
---|---|---|
type | Type of activity. | 'ExecutePipeline' (required) |
policy | Execute pipeline activity policy. | ExecutePipelineActivityPolicy |
typeProperties | Execute pipeline activity properties. | ExecutePipelineActivityTypeProperties (required) |
Name | Description | Value |
---|---|---|
secureInput | When set to true, Input from activity is considered as secure and will not be logged to monitoring. | bool |
{customized property} |
Name | Description | Value |
---|---|---|
parameters | Pipeline parameters. | ParameterValueSpecification |
pipeline | Pipeline reference. | PipelineReference (required) |
waitOnCompletion | Defines whether activity execution will wait for the dependent pipeline execution to finish. Default is false. | bool |
Name | Description | Value |
---|---|---|
name | Reference name. | string |
referenceName | Reference pipeline name. | string (required) |
type | Pipeline reference type. | 'PipelineReference' (required) |
Name | Description | Value |
---|---|---|
type | Type of activity. | 'ExecuteSSISPackage' (required) |
linkedServiceName | Linked service reference. | LinkedServiceReference |
policy | Activity policy. | ActivityPolicy |
typeProperties | Execute SSIS package activity properties. | ExecuteSsisPackageActivityTypeProperties (required) |
Name | Description | Value |
---|---|---|
connectVia | The integration runtime reference. | IntegrationRuntimeReference (required) |
environmentPath | The environment path to execute the SSIS package. Type: string (or Expression with resultType string). | |
executionCredential | The package execution credential. | SsisExecutionCredential |
loggingLevel | The logging level of SSIS package execution. Type: string (or Expression with resultType string). | |
logLocation | SSIS package execution log location. | SsisLogLocation |
packageConnectionManagers | The package level connection managers to execute the SSIS package. | ExecuteSsisPackageActivityTypePropertiesPackageConne... |
packageLocation | SSIS package location. | SsisPackageLocation (required) |
packageParameters | The package level parameters to execute the SSIS package. | ExecuteSsisPackageActivityTypePropertiesPackageParam... |
projectConnectionManagers | The project level connection managers to execute the SSIS package. | ExecuteSsisPackageActivityTypePropertiesProjectConne... |
projectParameters | The project level parameters to execute the SSIS package. | ExecuteSsisPackageActivityTypePropertiesProjectParam... |
propertyOverrides | The property overrides to execute the SSIS package. | ExecuteSsisPackageActivityTypePropertiesPropertyOver... |
runtime | Specifies the runtime to execute SSIS package. The value should be "x86" or "x64". Type: string (or Expression with resultType string). |
Name | Description | Value |
---|---|---|
domain | Domain for windows authentication. Type: string (or Expression with resultType string). | |
password | Password for windows authentication. | SecureString (required) |
userName | UseName for windows authentication. Type: string (or Expression with resultType string). |
Name | Description | Value |
---|---|---|
type | Type of the secret. | string (required) |
value | Value of secure string. | string (required) |
Name | Description | Value |
---|---|---|
logPath | The SSIS package execution log path. Type: string (or Expression with resultType string). | |
type | The type of SSIS log location. | 'File' (required) |
typeProperties | SSIS package execution log location properties. | SsisLogLocationTypeProperties (required) |
Name | Description | Value |
---|---|---|
accessCredential | The package execution log access credential. | SsisAccessCredential |
logRefreshInterval | Specifies the interval to refresh log. The default interval is 5 minutes. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
Name | Description | Value |
---|---|---|
domain | Domain for windows authentication. Type: string (or Expression with resultType string). | |
password | Password for windows authentication. | SecretBase (required) |
userName | UseName for windows authentication. Type: string (or Expression with resultType string). |
Name | Description | Value |
---|---|---|
type | Set the object type | AzureKeyVaultSecret SecureString (required) |
Name | Description | Value |
---|---|---|
secretName | The name of the secret in Azure Key Vault. Type: string (or Expression with resultType string). | |
secretVersion | The version of the secret in Azure Key Vault. The default value is the latest version of the secret. Type: string (or Expression with resultType string). | |
store | The Azure Key Vault linked service reference. | LinkedServiceReference (required) |
type | Type of the secret. | string (required) |
Name | Description | Value |
---|---|---|
{customized property} | SsisConnectionManager |
Name | Description | Value |
---|---|---|
{customized property} | SsisExecutionParameter |
Name | Description | Value |
---|---|---|
value | SSIS package execution parameter value. Type: string (or Expression with resultType string). |
Name | Description | Value |
---|---|---|
packagePath | The SSIS package path. Type: string (or Expression with resultType string). | |
type | The type of SSIS package location. | 'File' 'InlinePackage' 'PackageStore' 'SSISDB' |
typeProperties | SSIS package location properties. | SsisPackageLocationTypeProperties |
Name | Description | Value |
---|---|---|
accessCredential | The package access credential. | SsisAccessCredential |
childPackages | The embedded child package list. | SsisChildPackage[] |
configurationAccessCredential | The configuration file access credential. | SsisAccessCredential |
configurationPath | The configuration file of the package execution. Type: string (or Expression with resultType string). | |
packageContent | The embedded package content. Type: string (or Expression with resultType string). | |
packageLastModifiedDate | The embedded package last modified date. | string |
packageName | The package name. | string |
packagePassword | Password of the package. | SecretBase |
Name | Description | Value |
---|---|---|
packageContent | Content for embedded child package. Type: string (or Expression with resultType string). | |
packageLastModifiedDate | Last modified date for embedded child package. | string |
packageName | Name for embedded child package. | string |
packagePath | Path for embedded child package. Type: string (or Expression with resultType string). |
Name | Description | Value |
---|---|---|
{customized property} | SsisExecutionParameter |
Name | Description | Value |
---|---|---|
{customized property} | SsisConnectionManager |
Name | Description | Value |
---|---|---|
{customized property} | SsisExecutionParameter |
Name | Description | Value |
---|---|---|
{customized property} | SsisPropertyOverride |
Name | Description | Value |
---|---|---|
isSensitive | Whether SSIS package property override value is sensitive data. Value will be encrypted in SSISDB if it is true | bool |
value | SSIS package property override value. Type: string (or Expression with resultType string). |
Name | Description | Value |
---|---|---|
type | Type of activity. | 'ExecuteWranglingDataflow' (required) |
policy | Activity policy. | ActivityPolicy |
typeProperties | Execute power query activity properties. | ExecutePowerQueryActivityTypeProperties (required) |
Name | Description | Value |
---|---|---|
compute | Compute properties for data flow activity. | ExecuteDataFlowActivityTypePropertiesCompute |
continueOnError | Continue on error setting used for data flow execution. Enables processing to continue if a sink fails. Type: boolean (or Expression with resultType boolean) | |
dataFlow | Data flow reference. | DataFlowReference (required) |
integrationRuntime | The integration runtime reference. | IntegrationRuntimeReference |
queries | List of mapping for Power Query mashup query to sink dataset(s). | PowerQuerySinkMapping[] |
runConcurrently | Concurrent run setting used for data flow execution. Allows sinks with the same save order to be processed concurrently. Type: boolean (or Expression with resultType boolean) | |
sinks | (Deprecated. Please use Queries). List of Power Query activity sinks mapped to a queryName. | ExecutePowerQueryActivityTypePropertiesSinks |
sourceStagingConcurrency | Specify number of parallel staging for sources applicable to the sink. Type: integer (or Expression with resultType integer) | |
staging | Staging info for execute data flow activity. | DataFlowStagingInfo |
traceLevel | Trace level setting used for data flow monitoring output. Supported values are: 'coarse', 'fine', and 'none'. Type: string (or Expression with resultType string) |
Name | Description | Value |
---|---|---|
dataflowSinks | List of sinks mapped to Power Query mashup query. | PowerQuerySink[] |
queryName | Name of the query in Power Query mashup document. | string |
Name | Description | Value |
---|---|---|
dataset | Dataset reference. | DatasetReference |
description | Transformation description. | string |
flowlet | Flowlet Reference | DataFlowReference |
linkedService | Linked service reference. | LinkedServiceReference |
name | Transformation name. | string (required) |
rejectedDataLinkedService | Rejected data linked service reference. | LinkedServiceReference |
schemaLinkedService | Schema linked service reference. | LinkedServiceReference |
script | sink script. | string |
Name | Description | Value |
---|---|---|
{customized property} | PowerQuerySink |
Name | Description | Value |
---|---|---|
type | Type of activity. | 'Fail' (required) |
typeProperties | Fail activity properties. | FailActivityTypeProperties (required) |
Name | Description | Value |
---|---|---|
errorCode | The error code that categorizes the error type of the Fail activity. It can be dynamic content that's evaluated to a non empty/blank string at runtime. Type: string (or Expression with resultType string). | |
message | The error message that surfaced in the Fail activity. It can be dynamic content that's evaluated to a non empty/blank string at runtime. Type: string (or Expression with resultType string). |
Name | Description | Value |
---|---|---|
type | Type of activity. | 'Filter' (required) |
typeProperties | Filter activity properties. | FilterActivityTypeProperties (required) |
Name | Description | Value |
---|---|---|
condition | Condition to be used for filtering the input. | Expression (required) |
items | Input array on which filter should be applied. | Expression (required) |
Name | Description | Value |
---|---|---|
type | Expression type. | 'Expression' (required) |
value | Expression value. | string (required) |
Name | Description | Value |
---|---|---|
type | Type of activity. | 'ForEach' (required) |
typeProperties | ForEach activity properties. | ForEachActivityTypeProperties (required) |
Name | Description | Value |
---|---|---|
activities | List of activities to execute . | Activity[] (required) |
batchCount | Batch count to be used for controlling the number of parallel execution (when isSequential is set to false). | int Constraints: Max value = 50 |
isSequential | Should the loop be executed in sequence or in parallel (max 50) | bool |
items | Collection to iterate. | Expression (required) |
Name | Description | Value |
---|---|---|
type | Type of activity. | 'GetMetadata' (required) |
linkedServiceName | Linked service reference. | LinkedServiceReference |
policy | Activity policy. | ActivityPolicy |
typeProperties | GetMetadata activity properties. | GetMetadataActivityTypeProperties (required) |
Name | Description | Value |
---|---|---|
dataset | GetMetadata activity dataset reference. | DatasetReference (required) |
fieldList | Fields of metadata to get from dataset. | any[] |
formatSettings | GetMetadata activity format settings. | FormatReadSettings |
storeSettings | GetMetadata activity store settings. | StoreReadSettings |
Name | Description | Value |
---|---|---|
type | Set the object type | BinaryReadSettings DelimitedTextReadSettings JsonReadSettings ParquetReadSettings XmlReadSettings (required) |
Name | Description | Value |
---|---|---|
type | Type of activity. | 'HDInsightHive' (required) |
linkedServiceName | Linked service reference. | LinkedServiceReference |
policy | Activity policy. | ActivityPolicy |
typeProperties | HDInsight Hive activity properties. | HDInsightHiveActivityTypeProperties (required) |
Name | Description | Value |
---|---|---|
arguments | User specified arguments to HDInsightActivity. | any[] |
defines | Allows user to specify defines for Hive job request. | HDInsightHiveActivityTypePropertiesDefines |
getDebugInfo | Debug info option. | 'Always' 'Failure' 'None' |
queryTimeout | Query timeout value (in minutes). Effective when the HDInsight cluster is with ESP (Enterprise Security Package) | int |
scriptLinkedService | Script linked service reference. | LinkedServiceReference |
scriptPath | Script path. Type: string (or Expression with resultType string). | |
storageLinkedServices | Storage linked service references. | LinkedServiceReference[] |
variables | User specified arguments under hivevar namespace. | HDInsightHiveActivityTypePropertiesVariables |
Name | Description | Value |
---|---|---|
{customized property} |
Name | Description | Value |
---|---|---|
{customized property} |
Name | Description | Value |
---|---|---|
type | Type of activity. | 'HDInsightMapReduce' (required) |
linkedServiceName | Linked service reference. | LinkedServiceReference |
policy | Activity policy. | ActivityPolicy |
typeProperties | HDInsight MapReduce activity properties. | HDInsightMapReduceActivityTypeProperties (required) |
Name | Description | Value |
---|---|---|
arguments | User specified arguments to HDInsightActivity. | any[] |
className | Class name. Type: string (or Expression with resultType string). | |
defines | Allows user to specify defines for the MapReduce job request. | HDInsightMapReduceActivityTypePropertiesDefines |
getDebugInfo | Debug info option. | 'Always' 'Failure' 'None' |
jarFilePath | Jar path. Type: string (or Expression with resultType string). | |
jarLibs | Jar libs. | any[] |
jarLinkedService | Jar linked service reference. | LinkedServiceReference |
storageLinkedServices | Storage linked service references. | LinkedServiceReference[] |
Name | Description | Value |
---|---|---|
{customized property} |
Name | Description | Value |
---|---|---|
type | Type of activity. | 'HDInsightPig' (required) |
linkedServiceName | Linked service reference. | LinkedServiceReference |
policy | Activity policy. | ActivityPolicy |
typeProperties | HDInsight Pig activity properties. | HDInsightPigActivityTypeProperties (required) |
Name | Description | Value |
---|---|---|
arguments | User specified arguments to HDInsightActivity. Type: array (or Expression with resultType array). | |
defines | Allows user to specify defines for Pig job request. | HDInsightPigActivityTypePropertiesDefines |
getDebugInfo | Debug info option. | 'Always' 'Failure' 'None' |
scriptLinkedService | Script linked service reference. | LinkedServiceReference |
scriptPath | Script path. Type: string (or Expression with resultType string). | |
storageLinkedServices | Storage linked service references. | LinkedServiceReference[] |
Name | Description | Value |
---|---|---|
{customized property} |
Name | Description | Value |
---|---|---|
type | Type of activity. | 'HDInsightSpark' (required) |
linkedServiceName | Linked service reference. | LinkedServiceReference |
policy | Activity policy. | ActivityPolicy |
typeProperties | HDInsight spark activity properties. | HDInsightSparkActivityTypeProperties (required) |
Name | Description | Value |
---|---|---|
arguments | The user-specified arguments to HDInsightSparkActivity. | any[] |
className | The application's Java/Spark main class. | string |
entryFilePath | The relative path to the root folder of the code/package to be executed. Type: string (or Expression with resultType string). | |
getDebugInfo | Debug info option. | 'Always' 'Failure' 'None' |
proxyUser | The user to impersonate that will execute the job. Type: string (or Expression with resultType string). | |
rootPath | The root path in 'sparkJobLinkedService' for all the jobâs files. Type: string (or Expression with resultType string). | |
sparkConfig | Spark configuration property. | HDInsightSparkActivityTypePropertiesSparkConfig |
sparkJobLinkedService | The storage linked service for uploading the entry file and dependencies, and for receiving logs. | LinkedServiceReference |
Name | Description | Value |
---|---|---|
{customized property} |
Name | Description | Value |
---|---|---|
type | Type of activity. | 'HDInsightStreaming' (required) |
linkedServiceName | Linked service reference. | LinkedServiceReference |
policy | Activity policy. | ActivityPolicy |
typeProperties | HDInsight streaming activity properties. | HDInsightStreamingActivityTypeProperties (required) |
Name | Description | Value |
---|---|---|
arguments | User specified arguments to HDInsightActivity. | any[] |
combiner | Combiner executable name. Type: string (or Expression with resultType string). | |
commandEnvironment | Command line environment values. | any[] |
defines | Allows user to specify defines for streaming job request. | HDInsightStreamingActivityTypePropertiesDefines |
fileLinkedService | Linked service reference where the files are located. | LinkedServiceReference |
filePaths | Paths to streaming job files. Can be directories. | any[] (required) |
getDebugInfo | Debug info option. | 'Always' 'Failure' 'None' |
input | Input blob path. Type: string (or Expression with resultType string). | |
mapper | Mapper executable name. Type: string (or Expression with resultType string). | |
output | Output blob path. Type: string (or Expression with resultType string). | |
reducer | Reducer executable name. Type: string (or Expression with resultType string). | |
storageLinkedServices | Storage linked service references. | LinkedServiceReference[] |
Name | Description | Value |
---|---|---|
{customized property} |
Name | Description | Value |
---|---|---|
type | Type of activity. | 'IfCondition' (required) |
typeProperties | IfCondition activity properties. | IfConditionActivityTypeProperties (required) |
Name | Description | Value |
---|---|---|
expression | An expression that would evaluate to Boolean. This is used to determine the block of activities (ifTrueActivities or ifFalseActivities) that will be executed. | Expression (required) |
ifFalseActivities | List of activities to execute if expression is evaluated to false. This is an optional property and if not provided, the activity will exit without any action. | Activity[] |
ifTrueActivities | List of activities to execute if expression is evaluated to true. This is an optional property and if not provided, the activity will exit without any action. | Activity[] |
Name | Description | Value |
---|---|---|
type | Type of activity. | 'Lookup' (required) |
linkedServiceName | Linked service reference. | LinkedServiceReference |
policy | Activity policy. | ActivityPolicy |
typeProperties | Lookup activity properties. | LookupActivityTypeProperties (required) |
Name | Description | Value |
---|---|---|
dataset | Lookup activity dataset reference. | DatasetReference (required) |
firstRowOnly | Whether to return first row or all rows. Default value is true. Type: boolean (or Expression with resultType boolean). | |
source | Dataset-specific source properties, same as copy activity source. | CopySource (required) |
Name | Description | Value |
---|---|---|
type | Type of activity. | 'Script' (required) |
linkedServiceName | Linked service reference. | LinkedServiceReference |
policy | Activity policy. | ActivityPolicy |
typeProperties | Script activity properties. | ScriptActivityTypeProperties (required) |
Name | Description | Value |
---|---|---|
logSettings | Log settings of script activity. | ScriptActivityTypePropertiesLogSettings |
scriptBlockExecutionTimeout | ScriptBlock execution timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | |
scripts | Array of script blocks. Type: array. | ScriptActivityScriptBlock[] |
Name | Description | Value |
---|---|---|
logDestination | The destination of logs. Type: string. | 'ActivityOutput' 'ExternalStore' (required) |
logLocationSettings | Log location settings customer needs to provide when enabling log. | LogLocationSettings |
Name | Description | Value |
---|---|---|
parameters | Array of script parameters. Type: array. | ScriptActivityParameter[] |
text | The query text. Type: string (or Expression with resultType string). | |
type | The type of the query. Type: string. | 'NonQuery' 'Query' (required) |
Name | Description | Value |
---|---|---|
direction | The direction of the parameter. | 'Input' 'InputOutput' 'Output' |
name | The name of the parameter. Type: string (or Expression with resultType string). | |
size | The size of the output direction parameter. | int |
type | The type of the parameter. | 'Boolean' 'DateTime' 'DateTimeOffset' 'Decimal' 'Double' 'Guid' 'Int16' 'Int32' 'Int64' 'Single' 'String' 'Timespan' |
value | The value of the parameter. Type: string (or Expression with resultType string). |
Name | Description | Value |
---|---|---|
type | Type of activity. | 'SetVariable' (required) |
policy | Activity policy. | SecureInputOutputPolicy |
typeProperties | Set Variable activity properties. | SetVariableActivityTypeProperties (required) |
Name | Description | Value |
---|---|---|
secureInput | When set to true, Input from activity is considered as secure and will not be logged to monitoring. | bool |
secureOutput | When set to true, Output from activity is considered as secure and will not be logged to monitoring. | bool |
Name | Description | Value |
---|---|---|
setSystemVariable | If set to true, it sets the pipeline run return value. | bool |
value | Value to be set. Could be a static value or Expression. | |
variableName | Name of the variable whose value needs to be set. | string |
Name | Description | Value |
---|---|---|
type | Type of activity. | 'SparkJob' (required) |
linkedServiceName | Linked service reference. | LinkedServiceReference |
policy | Activity policy. | ActivityPolicy |
typeProperties | Execute spark job activity properties. | SynapseSparkJobActivityTypeProperties (required) |
Name | Description | Value |
---|---|---|
args | User specified arguments to SynapseSparkJobDefinitionActivity. | any[] |
className | The fully-qualified identifier or the main class that is in the main definition file, which will override the 'className' of the spark job definition you provide. Type: string (or Expression with resultType string). | |
conf | Spark configuration properties, which will override the 'conf' of the spark job definition you provide. | |
configurationType | The type of the spark config. | 'Artifact' 'Customized' 'Default' |
driverSize | Number of core and memory to be used for driver allocated in the specified Spark pool for the job, which will be used for overriding 'driverCores' and 'driverMemory' of the spark job definition you provide. Type: string (or Expression with resultType string). | |
executorSize | Number of core and memory to be used for executors allocated in the specified Spark pool for the job, which will be used for overriding 'executorCores' and 'executorMemory' of the spark job definition you provide. Type: string (or Expression with resultType string). | |
file | The main file used for the job, which will override the 'file' of the spark job definition you provide. Type: string (or Expression with resultType string). | |
files | (Deprecated. Please use pythonCodeReference and filesV2) Additional files used for reference in the main definition file, which will override the 'files' of the spark job definition you provide. | any[] |
filesV2 | Additional files used for reference in the main definition file, which will override the 'jars' and 'files' of the spark job definition you provide. | any[] |
numExecutors | Number of executors to launch for this job, which will override the 'numExecutors' of the spark job definition you provide. Type: integer (or Expression with resultType integer). | |
pythonCodeReference | Additional python code files used for reference in the main definition file, which will override the 'pyFiles' of the spark job definition you provide. | any[] |
scanFolder | Scanning subfolders from the root folder of the main definition file, these files will be added as reference files. The folders named 'jars', 'pyFiles', 'files' or 'archives' will be scanned, and the folders name are case sensitive. Type: boolean (or Expression with resultType boolean). | |
sparkConfig | Spark configuration property. | SynapseSparkJobActivityTypePropertiesSparkConfig |
sparkJob | Synapse spark job reference. | SynapseSparkJobReference (required) |
targetBigDataPool | The name of the big data pool which will be used to execute the spark batch job, which will override the 'targetBigDataPool' of the spark job definition you provide. | BigDataPoolParametrizationReference |
targetSparkConfiguration | The spark configuration of the spark job. | SparkConfigurationParametrizationReference |
Name | Description | Value |
---|---|---|
{customized property} |
Name | Description | Value |
---|---|---|
referenceName | Reference spark job name. Expression with resultType string. | |
type | Synapse spark job reference type. | 'SparkJobDefinitionReference' (required) |
Name | Description | Value |
---|---|---|
referenceName | Reference big data pool name. Type: string (or Expression with resultType string). | |
type | Big data pool reference type. | 'BigDataPoolReference' (required) |
Name | Description | Value |
---|---|---|
referenceName | Reference spark configuration name. Type: string (or Expression with resultType string). | |
type | Spark configuration reference type. | 'SparkConfigurationReference' (required) |
Name | Description | Value |
---|---|---|
type | Type of activity. | 'SqlServerStoredProcedure' (required) |
linkedServiceName | Linked service reference. | LinkedServiceReference |
policy | Activity policy. | ActivityPolicy |
typeProperties | SQL stored procedure activity properties. | SqlServerStoredProcedureActivityTypeProperties (required) |
Name | Description | Value |
---|---|---|
storedProcedureName | Stored procedure name. Type: string (or Expression with resultType string). | |
storedProcedureParameters | Value and type setting for stored procedure parameters. Example: "{Parameter1: {value: "1", type: "int"}}". |
Name | Description | Value |
---|---|---|
type | Type of activity. | 'Switch' (required) |
typeProperties | Switch activity properties. | SwitchActivityTypeProperties (required) |
Name | Description | Value |
---|---|---|
cases | List of cases that correspond to expected values of the 'on' property. This is an optional property and if not provided, the activity will execute activities provided in defaultActivities. | SwitchCase[] |
defaultActivities | List of activities to execute if no case condition is satisfied. This is an optional property and if not provided, the activity will exit without any action. | Activity[] |
on | An expression that would evaluate to a string or integer. This is used to determine the block of activities in cases that will be executed. | Expression (required) |
Name | Description | Value |
---|---|---|
activities | List of activities to execute for satisfied case condition. | Activity[] |
value | Expected value that satisfies the expression result of the 'on' property. | string |
Name | Description | Value |
---|---|---|
type | Type of activity. | 'SynapseNotebook' (required) |
linkedServiceName | Linked service reference. | LinkedServiceReference |
policy | Activity policy. | ActivityPolicy |
typeProperties | Execute Synapse notebook activity properties. | SynapseNotebookActivityTypeProperties (required) |
Name | Description | Value |
---|---|---|
conf | Spark configuration properties, which will override the 'conf' of the notebook you provide. | |
configurationType | The type of the spark config. | 'Artifact' 'Customized' 'Default' |
driverSize | Number of core and memory to be used for driver allocated in the specified Spark pool for the session, which will be used for overriding 'driverCores' and 'driverMemory' of the notebook you provide. Type: string (or Expression with resultType string). | |
executorSize | Number of core and memory to be used for executors allocated in the specified Spark pool for the session, which will be used for overriding 'executorCores' and 'executorMemory' of the notebook you provide. Type: string (or Expression with resultType string). | |
notebook | Synapse notebook reference. | SynapseNotebookReference (required) |
numExecutors | Number of executors to launch for this session, which will override the 'numExecutors' of the notebook you provide. Type: integer (or Expression with resultType integer). | |
parameters | Notebook parameters. | SynapseNotebookActivityTypePropertiesParameters |
sparkConfig | Spark configuration property. | SynapseNotebookActivityTypePropertiesSparkConfig |
sparkPool | The name of the big data pool which will be used to execute the notebook. | BigDataPoolParametrizationReference |
targetSparkConfiguration | The spark configuration of the spark job. | SparkConfigurationParametrizationReference |
Name | Description | Value |
---|---|---|
referenceName | Reference notebook name. Type: string (or Expression with resultType string). | |
type | Synapse notebook reference type. | 'NotebookReference' (required) |
Name | Description | Value |
---|---|---|
{customized property} | NotebookParameter |
Name | Description | Value |
---|---|---|
type | Notebook parameter type. | 'bool' 'float' 'int' 'string' |
value | Notebook parameter value. Type: string (or Expression with resultType string). |
Name | Description | Value |
---|---|---|
{customized property} |
Name | Description | Value |
---|---|---|
type | Type of activity. | 'Until' (required) |
typeProperties | Until activity properties. | UntilActivityTypeProperties (required) |
Name | Description | Value |
---|---|---|
activities | List of activities to execute. | Activity[] (required) |
expression | An expression that would evaluate to Boolean. The loop will continue until this expression evaluates to true | Expression (required) |
timeout | Specifies the timeout for the activity to run. If there is no value specified, it takes the value of TimeSpan.FromDays(7) which is 1 week as default. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
Name | Description | Value |
---|---|---|
type | Type of activity. | 'Validation' (required) |
typeProperties | Validation activity properties. | ValidationActivityTypeProperties (required) |
Name | Description | Value |
---|---|---|
childItems | Can be used if dataset points to a folder. If set to true, the folder must have at least one file. If set to false, the folder must be empty. Type: boolean (or Expression with resultType boolean). | |
dataset | Validation activity dataset reference. | DatasetReference (required) |
minimumSize | Can be used if dataset points to a file. The file must be greater than or equal in size to the value specified. Type: integer (or Expression with resultType integer). | |
sleep | A delay in seconds between validation attempts. If no value is specified, 10 seconds will be used as the default. Type: integer (or Expression with resultType integer). | |
timeout | Specifies the timeout for the activity to run. If there is no value specified, it takes the value of TimeSpan.FromDays(7) which is 1 week as default. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
Name | Description | Value |
---|---|---|
type | Type of activity. | 'Wait' (required) |
typeProperties | Wait activity properties. | WaitActivityTypeProperties (required) |
Name | Description | Value |
---|---|---|
waitTimeInSeconds | Duration in seconds. Type: integer (or Expression with resultType integer). |
Name | Description | Value |
---|---|---|
type | Type of activity. | 'WebActivity' (required) |
linkedServiceName | Linked service reference. | LinkedServiceReference |
policy | Activity policy. | ActivityPolicy |
typeProperties | Web activity properties. | WebActivityTypeProperties (required) |
Name | Description | Value |
---|---|---|
authentication | Authentication method used for calling the endpoint. | WebActivityAuthentication |
body | Represents the payload that will be sent to the endpoint. Required for POST/PUT method, not allowed for GET method Type: string (or Expression with resultType string). | |
connectVia | The integration runtime reference. | IntegrationRuntimeReference |
datasets | List of datasets passed to web endpoint. | DatasetReference[] |
disableCertValidation | When set to true, Certificate validation will be disabled. | bool |
headers | Represents the headers that will be sent to the request. For example, to set the language and type on a request: "headers" : { "Accept-Language": "en-us", "Content-Type": "application/json" }. Type: string (or Expression with resultType string). | WebActivityTypePropertiesHeaders |
httpRequestTimeout | Timeout for the HTTP request to get a response. Format is in TimeSpan (hh:mm:ss). This value is the timeout to get a response, not the activity timeout. The default value is 00:01:00 (1 minute). The range is from 1 to 10 minutes | |
linkedServices | List of linked services passed to web endpoint. | LinkedServiceReference[] |
method | Rest API method for target endpoint. | 'DELETE' 'GET' 'POST' 'PUT' (required) |
turnOffAsync | Option to disable invoking HTTP GET on location given in response header of a HTTP 202 Response. If set true, it stops invoking HTTP GET on http location given in response header. If set false then continues to invoke HTTP GET call on location given in http response headers. | bool |
url | Web activity target endpoint and path. Type: string (or Expression with resultType string). |
Name | Description | Value |
---|---|---|
credential | The credential reference containing authentication information. | CredentialReference |
password | Password for the PFX file or basic authentication / Secret when used for ServicePrincipal | SecretBase |
pfx | Base64-encoded contents of a PFX file or Certificate when used for ServicePrincipal | SecretBase |
resource | Resource for which Azure Auth token will be requested when using MSI Authentication. Type: string (or Expression with resultType string). | |
type | Web activity authentication (Basic/ClientCertificate/MSI/ServicePrincipal) | string |
username | Web activity authentication user name for basic authentication or ClientID when used for ServicePrincipal. Type: string (or Expression with resultType string). | |
userTenant | TenantId for which Azure Auth token will be requested when using ServicePrincipal Authentication. Type: string (or Expression with resultType string). |
Name | Description | Value |
---|---|---|
referenceName | Reference credential name. | string (required) |
type | Credential reference type. | 'CredentialReference' (required) |
{customized property} |
Name | Description | Value |
---|---|---|
{customized property} | string |
Name | Description | Value |
---|---|---|
type | Type of activity. | 'WebHook' (required) |
policy | Activity policy. | SecureInputOutputPolicy |
typeProperties | WebHook activity properties. | WebHookActivityTypeProperties (required) |
Name | Description | Value |
---|---|---|
authentication | Authentication method used for calling the endpoint. | WebActivityAuthentication |
body | Represents the payload that will be sent to the endpoint. Required for POST/PUT method, not allowed for GET method Type: string (or Expression with resultType string). | |
headers | Represents the headers that will be sent to the request. For example, to set the language and type on a request: "headers" : { "Accept-Language": "en-us", "Content-Type": "application/json" }. Type: string (or Expression with resultType string). | WebHookActivityTypePropertiesHeaders |
method | Rest API method for target endpoint. | 'POST' (required) |
reportStatusOnCallBack | When set to true, statusCode, output and error in callback request body will be consumed by activity. The activity can be marked as failed by setting statusCode >= 400 in callback request. Default is false. Type: boolean (or Expression with resultType boolean). | |
timeout | The timeout within which the webhook should be called back. If there is no value specified, it defaults to 10 minutes. Type: string. Pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | string |
url | WebHook activity target endpoint and path. Type: string (or Expression with resultType string). |
Name | Description | Value |
---|---|---|
{customized property} | string |
Name | Description | Value |
---|---|---|
name | The name of the folder that this Pipeline is in. | string |
Name | Description | Value |
---|---|---|
{customized property} | ParameterSpecification |
Name | Description | Value |
---|---|---|
defaultValue | Default value of parameter. | |
type | Parameter type. | 'Array' 'Bool' 'Float' 'Int' 'Object' 'SecureString' 'String' (required) |
Name | Description | Value |
---|---|---|
elapsedTimeMetric | Pipeline ElapsedTime Metric Policy. | PipelineElapsedTimeMetricPolicy |
Name | Description | Value |
---|---|---|
duration | TimeSpan value, after which an Azure Monitoring Metric is fired. |
Name | Description | Value |
---|---|---|
{customized property} |
Name | Description | Value |
---|---|---|
{customized property} | VariableSpecification |
Name | Description | Value |
---|---|---|
defaultValue | Default value of variable. | |
type | Variable type. | 'Array' 'Bool' 'String' (required) |
The following quickstart templates deploy this resource type.
Template | Description |
---|---|
More is possible with Azure Data Factory - One click to try Azure Data Factory |
This template creates a data factory pipeline for a copy activity from Azure Blob into another Azure Blob |
Create a Data Factory, copy from Salesforce to Blob |
This template uses Azure Data Factory to perform data movement from Salesforce to Azure Blobs |
Bulk copy using Azure Data Factory |
This template creates a V2 data factory that copies data in bulk an Azure SQL database to an Azure SQL data warehouse. |
Create a V2 data factory |
This template creates a V2 data factory that copies data from a folder in an Azure Blob Storage to another folder in the storage. |
Create a V2 data factory (MySQL) |
This template creates a V2 data factory that copies data from a folder in an Azure Blob Storage to a table in an Azure Database for MySQL. |
Create a V2 data factory (PostGre) |
This template creates a V2 data factory that copies data from a folder in an Azure Blob Storage to a table in an Azure Database for PostgreSQL. |
Create a V2 data factory (SQL) |
This template creates a V2 data factory that copies data from a folder in an Azure Blob Storage to a table in an Azure SQL database. |
Create a V2 data factory (SQL On-prem) |
This template creates a V2 data factory that copies data from an on-premises SQL Server to an Azure blob storage |
Create a V2 data factory (Spark) |
This template creates a V2 data factory that copies data from a folder in an Azure Blob Storage to another folder in the storage. |