PLATFORMS SOLUTIONS BLOGS CONTACT


DataZen HTTP API 2.0 Documentation

Enzo DataZen


Introduction

The DataZen HTTP API allows developers to control and program DataZen agents and jobs. The HTTP API provided in this documentation is the same used by the DataZen Manager application.

In order to successfully access DataZen agents and jobs using the HTTP API, you will need the following:

  • A DataZen agent installed and running
  • A valid license key
  • HTTP API Authentication Keys

Follow the steps provided in the Installation Guide to configure the agent and apply your license key.

License Key

A valid license key needs to be applied to the DataZen agent against which the HTTP APIs will be executed. The easiest way to apply the license key is to use the DataZen Manager application.

Security Configuration

The HTTP API requires the use of Access Keys. A DataZen agent can be configured to restrict access to the API using the following mechanisms:

  • API Access: the HTTP calls can be restriced to accessing either the Agent API or the Jobs API
  • HTTP Get: the HTTP calls can be restricted to access GET operations only (read-only)

Shared Access Keys are automatically generated when an agent is configured the first time. To view the Shared Access Keys, start the DataZen Manager, select the agent, and choose Configuration -> Agent Settings, and choose the Shared Access Keys tab.



Restrictions

A few HTTP API handlers are not available for security reasons and require the DataZen Manager to administer:

  • Agent Configuration: Any agent configuration setting, including setting the license key, cannot be modified using the HTTP API
  • Connection Strings: Connection Strings used in jobs cannot be created, updated, or deleted using the HTTP API

Sample HTTP Call

The following HTTP call shows you how to query an agent, listening on port 9559, authenticate using a Shared Access Key (using the Authorization Header) and returns a JSON payload that provides the current version of the agent:

All HTTP calls to a DataZen agent must be authenticated using the Authentication header. The only authentication mode supported in this release is SharedKey.

HTTP Request

GET http://localhost:9559/version  
Authorization: SharedKey 123456789426478564563123

Response

{
"URL":"http://localhost:9559",
"Running":true,
"Major":1,
"Minor":0,
"Revision":23770,
"Build":7866,
"JobsRunning":0
}

Parameters

If an API requires parameters, they are to be added as part of the Query String of the URL. The following example shows how to call the /job/output endpoint:

GET http://localhost:9559/job/output?guid=B427BC0D-57DB-4C88-A847-7B4D07AFBCC5 
Authorization: SharedKey 123456789426478564563123


Agent API Section

This section covers the HTTP APIs available to manage and monitor DataZen agents.


License

Parses a license file and returns detailed information about the license provided.

GET
/licenseinfo

Retrieves licensing details given a license file payload.

Response
object DataZenLicense show/hide
bool AcceptSourceWhen true, the agent can process Job Reader tasks
bool AcceptTargetWhen true, the agent can process Job Writer tasks
bool AllowDatabasesWhen true, the agent can process DB systems
bool SchedulerAllowedWhen true, the agent allows scheduling job execution
bool DataPipelineAllowedWhen true, the agent allows data pipelines
bool DirectJobsOnlyWhen true, the agent only accepts Direct Jobs (reader and writer as part of the same job)
bool AllowMessagingHubsWhen true, the agent allows the use of MESSAGING systems
bool AllowDriveFilesWhen true, the agent allows the use of DRIVE systems
bool AllowJobTriggersWhen true, the agent can start downstream jobs upon completion of a job
bool AllowPGPEncryptionWhen true, the agent allows using PGP encryption/decryption
bool AllowQueuingSourceWhen true, the agent can act as a Messaging Consumer against any supported messsaging platform
int MaxJobsWhen positive, the agent is limited to the specified number of jobs (includes valid Job Pack licenses)
bool AllowRBACWhen true, the agent allows RBAC rules to be modified (default: NT Authority\Authenticated)
bool AllowHeuristicsWhen true, the agent allows overriding the default Heuristic settings for jobs
string EditionThe edition of this license (professional, standard, advanced, enterprise)
string HeaderThe license file header information
string UserNameThe user name the license is assigned to
string CompanyNameThe company the license is assigned to
string EmailAddressThe email the license is assigned to
string LicenseFileThe content of the license file
bool IsValidWhen true, the license applied to the agent is valid
string LicenseValidationErrorThe license validation error, if any
string LastValidationMessageLast license-related validation error when saving a job, if any

request
POST /licenseinfo
John Miller
ACME
jmiller@acme.info
[SyncSource;SyncTarget;Enterprise]
AF2FFFFF62606FFFFF152B007BFFFFF0A3B28F
FFFFFF10F1731FFFFFFFFFFCCCBF8ABFFFFFFECBA6605E0FFFFFFFFFF1EEAFFF
response
{
    "DataZenLicense": {
        "AcceptSource":true,
        "AcceptTarget":true,
        "AllowDatabases":true,
        "SchedulerAllowed":true,
        "DataPipelineAllowed":true,
        "DirectJobsOnly":false,
        "AllowMessagingHubs":true,
        "AllowDriveFiles":true,
        "AllowPGPEncryption":true,
        "MaxJobs":-1,
        "AllowDataCopy":true,
        "AllowRBAC":true,
        "AllowHeuristics":true,
        "Edition":"Enterprise",
        "LicenseFile":"...",
        "IsValid":false,
        "LicenseValidationError":null,
        "LicenseValidationResult":4,
        "LastValidationMessage":"License is invalid"
    }
}

Agent Status

Provides the current status of a DataZen agent.

GET
/status

Retrieves configuration settings and status of a DataZen agent including licensing details.

Response
string URL The URL this agent is listening on (ex: http://localhost:9559)
bool Running When true, the agent is in a running status
string StageDatabaseServer The name of the database server where the Stage database is located
string StageDatabaseName The name of the Stage database
string StageLogin The UserId used for the Stage database
string LicenseFile The License File for the agent
object DataZenLicense show/hide
bool AcceptSourceWhen true, the agent can process Job Reader tasks
bool AcceptTargetWhen true, the agent can process Job Writer tasks
bool AllowDatabasesWhen true, the agent can process DB systems
bool SchedulerAllowedWhen true, the agent allows scheduling job execution
bool DataPipelineAllowedWhen true, the agent allows data pipelines
bool DirectJobsOnlyWhen true, the agent only accepts Direct Jobs (reader and writer as part of the same job)
bool AllowMessagingHubsWhen true, the agent allows the use of MESSAGING systems
bool AllowDriveFilesWhen true, the agent allows the use of DRIVE systems
bool AllowJobTriggersWhen true, the agent can start downstream jobs upon completion of a job
bool AllowPGPEncryptionWhen true, the agent allows using PGP encryption/decryption
bool AllowQueuingSourceWhen true, the agent can act as a Messaging Consumer against any supported messsaging platform
int MaxJobsWhen positive, the agent is limited to the specified number of jobs (includes valid Job Pack licenses)
bool AllowRBACWhen true, the agent allows RBAC rules to be modified (default: NT Authority\Authenticated)
bool AllowHeuristicsWhen true, the agent allows overriding the default Heuristic settings for jobs
string EditionThe edition of this license (professional, standard, advanced, enterprise)
string HeaderThe license file header information
string UserNameThe user name the license is assigned to
string CompanyNameThe company the license is assigned to
string EmailAddressThe email the license is assigned to
string LicenseFileThe content of the license file
bool IsValidWhen true, the license applied to the agent is valid
string LicenseValidationErrorThe license validation error, if any
string LastValidationMessageLast license-related validation error when saving a job, if any
object Version show/hide
int BuildThe agent build number
int JobsRunningNumber of jobs currently executing
int MajorThe agent major version number
int MinorThe agent minor version number
int RevisionThe agent revision version number
bool RunningWhen true, the agent is running and can process jobs
string URLThe URL the agent is listening on

request
GET /status
response
{
    "URL":"http://localhost:9559",
    "Running":true,
    "StageDatabaseServer":"devlap04",
    "StageDatabaseName":"EnzoStaging",
    "StageLogin":"",
    "LicenseFile":"...",
    "DataZenLicense": {
        "AcceptSource":true,
        "AcceptTarget":true,
        "AllowDatabases":true,
        "SchedulerAllowed":true,
        "DataPipelineAllowed":true,
        "DirectJobsOnly":false,
        "AllowMessagingHubs":true,
        "AllowDriveFiles":true,
        "AllowPGPEncryption":true,
        "MaxJobs":-1,
        "AllowDataCopy":true,
        "AllowRBAC":true,
        "AllowHeuristics":true,
        "Edition":"Enterprise",
        "LicenseFile":"...",
        "IsValid":true,
        "LicenseValidationError":null,
        "LicenseValidationResult":4,
        "LastValidationMessage":null
    },
    "Version": {
        "URL":"http://localhost:9559",
        "Running":true,
        "Major":1,
        "Minor":0,
        "Revision":23770,
        "Build":7866,
        "JobsRunning":0
    }
}

Version

Provides the version of an agent.

GET
/version

Retrieves the current version of a DataZen agent.

Response
object Version show/hide
int BuildThe agent build number
int JobsRunningNumber of jobs currently executing
int MajorThe agent major version number
int MinorThe agent minor version number
int RevisionThe agent revision version number
bool RunningWhen true, the agent is running and can process jobs
string URLThe URL the agent is listening on

request
GET /version
{
    "Version": {
        "URL":"http://localhost:9559",
        "Running":true,
        "Major":1,
        "Minor":0,
        "Revision":23770,
        "Build":7866,
        "JobsRunning":0
    }
}

Connection Strings

Manages connection strings stored securely in DataZen.

GET
/connections

Gets available connection strings or the connection string for the key provided. To retrieve connection string metata (without the actual connection string itself) call the /connections/metadata endpoint instead for better performance.

Query Parameters
string key The key of the connection string to retrieve (optional)
string decrypt When 1, decrypts the data node of the connection string (optional). By default, connection strings are returned encrypted in the 'data' node.
string type When provided, filters the list with the Connection Type provided (ex: DB) (optional)
 
Response
Key The unique friendly name of the Central Connection String
Data (always empty)
System The system this connection points to (ex: MySQL, Oracle, AWS S3 Bucket, Google BigQuery, HTTP...)
UseCount The number of jobs that reference this connection string
ConnectionType The type of connection string (HTTP, Drive, DB, Messaging, BigData)

request
GET /connections?key=MyAWSS3Connection
response
[
 {
    "Key":"MyAWSS3Connection",
    "Data": "{\"service\":\"awss3\",\"displayName\":\"AWS S3 Bucket\",\"displayConnectionString\":\"testdata\",\"bucket\":\"testdata\",\"accessKey\":\"...\",\"accessSecret\":\"...\",\"region\":\"us-east-1\"}",
    "System":"AWS S3 Bucket",
    "UseCount":4,
    "ConnectionType":"Drive"
 }
]

Connection Strings Metadata

Retrieves available connection strings information without the actual connection string itself (the 'data' node is always empty).

GET
/connections/metadata

Gets information about all available connection strings or for the key provided, without returning the actual connection string.

Query Parameters
string key The key of the connection string to retrieve (optional)
 
Response
Key The unique friendly name of the Central Connection String
Data (always empty)
System The system this connection points to (ex: MySQL, Oracle, AWS S3 Bucket, Google BigQuery, HTTP...)
UseCount The number of jobs that reference this connection string
ConnectionType The type of connection string (HTTP, Drive, DB, Messaging, BigData)

request
GET /connections/metadata
response
[
 {
    "Key":"AWSS3",
    "Data":"",
    "System":"AWS S3 Bucket",
    "UseCount":4,
    "ConnectionType":"Drive"
 },
 {
    "Key":"localtesttmp",
    "Data":"",
    "System":"Local Path",
    "UseCount":1,
    "ConnectionType":"Drive"
 },
 {
    "Key":"devlap04",
    "Data":"",
    "System":"SQLServer",
    "UseCount":3,
    "ConnectionType":"DB"
 }
]

Start

Changes the status of the agent to "running" and resumes job processing. Any active job that missed its processing window will start.

POST
/start

Allows the DataZen agent to resume processing scheduled jobs.

Response
OK Success response code

request
POST /start
response
OK

Stop

Changes the status of the agent to "stopped" and stops automatic job processing. Does not cancel jobs that are still running, but prevents future active jobs from starting on schedule.

POST
/stop

Stop job processing.

Response
OK Success response code

request
POST /stop
response
OK

Dynamic Job

Dynamic Jobs allows you to run data movement jobs without first having to create them in DataZen. Dynamic Jobs only exist in memory in DataZen; their definitions are not stored. However, the execution summary of a Dynamic Job can be viewed, including its detailed output even after is has been completed. The following options are available for Dynamic Jobs:

/jobssummary?dynamic=1
/job/stopprocessing
/job/dynamic/executions
/job/dynamic
/job/log?execguid=
/job/output

GET
/job/dynamic
/job/dynamic?guid=...
/job/dynamic?jobkey=...

Returns the definition of a dynamic job. If guid is provided, the jobkey parameter is ignored. If neither value is provided, all available Dynamic Jobs found in memory are returned. Restarting DataZen clears all Dynamic Jobs from memory.

Query Parameters
string guid The job guid to filter on (optional)
string jobkey The job key to filter on (optional)
 
Response
object[] object
where object can be either a jobReader or jobWriter
  jobReader show/hide
bool ActiveWhen true, the job will execute at the desired interval as specified by the CronSchedule parameter
bool AuditLogEnabledWhen true, the job execution output will be saved so it can be retrieved later (using job/log)
string ConnectionStringThe name of the Central Connection String to use for the source system
DateTime CreatedOnThe date/time the job was created
string CronScheduleWhen specified, the Cron expression determine the frequency of the execution of the job; blank when the job is started manually.
string GuidThe job unique identifier
bool Initialize(internal use only)
bool IsDBSourceWhen true, the source system is a relational database
bool IsDirectWhen true, the job is a Direct job (with source and target defined)
bool IsDirty(internal use only)
bool IsDriveSourceWhen true, the source system is a file stored on a drive (local, cloud, or FTP)
bool IsHttpSourceWhen true, the source system is an HTTP REST endpoint
bool IsQueueingSourceWhen true, the source system is a messaging hub and the job acts as a Messaging Consumer
string JobKeyThe name of the job
int JobType1 (Job Reader)
DateTime LastRunTimeThe last execution date/time
string LastTsDelPointerThe last High Water Mark value for deleted records
string LastTsPointerThe last High Water Mark value for inserted or updated columns
string PathThe path where the sync file will be created
bool PropagateDeleteWhen true, identify or capture deleted records
string SqlReadFor DB connections, the SQL command to execute that will fetch source records
string StatusThe status of the job (running, ready)
string SyncStrategy(internal use: full)
string SystemNameThe name of the source system (SQLServer, MySQL, Oracle, Teradata, Generic, AWS S3 Bucket, Azure Blob Container, Local Path...)
string TimestampColWhen provided, the date/time or numeric column to use as the High Watermark when reading from a database
string TrackLastTsWhen True, the job tracks a High Watermark
string UpsertColumnsComma-separated list of columns making up the unique key of the record
object JsonData show/hide
bool AuditLogEnabledWhen true, saves the execution log for auditing purposes
string EncryptionFileWhen provided, the file that contains the PGP public key used for encrypting Sync Files
bool BypassTopNWhen true, bypasses the TOP N heuristic method for DB data sources
int TopNOverrideWhen positive, uses this value for the TOP N heuristic (default: 10000)
string DirectWriterWhen a direct job, the name of the Job Writer (usually set to the name of the Job Reader, with two underscored as the prefix)
int ExecutionTimeoutWhen positive, uses this value for individual requests as the timeout value (in seconds)
bool BypassSchemaDetectionWhen true, the job should not extract detailed schema information from the data source; applies to DB data sources only
int DeletedRecordsMethodWhen PropagateDelete is true, determines the method used to identify deleted records (0: default, 1: custom SQL that returns all source records, 2: custom SQL that returns the IDs of the deleted records)
string DeletedRecordsSQLWhen DeletedRecordsMethod is greater than 0, specifies the SQL command used to identify deleted records
string CDCTokenDeletedFieldWhen DeletedRecordsMethod is 2, comma-separated column names returned by DeletedRecordsSQL that uniquely identify the deleted record
object MessagingOptions show/hide
int MessageProcessingThe messaging processing mode (0: passthrough, 1: Msg to Dataset, 2: Msg Payload to Dataset)
bool MessagingBatchEnabledWhen true, messages are batched before being sent to their intended target
int MessagingBatchCountWhen MessagingBatchEnabled is true, number of messages to include in a batch
int MessagingFlushTTLWhen positive and MessagingBatchEnabled is true, the maximum time to wait before messages are sent if the BatchCount hasn't been met
string MessagingSampleMsgSample message to use for testing purposes
string MessagingContentColumnWhen MessageProcessing is 2 and not empty, adds a dynamic column to the data of this name with the original content of the payload
string RootPathWhen MessageProcessing is 2 and not empty, XML or JSON path to use to extract the Dataset
bool IncludeHeadersWhen true, headers should be sent to the target system as much as possible
bool IncludeMetadataWhen true, message metadata should be sent to the target system as much as possible
string HeaderFieldsWhen not empty and MessageProcessing is NOT 2 and IncludeHeaders is true, the comma-separated list of header keys to send to the target system (blank sends all)
string MetadataFieldsWhen not empty and MessageProcessing is NOT 2 and IncludeHeaders is true, the comma-separated list of metadata fields to send to the target system (blank sends all)
string QueueNameOverrideWhen provided, the name of the queue/bus/topic to read messages from (if blank, uses the connection string default settings)
string MessagingGroupNameWhen relevant, the message group name to use when reading from the source messaging hub (if blank, reads from the default group)
string PartitionKeysWhen relevant, a comma-separated list of partition keys to use when reading from the source messaging hub (if blank, reads from all partitions)
object DriveSourceSettings show/hide
string DriveFileFormatThe messaging processing mode (0: passthrough, 1: Msg to Dataset, 2: Msg Payload to Dataset)
string DriveFilePatternWhen true, messages are batched before being sent to their intended target
string DriveSchemaFileWhen MessagingBatchEnabled is true, number of messages to include in a batch
string DriveSelectedColumnsWhen positive and MessagingBatchEnabled is true, the maximum time to wait before messages are sent if the BatchCount hasn't been met
string DriveUpdatedFilesOnlySample message to use for testing purposes
bool MessagingContentColumnWhen MessageProcessing is 2 and not empty, adds a dynamic column to the data of this name with the original content of the payload
bool DriveIncludeSubfoldersWhen MessageProcessing is 2 and not empty, XML or JSON path to use to extract the Dataset
string DriveFolderColumnNameWhen true, headers should be sent to the target system as much as possible
bool DriveAddFolderColumnWhen true, message metadata should be sent to the target system as much as possible
bool DriveUseRootPathWhen not empty and MessageProcessing is NOT 2 and IncludeHeaders is true, the list of header keys to send to the target system (blank sends all)
string DriveRootPathWhen not empty and MessageProcessing is NOT 2 and IncludeHeaders is true, the list of metadata fields to send to the target system (blank sends all)
bool SingleFileProcessingWhen provided, the name of the queue/bus/topic to read messages from (if blank, uses the connection string default settings)
object ExtendedOptions show/hide
CSVOptions
string ColumnDelimitersThe character that separate individual field values when reading from a delimited file (ignored if FieldWidths are provided)
bool HasHeaderRowWhen true, assumes the file has a header row and column names will be inferred from the header row
string FieldWidthsIf provided, a comma-separated list of integers representing the length of each column for a fixed-length file format
string[] CommentTokensAn array of string values to be treated as comment tokens in the file
bool QuotedIdentifierWhen true, will treat double-quotes as boundary tokens for field name and value
object Column2TableTxSettings show/hide
string TxDocumentMsgContentColThe column index or name to use from the dataset containing an XML or JSON string to be transformed into a dataset
string TxDocumentRootPathIf provided, the XML or JSON path to use as the starting point to read the document content
string TxDocumentSourceColumnIf provided, the name of a dynamic column that will be added to the dataset with the content of the JSON or XML record
object HttpSourceSettings show/hide
string HTTPContentTypeThe Content-Type header used when sending a POST or PUT request
string HTTPHeadersAn array of key-value pairs representing custom headers to set
string HTTPURIThe URI to call; if blank, will call the URI defined by the connection; may use a relative URL or a full URL
string HTTPVerbThe HTTP Verb to use (GET, POST, PUT, DELETE)
string HTTPPayloadWhen executing a POST or PUT request, the payload to send with the request
bool ApplyContentTxWhen true, converts the JSON or XML content of the payload into a dataset
int PagingStrategyThe paging strategy to use (0: none, 1: simple offset, 2: simple page, 3 simple link, 4: bookmark); when positive, the corrresponding OffsetStrategy object must be provided
string SelectedColumnsThe columns to return as part of the dataset; leave blank to return all available columns
string HighWaterMarkThe name of the field that contains the High Watermark for future calls
string HighWaterMarkParamThe name of the Query Parameter to use when adding a High Watermark value for future calls
dynamic OffsetStrategy show/hide
HttpSimpleOffsetStrategy
string URLOffsetParamThe name of the URL Query Parameter to use when specifying an offset
HttpSimplePageStrategy
string URLPageParamThe name of the URL Query Parameter to use when specifying a page number
HttpSimpleLinkStrategy
string URLPagePathThe JSON or XML document path where to find the next link to call within the response payload
HttpTokenPagingStrategy
string TokenPathThe JSON or XML document path where to find the next bookmark value to use within the response payload
string QueryParamNameThe namd of the URL Query Parameter to use when specifying the bookmark value
object PayloadTxSettings show/hide
string TxDocumentMsgContentColThe column index or name to use from the dataset containing an XML or JSON string to be transformed into a dataset
string TxDocumentRootPathIf provided, the XML or JSON path to use as the starting point to read the document content
string TxDocumentSourceColumnIf provided, the name of a dynamic column that will be added to the dataset with the content of the JSON or XML record
object[] JobTriggers show/hide
string guidThe Job GUID to trigger (must be provided)
string jobKeyThe JobKey of the job to trigger (informational only)
int eventIdThe condition that will fire the trigger (0: On Completion, 1: On Success (with wata), 2: On Success (always), 3: On Error)
string eventNameThe friendly name of the condition (optional)
true JobTriggersEnabledWhen true, executes the jobs specified in the JobTriggers collection upon completion of the job
dynamic[] PipelineDef show/hide
DataDynamicColumn
string NameThe name of the component: DataDynamicColumn
string ColumnNameThe column name to use
string DataTypeThe data type of the column (string, int, float, guid, datetime, long, double, char, bool, byte[])
string ExpressionTypeSQLExpression or EnzoFunction
string SQLExpressionThe expression (either a SQL expression, or an Enzo function call)
int MaxLengthColumn maximum length
bool IgnoreIfExistsWhen true, does not throw an error if the column already exists
bool DisabledThen true, bypass this component
DataHashing
string NameThe name of the component: DataHashing
string ColumnNameThe name of the column to hash
string HashNameThe hashing algorythm to use (MD5, SHA1, SHA256, SHA384, SHA512)
bool TruncateIfNeededWhen true, truncate the hash value to fit the column length
bool DisabledThen true, bypass this component
DataMasking
string NameThe name of the component: DataMasking
string ColumnNameThe name of the column to mask
string MaskingTypeThe masking operation (default, email, creditcard, randomnumber, customtext)
decimal RandomNbrMinWhen using RandomNumber masking, the minimum value to use
decimal RandomNbrMaxWhen using RandomNumber masking, the maximum value to use
int CustomPrefixWhen using CustomText, the prefix to use
int CustomSuffixWhen using CustomText, the suffix to use
string CustomPaddingWhen using CustomText, the padding to use
bool DisabledThen true, bypass this component
DataFilter
string NameThe name of the component: DataFilter
string FilterThe SQL filter to use to limit the records being processed
bool DisabledThen true, bypass this component
DataQuality
string NameThe name of the component: DataQuality
string ColumnNameThe column name to validate
string AllowedValuesComma-separated list of values to allow or deny
bool CheckListFlagWhen true, limit possible values to the list provided by the AllowedValues field
bool AllowNullWhen true, NULL values are allowed
bool AllowEmptyWhen true, empty strings are allowed
bool NumericRequiredWhen true, the value must be numeric
bool ThrowExceptionOnFailThen true, fail processing if the data validation rule failed
bool SkipRowOnFailWhen true, ignore the row if the data validation rule failed
bool OptOutListWhen true, the AllowedValues should be allowed (opt-in); else AllowedValues should be considered as opt-out.
bool DisabledThen true, bypass this component
  jobWriter show/hide
string GuidThe job unique identifier
string JobKeyThe name of the job
string PathThe path where the sync files are located
string UpsertColumnsComma-separated list of columns making up the unique key of the record
string ConnectionStringThe name of the Central Connection String to use for the target system
bool PropagateDeleteWhen true, identify or capture deleted records if available
string SourceJobKeyThe original name of the source job
string LastExecutionIdThe last ExecutionId extracted
string InitialExecutionIdThe initial ExecutionId used by this job
DateTime CreatedOnDate/time this job was created
string StatusThe status of this job (ready, running)
bool ActiveWhen true, the job will execute at the desired interval as specified by the CronSchedule parameter
DateTime LastRunTimeThe last execution date/time
string CronScheduleWhen specified, the Cron expression determine the frequency of the execution of the job
object JsonData
This setting is reserved and provided for information only
string EncryptionFileThe path to the file containing the PGP Private Key used to decrypt Sync Files
byte[] SaltThe cryptographic Salt used to encrypt the password
byte[] KEKThe encrypted password used to access the file
string TargetSystemThe target system (DB, DRIVE, MESSAGING, HTTP)
string TargetObjectThe name of the target object (for DB systems, the name of the table)
string UpsertScriptThe insert/update script to execute for DB systems (not needed for SQL Server (native) targets)
string DeleteScriptThe delete script to execute for DB systems (not needed for SQL Server (native) targets)
string LastSchemaHashThe hash of the source schema
object Options show/hide
bool AutoRecreateWhen true, for DB systems, determines if the target object should be recreated automatically
int ExecTimeoutWhen positive, controls the timeout of the write operation
int BatchCountWhen positive, sends the upserted or deleted records in batches of records
bool AuditLogEnabledWhen true, saves execution logs to the database for auditing purposes
HTTP Target Options show/hide
object HttpUpsertSettings
string HTTPUpsertURIFor HTTP targets, the URI of the HTTP endpoint for Upsert operations
string HTTPUpsertVerbFor HTTP targets, the HTTP Verb (GET, POST, DELETE, PUT) to use for Upsert operations
string HTTPUpsertContentTypeFor HTTP targets, the Content-Type header to use for Upsert operations
string HTTPUpsertHeadersFor HTTP targets, serialized list of headers and values for Upsert operations
object HttpDeleteSettings
string HTTPDeleteURIFor HTTP targets, the URI of the HTTP endpoint for Delete operations
string HTTPDeleteVerbFor HTTP targets, the HTTP Verb (GET, POST, DELETE, PUT) to use for Delete operations
string HTTPDeleteContentTypeFor HTTP targets, the Content-Type header to use for Delete operations
string HTTPDeleteHeadersFor HTTP targets, serialized list of headers and values for Delete operations
MESSAGING Target Options show/hide
object MessagingUpsertSettings
string SystemTypeA valid messaging system type (awssqs, rabbitmq, kafka, azeventhub)
rabbitmq settings:
string RoutingKeyThe RabbitMQ Routing Key
string ReplyToThe reply to queue
int PriorityThe priority of the message
string MessageIdThe message id (if any)
string AppIdThe application id (if any)
long expirationThe time to live of the message
bool persistWhen true, persist the message in the queue
awssqs settings:
string QueueNameThe AWS Queue Name (optional)
azeventhub settings:
string EventHubNameThe Azure Event Hub name (optional)
kafka settings:
string TopicNameThe Kafka topic name (optional)
string FieldKeyThe comma-separated list of key fields used to identify a unique message (the Upsert Key Columns)
string KeyFieldSeparatorThe Key values separator string to use (only needed when 2 or more key fields are provided)
object MessagingDeleteSettings
string SystemTypeA valid messaging system type (awssqs, rabbitmq, kafka, azeventhub)
refer to the MessageUpsertSettings section above
DRIVE Target Options show/hide
object TargetFileOptions
ContainerNameThe directory path, Azure Container name, or S3 Bucket Name
DistributionFieldNameThe field name used to dynamically distribute records across multiple files
FileNameThe file name to use (accepts field names and date tokens; ex: sales-[yyyy]-\{\{statecode\}\} )
UseRuntimeDateWhen true, use the execution date/time for date tokens
UseShadowCopyWhen true, use a shadow copy of remote files for faster updates
ShadowCopyTTLWhen UseShadowCopy is true, indicates how old the file can be before downloading a more recent version
ShadowCopyFrequencyThe ShadowCopyTTL value (hour, minute)
string SystemNameThe name of the target system (SQLServer, MySQL, Oracle, Teradata, Generic, AWS S3 Bucket, Azure Blob Container, Local Path, AWS SQS, Azure Event Hub, RabbitMQ)
dynamic[] PipelineDef show/hide
DataDynamicColumn
string NameThe name of the component: DataDynamicColumn
string ColumnNameThe column name to use
string DataTypeThe data type of the column (string, int, float, guid, datetime, long, double, char, bool, byte[])
string ExpressionTypeSQLExpression or EnzoFunction
string SQLExpressionThe expression (either a SQL expression, or an Enzo function call)
int MaxLengthColumn maximum length
bool IgnoreIfExistsWhen true, does not throw an error if the column already exists
bool DisabledThen true, bypass this component
DataHashing
string NameThe name of the component: DataHashing
string ColumnNameThe name of the column to hash
string HashNameThe hashing algorythm to use (MD5, SHA1, SHA256, SHA384, SHA512)
bool TruncateIfNeededWhen true, truncate the hash value to fit the column length
bool DisabledThen true, bypass this component
DataMasking
string NameThe name of the component: DataMasking
string ColumnNameThe name of the column to mask
string MaskingTypeThe masking operation (default, email, creditcard, randomnumber, customtext)
decimal RandomNbrMinWhen using RandomNumber masking, the minimum value to use
decimal RandomNbrMaxWhen using RandomNumber masking, the maximum value to use
int CustomPrefixWhen using CustomText, the prefix to use
int CustomSuffixWhen using CustomText, the suffix to use
string CustomPaddingWhen using CustomText, the padding to use
bool DisabledThen true, bypass this component
DataFilter
string NameThe name of the component: DataFilter
string FilterThe SQL filter to use to limit the records being processed
bool DisabledThen true, bypass this component
DataQuality
string NameThe name of the component: DataQuality
string ColumnNameThe column name to validate
string AllowedValuesComma-separated list of values to allow or deny
bool CheckListFlagWhen true, limit possible values to the list provided by the AllowedValues field
bool AllowNullWhen true, NULL values are allowed
bool AllowEmptyWhen true, empty strings are allowed
bool NumericRequiredWhen true, the value must be numeric
bool ThrowExceptionOnFailThen true, fail processing if the data validation rule failed
bool SkipRowOnFailWhen true, ignore the row if the data validation rule failed
bool OptOutListWhen true, the AllowedValues should be allowed (opt-in); else AllowedValues should be considered as opt-out.
bool DisabledThen true, bypass this component

request
GET /job/dynamic?guid=a3ff75315a4241d1958e1f9dbd53829f
response
[
 {
    "Guid":"a3ff75315a4241d1958e1f9dbd53829f",
    "JobKey":"a3ff75315a4241d1958e1f9dbd53829f",
    "SqlRead":"EXEC sharepoint._configuse 'default'; SELECT ID, Title, AccountNumber, CompanyId, Created, Modified FROM SharePoint.[list@companies]",
    "Path":"C:\\tmp\\datasync_reader",
    "UpsertColumns":"ID",
    "TimestampCol":null,
    "LastTsPointer":null,
    "LastTsDelPointer":null,
    "ConnectionString":"Data Source=localhost,9550;User Id=sa;password=.....",
    "PropagateDelete":false,
    "SyncStrategy":"Full",
    "CreatedOn":"2022-10-12T18:31:46.2373302Z",
    "Status":null,
    "Initialize":false,
    "Active":true,
    "LastRunTime":"2022-10-12T18:31:50.0782741Z",
    "CronSchedule":null,
    "JsonData": 
    {
        "DirectWriter":"__a3ff75315a4241d1958e1f9dbd53829f"
    },
    "PipelineDef":null,
    "SystemName":"Generic",
    "IsDBSource":true,
    "IsHttpSource":false,
    "IsDriveSource":false,
    "IsQueueingSource":false,
    "JobType":1,
    "IsDirty":false,
    "TrackLastTs":false,
    "AuditLogEnabled":false,
    "IsDirect":true,
    "IsCDCSource":false
 }
]
POST
/job/dynamic
/job/dynamic?resync=1

Executes a job using the definition provided. Dynamic jobs cannot be deleted or updated. Although the JobKey and Guid parameters are optional when creating a dynamic job, it is best to provide a value if you intend to query the other endpoints (such as the /log/output endpoint). Note that the Connection String is required, and can either be the key of an existing connection or a full connection string.
At a minimum either the JobReader or JobWriter node must be provided; if both are provided, the job is considered a Direct Job and the writer will start as soon as the reader completes successfully. If only the Job Reader is provided, a Change Log file will be created and stored in the Path provided if any records are available.
HINT: Use the Export JSON button on a Job Definition in DataZen Manager to obtain a valid template for creating dynamic jobs.

Query Parameters
resync (optional) When 1, runs a full resync of the source system (applies to Job Readers only)
recreate (optional) When 1, attempts to truncate or recreate the target object (applies to Job Writers only)
testrun (optional) When 1, reads a single record from the source system (applies to Job Readers only)
Payload
object JobCreation show/hide
object JobReader show/hide
bool AuditLogEnabledWhen true, the job execution output will be saved so it can be retrieved later (using job/log)
string ConnectionStringThe name of the Central Connection String to use for the source system
string CronScheduleWhen specified, the Cron expression determine the frequency of the execution of the job; blank when the job is started manually.
string GuidThe job unique identifier (when not provided, a random Guid will be assigned)
string JobKeyThe name of the job (when not provided, a random Guid will be assigned)
string LastTsDelPointerThe last High Water Mark value for deleted records
string LastTsPointerThe last High Water Mark value for inserted or updated columns
string PathThe path where the sync file will be created (if not provided, will be c:\)
bool PropagateDeleteWhen true, identify or capture deleted records
string SqlReadFor DB connections, the SQL command to execute that will fetch source records
string TimestampColWhen provided, the date/time or numeric column to use as the High Watermark when reading from a database
string TrackLastTsWhen True, the job tracks a High Watermark
string UpsertColumnsComma-separated list of columns making up the unique key of the record
object JsonData show/hide
bool AuditLogEnabledWhen true, saves the execution log for auditing purposes
string EncryptionFileWhen provided, the file that contains the PGP public key used for encrypting Sync Files
bool BypassTopNWhen true, bypasses the TOP N heuristic method for DB data sources
int TopNOverrideWhen positive, uses this value for the TOP N heuristic (default: 10000)
string DirectWriterWhen a direct job, the name of the Job Writer (usually set to the name of the Job Reader, with two underscored as the prefix)
int ExecutionTimeoutWhen positive, uses this value for individual requests as the timeout value (in seconds)
bool BypassSchemaDetectionWhen true, the job should not extract detailed schema information from the data source; applies to DB data sources only
int DeletedRecordsMethodWhen PropagateDelete is true, determines the method used to identify deleted records (0: default, 1: custom SQL that returns all source records, 2: custom SQL that returns the IDs of the deleted records)
string DeletedRecordsSQLWhen DeletedRecordsMethod is greater than 0, specifies the SQL command used to identify deleted records
string CDCTokenDeletedFieldWhen DeletedRecordsMethod is 2, comma-separated column names returned by DeletedRecordsSQL that uniquely identify the deleted record
object MessagingOptions show/hide
int MessageProcessingThe messaging processing mode (0: passthrough, 1: Msg to Dataset, 2: Msg Payload to Dataset)
bool MessagingBatchEnabledWhen true, messages are batched before being sent to their intended target
int MessagingBatchCountWhen MessagingBatchEnabled is true, number of messages to include in a batch
int MessagingFlushTTLWhen positive and MessagingBatchEnabled is true, the maximum time to wait before messages are sent if the BatchCount hasn't been met
string MessagingSampleMsgSample message to use for testing purposes
string MessagingContentColumnWhen MessageProcessing is 2 and not empty, adds a dynamic column to the data of this name with the original content of the payload
string RootPathWhen MessageProcessing is 2 and not empty, XML or JSON path to use to extract the Dataset
bool IncludeHeadersWhen true, headers should be sent to the target system as much as possible
bool IncludeMetadataWhen true, message metadata should be sent to the target system as much as possible
string HeaderFieldsWhen not empty and MessageProcessing is NOT 2 and IncludeHeaders is true, the comma-separated list of header keys to send to the target system (blank sends all)
string MetadataFieldsWhen not empty and MessageProcessing is NOT 2 and IncludeHeaders is true, the comma-separated list of metadata fields to send to the target system (blank sends all)
string QueueNameOverrideWhen provided, the name of the queue/bus/topic to read messages from (if blank, uses the connection string default settings)
string MessagingGroupNameWhen relevant, the message group name to use when reading from the source messaging hub (if blank, reads from the default group)
string PartitionKeysWhen relevant, a comma-separated list of partition keys to use when reading from the source messaging hub (if blank, reads from all partitions)
object DriveSourceSettings show/hide
string DriveFileFormatThe messaging processing mode (0: passthrough, 1: Msg to Dataset, 2: Msg Payload to Dataset)
string DriveFilePatternWhen true, messages are batched before being sent to their intended target
string DriveSchemaFileWhen MessagingBatchEnabled is true, number of messages to include in a batch
string DriveSelectedColumnsWhen positive and MessagingBatchEnabled is true, the maximum time to wait before messages are sent if the BatchCount hasn't been met
string DriveUpdatedFilesOnlySample message to use for testing purposes
bool MessagingContentColumnWhen MessageProcessing is 2 and not empty, adds a dynamic column to the data of this name with the original content of the payload
bool DriveIncludeSubfoldersWhen MessageProcessing is 2 and not empty, XML or JSON path to use to extract the Dataset
string DriveFolderColumnNameWhen true, headers should be sent to the target system as much as possible
bool DriveAddFolderColumnWhen true, message metadata should be sent to the target system as much as possible
bool DriveUseRootPathWhen not empty and MessageProcessing is NOT 2 and IncludeHeaders is true, the list of header keys to send to the target system (blank sends all)
string DriveRootPathWhen not empty and MessageProcessing is NOT 2 and IncludeHeaders is true, the list of metadata fields to send to the target system (blank sends all)
bool SingleFileProcessingWhen provided, the name of the queue/bus/topic to read messages from (if blank, uses the connection string default settings)
object ExtendedOptions show/hide
CSVOptions
string ColumnDelimitersThe character that separate individual field values when reading from a delimited file (ignored if FieldWidths are provided)
bool HasHeaderRowWhen true, assumes the file has a header row and column names will be inferred from the header row
string FieldWidthsIf provided, a comma-separated list of integers representing the length of each column for a fixed-length file format
string[] CommentTokensAn array of string values to be treated as comment tokens in the file
bool QuotedIdentifierWhen true, will treat double-quotes as boundary tokens for field name and value
object Column2TableTxSettings show/hide
string TxDocumentMsgContentColThe column index or name to use from the dataset containing an XML or JSON string to be transformed into a dataset
string TxDocumentRootPathIf provided, the XML or JSON path to use as the starting point to read the document content
string TxDocumentSourceColumnIf provided, the name of a dynamic column that will be added to the dataset with the content of the JSON or XML record
object HttpSourceSettings show/hide
string HTTPContentTypeThe Content-Type header used when sending a POST or PUT request
string HTTPHeadersAn array of key-value pairs representing custom headers to set
string HTTPURIThe URI to call; if blank, will call the URI defined by the connection; may use a relative URL or a full URL
string HTTPVerbThe HTTP Verb to use (GET, POST, PUT, DELETE)
string HTTPPayloadWhen executing a POST or PUT request, the payload to send with the request
bool ApplyContentTxWhen true, converts the JSON or XML content of the payload into a dataset
int PagingStrategyThe paging strategy to use (0: none, 1: simple offset, 2: simple page, 3 simple link, 4: bookmark); when positive, the corrresponding OffsetStrategy object must be provided
string SelectedColumnsThe columns to return as part of the dataset; leave blank to return all available columns
string HighWaterMarkThe name of the field that contains the High Watermark for future calls
string HighWaterMarkParamThe name of the Query Parameter to use when adding a High Watermark value for future calls
dynamic OffsetStrategy show/hide
HttpSimpleOffsetStrategy
string URLOffsetParamThe name of the URL Query Parameter to use when specifying an offset
HttpSimplePageStrategy
string URLPageParamThe name of the URL Query Parameter to use when specifying a page number
HttpSimpleLinkStrategy
string URLPagePathThe JSON or XML document path where to find the next link to call within the response payload
HttpTokenPagingStrategy
string TokenPathThe JSON or XML document path where to find the next bookmark value to use within the response payload
string QueryParamNameThe namd of the URL Query Parameter to use when specifying the bookmark value
object PayloadTxSettings show/hide
string TxDocumentMsgContentColThe column index or name to use from the dataset containing an XML or JSON string to be transformed into a dataset
string TxDocumentRootPathIf provided, the XML or JSON path to use as the starting point to read the document content
string TxDocumentSourceColumnIf provided, the name of a dynamic column that will be added to the dataset with the content of the JSON or XML record
object[] JobTriggers show/hide
string guidThe Job GUID to trigger (must be provided)
string jobKeyThe JobKey of the job to trigger (informational only)
int eventIdThe condition that will fire the trigger (0: On Completion, 1: On Success (with wata), 2: On Success (always), 3: On Error)
string eventNameThe friendly name of the condition (optional)
true JobTriggersEnabledWhen true, executes the jobs specified in the JobTriggers collection upon completion of the job
dynamic[] PipelineDef show/hide
DataDynamicColumn
string NameThe name of the component: DataDynamicColumn
string ColumnNameThe column name to use
string DataTypeThe data type of the column (string, int, float, guid, datetime, long, double, char, bool, byte[])
string ExpressionTypeSQLExpression or EnzoFunction
string SQLExpressionThe expression (either a SQL expression, or an Enzo function call)
int MaxLengthColumn maximum length
bool IgnoreIfExistsWhen true, does not throw an error if the column already exists
bool DisabledThen true, bypass this component
DataHashing
string NameThe name of the component: DataHashing
string ColumnNameThe name of the column to hash
string HashNameThe hashing algorythm to use (MD5, SHA1, SHA256, SHA384, SHA512)
bool TruncateIfNeededWhen true, truncate the hash value to fit the column length
bool DisabledThen true, bypass this component
DataMasking
string NameThe name of the component: DataMasking
string ColumnNameThe name of the column to mask
string MaskingTypeThe masking operation (default, email, creditcard, randomnumber, customtext)
decimal RandomNbrMinWhen using RandomNumber masking, the minimum value to use
decimal RandomNbrMaxWhen using RandomNumber masking, the maximum value to use
int CustomPrefixWhen using CustomText, the prefix to use
int CustomSuffixWhen using CustomText, the suffix to use
string CustomPaddingWhen using CustomText, the padding to use
bool DisabledThen true, bypass this component
DataFilter
string NameThe name of the component: DataFilter
string FilterThe SQL filter to use to limit the records being processed
bool DisabledThen true, bypass this component
DataQuality
string NameThe name of the component: DataQuality
string ColumnNameThe column name to validate
string AllowedValuesComma-separated list of values to allow or deny
bool CheckListFlagWhen true, limit possible values to the list provided by the AllowedValues field
bool AllowNullWhen true, NULL values are allowed
bool AllowEmptyWhen true, empty strings are allowed
bool NumericRequiredWhen true, the value must be numeric
bool ThrowExceptionOnFailThen true, fail processing if the data validation rule failed
bool SkipRowOnFailWhen true, ignore the row if the data validation rule failed
bool OptOutListWhen true, the AllowedValues should be allowed (opt-in); else AllowedValues should be considered as opt-out.
bool DisabledThen true, bypass this component
object JobWriter show/hide
string GuidThe job unique identifier (when not provided, a random Guid will be assigned)
string JobKeyThe name of the job (when not provided, a random Guid will be assigned)
string PathThe path where the sync files are located
string UpsertColumnsComma-separated list of columns making up the unique key of the record
string ConnectionStringThe name of the Central Connection String to use for the target system
bool PropagateDeleteWhen true, identify or capture deleted records if available
string SourceJobKeyThe original name of the source job (ignored for Direct Jobs)
string InitialExecutionIdThe initial ExecutionId used by this job
string CronScheduleWhen specified, the Cron expression determine the frequency of the execution of the job
object JsonData
This setting is reserved and provided for information only
string EncryptionFileThe path to the file containing the PGP Private Key used to decrypt Sync Files
byte[] SaltThe cryptographic Salt used to encrypt the password
byte[] KEKThe encrypted password used to access the file
string TargetSystemThe target system (DB, DRIVE, MESSAGING, HTTP)
string TargetObjectThe name of the target object (for DB systems, the name of the table)
string UpsertScriptThe insert/update script to execute for DB systems (not needed for SQL Server (native) targets)
string DeleteScriptThe delete script to execute for DB systems (not needed for SQL Server (native) targets)
string LastSchemaHashThe hash of the source schema
object Options show/hide
bool AutoRecreateWhen true, for DB systems, determines if the target object should be recreated automatically
int ExecTimeoutWhen positive, controls the timeout of the write operation
int BatchCountWhen positive, sends the upserted or deleted records in batches of records
bool AuditLogEnabledWhen true, saves execution logs to the database for auditing purposes
HTTP Target Options show/hide
object HttpUpsertSettings
string HTTPUpsertURIFor HTTP targets, the URI of the HTTP endpoint for Upsert operations
string HTTPUpsertVerbFor HTTP targets, the HTTP Verb (GET, POST, DELETE, PUT) to use for Upsert operations
string HTTPUpsertContentTypeFor HTTP targets, the Content-Type header to use for Upsert operations
string HTTPUpsertHeadersFor HTTP targets, serialized list of headers and values for Upsert operations
object HttpDeleteSettings
string HTTPDeleteURIFor HTTP targets, the URI of the HTTP endpoint for Delete operations
string HTTPDeleteVerbFor HTTP targets, the HTTP Verb (GET, POST, DELETE, PUT) to use for Delete operations
string HTTPDeleteContentTypeFor HTTP targets, the Content-Type header to use for Delete operations
string HTTPDeleteHeadersFor HTTP targets, serialized list of headers and values for Delete operations
MESSAGING Target Options show/hide
object MessagingUpsertSettings
string SystemTypeA valid messaging system type (awssqs, rabbitmq, kafka, azeventhub)
rabbitmq settings:
string RoutingKeyThe RabbitMQ Routing Key
string ReplyToThe reply to queue
int PriorityThe priority of the message
string MessageIdThe message id (if any)
string AppIdThe application id (if any)
long expirationThe time to live of the message
bool persistWhen true, persist the message in the queue
awssqs settings:
string QueueNameThe AWS Queue Name (optional)
azeventhub settings:
string EventHubNameThe Azure Event Hub name (optional)
kafka settings:
string TopicNameThe Kafka topic name (optional)
string FieldKeyThe comma-separated list of key fields used to identify a unique message (the Upsert Key Columns)
string KeyFieldSeparatorThe Key values separator string to use (only needed when 2 or more key fields are provided)
object MessagingDeleteSettings
string SystemTypeA valid messaging system type (awssqs, rabbitmq, kafka, azeventhub)
refer to the MessageUpsertSettings section above
DRIVE Target Options show/hide
object TargetFileOptions
ContainerNameThe directory path, Azure Container name, or S3 Bucket Name
DistributionFieldNameThe field name used to dynamically distribute records across multiple files
FileNameThe file name to use (accepts field names and date tokens; ex: sales-[yyyy]-\{\{statecode\}\} )
UseRuntimeDateWhen true, use the execution date/time for date tokens
UseShadowCopyWhen true, use a shadow copy of remote files for faster updates
ShadowCopyTTLWhen UseShadowCopy is true, indicates how old the file can be before downloading a more recent version
ShadowCopyFrequencyThe ShadowCopyTTL value (hour, minute)
string SystemNameThe name of the target system (SQLServer, MySQL, Oracle, Teradata, Generic, AWS S3 Bucket, Azure Blob Container, Local Path, AWS SQS, Azure Event Hub, RabbitMQ)
dynamic[] PipelineDef show/hide
DataDynamicColumn
string NameThe name of the component: DataDynamicColumn
string ColumnNameThe column name to use
string DataTypeThe data type of the column (string, int, float, guid, datetime, long, double, char, bool, byte[])
string ExpressionTypeSQLExpression or EnzoFunction
string SQLExpressionThe expression (either a SQL expression, or an Enzo function call)
int MaxLengthColumn maximum length
bool IgnoreIfExistsWhen true, does not throw an error if the column already exists
bool DisabledThen true, bypass this component
DataHashing
string NameThe name of the component: DataHashing
string ColumnNameThe name of the column to hash
string HashNameThe hashing algorythm to use (MD5, SHA1, SHA256, SHA384, SHA512)
bool TruncateIfNeededWhen true, truncate the hash value to fit the column length
bool DisabledThen true, bypass this component
DataMasking
string NameThe name of the component: DataMasking
string ColumnNameThe name of the column to mask
string MaskingTypeThe masking operation (default, email, creditcard, randomnumber, customtext)
decimal RandomNbrMinWhen using RandomNumber masking, the minimum value to use
decimal RandomNbrMaxWhen using RandomNumber masking, the maximum value to use
int CustomPrefixWhen using CustomText, the prefix to use
int CustomSuffixWhen using CustomText, the suffix to use
string CustomPaddingWhen using CustomText, the padding to use
bool DisabledThen true, bypass this component
DataFilter
string NameThe name of the component: DataFilter
string FilterThe SQL filter to use to limit the records being processed
bool DisabledThen true, bypass this component
DataQuality
string NameThe name of the component: DataQuality
string ColumnNameThe column name to validate
string AllowedValuesComma-separated list of values to allow or deny
bool CheckListFlagWhen true, limit possible values to the list provided by the AllowedValues field
bool AllowNullWhen true, NULL values are allowed
bool AllowEmptyWhen true, empty strings are allowed
bool NumericRequiredWhen true, the value must be numeric
bool ThrowExceptionOnFailThen true, fail processing if the data validation rule failed
bool SkipRowOnFailWhen true, ignore the row if the data validation rule failed
bool OptOutListWhen true, the AllowedValues should be allowed (opt-in); else AllowedValues should be considered as opt-out.
bool DisabledThen true, bypass this component
 
Response
OK Success response code

request
POST /job/dynamic
{
	"JobReader": {
		"SqlRead": "SELECT ID, Title, AccountNumber, CompanyId, Created, Modified FROM SharePoint.[list@companies]",
		"Path": "C:\\tmp\\datasync_reader",
		"UpsertColumns": "ID",
		"ConnectionString": "enzo9550"
	},
	"JobWriter": {
		"UpsertColumns": "ID",
		"ConnectionString": "sql2017",
		"TargetObject": "[regression].SPCompanies2"
	}
}
response
OK

Dynamic Job Executions

Returns the history of executions available of a specific job given its unique identifier (guid). Although Dynamic Jobs are only stored in memory by DataZen, their executions are recorded in history and available even after restarting DataZen.

GET
/job/dynamic/executions?guid=...

NOTE: The Unique Identifier of a Dynamic Job can be specifyed when it is started; this can make it easier to track the executions of related dynamic jobs over time. To view all available Dynamic Jobs still in memory, call the /jobssummary endpoint with the dynamic=1 option.

Query Parameters
string guid The job guid to query(required)
 
Response
guid The guid of the dynamic job
execGuid The execution id of the dynamic job
lastRunTime The last run date/time of the dynamic job
errors Number of errors detected
warnings Number of warnings detected

request
GET /job/dynamic/executions?guid=a3ff75315a4241d1958e1f9dbd53829f
response
[
 {
    "guid":"a3ff75315a4241d1958e1f9dbd53829f",
    "execGuid":"a3ff75315a4241d1958e1f9dbd53829f",
    "lastRunTime":"2022-10-12T18:31:50.077",
    "errors":0,
    "warnings":0
    }
]

Enable/Disable Job

Enables or disables a job. When re-enabled, job processing will resume at the next scheduled interval; if the last interval was missed, the job will start immediately.
This operation does not support Dynamic Jobs.

/job/enable?guid=123456789&active=0
POST

Enables or disables a job.

Input Parameters
string guid The job identifier
string active The flag to disable or enable a job (0:disable, 1:enable)
 
Response
OK Success response code

request
GET /job/enable?guid=123456789&active=0
response
OK

Job

Creates, updates or deletes a Job Reader, Job Writer, or Direct Job. A Direct Job contains both a reader and writer, and the job is considered successful when both are completed successfully.
This operation does not support Dynamic Jobs.

POST

Creates a Job Reader, Job Writer, or Direct Job. When updating a DirectJob, you need to provide both the JobReader and the JobWriter.

Input Parameters
object JobCreation show/hide
object JobReader show/hide
bool ActiveWhen true, the job will execute at the desired interval as specified by the CronSchedule parameter
bool AuditLogEnabledWhen true, the job execution output will be saved so it can be retrieved later (using job/log)
string ConnectionStringThe name of the Central Connection String to use for the source system
DateTime CreatedOnThe date/time the job was created
string CronScheduleWhen specified, the Cron expression determine the frequency of the execution of the job; blank when the job is started manually.
string GuidThe job unique identifier
bool Initialize(internal use only)
bool IsDBSourceWhen true, the source system is a relational database
bool IsDirectWhen true, the job is a Direct job (with source and target defined)
bool IsDirty(internal use only)
bool IsDriveSourceWhen true, the source system is a file stored on a drive (local, cloud, or FTP)
bool IsHttpSourceWhen true, the source system is an HTTP REST endpoint
bool IsQueueingSourceWhen true, the source system is a messaging hub and the job acts as a Messaging Consumer
string JobKeyThe name of the job
int JobType1 (Job Reader)
DateTime LastRunTimeThe last execution date/time
string LastTsDelPointerThe last High Water Mark value for deleted records
string LastTsPointerThe last High Water Mark value for inserted or updated columns
string PathThe path where the sync file will be created
bool PropagateDeleteWhen true, identify or capture deleted records
string SqlReadFor DB connections, the SQL command to execute that will fetch source records
string StatusThe status of the job (running, ready)
string SyncStrategy(internal use: full)
string SystemNameThe name of the source system (SQLServer, MySQL, Oracle, Teradata, Generic, AWS S3 Bucket, Azure Blob Container, Local Path...)
string TimestampColWhen provided, the date/time or numeric column to use as the High Watermark when reading from a database
string TrackLastTsWhen True, the job tracks a High Watermark
string UpsertColumnsComma-separated list of columns making up the unique key of the record
object JsonData show/hide
bool AuditLogEnabledWhen true, saves the execution log for auditing purposes
string EncryptionFileWhen provided, the file that contains the PGP public key used for encrypting Sync Files
bool BypassTopNWhen true, bypasses the TOP N heuristic method for DB data sources
int TopNOverrideWhen positive, uses this value for the TOP N heuristic (default: 10000)
string DirectWriterWhen a direct job, the name of the Job Writer (usually set to the name of the Job Reader, with two underscored as the prefix)
int ExecutionTimeoutWhen positive, uses this value for individual requests as the timeout value (in seconds)
bool BypassSchemaDetectionWhen true, the job should not extract detailed schema information from the data source; applies to DB data sources only
int DeletedRecordsMethodWhen PropagateDelete is true, determines the method used to identify deleted records (0: default, 1: custom SQL that returns all source records, 2: custom SQL that returns the IDs of the deleted records)
string DeletedRecordsSQLWhen DeletedRecordsMethod is greater than 0, specifies the SQL command used to identify deleted records
string CDCTokenDeletedFieldWhen DeletedRecordsMethod is 2, comma-separated column names returned by DeletedRecordsSQL that uniquely identify the deleted record
object MessagingOptions show/hide
int MessageProcessingThe messaging processing mode (0: passthrough, 1: Msg to Dataset, 2: Msg Payload to Dataset)
bool MessagingBatchEnabledWhen true, messages are batched before being sent to their intended target
int MessagingBatchCountWhen MessagingBatchEnabled is true, number of messages to include in a batch
int MessagingFlushTTLWhen positive and MessagingBatchEnabled is true, the maximum time to wait before messages are sent if the BatchCount hasn't been met
string MessagingSampleMsgSample message to use for testing purposes
string MessagingContentColumnWhen MessageProcessing is 2 and not empty, adds a dynamic column to the data of this name with the original content of the payload
string RootPathWhen MessageProcessing is 2 and not empty, XML or JSON path to use to extract the Dataset
bool IncludeHeadersWhen true, headers should be sent to the target system as much as possible
bool IncludeMetadataWhen true, message metadata should be sent to the target system as much as possible
string HeaderFieldsWhen not empty and MessageProcessing is NOT 2 and IncludeHeaders is true, the comma-separated list of header keys to send to the target system (blank sends all)
string MetadataFieldsWhen not empty and MessageProcessing is NOT 2 and IncludeHeaders is true, the comma-separated list of metadata fields to send to the target system (blank sends all)
string QueueNameOverrideWhen provided, the name of the queue/bus/topic to read messages from (if blank, uses the connection string default settings)
string MessagingGroupNameWhen relevant, the message group name to use when reading from the source messaging hub (if blank, reads from the default group)
string PartitionKeysWhen relevant, a comma-separated list of partition keys to use when reading from the source messaging hub (if blank, reads from all partitions)
object DriveSourceSettings show/hide
string DriveFileFormatThe messaging processing mode (0: passthrough, 1: Msg to Dataset, 2: Msg Payload to Dataset)
string DriveFilePatternWhen true, messages are batched before being sent to their intended target
string DriveSchemaFileWhen MessagingBatchEnabled is true, number of messages to include in a batch
string DriveSelectedColumnsWhen positive and MessagingBatchEnabled is true, the maximum time to wait before messages are sent if the BatchCount hasn't been met
string DriveUpdatedFilesOnlySample message to use for testing purposes
bool MessagingContentColumnWhen MessageProcessing is 2 and not empty, adds a dynamic column to the data of this name with the original content of the payload
bool DriveIncludeSubfoldersWhen MessageProcessing is 2 and not empty, XML or JSON path to use to extract the Dataset
string DriveFolderColumnNameWhen true, headers should be sent to the target system as much as possible
bool DriveAddFolderColumnWhen true, message metadata should be sent to the target system as much as possible
bool DriveUseRootPathWhen not empty and MessageProcessing is NOT 2 and IncludeHeaders is true, the list of header keys to send to the target system (blank sends all)
string DriveRootPathWhen not empty and MessageProcessing is NOT 2 and IncludeHeaders is true, the list of metadata fields to send to the target system (blank sends all)
bool SingleFileProcessingWhen provided, the name of the queue/bus/topic to read messages from (if blank, uses the connection string default settings)
object ExtendedOptions show/hide
CSVOptions
string ColumnDelimitersThe character that separate individual field values when reading from a delimited file (ignored if FieldWidths are provided)
bool HasHeaderRowWhen true, assumes the file has a header row and column names will be inferred from the header row
string FieldWidthsIf provided, a comma-separated list of integers representing the length of each column for a fixed-length file format
string[] CommentTokensAn array of string values to be treated as comment tokens in the file
bool QuotedIdentifierWhen true, will treat double-quotes as boundary tokens for field name and value
object Column2TableTxSettings show/hide
string TxDocumentMsgContentColThe column index or name to use from the dataset containing an XML or JSON string to be transformed into a dataset
string TxDocumentRootPathIf provided, the XML or JSON path to use as the starting point to read the document content
string TxDocumentSourceColumnIf provided, the name of a dynamic column that will be added to the dataset with the content of the JSON or XML record
object HttpSourceSettings show/hide
string HTTPContentTypeThe Content-Type header used when sending a POST or PUT request
string HTTPHeadersAn array of key-value pairs representing custom headers to set
string HTTPURIThe URI to call; if blank, will call the URI defined by the connection; may use a relative URL or a full URL
string HTTPVerbThe HTTP Verb to use (GET, POST, PUT, DELETE)
string HTTPPayloadWhen executing a POST or PUT request, the payload to send with the request
bool ApplyContentTxWhen true, converts the JSON or XML content of the payload into a dataset
int PagingStrategyThe paging strategy to use (0: none, 1: simple offset, 2: simple page, 3 simple link, 4: bookmark); when positive, the corrresponding OffsetStrategy object must be provided
string SelectedColumnsThe columns to return as part of the dataset; leave blank to return all available columns
string HighWaterMarkThe name of the field that contains the High Watermark for future calls
string HighWaterMarkParamThe name of the Query Parameter to use when adding a High Watermark value for future calls
dynamic OffsetStrategy show/hide
HttpSimpleOffsetStrategy
string URLOffsetParamThe name of the URL Query Parameter to use when specifying an offset
HttpSimplePageStrategy
string URLPageParamThe name of the URL Query Parameter to use when specifying a page number
HttpSimpleLinkStrategy
string URLPagePathThe JSON or XML document path where to find the next link to call within the response payload
HttpTokenPagingStrategy
string TokenPathThe JSON or XML document path where to find the next bookmark value to use within the response payload
string QueryParamNameThe namd of the URL Query Parameter to use when specifying the bookmark value
object PayloadTxSettings show/hide
string TxDocumentMsgContentColThe column index or name to use from the dataset containing an XML or JSON string to be transformed into a dataset
string TxDocumentRootPathIf provided, the XML or JSON path to use as the starting point to read the document content
string TxDocumentSourceColumnIf provided, the name of a dynamic column that will be added to the dataset with the content of the JSON or XML record
object[] JobTriggers show/hide
string guidThe Job GUID to trigger (must be provided)
string jobKeyThe JobKey of the job to trigger (informational only)
int eventIdThe condition that will fire the trigger (0: On Completion, 1: On Success (with wata), 2: On Success (always), 3: On Error)
string eventNameThe friendly name of the condition (optional)
true JobTriggersEnabledWhen true, executes the jobs specified in the JobTriggers collection upon completion of the job
dynamic[] PipelineDef show/hide
DataDynamicColumn
string NameThe name of the component: DataDynamicColumn
string ColumnNameThe column name to use
string DataTypeThe data type of the column (string, int, float, guid, datetime, long, double, char, bool, byte[])
string ExpressionTypeSQLExpression or EnzoFunction
string SQLExpressionThe expression (either a SQL expression, or an Enzo function call)
int MaxLengthColumn maximum length
bool IgnoreIfExistsWhen true, does not throw an error if the column already exists
bool DisabledThen true, bypass this component
DataHashing
string NameThe name of the component: DataHashing
string ColumnNameThe name of the column to hash
string HashNameThe hashing algorythm to use (MD5, SHA1, SHA256, SHA384, SHA512)
bool TruncateIfNeededWhen true, truncate the hash value to fit the column length
bool DisabledThen true, bypass this component
DataMasking
string NameThe name of the component: DataMasking
string ColumnNameThe name of the column to mask
string MaskingTypeThe masking operation (default, email, creditcard, randomnumber, customtext)
decimal RandomNbrMinWhen using RandomNumber masking, the minimum value to use
decimal RandomNbrMaxWhen using RandomNumber masking, the maximum value to use
int CustomPrefixWhen using CustomText, the prefix to use
int CustomSuffixWhen using CustomText, the suffix to use
string CustomPaddingWhen using CustomText, the padding to use
bool DisabledThen true, bypass this component
DataFilter
string NameThe name of the component: DataFilter
string FilterThe SQL filter to use to limit the records being processed
bool DisabledThen true, bypass this component
DataQuality
string NameThe name of the component: DataQuality
string ColumnNameThe column name to validate
string AllowedValuesComma-separated list of values to allow or deny
bool CheckListFlagWhen true, limit possible values to the list provided by the AllowedValues field
bool AllowNullWhen true, NULL values are allowed
bool AllowEmptyWhen true, empty strings are allowed
bool NumericRequiredWhen true, the value must be numeric
bool ThrowExceptionOnFailThen true, fail processing if the data validation rule failed
bool SkipRowOnFailWhen true, ignore the row if the data validation rule failed
bool OptOutListWhen true, the AllowedValues should be allowed (opt-in); else AllowedValues should be considered as opt-out.
bool DisabledThen true, bypass this component
object JobWriter show/hide
string GuidThe job unique identifier
string JobKeyThe name of the job
string PathThe path where the sync files are located
string UpsertColumnsComma-separated list of columns making up the unique key of the record
string ConnectionStringThe name of the Central Connection String to use for the target system
bool PropagateDeleteWhen true, identify or capture deleted records if available
string SourceJobKeyThe original name of the source job
string LastExecutionIdThe last ExecutionId extracted
string InitialExecutionIdThe initial ExecutionId used by this job
DateTime CreatedOnDate/time this job was created
string StatusThe status of this job (ready, running)
bool ActiveWhen true, the job will execute at the desired interval as specified by the CronSchedule parameter
DateTime LastRunTimeThe last execution date/time
string CronScheduleWhen specified, the Cron expression determine the frequency of the execution of the job
object JsonData
This setting is reserved and provided for information only
string EncryptionFileThe path to the file containing the PGP Private Key used to decrypt Sync Files
byte[] SaltThe cryptographic Salt used to encrypt the password
byte[] KEKThe encrypted password used to access the file
string TargetSystemThe target system (DB, DRIVE, MESSAGING, HTTP)
string TargetObjectThe name of the target object (for DB systems, the name of the table)
string UpsertScriptThe insert/update script to execute for DB systems (not needed for SQL Server (native) targets)
string DeleteScriptThe delete script to execute for DB systems (not needed for SQL Server (native) targets)
string LastSchemaHashThe hash of the source schema
object Options show/hide
bool AutoRecreateWhen true, for DB systems, determines if the target object should be recreated automatically
int ExecTimeoutWhen positive, controls the timeout of the write operation
int BatchCountWhen positive, sends the upserted or deleted records in batches of records
bool AuditLogEnabledWhen true, saves execution logs to the database for auditing purposes
HTTP Target Options show/hide
object HttpUpsertSettings
string HTTPUpsertURIFor HTTP targets, the URI of the HTTP endpoint for Upsert operations
string HTTPUpsertVerbFor HTTP targets, the HTTP Verb (GET, POST, DELETE, PUT) to use for Upsert operations
string HTTPUpsertContentTypeFor HTTP targets, the Content-Type header to use for Upsert operations
string HTTPUpsertHeadersFor HTTP targets, serialized list of headers and values for Upsert operations
object HttpDeleteSettings
string HTTPDeleteURIFor HTTP targets, the URI of the HTTP endpoint for Delete operations
string HTTPDeleteVerbFor HTTP targets, the HTTP Verb (GET, POST, DELETE, PUT) to use for Delete operations
string HTTPDeleteContentTypeFor HTTP targets, the Content-Type header to use for Delete operations
string HTTPDeleteHeadersFor HTTP targets, serialized list of headers and values for Delete operations
MESSAGING Target Options show/hide
object MessagingUpsertSettings
string SystemTypeA valid messaging system type (awssqs, rabbitmq, kafka, azeventhub)
rabbitmq settings:
string RoutingKeyThe RabbitMQ Routing Key
string ReplyToThe reply to queue
int PriorityThe priority of the message
string MessageIdThe message id (if any)
string AppIdThe application id (if any)
long expirationThe time to live of the message
bool persistWhen true, persist the message in the queue
awssqs settings:
string QueueNameThe AWS Queue Name (optional)
azeventhub settings:
string EventHubNameThe Azure Event Hub name (optional)
kafka settings:
string TopicNameThe Kafka topic name (optional)
string FieldKeyThe comma-separated list of key fields used to identify a unique message (the Upsert Key Columns)
string KeyFieldSeparatorThe Key values separator string to use (only needed when 2 or more key fields are provided)
object MessagingDeleteSettings
string SystemTypeA valid messaging system type (awssqs, rabbitmq, kafka, azeventhub)
refer to the MessageUpsertSettings section above
DRIVE Target Options show/hide
object TargetFileOptions
ContainerNameThe directory path, Azure Container name, or S3 Bucket Name
DistributionFieldNameThe field name used to dynamically distribute records across multiple files
FileNameThe file name to use (accepts field names and date tokens; ex: sales-[yyyy]-\{\{statecode\}\} )
UseRuntimeDateWhen true, use the execution date/time for date tokens
UseShadowCopyWhen true, use a shadow copy of remote files for faster updates
ShadowCopyTTLWhen UseShadowCopy is true, indicates how old the file can be before downloading a more recent version
ShadowCopyFrequencyThe ShadowCopyTTL value (hour, minute)
string SystemNameThe name of the target system (SQLServer, MySQL, Oracle, Teradata, Generic, AWS S3 Bucket, Azure Blob Container, Local Path, AWS SQS, Azure Event Hub, RabbitMQ)
dynamic[] PipelineDef show/hide
DataDynamicColumn
string NameThe name of the component: DataDynamicColumn
string ColumnNameThe column name to use
string DataTypeThe data type of the column (string, int, float, guid, datetime, long, double, char, bool, byte[])
string ExpressionTypeSQLExpression or EnzoFunction
string SQLExpressionThe expression (either a SQL expression, or an Enzo function call)
int MaxLengthColumn maximum length
bool IgnoreIfExistsWhen true, does not throw an error if the column already exists
bool DisabledThen true, bypass this component
DataHashing
string NameThe name of the component: DataHashing
string ColumnNameThe name of the column to hash
string HashNameThe hashing algorythm to use (MD5, SHA1, SHA256, SHA384, SHA512)
bool TruncateIfNeededWhen true, truncate the hash value to fit the column length
bool DisabledThen true, bypass this component
DataMasking
string NameThe name of the component: DataMasking
string ColumnNameThe name of the column to mask
string MaskingTypeThe masking operation (default, email, creditcard, randomnumber, customtext)
decimal RandomNbrMinWhen using RandomNumber masking, the minimum value to use
decimal RandomNbrMaxWhen using RandomNumber masking, the maximum value to use
int CustomPrefixWhen using CustomText, the prefix to use
int CustomSuffixWhen using CustomText, the suffix to use
string CustomPaddingWhen using CustomText, the padding to use
bool DisabledThen true, bypass this component
DataFilter
string NameThe name of the component: DataFilter
string FilterThe SQL filter to use to limit the records being processed
bool DisabledThen true, bypass this component
DataQuality
string NameThe name of the component: DataQuality
string ColumnNameThe column name to validate
string AllowedValuesComma-separated list of values to allow or deny
bool CheckListFlagWhen true, limit possible values to the list provided by the AllowedValues field
bool AllowNullWhen true, NULL values are allowed
bool AllowEmptyWhen true, empty strings are allowed
bool NumericRequiredWhen true, the value must be numeric
bool ThrowExceptionOnFailThen true, fail processing if the data validation rule failed
bool SkipRowOnFailWhen true, ignore the row if the data validation rule failed
bool OptOutListWhen true, the AllowedValues should be allowed (opt-in); else AllowedValues should be considered as opt-out.
bool DisabledThen true, bypass this component
 
Response
OK Success response code

request
POST /job

{
	"JobReader": {
		"Guid": null,
		"JobKey": "TEST-JOB-ALLDATABASES",
		"SqlRead": "SELECT * FROM sys.databases",
		"Path": "C:\\Tmp\\datasync_reader\\",
		"UpsertColumns": "database_id",
		"TimestampCol": "",
		"LastTsPointer": null,
		"LastTsDelPointer": null,
		"ConnectionString": "sql2017",
		"PropagateDelete": false,
		"SyncStrategy": "Full",
		"CreatedOn": "2022-10-12T19:24:14.3440036Z",
		"Status": null,
		"Initialize": false,
		"Active": true,
		"LastRunTime": null,
		"CronSchedule": "0 */5 * ? * *",
		"JsonData": {
			"AuditLogEnabled": true,
			"KeepChangedLogFiles": true,
			"EncryptionFile": "",
			"BypassTopN": false,
			"TopNOverride": 0,
			"DirectWriter": null,
			"ExecutionTimeout": 0,
			"BypassSchemaDetection": false,
			"DeletedRecordsMethod": 0,
			"DeletedRecordsSQL": null,
			"CDCTokenDeletedField": "",
			"MessagingOptions": null,
			"DriveSourceSettings": null,
			"Column2TableTxSettings": null,
			"HttpSourceSettings": null,
			"BigDataSourceSettings": null,
			"JobTriggers": [],
			"JobTriggersEnabled": false,
			"CDCSettings": null,
			"DynamicParameter": null
		},
		"SystemName": null,
		"PipelineDef": []
	},
	"JobWriter": null
}
response
OK
 
PUT
/job

Updates a Job Reader, Job Writer, or Direct Job. When updating a DirectJob, you need to provide both the JobReader and the JobWriter. The Guid nodes must be provided.

Input Parameters
object JobCreation show/hide
object JobReader show/hide
bool ActiveWhen true, the job will execute at the desired interval as specified by the CronSchedule parameter
bool AuditLogEnabledWhen true, the job execution output will be saved so it can be retrieved later (using job/log)
string ConnectionStringThe name of the Central Connection String to use for the source system
DateTime CreatedOnThe date/time the job was created
string CronScheduleWhen specified, the Cron expression determine the frequency of the execution of the job; blank when the job is started manually.
string GuidThe job unique identifier
bool Initialize(internal use only)
bool IsDBSourceWhen true, the source system is a relational database
bool IsDirectWhen true, the job is a Direct job (with source and target defined)
bool IsDirty(internal use only)
bool IsDriveSourceWhen true, the source system is a file stored on a drive (local, cloud, or FTP)
bool IsHttpSourceWhen true, the source system is an HTTP REST endpoint
bool IsQueueingSourceWhen true, the source system is a messaging hub and the job acts as a Messaging Consumer
string JobKeyThe name of the job
int JobType1 (Job Reader)
DateTime LastRunTimeThe last execution date/time
string LastTsDelPointerThe last High Water Mark value for deleted records
string LastTsPointerThe last High Water Mark value for inserted or updated columns
string PathThe path where the sync file will be created
bool PropagateDeleteWhen true, identify or capture deleted records
string SqlReadFor DB connections, the SQL command to execute that will fetch source records
string StatusThe status of the job (running, ready)
string SyncStrategy(internal use: full)
string SystemNameThe name of the source system (SQLServer, MySQL, Oracle, Teradata, Generic, AWS S3 Bucket, Azure Blob Container, Local Path...)
string TimestampColWhen provided, the date/time or numeric column to use as the High Watermark when reading from a database
string TrackLastTsWhen True, the job tracks a High Watermark
string UpsertColumnsComma-separated list of columns making up the unique key of the record
object JsonData show/hide
bool AuditLogEnabledWhen true, saves the execution log for auditing purposes
string EncryptionFileWhen provided, the file that contains the PGP public key used for encrypting Sync Files
bool BypassTopNWhen true, bypasses the TOP N heuristic method for DB data sources
int TopNOverrideWhen positive, uses this value for the TOP N heuristic (default: 10000)
string DirectWriterWhen a direct job, the name of the Job Writer (usually set to the name of the Job Reader, with two underscored as the prefix)
int ExecutionTimeoutWhen positive, uses this value for individual requests as the timeout value (in seconds)
bool BypassSchemaDetectionWhen true, the job should not extract detailed schema information from the data source; applies to DB data sources only
int DeletedRecordsMethodWhen PropagateDelete is true, determines the method used to identify deleted records (0: default, 1: custom SQL that returns all source records, 2: custom SQL that returns the IDs of the deleted records)
string DeletedRecordsSQLWhen DeletedRecordsMethod is greater than 0, specifies the SQL command used to identify deleted records
string CDCTokenDeletedFieldWhen DeletedRecordsMethod is 2, comma-separated column names returned by DeletedRecordsSQL that uniquely identify the deleted record
object MessagingOptions show/hide
int MessageProcessingThe messaging processing mode (0: passthrough, 1: Msg to Dataset, 2: Msg Payload to Dataset)
bool MessagingBatchEnabledWhen true, messages are batched before being sent to their intended target
int MessagingBatchCountWhen MessagingBatchEnabled is true, number of messages to include in a batch
int MessagingFlushTTLWhen positive and MessagingBatchEnabled is true, the maximum time to wait before messages are sent if the BatchCount hasn't been met
string MessagingSampleMsgSample message to use for testing purposes
string MessagingContentColumnWhen MessageProcessing is 2 and not empty, adds a dynamic column to the data of this name with the original content of the payload
string RootPathWhen MessageProcessing is 2 and not empty, XML or JSON path to use to extract the Dataset
bool IncludeHeadersWhen true, headers should be sent to the target system as much as possible
bool IncludeMetadataWhen true, message metadata should be sent to the target system as much as possible
string HeaderFieldsWhen not empty and MessageProcessing is NOT 2 and IncludeHeaders is true, the comma-separated list of header keys to send to the target system (blank sends all)
string MetadataFieldsWhen not empty and MessageProcessing is NOT 2 and IncludeHeaders is true, the comma-separated list of metadata fields to send to the target system (blank sends all)
string QueueNameOverrideWhen provided, the name of the queue/bus/topic to read messages from (if blank, uses the connection string default settings)
string MessagingGroupNameWhen relevant, the message group name to use when reading from the source messaging hub (if blank, reads from the default group)
string PartitionKeysWhen relevant, a comma-separated list of partition keys to use when reading from the source messaging hub (if blank, reads from all partitions)
object DriveSourceSettings show/hide
string DriveFileFormatThe messaging processing mode (0: passthrough, 1: Msg to Dataset, 2: Msg Payload to Dataset)
string DriveFilePatternWhen true, messages are batched before being sent to their intended target
string DriveSchemaFileWhen MessagingBatchEnabled is true, number of messages to include in a batch
string DriveSelectedColumnsWhen positive and MessagingBatchEnabled is true, the maximum time to wait before messages are sent if the BatchCount hasn't been met
string DriveUpdatedFilesOnlySample message to use for testing purposes
bool MessagingContentColumnWhen MessageProcessing is 2 and not empty, adds a dynamic column to the data of this name with the original content of the payload
bool DriveIncludeSubfoldersWhen MessageProcessing is 2 and not empty, XML or JSON path to use to extract the Dataset
string DriveFolderColumnNameWhen true, headers should be sent to the target system as much as possible
bool DriveAddFolderColumnWhen true, message metadata should be sent to the target system as much as possible
bool DriveUseRootPathWhen not empty and MessageProcessing is NOT 2 and IncludeHeaders is true, the list of header keys to send to the target system (blank sends all)
string DriveRootPathWhen not empty and MessageProcessing is NOT 2 and IncludeHeaders is true, the list of metadata fields to send to the target system (blank sends all)
bool SingleFileProcessingWhen provided, the name of the queue/bus/topic to read messages from (if blank, uses the connection string default settings)
object ExtendedOptions show/hide
CSVOptions
string ColumnDelimitersThe character that separate individual field values when reading from a delimited file (ignored if FieldWidths are provided)
bool HasHeaderRowWhen true, assumes the file has a header row and column names will be inferred from the header row
string FieldWidthsIf provided, a comma-separated list of integers representing the length of each column for a fixed-length file format
string[] CommentTokensAn array of string values to be treated as comment tokens in the file
bool QuotedIdentifierWhen true, will treat double-quotes as boundary tokens for field name and value
object Column2TableTxSettings show/hide
string TxDocumentMsgContentColThe column index or name to use from the dataset containing an XML or JSON string to be transformed into a dataset
string TxDocumentRootPathIf provided, the XML or JSON path to use as the starting point to read the document content
string TxDocumentSourceColumnIf provided, the name of a dynamic column that will be added to the dataset with the content of the JSON or XML record
object HttpSourceSettings show/hide
string HTTPContentTypeThe Content-Type header used when sending a POST or PUT request
string HTTPHeadersAn array of key-value pairs representing custom headers to set
string HTTPURIThe URI to call; if blank, will call the URI defined by the connection; may use a relative URL or a full URL
string HTTPVerbThe HTTP Verb to use (GET, POST, PUT, DELETE)
string HTTPPayloadWhen executing a POST or PUT request, the payload to send with the request
bool ApplyContentTxWhen true, converts the JSON or XML content of the payload into a dataset
int PagingStrategyThe paging strategy to use (0: none, 1: simple offset, 2: simple page, 3 simple link, 4: bookmark); when positive, the corrresponding OffsetStrategy object must be provided
string SelectedColumnsThe columns to return as part of the dataset; leave blank to return all available columns
string HighWaterMarkThe name of the field that contains the High Watermark for future calls
string HighWaterMarkParamThe name of the Query Parameter to use when adding a High Watermark value for future calls
dynamic OffsetStrategy show/hide
HttpSimpleOffsetStrategy
string URLOffsetParamThe name of the URL Query Parameter to use when specifying an offset
HttpSimplePageStrategy
string URLPageParamThe name of the URL Query Parameter to use when specifying a page number
HttpSimpleLinkStrategy
string URLPagePathThe JSON or XML document path where to find the next link to call within the response payload
HttpTokenPagingStrategy
string TokenPathThe JSON or XML document path where to find the next bookmark value to use within the response payload
string QueryParamNameThe namd of the URL Query Parameter to use when specifying the bookmark value
object PayloadTxSettings show/hide
string TxDocumentMsgContentColThe column index or name to use from the dataset containing an XML or JSON string to be transformed into a dataset
string TxDocumentRootPathIf provided, the XML or JSON path to use as the starting point to read the document content
string TxDocumentSourceColumnIf provided, the name of a dynamic column that will be added to the dataset with the content of the JSON or XML record
object[] JobTriggers show/hide
string guidThe Job GUID to trigger (must be provided)
string jobKeyThe JobKey of the job to trigger (informational only)
int eventIdThe condition that will fire the trigger (0: On Completion, 1: On Success (with wata), 2: On Success (always), 3: On Error)
string eventNameThe friendly name of the condition (optional)
true JobTriggersEnabledWhen true, executes the jobs specified in the JobTriggers collection upon completion of the job
dynamic[] PipelineDef show/hide
DataDynamicColumn
string NameThe name of the component: DataDynamicColumn
string ColumnNameThe column name to use
string DataTypeThe data type of the column (string, int, float, guid, datetime, long, double, char, bool, byte[])
string ExpressionTypeSQLExpression or EnzoFunction
string SQLExpressionThe expression (either a SQL expression, or an Enzo function call)
int MaxLengthColumn maximum length
bool IgnoreIfExistsWhen true, does not throw an error if the column already exists
bool DisabledThen true, bypass this component
DataHashing
string NameThe name of the component: DataHashing
string ColumnNameThe name of the column to hash
string HashNameThe hashing algorythm to use (MD5, SHA1, SHA256, SHA384, SHA512)
bool TruncateIfNeededWhen true, truncate the hash value to fit the column length
bool DisabledThen true, bypass this component
DataMasking
string NameThe name of the component: DataMasking
string ColumnNameThe name of the column to mask
string MaskingTypeThe masking operation (default, email, creditcard, randomnumber, customtext)
decimal RandomNbrMinWhen using RandomNumber masking, the minimum value to use
decimal RandomNbrMaxWhen using RandomNumber masking, the maximum value to use
int CustomPrefixWhen using CustomText, the prefix to use
int CustomSuffixWhen using CustomText, the suffix to use
string CustomPaddingWhen using CustomText, the padding to use
bool DisabledThen true, bypass this component
DataFilter
string NameThe name of the component: DataFilter
string FilterThe SQL filter to use to limit the records being processed
bool DisabledThen true, bypass this component
DataQuality
string NameThe name of the component: DataQuality
string ColumnNameThe column name to validate
string AllowedValuesComma-separated list of values to allow or deny
bool CheckListFlagWhen true, limit possible values to the list provided by the AllowedValues field
bool AllowNullWhen true, NULL values are allowed
bool AllowEmptyWhen true, empty strings are allowed
bool NumericRequiredWhen true, the value must be numeric
bool ThrowExceptionOnFailThen true, fail processing if the data validation rule failed
bool SkipRowOnFailWhen true, ignore the row if the data validation rule failed
bool OptOutListWhen true, the AllowedValues should be allowed (opt-in); else AllowedValues should be considered as opt-out.
bool DisabledThen true, bypass this component
object JobWriter show/hide
string GuidThe job unique identifier
string JobKeyThe name of the job
string PathThe path where the sync files are located
string UpsertColumnsComma-separated list of columns making up the unique key of the record
string ConnectionStringThe name of the Central Connection String to use for the target system
bool PropagateDeleteWhen true, identify or capture deleted records if available
string SourceJobKeyThe original name of the source job
string LastExecutionIdThe last ExecutionId extracted
string InitialExecutionIdThe initial ExecutionId used by this job
DateTime CreatedOnDate/time this job was created
string StatusThe status of this job (ready, running)
bool ActiveWhen true, the job will execute at the desired interval as specified by the CronSchedule parameter
DateTime LastRunTimeThe last execution date/time
string CronScheduleWhen specified, the Cron expression determine the frequency of the execution of the job
object JsonData
This setting is reserved and provided for information only
string EncryptionFileThe path to the file containing the PGP Private Key used to decrypt Sync Files
byte[] SaltThe cryptographic Salt used to encrypt the password
byte[] KEKThe encrypted password used to access the file
string TargetSystemThe target system (DB, DRIVE, MESSAGING, HTTP)
string TargetObjectThe name of the target object (for DB systems, the name of the table)
string UpsertScriptThe insert/update script to execute for DB systems (not needed for SQL Server (native) targets)
string DeleteScriptThe delete script to execute for DB systems (not needed for SQL Server (native) targets)
string LastSchemaHashThe hash of the source schema
object Options show/hide
bool AutoRecreateWhen true, for DB systems, determines if the target object should be recreated automatically
int ExecTimeoutWhen positive, controls the timeout of the write operation
int BatchCountWhen positive, sends the upserted or deleted records in batches of records
bool AuditLogEnabledWhen true, saves execution logs to the database for auditing purposes
HTTP Target Options show/hide
object HttpUpsertSettings
string HTTPUpsertURIFor HTTP targets, the URI of the HTTP endpoint for Upsert operations
string HTTPUpsertVerbFor HTTP targets, the HTTP Verb (GET, POST, DELETE, PUT) to use for Upsert operations
string HTTPUpsertContentTypeFor HTTP targets, the Content-Type header to use for Upsert operations
string HTTPUpsertHeadersFor HTTP targets, serialized list of headers and values for Upsert operations
object HttpDeleteSettings
string HTTPDeleteURIFor HTTP targets, the URI of the HTTP endpoint for Delete operations
string HTTPDeleteVerbFor HTTP targets, the HTTP Verb (GET, POST, DELETE, PUT) to use for Delete operations
string HTTPDeleteContentTypeFor HTTP targets, the Content-Type header to use for Delete operations
string HTTPDeleteHeadersFor HTTP targets, serialized list of headers and values for Delete operations
MESSAGING Target Options show/hide
object MessagingUpsertSettings
string SystemTypeA valid messaging system type (awssqs, rabbitmq, kafka, azeventhub)
rabbitmq settings:
string RoutingKeyThe RabbitMQ Routing Key
string ReplyToThe reply to queue
int PriorityThe priority of the message
string MessageIdThe message id (if any)
string AppIdThe application id (if any)
long expirationThe time to live of the message
bool persistWhen true, persist the message in the queue
awssqs settings:
string QueueNameThe AWS Queue Name (optional)
azeventhub settings:
string EventHubNameThe Azure Event Hub name (optional)
kafka settings:
string TopicNameThe Kafka topic name (optional)
string FieldKeyThe comma-separated list of key fields used to identify a unique message (the Upsert Key Columns)
string KeyFieldSeparatorThe Key values separator string to use (only needed when 2 or more key fields are provided)
object MessagingDeleteSettings
string SystemTypeA valid messaging system type (awssqs, rabbitmq, kafka, azeventhub)
refer to the MessageUpsertSettings section above
DRIVE Target Options show/hide
object TargetFileOptions
ContainerNameThe directory path, Azure Container name, or S3 Bucket Name
DistributionFieldNameThe field name used to dynamically distribute records across multiple files
FileNameThe file name to use (accepts field names and date tokens; ex: sales-[yyyy]-\{\{statecode\}\} )
UseRuntimeDateWhen true, use the execution date/time for date tokens
UseShadowCopyWhen true, use a shadow copy of remote files for faster updates
ShadowCopyTTLWhen UseShadowCopy is true, indicates how old the file can be before downloading a more recent version
ShadowCopyFrequencyThe ShadowCopyTTL value (hour, minute)
string SystemNameThe name of the target system (SQLServer, MySQL, Oracle, Teradata, Generic, AWS S3 Bucket, Azure Blob Container, Local Path, AWS SQS, Azure Event Hub, RabbitMQ)
dynamic[] PipelineDef show/hide
DataDynamicColumn
string NameThe name of the component: DataDynamicColumn
string ColumnNameThe column name to use
string DataTypeThe data type of the column (string, int, float, guid, datetime, long, double, char, bool, byte[])
string ExpressionTypeSQLExpression or EnzoFunction
string SQLExpressionThe expression (either a SQL expression, or an Enzo function call)
int MaxLengthColumn maximum length
bool IgnoreIfExistsWhen true, does not throw an error if the column already exists
bool DisabledThen true, bypass this component
DataHashing
string NameThe name of the component: DataHashing
string ColumnNameThe name of the column to hash
string HashNameThe hashing algorythm to use (MD5, SHA1, SHA256, SHA384, SHA512)
bool TruncateIfNeededWhen true, truncate the hash value to fit the column length
bool DisabledThen true, bypass this component
DataMasking
string NameThe name of the component: DataMasking
string ColumnNameThe name of the column to mask
string MaskingTypeThe masking operation (default, email, creditcard, randomnumber, customtext)
decimal RandomNbrMinWhen using RandomNumber masking, the minimum value to use
decimal RandomNbrMaxWhen using RandomNumber masking, the maximum value to use
int CustomPrefixWhen using CustomText, the prefix to use
int CustomSuffixWhen using CustomText, the suffix to use
string CustomPaddingWhen using CustomText, the padding to use
bool DisabledThen true, bypass this component
DataFilter
string NameThe name of the component: DataFilter
string FilterThe SQL filter to use to limit the records being processed
bool DisabledThen true, bypass this component
DataQuality
string NameThe name of the component: DataQuality
string ColumnNameThe column name to validate
string AllowedValuesComma-separated list of values to allow or deny
bool CheckListFlagWhen true, limit possible values to the list provided by the AllowedValues field
bool AllowNullWhen true, NULL values are allowed
bool AllowEmptyWhen true, empty strings are allowed
bool NumericRequiredWhen true, the value must be numeric
bool ThrowExceptionOnFailThen true, fail processing if the data validation rule failed
bool SkipRowOnFailWhen true, ignore the row if the data validation rule failed
bool OptOutListWhen true, the AllowedValues should be allowed (opt-in); else AllowedValues should be considered as opt-out.
bool DisabledThen true, bypass this component
 
Response
OK Success response code

request
PUT /job

{
	"JobReader": {
		"Guid": a3ff75315a4241d1958e1f9dbd53829f,
		"JobKey": "TEST-JOB-ALLDATABASES",
		"SqlRead": "SELECT * FROM sys.databases",
		"Path": "C:\\Tmp\\datasync_reader\\",
		"UpsertColumns": "database_id",
		"TimestampCol": "",
		"LastTsPointer": null,
		"LastTsDelPointer": null,
		"ConnectionString": "sql2017",
		"PropagateDelete": false,
		"SyncStrategy": "Full",
		"CreatedOn": "2022-10-12T19:24:14.3440036Z",
		"Status": null,
		"Initialize": false,
		"Active": true,
		"LastRunTime": null,
		"CronSchedule": "0 */5 * ? * *",
		"JsonData": {
			"AuditLogEnabled": true,
			"KeepChangedLogFiles": true,
			"EncryptionFile": "",
			"BypassTopN": false,
			"TopNOverride": 0,
			"DirectWriter": null,
			"ExecutionTimeout": 0,
			"BypassSchemaDetection": false,
			"DeletedRecordsMethod": 0,
			"DeletedRecordsSQL": null,
			"CDCTokenDeletedField": "",
			"MessagingOptions": null,
			"DriveSourceSettings": null,
			"Column2TableTxSettings": null,
			"HttpSourceSettings": null,
			"BigDataSourceSettings": null,
			"JobTriggers": [],
			"JobTriggersEnabled": false,
			"CDCSettings": null,
			"DynamicParameter": null
		},
		"SystemName": null,
		"PipelineDef": []
	},
	"JobWriter": null
}
response
OK
 
DELETE

Deletes a Job Reader, Job Writer, or Direct Job. When deleting a DirectJob, only the JobReader identifier is needed. This operation does not delete the execution log of the job.

Query Parameters
string guid The job identifier
 
Response
OK Success response code

request
DELETE /job?guid=a3ff75315a4241d1958e1f9dbd53829f
response
OK

Job Output

Provides the latest, in-memory, log output for of given job. The output is limited to the last available 200 entries. For job-specific output, see job/log.

GET
/job/output?guid=123456789

Gets the last 200 debug messages of a job given its job identifier.

Input Parameters
string guid The job identifier
 
Response
object[] OutputMessage show/hide
long SequenceA session unique identifier for the message; resets to 1 when the agent restarts
string GuidThe unique identifier for the job
DateTime DateAddedDate/time when the message was recorded
string MessageThe output message
int LevelThe level of the message (0: debug, 1: info, 2: warning, 3: error)
string ExecGuidThe execution unique identifier

request
GET /job/output?guid=2b76c83008174135849ab0c7619a6d9e
response
[
    {
        "Sequence":1,"
        Guid":"2b76c83008174135849ab0c7619a6d9e",
        "DateAdded":"2021-07-16T14:16:49.5078227Z",
        "Message":"Reader starting...",
        "Level":1,
        "ExecGuid": "b423b7621b2a4da1968c5372612000a6"
    },
    {
        "Sequence":2,"
        Guid":"2b76c83008174135849ab0c7619a6d9e",
        "DateAdded":"2021-07-16T14:16:49.6019398Z",
        "Message":"Strategy: full",
        "Level":1,
        "ExecGuid": "b423b7621b2a4da1968c5372612000a6"
    },
    {
        "Sequence":3,
        "Guid":"2b76c83008174135849ab0c7619a6d9e",
        "DateAdded":"2021-07-16T14:16:49.7131806Z",
        "Message":"Schema extracted successfully",
        "Level":1,
        "ExecGuid": "b423b7621b2a4da1968c5372612000a6"
    }
]

Job History

Provides the execution history of a job. When the job is a Direct Job (with both a reader and writer), the response object contains a separate entry for each job, and the RecordsDirectWritten and RecordsDirectDeleted attributes on the Job Reader represent the Direct Writer's changes.
This operation does not support Dynamic Jobs.

GET
/job/history?guid=123456789
/job/history?guid=123456789&limit=500&listenerType=writer&filter=all

Gets the execution history of a job.

Query Parameters
string guid The job identifier
limit Maximum number of rows to return (default: 100)
listenerType The listener type (reader, writer) (default: reader). Direct Jobs are considered reader jobs.
filter When empty or not provided (default), only returns execution history that identified or processed records, or that contain errors. When the filter is set to all, includes execution history that did not have any records identified or processed.
 
Response
object[] SyncStatus show/hide
string GuidThe job unique identifier
string ListenerTypeThe type of listener (reader or writer)
double ElapsedSQLTimeElapsed time reading from the source system (CPU seconds)
double ElapsedMapRecudeTimeElapsed time processing the map reduce logic (CPU seconds)
long RecordCountTotal number of records processed
long RecordsAvailableTotal number of records available (upsert and deleted)
long DeleteCountTotal number of deleted records available
string ErrorMessageProcessing error message
bool SuccessSuccess flag
string JobKeyThe name of the job
string ExecutionIdExecution Id of the job
string PackageFileThe name of the file where the available are found
DateTime CreatedOnDate/time this status was created
bool IsConsumerDeamonErrorWhen true, the error was thrown from the underlying messaging listener
long RecordsDirectWrittenRecords written by the Direct Writer job
long RecordsDirectDeletedRecords deleted by the Direct Writer job
string ExecGuidThe job execution unique identifier (shared by readers and direct writers)
double ExecRunTimeThe total duration of the job (reader and direct writer)
DateTime ExecStartTimeThe start time of the job
DateTime ExecStopTimeThe job completion time

request
GET /job/history?guid=0c843d0f69dd4b449d3c4c43d47bf7c7
response
[
    {
		"Guid": "0c843d0f69dd4b449d3c4c43d47bf7c7",
		"ListenerType": "reader",
		"ElapsedSQLTime": 0.0361273,
		"ElapsedMapRecudeTime": 0.1177181,
		"RecordCount": 5,
		"RecordsAvailable": 5,
		"DeleteCount": 0,
		"LastTsPointer": null,
		"LastDelTsPointer": null,
		"ErrorMessage": "",
		"Success": true,
		"JobKey": "SQL-2-SQLTest",
		"ExecutionId": "1660662334171",
		"PackageFile": "C:\\tmp\\datasync_reader\\SQL-2-SQLTest_1660662334171.eds",
		"CreatedOn": "2022-08-16T15:05:32.94",
		"IsConsumerDeamonError": false,
		"RecordsDirectWritten": 5,
		"RecordsDirectDeleted": 0,
		"ExecGuid": "b5b37cd4190d4e97a01d33e00c82fa74",
		"ExecRunTime": 4.83,
		"ExecStartTime": "2022-08-16T15:05:32.94",
		"ExecStopTime": "2022-08-16T15:05:36.77"
	},
	{
		"Guid": "c920da3f3cac4bb99f8cde2f8b2f231c",
		"ListenerType": "writer",
		"ElapsedSQLTime": 0.1596194,
		"ElapsedMapRecudeTime": 0.0,
		"RecordCount": 197,
		"RecordsAvailable": 197,
		"DeleteCount": 0,
		"LastTsPointer": null,
		"LastDelTsPointer": null,
		"ErrorMessage": "",
		"Success": true,
		"JobKey": "__SQL-2-SQLTest",
		"ExecutionId": "1647274246435",
		"PackageFile": "C:\\tmp\\datasync_reader\\SQL-2-SQLTest_1647274246435.eds",
		"CreatedOn": "2022-03-14T16:10:46.633",
		"IsConsumerDeamonError": false,
		"RecordsDirectWritten": 0,
		"RecordsDirectDeleted": 0,
		"ExecGuid": null,
		"ExecRunTime": -1.0,
		"ExecStartTime": "0001-01-01T00:00:00",
		"ExecStopTime": "0001-01-01T00:00:00"
	}
]

Jobs Info

Provides the configuration settings of a job, or all jobs, including any direct writer settings and last job status.
This operation does not support Dynamic Jobs.

GET
/jobs/info
/jobs/info?guid=123456789

Provides the configuration settings of a job.

Query Parameters
string guid The job identifier (optional)
 
Response
object[] JobInfoPackage show/hide
string GuidThe job unique identifier
int JobTypeThe type of job (1: reader, 2: writer)
string JobKeyThe name of the job
bool ActiveFlag indicating whether the job is active
object JobReader show/hide
bool ActiveWhen true, the job will execute at the desired interval as specified by the CronSchedule parameter
bool AuditLogEnabledWhen true, the job execution output will be saved so it can be retrieved later (using job/log)
string ConnectionStringThe name of the Central Connection String to use for the source system
DateTime CreatedOnThe date/time the job was created
string CronScheduleWhen specified, the Cron expression determine the frequency of the execution of the job; blank when the job is started manually.
string GuidThe job unique identifier
bool Initialize(internal use only)
bool IsDBSourceWhen true, the source system is a relational database
bool IsDirectWhen true, the job is a Direct job (with source and target defined)
bool IsDirty(internal use only)
bool IsDriveSourceWhen true, the source system is a file stored on a drive (local, cloud, or FTP)
bool IsHttpSourceWhen true, the source system is an HTTP REST endpoint
bool IsQueueingSourceWhen true, the source system is a messaging hub and the job acts as a Messaging Consumer
string JobKeyThe name of the job
int JobType1 (Job Reader)
DateTime LastRunTimeThe last execution date/time
string LastTsDelPointerThe last High Water Mark value for deleted records
string LastTsPointerThe last High Water Mark value for inserted or updated columns
string PathThe path where the sync file will be created
bool PropagateDeleteWhen true, identify or capture deleted records
string SqlReadFor DB connections, the SQL command to execute that will fetch source records
string StatusThe status of the job (running, ready)
string SyncStrategy(internal use: full)
string SystemNameThe name of the source system (SQLServer, MySQL, Oracle, Teradata, Generic, AWS S3 Bucket, Azure Blob Container, Local Path...)
string TimestampColWhen provided, the date/time or numeric column to use as the High Watermark when reading from a database
string TrackLastTsWhen True, the job tracks a High Watermark
string UpsertColumnsComma-separated list of columns making up the unique key of the record
object JsonData show/hide
bool AuditLogEnabledWhen true, saves the execution log for auditing purposes
string EncryptionFileWhen provided, the file that contains the PGP public key used for encrypting Sync Files
bool BypassTopNWhen true, bypasses the TOP N heuristic method for DB data sources
int TopNOverrideWhen positive, uses this value for the TOP N heuristic (default: 10000)
string DirectWriterWhen a direct job, the name of the Job Writer (usually set to the name of the Job Reader, with two underscored as the prefix)
int ExecutionTimeoutWhen positive, uses this value for individual requests as the timeout value (in seconds)
bool BypassSchemaDetectionWhen true, the job should not extract detailed schema information from the data source; applies to DB data sources only
int DeletedRecordsMethodWhen PropagateDelete is true, determines the method used to identify deleted records (0: default, 1: custom SQL that returns all source records, 2: custom SQL that returns the IDs of the deleted records)
string DeletedRecordsSQLWhen DeletedRecordsMethod is greater than 0, specifies the SQL command used to identify deleted records
string CDCTokenDeletedFieldWhen DeletedRecordsMethod is 2, comma-separated column names returned by DeletedRecordsSQL that uniquely identify the deleted record
object MessagingOptions show/hide
int MessageProcessingThe messaging processing mode (0: passthrough, 1: Msg to Dataset, 2: Msg Payload to Dataset)
bool MessagingBatchEnabledWhen true, messages are batched before being sent to their intended target
int MessagingBatchCountWhen MessagingBatchEnabled is true, number of messages to include in a batch
int MessagingFlushTTLWhen positive and MessagingBatchEnabled is true, the maximum time to wait before messages are sent if the BatchCount hasn't been met
string MessagingSampleMsgSample message to use for testing purposes
string MessagingContentColumnWhen MessageProcessing is 2 and not empty, adds a dynamic column to the data of this name with the original content of the payload
string RootPathWhen MessageProcessing is 2 and not empty, XML or JSON path to use to extract the Dataset
bool IncludeHeadersWhen true, headers should be sent to the target system as much as possible
bool IncludeMetadataWhen true, message metadata should be sent to the target system as much as possible
string HeaderFieldsWhen not empty and MessageProcessing is NOT 2 and IncludeHeaders is true, the comma-separated list of header keys to send to the target system (blank sends all)
string MetadataFieldsWhen not empty and MessageProcessing is NOT 2 and IncludeHeaders is true, the comma-separated list of metadata fields to send to the target system (blank sends all)
string QueueNameOverrideWhen provided, the name of the queue/bus/topic to read messages from (if blank, uses the connection string default settings)
string MessagingGroupNameWhen relevant, the message group name to use when reading from the source messaging hub (if blank, reads from the default group)
string PartitionKeysWhen relevant, a comma-separated list of partition keys to use when reading from the source messaging hub (if blank, reads from all partitions)
object DriveSourceSettings show/hide
string DriveFileFormatThe messaging processing mode (0: passthrough, 1: Msg to Dataset, 2: Msg Payload to Dataset)
string DriveFilePatternWhen true, messages are batched before being sent to their intended target
string DriveSchemaFileWhen MessagingBatchEnabled is true, number of messages to include in a batch
string DriveSelectedColumnsWhen positive and MessagingBatchEnabled is true, the maximum time to wait before messages are sent if the BatchCount hasn't been met
string DriveUpdatedFilesOnlySample message to use for testing purposes
bool MessagingContentColumnWhen MessageProcessing is 2 and not empty, adds a dynamic column to the data of this name with the original content of the payload
bool DriveIncludeSubfoldersWhen MessageProcessing is 2 and not empty, XML or JSON path to use to extract the Dataset
string DriveFolderColumnNameWhen true, headers should be sent to the target system as much as possible
bool DriveAddFolderColumnWhen true, message metadata should be sent to the target system as much as possible
bool DriveUseRootPathWhen not empty and MessageProcessing is NOT 2 and IncludeHeaders is true, the list of header keys to send to the target system (blank sends all)
string DriveRootPathWhen not empty and MessageProcessing is NOT 2 and IncludeHeaders is true, the list of metadata fields to send to the target system (blank sends all)
bool SingleFileProcessingWhen provided, the name of the queue/bus/topic to read messages from (if blank, uses the connection string default settings)
object ExtendedOptions show/hide
CSVOptions
string ColumnDelimitersThe character that separate individual field values when reading from a delimited file (ignored if FieldWidths are provided)
bool HasHeaderRowWhen true, assumes the file has a header row and column names will be inferred from the header row
string FieldWidthsIf provided, a comma-separated list of integers representing the length of each column for a fixed-length file format
string[] CommentTokensAn array of string values to be treated as comment tokens in the file
bool QuotedIdentifierWhen true, will treat double-quotes as boundary tokens for field name and value
object Column2TableTxSettings show/hide
string TxDocumentMsgContentColThe column index or name to use from the dataset containing an XML or JSON string to be transformed into a dataset
string TxDocumentRootPathIf provided, the XML or JSON path to use as the starting point to read the document content
string TxDocumentSourceColumnIf provided, the name of a dynamic column that will be added to the dataset with the content of the JSON or XML record
object HttpSourceSettings show/hide
string HTTPContentTypeThe Content-Type header used when sending a POST or PUT request
string HTTPHeadersAn array of key-value pairs representing custom headers to set
string HTTPURIThe URI to call; if blank, will call the URI defined by the connection; may use a relative URL or a full URL
string HTTPVerbThe HTTP Verb to use (GET, POST, PUT, DELETE)
string HTTPPayloadWhen executing a POST or PUT request, the payload to send with the request
bool ApplyContentTxWhen true, converts the JSON or XML content of the payload into a dataset
int PagingStrategyThe paging strategy to use (0: none, 1: simple offset, 2: simple page, 3 simple link, 4: bookmark); when positive, the corrresponding OffsetStrategy object must be provided
string SelectedColumnsThe columns to return as part of the dataset; leave blank to return all available columns
string HighWaterMarkThe name of the field that contains the High Watermark for future calls
string HighWaterMarkParamThe name of the Query Parameter to use when adding a High Watermark value for future calls
dynamic OffsetStrategy show/hide
HttpSimpleOffsetStrategy
string URLOffsetParamThe name of the URL Query Parameter to use when specifying an offset
HttpSimplePageStrategy
string URLPageParamThe name of the URL Query Parameter to use when specifying a page number
HttpSimpleLinkStrategy
string URLPagePathThe JSON or XML document path where to find the next link to call within the response payload
HttpTokenPagingStrategy
string TokenPathThe JSON or XML document path where to find the next bookmark value to use within the response payload
string QueryParamNameThe namd of the URL Query Parameter to use when specifying the bookmark value
object PayloadTxSettings show/hide
string TxDocumentMsgContentColThe column index or name to use from the dataset containing an XML or JSON string to be transformed into a dataset
string TxDocumentRootPathIf provided, the XML or JSON path to use as the starting point to read the document content
string TxDocumentSourceColumnIf provided, the name of a dynamic column that will be added to the dataset with the content of the JSON or XML record
object[] JobTriggers show/hide
string guidThe Job GUID to trigger (must be provided)
string jobKeyThe JobKey of the job to trigger (informational only)
int eventIdThe condition that will fire the trigger (0: On Completion, 1: On Success (with wata), 2: On Success (always), 3: On Error)
string eventNameThe friendly name of the condition (optional)
true JobTriggersEnabledWhen true, executes the jobs specified in the JobTriggers collection upon completion of the job
dynamic[] PipelineDef show/hide
DataDynamicColumn
string NameThe name of the component: DataDynamicColumn
string ColumnNameThe column name to use
string DataTypeThe data type of the column (string, int, float, guid, datetime, long, double, char, bool, byte[])
string ExpressionTypeSQLExpression or EnzoFunction
string SQLExpressionThe expression (either a SQL expression, or an Enzo function call)
int MaxLengthColumn maximum length
bool IgnoreIfExistsWhen true, does not throw an error if the column already exists
bool DisabledThen true, bypass this component
DataHashing
string NameThe name of the component: DataHashing
string ColumnNameThe name of the column to hash
string HashNameThe hashing algorythm to use (MD5, SHA1, SHA256, SHA384, SHA512)
bool TruncateIfNeededWhen true, truncate the hash value to fit the column length
bool DisabledThen true, bypass this component
DataMasking
string NameThe name of the component: DataMasking
string ColumnNameThe name of the column to mask
string MaskingTypeThe masking operation (default, email, creditcard, randomnumber, customtext)
decimal RandomNbrMinWhen using RandomNumber masking, the minimum value to use
decimal RandomNbrMaxWhen using RandomNumber masking, the maximum value to use
int CustomPrefixWhen using CustomText, the prefix to use
int CustomSuffixWhen using CustomText, the suffix to use
string CustomPaddingWhen using CustomText, the padding to use
bool DisabledThen true, bypass this component
DataFilter
string NameThe name of the component: DataFilter
string FilterThe SQL filter to use to limit the records being processed
bool DisabledThen true, bypass this component
DataQuality
string NameThe name of the component: DataQuality
string ColumnNameThe column name to validate
string AllowedValuesComma-separated list of values to allow or deny
bool CheckListFlagWhen true, limit possible values to the list provided by the AllowedValues field
bool AllowNullWhen true, NULL values are allowed
bool AllowEmptyWhen true, empty strings are allowed
bool NumericRequiredWhen true, the value must be numeric
bool ThrowExceptionOnFailThen true, fail processing if the data validation rule failed
bool SkipRowOnFailWhen true, ignore the row if the data validation rule failed
bool OptOutListWhen true, the AllowedValues should be allowed (opt-in); else AllowedValues should be considered as opt-out.
bool DisabledThen true, bypass this component
object JobWriter show/hide
string GuidThe job unique identifier
string JobKeyThe name of the job
string PathThe path where the sync files are located
string UpsertColumnsComma-separated list of columns making up the unique key of the record
string ConnectionStringThe name of the Central Connection String to use for the target system
bool PropagateDeleteWhen true, identify or capture deleted records if available
string SourceJobKeyThe original name of the source job
string LastExecutionIdThe last ExecutionId extracted
string InitialExecutionIdThe initial ExecutionId used by this job
DateTime CreatedOnDate/time this job was created
string StatusThe status of this job (ready, running)
bool ActiveWhen true, the job will execute at the desired interval as specified by the CronSchedule parameter
DateTime LastRunTimeThe last execution date/time
string CronScheduleWhen specified, the Cron expression determine the frequency of the execution of the job
object JsonData
This setting is reserved and provided for information only
string EncryptionFileThe path to the file containing the PGP Private Key used to decrypt Sync Files
byte[] SaltThe cryptographic Salt used to encrypt the password
byte[] KEKThe encrypted password used to access the file
string TargetSystemThe target system (DB, DRIVE, MESSAGING, HTTP)
string TargetObjectThe name of the target object (for DB systems, the name of the table)
string UpsertScriptThe insert/update script to execute for DB systems (not needed for SQL Server (native) targets)
string DeleteScriptThe delete script to execute for DB systems (not needed for SQL Server (native) targets)
string LastSchemaHashThe hash of the source schema
object Options show/hide
bool AutoRecreateWhen true, for DB systems, determines if the target object should be recreated automatically
int ExecTimeoutWhen positive, controls the timeout of the write operation
int BatchCountWhen positive, sends the upserted or deleted records in batches of records
bool AuditLogEnabledWhen true, saves execution logs to the database for auditing purposes
HTTP Target Options show/hide
object HttpUpsertSettings
string HTTPUpsertURIFor HTTP targets, the URI of the HTTP endpoint for Upsert operations
string HTTPUpsertVerbFor HTTP targets, the HTTP Verb (GET, POST, DELETE, PUT) to use for Upsert operations
string HTTPUpsertContentTypeFor HTTP targets, the Content-Type header to use for Upsert operations
string HTTPUpsertHeadersFor HTTP targets, serialized list of headers and values for Upsert operations
object HttpDeleteSettings
string HTTPDeleteURIFor HTTP targets, the URI of the HTTP endpoint for Delete operations
string HTTPDeleteVerbFor HTTP targets, the HTTP Verb (GET, POST, DELETE, PUT) to use for Delete operations
string HTTPDeleteContentTypeFor HTTP targets, the Content-Type header to use for Delete operations
string HTTPDeleteHeadersFor HTTP targets, serialized list of headers and values for Delete operations
MESSAGING Target Options show/hide
object MessagingUpsertSettings
string SystemTypeA valid messaging system type (awssqs, rabbitmq, kafka, azeventhub)
rabbitmq settings:
string RoutingKeyThe RabbitMQ Routing Key
string ReplyToThe reply to queue
int PriorityThe priority of the message
string MessageIdThe message id (if any)
string AppIdThe application id (if any)
long expirationThe time to live of the message
bool persistWhen true, persist the message in the queue
awssqs settings:
string QueueNameThe AWS Queue Name (optional)
azeventhub settings:
string EventHubNameThe Azure Event Hub name (optional)
kafka settings:
string TopicNameThe Kafka topic name (optional)
string FieldKeyThe comma-separated list of key fields used to identify a unique message (the Upsert Key Columns)
string KeyFieldSeparatorThe Key values separator string to use (only needed when 2 or more key fields are provided)
object MessagingDeleteSettings
string SystemTypeA valid messaging system type (awssqs, rabbitmq, kafka, azeventhub)
refer to the MessageUpsertSettings section above
DRIVE Target Options show/hide
object TargetFileOptions
ContainerNameThe directory path, Azure Container name, or S3 Bucket Name
DistributionFieldNameThe field name used to dynamically distribute records across multiple files
FileNameThe file name to use (accepts field names and date tokens; ex: sales-[yyyy]-\{\{statecode\}\} )
UseRuntimeDateWhen true, use the execution date/time for date tokens
UseShadowCopyWhen true, use a shadow copy of remote files for faster updates
ShadowCopyTTLWhen UseShadowCopy is true, indicates how old the file can be before downloading a more recent version
ShadowCopyFrequencyThe ShadowCopyTTL value (hour, minute)
string SystemNameThe name of the target system (SQLServer, MySQL, Oracle, Teradata, Generic, AWS S3 Bucket, Azure Blob Container, Local Path, AWS SQS, Azure Event Hub, RabbitMQ)
dynamic[] PipelineDef show/hide
DataDynamicColumn
string NameThe name of the component: DataDynamicColumn
string ColumnNameThe column name to use
string DataTypeThe data type of the column (string, int, float, guid, datetime, long, double, char, bool, byte[])
string ExpressionTypeSQLExpression or EnzoFunction
string SQLExpressionThe expression (either a SQL expression, or an Enzo function call)
int MaxLengthColumn maximum length
bool IgnoreIfExistsWhen true, does not throw an error if the column already exists
bool DisabledThen true, bypass this component
DataHashing
string NameThe name of the component: DataHashing
string ColumnNameThe name of the column to hash
string HashNameThe hashing algorythm to use (MD5, SHA1, SHA256, SHA384, SHA512)
bool TruncateIfNeededWhen true, truncate the hash value to fit the column length
bool DisabledThen true, bypass this component
DataMasking
string NameThe name of the component: DataMasking
string ColumnNameThe name of the column to mask
string MaskingTypeThe masking operation (default, email, creditcard, randomnumber, customtext)
decimal RandomNbrMinWhen using RandomNumber masking, the minimum value to use
decimal RandomNbrMaxWhen using RandomNumber masking, the maximum value to use
int CustomPrefixWhen using CustomText, the prefix to use
int CustomSuffixWhen using CustomText, the suffix to use
string CustomPaddingWhen using CustomText, the padding to use
bool DisabledThen true, bypass this component
DataFilter
string NameThe name of the component: DataFilter
string FilterThe SQL filter to use to limit the records being processed
bool DisabledThen true, bypass this component
DataQuality
string NameThe name of the component: DataQuality
string ColumnNameThe column name to validate
string AllowedValuesComma-separated list of values to allow or deny
bool CheckListFlagWhen true, limit possible values to the list provided by the AllowedValues field
bool AllowNullWhen true, NULL values are allowed
bool AllowEmptyWhen true, empty strings are allowed
bool NumericRequiredWhen true, the value must be numeric
bool ThrowExceptionOnFailThen true, fail processing if the data validation rule failed
bool SkipRowOnFailWhen true, ignore the row if the data validation rule failed
bool OptOutListWhen true, the AllowedValues should be allowed (opt-in); else AllowedValues should be considered as opt-out.
bool DisabledThen true, bypass this component
object LastStatus show/hide
string GuidThe job unique identifier
string ListenerTypeThe type of listener (reader or writer)
double ElapsedSQLTimeElapsed time reading from the source system (CPU seconds)
double ElapsedMapRecudeTimeElapsed time processing the map reduce logic (CPU seconds)
long RecordCountTotal number of records processed
long RecordsAvailableTotal number of records available (upsert and deleted)
long DeleteCountTotal number of deleted records available
string ErrorMessageProcessing error message
bool SuccessSuccess flag
string JobKeyThe name of the job
string ExecutionIdExecution Id of the job
string PackageFileThe name of the file where the available are found
DateTime CreatedOnDate/time this status was created
bool IsConsumerDeamonErrorWhen true, the error was thrown from the underlying messaging listener
long RecordsDirectWrittenRecords written by the Direct Writer job
long RecordsDirectDeletedRecords deleted by the Direct Writer job
string ExecGuidThe job execution unique identifier (shared by readers and direct writers)
double ExecRunTimeThe total duration of the job (reader and direct writer)
DateTime ExecStartTimeThe start time of the job
DateTime ExecStopTimeThe job completion time
bool IsDirectModeFlag indicating whether this is a Direct Job
object DirectLastStatus show/hide
string GuidThe job unique identifier
string ListenerTypeThe type of listener (reader or writer)
double ElapsedSQLTimeElapsed time reading from the source system (CPU seconds)
double ElapsedMapRecudeTimeElapsed time processing the map reduce logic (CPU seconds)
long RecordCountTotal number of records processed
long RecordsAvailableTotal number of records available (upsert and deleted)
long DeleteCountTotal number of deleted records available
string ErrorMessageProcessing error message
bool SuccessSuccess flag
string JobKeyThe name of the job
string ExecutionIdExecution Id of the job
string PackageFileThe name of the file where the available are found
DateTime CreatedOnDate/time this status was created
bool IsConsumerDeamonErrorWhen true, the error was thrown from the underlying messaging listener
long RecordsDirectWrittenRecords written by the Direct Writer job
long RecordsDirectDeletedRecords deleted by the Direct Writer job
string ExecGuidThe job execution unique identifier (shared by readers and direct writers)
double ExecRunTimeThe total duration of the job (reader and direct writer)
DateTime ExecStartTimeThe start time of the job
DateTime ExecStopTimeThe job completion time

request
GET /jobs/info?guid=0c843d0f69dd4b449d3c4c43d47bf7c7
Sample GET Output: Direct Job (MySQL Reader and Azure Event Hub Target)
response
show/hide
[
	{
		"Guid": "5945167f7aef40aa9c8a7873432d2ebb",
		"JobType": 1,
		"JobKey": "FILMS-AzEventHub",
		"Active": true,
		"JobReader": {
			"Guid": "5945167f7aef40aa9c8a7873432d2ebb",
			"JobKey": "FILMS-AzEventHub",
			"SqlRead": "SELECT * FROM Film LIMIT 10",
			"Path": "C:\\tmp\\datasync_reader",
			"UpsertColumns": "film_id",
			"TimestampCol": "",
			"LastTsPointer": "",
			"LastTsDelPointer": "",
			"ConnectionString": "mysql-sakila",
			"PropagateDelete": false,
			"SyncStrategy": "Full",
			"CreatedOn": "2021-10-19T18:51:12",
			"Status": "ready",
			"Initialize": false,
			"Active": true,
			"LastRunTime": "2021-11-11T18:58:14",
			"CronSchedule": "",
			"JsonData": {
				"AuditLogEnabled": true,
				"EncryptionFile": "",
				"BypassTopN": true,
				"TopNOverride": 0,
				"DirectWriter": "__FILMS-AzEventHub",
				"ExecutionTimeout": 0,
				"BypassSchemaDetection": true,
				"DeletedRecordsMethod": 0,
				"DeletedRecordsSQL": null,
				"CDCTokenDeletedField": "",
				"DriveFileFormat": "",
				"DriveFilePattern": "",
				"DriveSchemaFile": "",
				"DriveFileSelectedColumns": "",
				"DriveUpdatedFilesOnly": false,
				"DriveIncludeSubfolders": false,
				"DriveFolderColumnName": "",
				"DriveAddFolderColumn": false,
				"Column2TableTxSettings": null
			},
			"PipelineDef": [],
			"SystemName": "MySQL",
			"IsDBSource": true,
			"IsHttpSource": false,
			"IsDriveSource": false,
			"IsQueueingSource": false,
			"JobType": 1,
			"IsDirty": false,
			"TrackLastTs": false,
			"AuditLogEnabled": true,
			"IsDirect": true
		},
		"JobWriter": {
			"Guid": "6e3c31f6d2d94efe96283bcc30195b44",
			"JobKey": "__FILMS-AzEventHub",
			"Path": "C:\\tmp\\datasync_reader",
			"UpsertColumns": "film_id",
			"ConnectionString": "datazenhub",
			"PropagateDelete": false,
			"SourceJobKey": "FILMS-AzEventHub",
			"LastExecutionId": "1636657095008",
			"CreatedOn": "2021-10-19T18:51:12",
			"Status": "ready",
			"Active": true,
			"LastRunTime": "2021-11-11T18:58:15",
			"CronSchedule": "",
			"JsonData": null,
			"TargetSystem": "MESSAGING",
			"TargetObject": "",
			"InitialExecutionId": "",
			"UpsertScript": "{\"film_id\": {{film_id}}, \n\"title\": \"{{title}}\", \n\"description\": \"{{description}}\", \n\"release_year\": {{release_year}}, \n\"language_id\": {{language_id}}, \n\"original_language_id\": {{original_language_id}}, \n\"rental_duration\": {{rental_duration}}, \n\"rental_rate\": {{rental_rate}}, \n\"length\": {{length}}, \n\"replacement_cost\": {{replacement_cost}}, \n\"rating\": \"{{rating}}\", \n\"special_features\": \"{{special_features}}\", \n\"last_update\": \"{{last_update}}\"}",
			"DeleteScript": "",
			"LastSchemaHash": "",
			"Options": {
				"AutoRecreate": false,
				"ExecTimeout": 0,
				"BatchCount": 1000,
				"MessagingUpsertSettings": {
					"SystemType": "azeventhub",
					"SystemTypeDisplay": "Azure Event Hub",
					"EventHubName": "datazen1"
				},
				"MessagingDeleteSettings": {
					"SystemType": "azeventhub",
					"SystemTypeDisplay": "Azure Event Hub",
					"EventHubName": ""
				}
			},
			"SystemName": "Generic",
			"PipelineDef": null,
			"JobType": 2,
			"IsDirty": false,
			"IsDirect": false,
			"AuditLogEnabled": false
		},
		"LastStatus": {
			"Guid": "5945167f7aef40aa9c8a7873432d2ebb",
			"ListenerType": "reader",
			"ElapsedSQLTime": 0.0465737,
			"ElapsedMapRecudeTime": 0.0006172,
			"RecordCount": 1,
			"RecordsAvailable": 1,
			"DeleteCount": 0,
			"ErrorMessage": "",
			"Success": true,
			"JobKey": "FILMS-AzEventHub",
			"ExecutionId": "1636657095008",
			"PackageFile": "C:\\tmp\\datasync_reader\\FILMS-AzEventHub_1636657095008.eds",
			"CreatedOn": "2021-11-11T18:58:14",
			"IsConsumerDeamonError": false,
			"RecordsDirectWritten": 1,
			"RecordsDirectDeleted": 0
		},
		"IsDirectMode": true,
		"DirectLastStatus": {
			"Guid": "6e3c31f6d2d94efe96283bcc30195b44",
			"ListenerType": "writer",
			"ElapsedSQLTime": 0.7656911,
			"ElapsedMapRecudeTime": 0.0,
			"RecordCount": 1,
			"RecordsAvailable": 1,
			"DeleteCount": 0,
			"ErrorMessage": "",
			"Success": true,
			"JobKey": "__FILMS-AzEventHub",
			"ExecutionId": "1636657095008",
			"PackageFile": "C:\\tmp\\datasync_reader\\FILMS-AzEventHub_1636657095008.eds",
			"CreatedOn": "2021-11-11T18:58:15",
			"IsConsumerDeamonError": false,
			"RecordsDirectWritten": 0,
			"RecordsDirectDeleted": 0
		}
	}
]


GET /jobs/info?guid=c920da3f3cac4bb99f8cde2f8b2f231c
Sample GET Output: Direct Job (RabbitMQ Consumer, Kafka Producer, Pipeline Tx)
response
show/hide
[
	{
		"Guid": "fd87f7b599084360a94b0d5fba2e5c5c",
		"JobType": 1,
		"JobKey": "Rabbit-Kafka",
		"Active": false,
		"JobReader": {
			"Guid": "fd87f7b599084360a94b0d5fba2e5c5c",
			"JobKey": "Rabbit-Kafka",
			"SqlRead": "",
			"Path": "C:\\tmp\\datasync_reader\\MsgConsumers",
			"UpsertColumns": "",
			"TimestampCol": "",
			"LastTsPointer": "",
			"LastTsDelPointer": "",
			"ConnectionString": "rabbitmq1",
			"PropagateDelete": false,
			"SyncStrategy": "Full",
			"CreatedOn": "2021-10-11T17:57:58",
			"Status": "ready",
			"Initialize": false,
			"Active": false,
			"LastRunTime": "2021-12-09T20:49:29",
			"CronSchedule": "",
			"JsonData": {
				"AuditLogEnabled": true,
				"BypassSchemaDetection": true,
				"DirectWriter": "__Rabbit-Kafka",
				"IsQueuingSource": true,
				"MessagingOptions": {
					"MessageProcessing": 0,
					"MessagingColumns": "",
					"MessagingFlushTTL": 1,
					"RootPath": "",
					"MessagingContentColumn": "_raw",
					"MessagingBatchEnabled": true,
					"MessagingBatchCount": 20,
					"MessagingSampleMsg": "{\"film_id\": 1, \n\"title\": \"ACADEMY DINOSAUR\", \n\"description\": \"A Epic Drama of a Feminist And a Mad Scientist who must Battle a Teacher in The Canadian Rockies\", \n\"release_year\": 2006, \n\"language_id\": 1, \n\"original_language_id\": null, \n\"rental_duration\": 6, \n\"rental_rate\": 0.99, \n\"length\": 86, \n\"replacement_cost\": 20.99, \n\"rating\": \"PG\", \n\"special_features\": \"Deleted Scenes,Behind the Scenes\", \n\"last_update\": \"2006-02-15T05:03:42.0000000\"}",
					"QueueNameOverride": "ttt",
					"MessagingGroupName": null,
					"PartitionKeys": null,
					"IncludeMetadata": true,
					"IncludeHeaders": true,
					"MetadataFields": "",
					"HeaderFields": ""
				}
			},
			"PipelineDef": [
				{
					"ColumnName": "Engine",
					"DataType": "String",
					"ExpressionType": "SQL Expression (SQL Only)",
					"SQLExpression": "'DATAZEN'",
					"MaxLength": 0,
					"IgnoreIfExists": false,
					"Name": "DataDynamicColumn",
					"Disabled": false
				},
				{
					"Filter": "\\\"film_id\\\".+\\b\\d+.+",
					"Column": "0",
					"KeepRowOnConvertionError": false,
					"KeepRowOnNullOrEmpty": false,
					"RegexIgnoreCase": true,
					"RegexSingleLine": true,
					"FilterType": 3,
					"Name": "DataFilter",
					"Disabled": false
				}
			],
			"SystemName": "RabbitMQ",
			"IsDBSource": false,
			"IsHttpSource": false,
			"IsDriveSource": false,
			"IsQueueingSource": true,
			"JobType": 1,
			"IsDirty": false,
			"TrackLastTs": false,
			"AuditLogEnabled": true,
			"IsDirect": true
		},
		"JobWriter": {
			"Guid": "9dce102385804200ac3ff1e015e8cd9a",
			"JobKey": "__Rabbit-Kafka",
			"Path": "C:\\tmp\\datasync_reader\\MsgConsumers",
			"UpsertColumns": "",
			"ConnectionString": "kafka-cloud",
			"PropagateDelete": false,
			"SourceJobKey": "Rabbit-Kafka",
			"LastExecutionId": "1636649966604",
			"CreatedOn": "2021-10-11T17:57:58",
			"Status": "ready",
			"Active": true,
			"LastRunTime": "2021-12-09T20:49:30",
			"CronSchedule": "",
			"JsonData": null,
			"TargetSystem": "MESSAGING",
			"TargetObject": "",
			"InitialExecutionId": "",
			"UpsertScript": "",
			"DeleteScript": "",
			"LastSchemaHash": "",
			"Options": {
				"AutoRecreate": false,
				"ExecTimeout": 0,
				"BatchCount": 1000,
				"MessagingUpsertSettings": {
					"SystemType": "kafka",
					"SystemTypeDisplay": "Kafka",
					"TopicName": "",
					"FieldKey": null,
					"KeyFieldSeparator": "",
					"Headers": null
				},
				"MessagingDeleteSettings": null,
				"IsPassthroughMessaging": true,
				"ErrorHandling": {
					"ActionOnFail": 0,
					"RetryCount": 0,
					"RetryTTLSeconds": 0,
					"RetryExponential": false,
					"DeadLetterQueuePath": ""
				}
			},
			"SystemName": "Generic",
			"PipelineDef": null,
			"JobType": 2,
			"IsDirty": false,
			"IsDirect": false,
			"AuditLogEnabled": false
		},
		"LastStatus": {
			"Guid": "fd87f7b599084360a94b0d5fba2e5c5c",
			"ListenerType": "reader",
			"ElapsedSQLTime": 0.0,
			"ElapsedMapRecudeTime": 0.0467424,
			"RecordCount": 1,
			"RecordsAvailable": 1,
			"DeleteCount": 0,
			"ErrorMessage": "",
			"Success": true,
			"JobKey": "Rabbit-Kafka",
			"ExecutionId": "1636649966604",
			"PackageFile": "C:\\tmp\\datasync_reader\\MsgConsumers\\Rabbit-Kafka_1636649966604.eds",
			"CreatedOn": "2021-11-11T16:59:26",
			"IsConsumerDeamonError": false,
			"RecordsDirectWritten": 1,
			"RecordsDirectDeleted": 0
		},
		"IsDirectMode": true,
		"DirectLastStatus": {
			"Guid": "9dce102385804200ac3ff1e015e8cd9a",
			"ListenerType": "writer",
			"ElapsedSQLTime": 1.1952159,
			"ElapsedMapRecudeTime": 0.0,
			"RecordCount": 1,
			"RecordsAvailable": 1,
			"DeleteCount": 0,
			"ErrorMessage": "",
			"Success": true,
			"JobKey": "__Rabbit-Kafka",
			"ExecutionId": "1636649966604",
			"PackageFile": "C:\\tmp\\datasync_reader\\MsgConsumers\\Rabbit-Kafka_1636649966604.eds",
			"CreatedOn": "2021-11-11T16:59:26",
			"IsConsumerDeamonError": false,
			"RecordsDirectWritten": 0,
			"RecordsDirectDeleted": 0
		}
	}
]



Job Log

Provides access to a job execution output log. A job can be executed multiple times; however, each execution is assigned a unique identifier. This endpoint provides access to a specific job execution.

GET
/job/log?execguid=123456789

Returns the execution log of a job given its execution guid. An execution guid is assigned to a job when it start running. When a Direct job is executed, both the reader and writer share the same execution guid.

Query Parameters
string execguid The job execution guid (required)
 
Response
object[] OutputMessage show/hide
long SequenceThe sequence number of a log entry
string GuidThe job Guid
DateTime DateAddedThe date/time of the log entry
string MessageThe log message
int LevelThe log level (0: debug, 1: info, 2: warning, 3: error)
string ExecGuidThe execution unique identifier

request
GET /job/log?execguid=c920da3f3cac4bb99f8cde2f8b2f231c
response
[
 {
    "Sequence": 1,
    "Guid": "5945167f7aef40aa9c8a7873432d2ebb",
    "DateAdded": "2021-10-19T20:41:04",
    "Message": "Reader starting...",
    "Level": 0,
    "ExecGuid": "c920da3f3cac4bb99f8cde2f8b2f231c"
 },
 {
    "Sequence": 2,
    "Guid": "5945167f7aef40aa9c8a7873432d2ebb",
    "DateAdded": "2021-10-19T20:41:04",
    "Message": "Strategy: full",
    "Level": 0,
    "ExecGuid": "c920da3f3cac4bb99f8cde2f8b2f231c"
 },
 {
    "Sequence": 3,
    "Guid": "5945167f7aef40aa9c8a7873432d2ebb",
    "DateAdded": "2021-10-19T20:41:04",
    "Message": "Source data extraction starting...",
    "Level": 0,
    "ExecGuid": "c920da3f3cac4bb99f8cde2f8b2f231c"
 }
]


Job Status

Returns the Job Status for a given execution id.

GET
/job/status

Returns all available jobs defined under the agent including high level summary information.
This operation does not return Dynamic Jobs.

Input Parameters
string execguid
 
Response
object SyncStatus show/hide
string GuidThe job unique identifier
string ListenerTypeThe type of listener (reader or writer)
double ElapsedSQLTimeElapsed time reading from the source system (CPU seconds)
double ElapsedMapRecudeTimeElapsed time processing the map reduce logic (CPU seconds)
long RecordCountTotal number of records processed
long RecordsAvailableTotal number of records available (upsert and deleted)
long DeleteCountTotal number of deleted records available
string ErrorMessageProcessing error message
bool SuccessSuccess flag
string JobKeyThe name of the job
string ExecutionIdExecution Id of the job
string PackageFileThe name of the file where the available are found
DateTime CreatedOnDate/time this status was created
bool IsConsumerDeamonErrorWhen true, the error was thrown from the underlying messaging listener
long RecordsDirectWrittenRecords written by the Direct Writer job
long RecordsDirectDeletedRecords deleted by the Direct Writer job
string ExecGuidThe job execution unique identifier (shared by readers and direct writers)
double ExecRunTimeThe total duration of the job (reader and direct writer)
DateTime ExecStartTimeThe start time of the job
DateTime ExecStopTimeThe job completion time

GET
/jobs/status?execguid=5d66ebae-42d8-4a0d-b0c7-c4f323a5bb63
{
    "SyncStatus": {
        "Guid": "9dce102385804200ac3ff1e015e8cd9a",
        "ListenerType": "reader",
        "ElapsedSQLTime": 1.1952159,
        "ElapsedMapRecudeTime": 0.0,
        "RecordCount": 1,
        "RecordsAvailable": 1,
        "DeleteCount": 0,
        "ErrorMessage": "",
        "LastTsPointer": null,
        "LastDelTsPointer": null,
        "Success": true,
        "JobKey": "TestReader",
        "ExecutionId": "1704305959599",
        "PackageFile": "C:\\tmp\\datasync_reader\\MsgConsumers\\TestReader_1704305959599.eds",
        "CreatedOn": "2024-03-01T18:19:14",
        "IsConsumerDeamonError": false,
        "RecordsDirectWritten": 0,
        "RecordsDirectDeleted": 0,
        "ExecGuid": "5d66ebae-42d8-4a0d-b0c7-c4f323a5bb63",
        "ExecRunTime": "5.00",
        "ExecStartTime": "2024-01-11T16:59:26",
        "ExecStopTime": "2024-01-11T17:00:01",
    }
}

Jobs Summary

Provides summary information regarding all the jobs defined in this agent.

GET
/jobs/summary

Returns all available jobs defined under the agent including high level summary information.
This operation does not return Dynamic Jobs.

Input Parameters
(none)
 
Response
Guid The unique Guid of the job
TargetGuid The unique Guid of the target job (Direct Jobs only)
JobType 1: Job Reader, 2: Job Writer
JobKey The JobKey of the job
Active When true, the job is active and can be executed (manually or through a schedule)
IsDirectMode When true, indicates a Direct job (a reader and writer defined); when false, either a Reader or Writer is defined
Path The job path where the intermidate Sync Files will be stored
SourceSystemName The name of the source system (ex: MySQL)
TargetSystemName The name of the target system (ex: AzEventHub)
LastExecutionId The last Execution Id of the job
LastTsPointer The last High Watermark value (if any defined) for the job reader
LastTsDelPointer The last High Watermark value for Deleted records identification (if any defined)
UpsertColumns Comma-separated list of key columns, if any. If not specified, the target will use an 'append-only' mode
TimestampCol The High Watermark column name to use (if any defined) for the job reader
CreatedOn The date/time the job was created
CronSchedule The Cron schedule that defines the job frequency (if blank, the job runs manually or through a Trigger)
LastRunTime The last date/time the job was executed
Status The current status of the job (Ready, Running, Listening)
IsQueueingSource When True, the job is a Message Consumer
IsHttpSource When True, the job accesses an HTTP REST endpoint
IsDriveSource When True, the job accesses files on a Local Drive, Cloud Drive, or FTP site
IsDBSource When True, the job reads data from a database
TargetSystem The type of target system where data is being sent (Direct Jobs or Job Writers)
IsConsumerDeamonError When True, the error identified is the result of an underlying Messaging platform error
Success When True, the last job finished successfully
ErrorMessage When Success is False, contains the error message of the source system
DirectSuccess When a Direct Job, determines if the target operation was successful
DirectErrorMessage When DirectSuccess is false, contains the error message of the target system
TrackLastTs When True, indicates that the job tracks High Watermark values
ConnectionString The Connection String Key being used
SourceSystem The system name used by the Job Reader
IsCDCSource When true, the source system is a Change Data Capture stream

GET
/jobs/summary
[
	{
		"Guid": "2f7e8ec3258c456eb3ca9a9b6ef383b1",
		"TargetGuid": "71e6b0ab03a14f2dba5020e924713ea7",
		"JobType": 1,
		"JobKey": "FILMS-RabbitMQ",
		"Active": false,
		"IsDirectMode": true,
		"Path": "C:\\tmp\\films",
		"SourceSystemName": "MySQL",
		"TargetSystemName": "RabbitMQ",
		"LastExecutionId": "1636649963336",
		"LastTsPointer": "",
		"LastTsDelPointer": "",
		"UpsertColumns": "film_id",
		"TimestampCol": "",
		"CreatedOn": "2021-10-11T17:46:25",
		"CronSchedule": "0 * * ? * *",
		"LastRunTime": "2021-11-11T23:32:01",
		"Status": "ready",
		"IsQueueingSource": false,
		"IsHttpSource": false,
		"IsDriveSource": false,
		"IsDBSource": true,
		"TargetSystem": "MESSAGING",
		"IsConsumerDeamonError": false,
		"Success": true,
		"ErrorMessage": "",
		"DirectSuccess": false,
		"DirectErrorMessage": null,
		"TrackLastTs": false,
		"ConnectionString": "mysql-sakila",
		"SourceSystem": "MySQL",
		"IsCDCSource": false
	},
	{
		"Guid": "708f8f348a514dbbb394f95aef6f373e",
		"TargetGuid": "a71ab5bd3d324708b7c122bfecfef0b5",
		"JobType": 1,
		"JobKey": "Kafka-MSMQ",
		"Active": false,
		"IsDirectMode": true,
		"Path": "C:\\tmp",
		"SourceSystemName": "Kafka",
		"TargetSystemName": "MSMQ",
		"LastExecutionId": "1636653527797",
		"LastTsPointer": "{\"0\":3331}",
		"LastTsDelPointer": "",
		"UpsertColumns": "",
		"TimestampCol": "",
		"CreatedOn": "2021-10-14T18:51:14",
		"CronSchedule": "",
		"LastRunTime": "2021-12-09T20:49:07",
		"Status": "ready",
		"IsQueueingSource": true,
		"IsHttpSource": false,
		"IsDriveSource": false,
		"IsDBSource": false,
		"TargetSystem": "MESSAGING",
		"IsConsumerDeamonError": false,
		"Success": true,
		"ErrorMessage": "",
		"DirectSuccess": true,
		"DirectErrorMessage": "",
		"TrackLastTs": false,
		"ConnectionString": "kafka-cloud",
		"SourceSystem": "Kafka",
		"IsCDCSource": false
	}
]

Job Ts Pointer

Updates a Job Reader's Timestamp pointer used to determine the next starting time window for either read or delete operations. The Timestamp window is used when a Timestamp Column is used in a Job Reader to select records from a known last High Water Mark, or when a Job Reader uses a Change Data Capture stream to identify deleted records from a source system.
This operation does not support Dynamic Jobs.

POST
/job/settspointer?guid=123456789&type=read

Updates the last Timestamp value of a Job Reader for either read or delete operations.

Query Parameters
string guid The job identifier
string type The type of window to modify (read, delete)
string val The new value for the time window (optional); if empty or not provided, the Timestamp is set to an empty value.
 
Response
OK Success response code

request
POST /job/settspointer?guid=123456789&type=read
response
OK

Start Job

Starts a job immediately if it is enabled (active). If the job is already running, this operation does nothing.

POST
/job/start?guid=123456789

Start a job immediately. If the job is already running, this command is ignored and returns OK. The job must be enabled for this operation to work.
This operation does not support Dynamic Jobs.

Query Parameters
string guid The job identifier
 
Response
OK Success response code

request
POST /job/start?guid=c920da3f3cac4bb99f8cde2f8b2f231c
response
OK

Stop Job

Stops a job as soon as possible. If the job is not currently running, this operation does nothing. If the job is currently running, this operation sends cancellation requests to all pending operations and returns as soon as possible.

POST
/job/stop?guid=123456789

Stops a job.

Query Parameters
string guid The job identifier
 
Response
OK Success response code

request
POST /job/stop?guid=c920da3f3cac4bb99f8cde2f8b2f231c
response
OK

601 21st St Suite 300
Vero Beach, FL 32960
United States

(561) 921-8669
info@enzounified.com
terms of service
privacy policy

PLATFORM

ENZO SERVER
ENZO DATAZEN

SOLUTIONS

SOLUTIONS OVERVIEW
INTEGRATION
SaaS
CLOUD ANALYTICS

RESOURCES

DOWNLOAD
BLOGS & VIDEOS
IN THE NEWS
ENZO ADAPTERS
ONLINE DOCUMENTATION
TCO CALCULATOR

COMPANY

LEADERSHIP TEAM
PARTNERS


© 2023 - Enzo Unified