Data Hub Gradle Tasks
A complete list of all of the Gradle tasks available in Data Hub Gradle Plugin (ml-data-hub).
Using Gradle in Data Hub
To use Data Hub Gradle Plugin in the Data Hub flows, see Data Hub Gradle Plugin.
To pass parameters to Gradle tasks, use the -P option.
You can use Gradle's -i
option to enable info-level logging.
- Tasks with names starting with
ml
are customized for Data Hub from the ml-gradle implementation. - Tasks with names starting with
hub
are created specifically for Data Hub.
gradle tasks
.MarkLogic Data Hub Setup Tasks
These tasks are used to configure MarkLogic Data Hub and manage the data hub.
Uses hubPreinstallCheck
to deploy your Data Hub project.
Extends ml-gradle's WatchTask by ensuring that modules in Data Hub-specific folders (plugins
and entity-config
) are monitored.
Updates the properties of every database without creating or updating forests. Many properties of a database are related to indexing.
Updates your Data Hub instance to a newer version.
Before you run the hubUpdate
task, edit the build.gradle
file. Under plugins
, change the value of 'com.marklogic.ml-data-hub' version to the new Data Hub version.
plugins {
id 'com.marklogic.ml-data-hub' version '5.0.4'
}
For complete instructions on upgrading to a newer Data Hub version, see Upgrading Data Hub.
Running the hubUpdate task with the -i option (info mode) displays specifically what the task does, including configuration settings that changed.
Prints out basic info about the Data Hub configuration.
Installs user artifacts, such as entities and mappings, to the MarkLogic server. (Data Hub 4.2 or later)
MarkLogic Data Hub Scaffolding Tasks
These tasks allow you to scaffold projects, entities, flows, and steps.
Initializes the current directory as a Data Hub project.
Creates a boilerplate entity.
- entityName
- (Required) The name of the entity to create.
Creates a boilerplate flow definition file.
- flowName
- (Required) The name of the flow to create.
Creates a custom step definition that can be added to a flow as a step.
- stepDefName
- (Required) The name of the custom step definition to create.
- stepDefType
- The type of the step definition to create:
ingestion
,mapping
,mastering
, orcustom
. Default iscustom
. - format
- The format of the module to associate with the new step definition:
xqy
for XQuery orsjs
for JavaScript. Default issjs
.
A JavaScript module (main.sjs) is created and associated with the step definition to perform the processes required for the step.
- If
-Pformat=sjs
or if the option is not specified, only the main.sjs file is created, and it contains the processes required for the step. - If
-Pformat=xqy
, two files are created:- lib.xqy, which is the XQuery module that you must customize. It contains the processes required for the step; for example, custom code to create an envelope.
- main.sjs, which acts as a wrapper around lib.xqy.
These modules can be found under your-project-root/src/main/ml-modules.
hubCreateFlow
. The example steps use the predefined default-ingestion
, default-mapping
, and default-mastering
step definitions, so you won't need to create a new one.Generates security configuration files for protecting entity properties designated as Personally Identifiable Information (PII). For details, see Managing Personally Identifiable Information.
MarkLogic Data Hub Flow Management Tasks
These tasks allow you to run flows and clean up.
Runs a flow.
- flowName
- (Required) The name of the harmonize flow to run.
- entityName
- (Required if the flow includes a mapping step) The name of the entity used with the mapping step.
- batchSize
- The number of items to include in a batch. Default is 100.
- threadCount
- The number of threads to run. Default is 4.
- showOptions
- If
true
, options that were passed to the command are printed out. Default isfalse
. - failHard
- If
true
, the flow's execution is ended immediately if a step fails. Default isfalse
. - steps
- The comma-separated numbers of the steps to run. If not provided, the entire flow is run.
- jobId
- A unique job ID to associate with the flow run. This option can be used if the flow run is part of a larger process (e.g., a process orchestrated by NiFi with its own job/process ID). Must not be the same as an existing Data Hub job ID. If not provided, a unique Data Hub job ID will be assigned.
- options
- A JSON structure containing key-value pairs to be passed as custom parameters to your step modules.
- optionsFile
- The path to a JSON file containing key-value pairs to be passed as custom parameters to your step modules.
The custom key-value parameters passed to your step module are available through the $options (xqy) or options (sjs) variables inside your step module.
Exports job records. This task does not affect the contents of the staging or final databases.
- jobIds
- A comma-separated list of job IDs to export.
- filename
- The name of the zip file to generated, including the file extension. Default is
jobexport.zip
.
Deletes job records. This task does not affect the contents of the staging or final databases.
- jobIds
- (Required) A comma-separated list of job IDs to delete.
MarkLogic Data Hub Uninstall Tasks
Removes all components of your data hub from the MarkLogic server, including databases, application servers, forests, and users.
Legacy (DHF 4.x) Tasks
Creates a legacy (DHF 4.x) input flow. The resulting DHF 4.x flow must be executed using hubRunLegacyFlow.
- entityName
- (Required) The name of the entity that owns the flow.
- flowName
- (Required) The name of the input flow to create.
- dataFormat
xml
orjson
. Default isjson
.- pluginFormat
xqy
orsjs
. The plugin programming language.
Creates a legacy (DHF 4.x) harmonization flow. The resulting DHF 4.x flow must be executed using hubRunLegacyFlow.
- entityName
- (Required) The name of the entity that owns the flow.
- flowName
- (Required) The name of the harmonize flow to create.
- dataFormat
xml
orjson
. Default isjson
.- pluginFormat
xqy
orsjs
. The plugin programming language.- mappingName
- The name of a model-to-model mapping to use during code generation.
Runs a (legacy) DHF 4.x harmonization flow.
- entityName
- (Required) The name of the entity containing the harmonize flow.
- flowName
- (Required) The name of the harmonize flow to run.
- batchSize
- The number of items to include in a batch. Default is 100.
- threadCount
- The number of threads to run. Default is 4.
- sourceDB
- The name of the database to run against. Default is the name of your staging database.
- destDB
- The name of the database to put harmonized results into. Default is the name of your final database.
- showOptions
- Whether to print out options that were passed in to the command. Default is
false
. - dhf.YourKey
- The value to associate with your key. These key-value pairs are passed as custom parameters to your flow. You can pass additional key-value pairs as separate options:
hubrunlegacyflow ... -Pdhf.YourKeyA=YourValueA -Pdhf.YourKeyB=YourValueB ...
The custom key-value parameters passed to your step module are available through the $options (xqy) or options (sjs) variables inside your step module.