Skip to main content
Version: 1.14.0

Enable Preview Features

Preview features are LiveData Migrator features that are still under development and are subject to improvement. These are disabled by default, and must be enabled in the properties file:

/etc/wandisco/livedata-migrator/application.properties

Preview features must be enabled by finding the corresponding property with the prefix preview.feature. and changing the OFF value to ON. For example:

Example for enabling Backup & Restore
preview.feature.backup-restore=ON

LiveData Migrator must then be restarted for the changes to take effect:

service livedata-migrator restart

Current preview features#

You may check which preview features are currently active with the following curl command:

Command
curl localhost:18080/preview

The command will return information similar to the following:

Output
curl localhost:18080/preview{  "preview/verifications" : "OFF",  "preview/backups" : "OFF"}

The following preview features are currently available in LiveData Migrator.

Backup and Restore#

This feature allows you to save LiveData Migrator's current state - including migrations, filesystems, path mappings and configuration - and restore it later. View the full list of properties this feature backs up and restores here.

Enable this feature with the following property:

preview.feature.backup-restore=ON

Migration verifications#

This feature allows you to run a verification job on a Live, Running or Completed data migration. The verification job will scan your source and target filesystems (in the migration path) and compare them for any discrepancies.

Enable this feature with the following property:

preview.feature.verifications=ON

Preview status#

These features do not require enablement in the properties file. They are immediately available for use.

S3 as a source filesystem#

LiveData Migrator supports Amazon Simple Storage Service (Amazon S3) buckets as source storage.

Creating an Amazon S3 source in the UI#

To configure an Amazon S3 bucket as a source filesystem, select Amazon S3 in the Storage Type dropdown menu when configuring storage with the UI.

Creating an Amazon S3 source through the CLI#

Supply the --source parameter to the filesystem add s3a command.

Example
filesystem add s3a --file-system-id mytarget --bucket-name mybucket1 --credentials-provider org.apache.hadoop.fs.s3a.SimpleAWSCredentialsProvider --access-key B6ZEXAMPLEACCESSKEYA --secret-key OP87ChoExampleSecretKeyI904lT7AaDBGJpp0D --source

Databricks metadata agent#

LiveData Migrator supports metadata migration to Databricks Delta Lake.

Creating a Databricks metadata agent in the UI#

To configure Databricks Delta Lake as a metadata agent, select Databricks in the Agent Type dropdown menu when connecting metastores with the UI.

Creating a Databricks metadata agent through the CLI#

Use the hive agent add databricks command to set up a Databricks agent in the CLI.

Example for remote Databricks agent
hive agent add databricks --name databricksAgent --jdbc-server-hostname mydbcluster.cloud.databricks.com  --jdbc-port 443 --jdbc-http-path sql/protocolv1/o/8445611123456789/0234-125567-testy978 --access-token daexamplefg123456789t6f0b57dfdtoken4 --file-system-id mys3bucket --default-fs-override dbfs: --fs-mount-point /mnt/mybucket --convert-to-delta --host myRemoteHost.example.com --port 5552

See the command reference page for more information on how to configure Delta Lake and set up a metadata agent.