Skip to main content
Version: 1.20.0 (latest)

Configure source filesystems

Configure source filesystems for each product to migrate your data from depending on what your environment is:

  • Hadoop Distributed File System (HDFS) - Add one source filesystem only for each product.
  • S3 sources (IBM Cloud Object Storage, Amazon S3) - Add one or more source filesystems.

Data Migrator supports the following filesystems as a source:

Configure source filesystems with the UI#

The Filesystems panel shows the source and target filesystems Data Migrator can use for data migrations.

Use the Filesystems panel to:

  • View and configure source and target filesystems.
  • Add or remove targets.

Add a source filesystem#

To add a source filesystem from your dashboard, select the following:

  1. The relevant instance from the Products panel.
  2. Add source filesystem in the Filesystem Configuration page.
info

If you have HDFS in your environment, Data Migrator automatically detects it as your source filesystem. However, if Kerberos is enabled, or if your Hadoop configuration doesn't contain the configuration file information required for Data Migrator to connect to Hadoop, configure the source filesystem manually with additional Kerberos configuration settings.

If you want to configure a new source manually, delete any existing source first, and then manually add a new source.

note

If you deleted the HDFS source that Data Migrator detected automatically, and you want to redetect it, go to the CLI and run the command filesystem auto-discover-hdfs.

Configure HDFS as a source#

See Configure a HDFS source filesystem.

Configure source filesystems with the CLI#

Data Migrator migrates data from a single source filesystem. Data Migrator automatically detects the Hadoop Distributed File System (HDFS) it's installed on and configures it as the source filesystem. If it doesn't detect the HDFS source automatically, you can validate the source. You can override auto-discovery of any HDFS source by manually adding a source filesystem.

note

At this time, Azure Data Lake Storage (ADLS) Gen2 source filesystems can only be used for one-time migrations.

Use the following CLI commands to add source filesystems:

CommandAction
filesystem add adls2 oauthAdd an ADLS Gen 2 filesystem resource using a service principal and oauth credentials
filesystem add adls2 sharedKeyAdd an ADLS Gen 2 filesystem resource using access key credentials
filesystem add gcsAdd a Google Cloud Storage filesystem resource
filesystem add hdfsAdd a HDFS resource
filesystem add s3aAdd an S3 filesystem resource (choose this when using Amazon S3 and IBM Cloud Object Storage)
filesystem add localAdd a local filesystem resource

Validate your source filesystem#

Verify that the correct source filesystem is registered or delete the existing one (you define a new source in the step Add a source filesystem.

If Kerberos is enabled or your Hadoop configuration does not contain the information needed to connect to the Hadoop filesystem, use the filesystem auto-discover-source hdfs command to enter your Kerberos credentials and auto-discover your source HDFS configuration.

note

If Kerberos is disabled, and Hadoop configuration is on the host, Data Migrator will detect the source filesystem automatically on startup.

Manage your source filesystem#

Manage the source filesystem with the following commands:

CommandAction
source clearDelete all sources
source delDelete one source
filesystem auto-discover-hdfsAutomatically detect a HDFS source
source showView the source filesystem configuration
filesystem auto-discover-source hdfsEnter your Kerberos credentials to access your source HDFS configuration
note

To update existing filesystems, first stop all migrations associated with them. After saving updates to your configuration, you'll need to restart the Data Migrator service for your updates to take effect. In most supported Linux distributions, run the command service livedata-migrator restart.