Release Notes
WD Fusion 2.2 Build 107 - May 27 2015
The 2.2 release provides a number of important improvements in terms of operating and ease-of-deployment.
Available Packages
TBA. The engineering team are looking at providing an up-to-date list of the available packages.
System Requirements
Before you install this software you must ensure that you meet the necessary system, software and hardware requirements for running WD Fusion. See a full checklist in our online user guide: docs.wandisco.com/bigdata/wdfusion/
Supported Hadoop Packages:
View our list of supported versions of Hadoop: - docs.wandisco.com/bigdata/wdfusion/install.html#supportedCertified Platforms / DBMS & DConE Support:
- HDP2.1.2 - 2.2.X
- PHD3.0
- CDH5.15 5.4
- EMC Isilon 7.2
- MapR M5,7
- DConE 1.3
- MySQL (Hive MetaStore)
Client Applications Supported:
- Hive
- SparkSQL
- Impala* Some current limitations - contact WANdisco's solutions team
- HBase SQOOP
- Flume
- Kafka
- Storm
Installation
You can find detailed instructions on how to get up and running in our
online user guide:
- docs.wandisco.com/bigdata/wdfusion/install.html#procedure
Upgrades from an earlier version:
It's essential that you remove previously installed versions of WD Fusion before you complete a new installation:
- docs.wandisco.com/bigdata/wdfusion/install.html#cleanup
Previous Release Notes:
You can view the release notes for previous releases in the release archive:
- docs.wandisco.com/bigdata/wdfusion/archive.html
Known issues
- The WD Fusion package has large file size (676 MB), this doesn't reflect the size of the installed application, the package currently includes separate install files for each of the 9 supported version and flavour of Hadoop.
- Hive from CDH 5.3 does not work with WD Fusion, as a result of HIVE-9991. The issue will be addressed once this fix for Hive is released.
-
Hive from CDH 5.4 will also fail with WD Fusion, resulting in the following message:
(FUS-455)FAILED: Execution Error, return code 1 from org.apache.hadoop.hive.ql.exec.DDLTask. MetaException(message:java.lang.RuntimeException: java.lang.ClassNotFoundException: Class com.wandisco.fs.client.FusionFs not found)
- Do not locate WD Fusion on the same server as other HDFS processors, especially DataNodes. HDFS's default block placement policy dictates that if a client is co-located on a DataNode, then that co-located DataNode will receive 1 block of whatever file is being put into HDFS from that client. This means that if the WD Fusion Server (where all transfers go through) is co-located on a DataNode, then all incoming transfers will place 1 block onto that DataNode. In which case the DataNode is likely to consumes lots of disk space in a transfer-heavy cluster, potentially forcing the WD Fusion Server to shut down in order to keep the Prevaylers from getting corrupted. (FUS-453)
-
Impala does not work with WD Fusion. Impala is not able to read data from non-HDFS file systems, which is how Impala sees Fusion. With Fusion 2.3 a Fusion Impala client will be available that will enable Fusion to support Impala reads without modifications to CDH, the Hive Metastore, or the customer's Impala applications. This should satisfy the vast majority of customer use cases, since Impala is typically used for fast read performance. Most customers use Hive to write data to HDFS and modify the Hive metastore, which Impala also uses.
If the customer is using Impala for both reads and writes a workaround is available to the field upon request. (FUS-476)
New Features
- WD Fusion now supports the deployment of multiple servers per zone, which:
- eliminates any single points of failure.
- provides scaling across multiple machines as needed.
- allows for clients to fail over from a crashed/disconnected WD Fusion server to one that is up.
- load balances traffic across multiple machines. (FUS-521)
- The WD Fusion now has a unified installer that handles the installation of all server-side elements: WD Fusion server, IHC servers and the WD Fusion UI. A browser-based UI is now included to enhance the speed and ease-of-use of the installer. (FUI-140)
- WD Fusion's REST API can now see all IHC servers on the cluster. (FUS-496)
- Additional logging has been added around the Consistency Check tool and for attempts made to login through Kerberos. (FUS-399) (FUS-358)
- WD Fusion's underlying replication engine has been updated to version 1.3.1. (FUS-549)
- Improvements to the replication engine include the "Early file pull" feature which can greatly improve the syncing between zones by allowing data to replicate as the file is still getting written by the originating client" (FUS-431)
- We've enabled password encryption mechanism for when DConE is run over SSL. (FUS-531)
Fixed Issues
- We have previously stated that Oracle's JDK is a prerequisite for installing WD Fusion. We can now confirm that WD Fusion will run with only the JRE, although we still specify the Oracle version as this is the version that we test. (FUS-545)
- Fixed an issue that caused the WD Fusion UI to have insufficient file permissions to run the stop/start server script. (FUS-174)