Introduction
The Anypoint Connector for the Hadoop Distributed File System (HDFS) is used as a bi-directional gateway between Mule applications and HDFS.
Read through this user guide to understand how to set up and configure a basic flow using the connector. Track feature additions, compatibility, limitations and API version updates with each release of the connector using the Connector Release Notes. Review the connector operations and functionality using the Technical Reference alongside the demo applications.
MuleSoft maintains this connector under the Select support policy.
Prerequisites
This document assumes that you are familiar with Mule, Anypoint Connectors, and Anypoint Studio Essentials. To increase your familiarity with Studio, consider completing a Anypoint Studio Tutorial. This page requires some basic knowledge of Mule Concepts, Elements in a Mule Flow, and Global Elements.
To use the HDFS connector, you need:
-
Anypoint Studio - An instance of Anypoint Studio. If you do not use Anypoint Studio for development, follow the instructions in Configuring Maven Dependencies for your project.
-
An instance of Hadoop Distributed File System up and running. It can be downloaded from here.
Hardware and Software Requirements
For hardware and software requirements, please visit the Hardware and Software Requirements page.
Compatibility
HDFS Hadoop connector is compatible with the following:
Application/Service | Version |
---|---|
Mule Runtime |
3.6 or newer |
Apache Hadoop |
2.7.1 or newer |
Starting with v5.0.0, the HDFS connector is licensed commercially with Anypoint Platform as are other Select connectors. Prior versions will remain freely available to the community. |
Install the Connector
You can install the connector in Anypoint Studio using the instructions in Installing a Connector from Anypoint Exchange.
Upgrading from an Older Version
If you’re currently using an older version of the connector, a small popup appears in the bottom right corner of Anypoint Studio with an "Updates Available" message.
-
Click the popup and check for available updates.
-
Click the Connector version checkbox and click Next and follow the instructions provided by the user interface.
-
Restart Studio when prompted.
-
After restarting, when creating a flow and using the HDFS Connector, if you have several versions of the connector installed, you may be asked which version you would like to use. Choose the version you would like to use.
Additionally, we recommend that you keep Studio up to date with its latest version.
Configure the Connector Global Element
To use the HDFS connector in your Mule application, you must configure a global HDFS element that can be used by the connector (read more about Global Elements). The HDFS connector offers the following global configuration, requiring the following credentials:
-
Simple authetication configuration:
Field Description NameNode URI
The URI of the file system to connect to.
It is passed to HDFS client as the FileSystem#FS_DEFAULT_NAME_KEY configuration entry. It can be overriden by values in configurationResources and configurationEntries. Username
A simple user identity of a client process.
It is passed to HDFS client as the "hadoop.job.ugi" configuration entry. It can be overriden by values in configurationResources and configurationEntries. If not provided it will use the currently logged in user. Configuration Resources
A list of configuration resource files to be loaded by the HDFS client. Here you can provide additional configuration files. (e.g core-site.xml)
Configuration Entries
A map of configuration entries to be used by the HDFS client. Here you can provide additional configuration entries as key/value pairs.
-
Kerberos authentication configuration:
Field Description NameNode URI
The URI of the file system to connect to.
It is passed to HDFS client as the FileSystem#FS_DEFAULT_NAME_KEY configuration entry. It can be overriden by values in configurationResources and configurationEntries. Username
A simple user identity of a client process.
It is passed to HDFS client as the "hadoop.job.ugi" configuration entry. It can be overriden by values in configurationResources and configurationEntries. If not provided it will use the currently logged in user. KeytabPath
Path to the keytab file associated with username.
It is used in order to obtain TGT from "Authorization server". If not provided it will look for a TGT associated to username within your local kerberos cache. Configuration Resources
A list of configuration resource files to be loaded by the HDFS client. Here you can provide additional configuration files. (e.g core-site.xml)
Configuration Entries
A map of configuration entries to be used by the HDFS client. Here you can provide additional configuration entries as key/value pairs.
Using the Connector
You can use this connector as an inbound endpoint for polling content of a file at a configurable rate (interval) or as an outbound connector for manipulating data into the HDFS server.
See a full list of operations for any version of the connector here. |
Connector Namespace and Schema
When designing your application in Studio, the act of dragging the connector from the palette onto the Anypoint Studio canvas should automatically populate the XML code with the connector namespace and schema location.
Namespace: http://www.mulesoft.org/schema/mule/hdfs
Schema Location: http://www.mulesoft.org/schema/mule/connector/current/mule-hdfs.xsd
If you are manually coding the Mule application in Studio’s XML editor or other text editor, define the namespace and schema location in the header of your Configuration XML, inside the <mule> tag.
|
1
2
3
4
5
6
7
8
9
10
11
12
<mule xmlns="http://www.mulesoft.org/schema/mule/core"
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"
xmlns:connector="http://www.mulesoft.org/schema/mule/hdfs"
xsi:schemaLocation="
http://www.mulesoft.org/schema/mule/core
http://www.mulesoft.org/schema/mule/core/current/mule.xsd
http://www.mulesoft.org/schema/mule/connector
http://www.mulesoft.org/schema/mule/connector/current/mule-hdfs.xsd">
<!-- put your global configuration elements and flows here -->
</mule>
Using the Connector in a Mavenized Mule App
If you are coding a Mavenized Mule application, this XML snippet must be included in your pom.xml
file.
1
2
3
4
5
<dependency>
<groupId></groupId>
<artifactId></artifactId>
<version></version>
</dependency>
Inside the
|
Demo Mule Applications Using Connector
Exsting demos demontrate how to use connector for basic file system operations and how to poll data from a file at specific interval.
Example Use Case
The following example shows how to create a text file into HDFS through connector:
-
In Anypoint Studio, click File > New > Mule Project, name the project, and click OK.
-
In the search field, type "http" and drag the HTTP connector to the canvas, click the green plus sign to the right of Connector Configuration, and in the next screen, click OK to accept the default settings. Name the endpoint /createFile.
-
In the Search bar type "HDFS" and drag the HDFS connector onto the canvas. Configure as explained Configure the Connector Global Element
-
Choose Write to path as an operation. Set Path to "/test.txt" (this is the path of the file that is going to be created into HDFS) and leave other options with default values.
-
The flow should look like this:
-
Run the application and from your favorite http client make a POST request having "Content-type:plain/text" to locahost:8081/createFile with content that you want to write as payload. (e.g. curl -X POST -H "Content-Type:plain/text" -d "payload to write to file" localhost:8090/createFile)
-
Check that /test.txt has been created and has your content by using Hadoop explorer.
This and other basic file systen operations are demonstrated within this demo.
Example Use Case - XML
Paste this into Anypoint Studio to interact with the example use case application discussed in this guide.
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
<?xml version="1.0" encoding="UTF-8"?>
<mule xmlns:hdfs="http://www.mulesoft.org/schema/mule/hdfs" xmlns:http="http://www.mulesoft.org/schema/mule/http" xmlns="http://www.mulesoft.org/schema/mule/core" xmlns:doc="http://www.mulesoft.org/schema/mule/documentation"
xmlns:spring="http://www.springframework.org/schema/beans"
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"
xsi:schemaLocation="http://www.springframework.org/schema/beans http://www.springframework.org/schema/beans/spring-beans-current.xsd
http://www.mulesoft.org/schema/mule/core http://www.mulesoft.org/schema/mule/core/current/mule.xsd
http://www.mulesoft.org/schema/mule/http http://www.mulesoft.org/schema/mule/http/current/mule-http.xsd
http://www.mulesoft.org/schema/mule/hdfs http://www.mulesoft.org/schema/mule/hdfs/current/mule-hdfs.xsd">
<http:listener-config name="HTTP_Listener_Configuration" host="0.0.0.0" port="8081" doc:name="HTTP Listener Configuration"/>
<hdfs:config name="HDFS__Configuration" nameNodeUri="hdfs://localhost:9000" doc:name="HDFS: Configuration"/>
<flow name="hdfs-example-use-caseFlow">
<http:listener config-ref="HTTP_Listener_Configuration" path="/createFile" doc:name="HTTP"/>
<hdfs:write config-ref="HDFS__Configuration" path="/test.txt" doc:name="HDFS"/>
</flow>
</mule>
Connector Performance
To define the pooling profile for the connector manually, access the Pooling Profile tab in the applicable global element for the connector.
For background information on pooling, see Tuning Performance.
Resources
-
Access the HDFS Connector Release Notes.