Use HttpFS

HDFS provides a built-in HttpFS server, which can be used to transfer data between clusters running different Hadoop versions, access data through firewall, use HTTP utilities and more.

In Arenadata Hadoop, HttpFS server comes preinstalled and available as a HDFS UI. For more information, see HDFS UI overview.

You can look up the default HttpFS server settings in the following sections on the HDFS configuration page: httpfs-site.xml, httpfs-env.sh, and HDFS Heap Memory settings.

To find these parameters in ADCM, go to the Clusters page, select the desired cluster, navigate to the Services tab, and click at HDFS.

An example command that calls the ls function for the user/tmp directory as the admin user:

$ curl 'http://<httpfs-host>:14000/webhdfs/v1/user/tmp?op=LISTSTATUS&user.name=admin'

Where <httpfs-host> is the IP address of the host with installed HttpFS server.

Found a mistake? Seleсt text and press Ctrl+Enter to report it