Accessing HDFS web UI from another machine - windows

I can access my Hadoop File System from Web UI <IP address>:50070 . I want to access the same from a machine present outside of cluster. I tried to do it but it says webpage not found error. How can I access my HDFS from windows machine through a URL?

Related

How to access Hadoop from another machine on same network?

I have one machine running Hadoop and I do want to access Ambari by another machine on the same network.
How can I do that?
For security reasons, ports used by Hadoop cannot be accessed over a public IP address. To connect to Hadoop from a different machine, you must the open port of the service you want to access remotely. I believe Openstack provides the option for assigning Floating IPs. That's one of the options to enable you to access remotely. You must create a port redirect rule on your VM program.

Is that possible to connect Ubuntu HDFS using C# application

I am having HDFS in Ubuntu Environment, Is that possible to connect Ubuntu HDFS using C# application (Windows OS).
All the systems are connected via LAN.
I want to read simple CSV file from HDFS.
I want to know whether it is possible or not.
If you are using Hortonworks Azure HDInsight you can directly use C# to access the HDFS. In your case you are trying to read from windows OS. Please try using webhdfs. But it need configuration. Please check the below url for details.
URL: http://hadoop.apache.org/docs/r2.4.1/hadoop-hdfs-httpfs/

Downloading Hadoop Data from other PC

I have Hadoop v2.6 installed in my one PC in Ubuntu OS 14.04. I have added lots of unstructured data using Hadoop -put command into HDFS.
Can someone tell me how to download this data from another PC which is not in Hadoop Cluster using the Web User Interface provided by Hadoop??
I can access the data from other PC by typing in the address bar of browser (the IP address of HDFS server):Port Number
Like this: 192.168.x.x:50070
The problem is, I am not able to download the data as it gives the error "Webpage Not Available". I also tried other browsers, but still no luck.
Port 50070 is the default name node port. You should try port 14000 which is the default HttpFS port. If it still doesn't work try using the example from the manual:
http://192.168.x.x:14000?user.name=babu&op=homedir

Connecting to a Hadoop HDFS within a cloudera virtual machine via IP

I'm looking to connect to a hadoop instance on a cloudera virtual machine via the server address of the hdfs. Would anyone know how to find the IP address of this?
If not, how could I connect locally to the hdfs within virtual machine, as they are both running on the same computer.
Need to do this for a pentaho kettle connection.
If you're trying to configure Pentaho Data Integration (Kettle) to use HDFS as an input data source, then first you'll need to get the hostname/IP address and port number of the HDFS NameNode service, which you will then enter into (Pentaho) Spoon (the GUI to Kettle).
Getting HDFS NameNode IP/port number
The default port of the Hadoop HDFS NameNode service is 8020 in both CDH4 and 5 (source).
If for some reason you're not using the defaults, then the hostname/port of the HDFS NameNode service can be found in Cloudera Manager (which should be installed if you're using the Cloudera Quickstart VM, for example):
Click on the HDFS service on the main Cloudera Manager page
Click on Configuration - View and Edit
Click on NameNode - Ports and Addresses.
"NameNode Port" is the one you want, i.e. not "NameNode Web UI Port". As PDI needs the NameNode port.
Browse HDFS files in PDI to confirm
Test by opening Pentaho Data Integration (Spoon) and creating a "Hadoop Copy Files" transformation step, as an example, and then enter in your HDFS details in the "browse files" area and check if a directory list shows up.

Upload file from my virtual machine to another virtual machine using hadoop hdfs

Can any one please tell me, how can I upload a txt file from my local machine to another virtual machine based on IP Address in hdfs.
Regards,
Baskar.V
You might find webHDFS REST API useful. I have tried it to write content from my local FS to the HDFS and it works fine. But being REST based it should work fine from local FS of a remote machine as well, provided both machines are connected.

Resources