I am using Cloudera Manager with CDH4.2.2 for my 3+1 cluster. On starting the installation with cloudera manager, it automatically downloads and installs JDK1.6. I want to use JDK1.7 with CDH for my convinience. Is it possible or is there any version of CDH which while installating Hadoop in the cluster automatically downloads and installs and successfully runs Hadoop with JDK1.7?
If yes, may I know which version of CDH is it and where do i get to download it from?
I want to work with JDK1.7 instead of 1.6 because i want to install Apache Giraph on CDH but it seems Giraph does not fit fine with JDK1.6 and needs the JDK1.7.
With Regards,
JDK 1.7 is supported for all CDH applications as of CDH 4.4 and Cloudera Manager 4.7.
That being said, no version of Cloudera Manager 4.x installs JDK 1.7 during the installation (latest version is 4.8.2). The only version of Cloudera Manager that installs JDK 1.7 automatically is 5.0.0.
To summarize: If you want an automated installation of JDK 1.7 via Cloudera Manager, you need to upgrade to CDH 5, and CM 5.0.0. Alternatively, you could upgrade to CDH4.4, and then perform a manual installation of JDK 1.7.
Related
I installed & built Apache Ambari 2.7.6 from:
Installation Guide for Ambari 2.7.6 .In launch install wizard step 1, Select Version is empty. How can i pass this step without using HDP/HDF and use own Apache?
Thanks
Installer step 1
To use open source with Ambari you'd need to build ambari with this patch:
https://issues.apache.org/jira/browse/AMBARI-25366
You can also use the Mpack that is provided by BigTop in this
pull: https://github.com/apache/bigtop/pull/669
It installs a management pack that could then be used to to install BigTop which is an opensource version of hadoop that packages versions of the Hadoop Zoo.
Build Ambari (with that patch) in ticket above, or download the management pack from BigTop
Install the Bigtop management pack
Then the 'bigTop' installed version will become available in the wizard.
I am new to HDP installation using Ambari. I want to install Hadoop 2.9.0 using Ambari web installation. My Ambari version is 2.7.0.0 and I am using HDP 3.0 which has Hadoop 3.1.0. But I need to install Hadoop 2.9.0. Can someone please let me know if this can be done? And how can this be achieved?
I have not started the cluster installation yet and I'm done with Ambari installation.
Ambari uses pre-defined software stacks.
HDP does not offer any stack with Hadoop 2.9.0
You would therefore need to manually install that version of Hadoop yourself, although you can still manage the servers (but not the Hadoop configuration) using Ambari
In any case, there's little benefit to installing a lower version of the software, plus you won't get Hortonworks support if you do that
I'm trying to install Scala IDE 4.7 in my Cloudera VM 5.10, Which is preconfigured with JDK 1.7, Spark 1.6 version respectively.
So, I have installed jdk version 1.8 in /opt/ location by uninstalling the default JDK located at /usr/java/jdk1.7 given by Cloudera. I added the java environment variables in .bash_profile. I was successfully able to install Scala IDE.
But now all the ecosystem of Hadoop are pointing towards old JDK 1.7 which I have uninstalled and throws an error when running. Can anyone let me know where I can config java variables for Hadoop ecosystem to work with new 1.8 JDK?
Here is the screen shot for reference
Step 1: Stop all the hadoop servers
Step 2: edit bigtop-utils file.
I've added the location of jdk 1.8.
Create a variable BIGTOP_MAJOR = 8.
the order of preference for it to choose the jdk is 6,7,8 and open JDK.
Step 3: save and reboot.
I have an installed CDH cluster and used hadoop version, but it returns only with Hadoop version. Is there any way to get maybe all installed components version number on a graphical interface? Which command can get for example Spark version number?
Open CM (hostname:portnumber) -> Hosts tab -> Host Inspector to find what version of
CM and CDH is installed across all hosts in the cluster, as well as installed cdh components list with version details
Spark version can checked in using
spark-submit --version
Spark was developed separately from Hadoop-hdfs and Hadoop-mapreduce as a standalone tool which can be be used along with Hadoop, as such most of its interfaces are are different from hadoop.
As I can see HDP 2.2 needs Centos 6.5 as an operation system, probably because Ambari needs Centos 6.5. My question is if anyone has installed it on Centos 7. Is there any hard dependencies that will not allow me to complete the installation successfully?
Ambari 2.2+ can be installed successfully and works fine on CentOS 7. Then you can install HDP 2.0+.
As far as I'm aware there are no hard dependencies, per se. However Ambari itself looks at the operating system version, and if its CentOS 7, it'll stop the install.
In order to work around that you'd need to edit Ambari's source code.
Just consult the official Installation Guide for a relevant Ambari version
https://cwiki.apache.org/confluence/display/AMBARI/Install+Ambari+2.2.0+from+Public+Repositories
It's an up-to-date source of OS compatibility information.
Here you can see that Centos 7 is officially supported.