How to reinstall crashed ambari - hortonworks-data-platform

I installed Ambari and Hortonworks on Centos 7.
I had many problems with the Kerberos configuration, so I tried to reinstall Ambari.
I followed this link, but I always get errors in the Install and test step.
So what is the best way to reinstall ambari server?

Errors during "Install&Test" step are usually caused by conflicts with already installed HDP packages.
Did you try to follow this link? http://web.archive.org/web/20170816163504/http://www.yourtechchick.com/hadoop/how-to-completely-remove-and-uninstall-hdp-components-hadoop-uninstall-on-linux-system/

Related

Install specific version of hadoop using Ambari

I am new to HDP installation using Ambari. I want to install Hadoop 2.9.0 using Ambari web installation. My Ambari version is 2.7.0.0 and I am using HDP 3.0 which has Hadoop 3.1.0. But I need to install Hadoop 2.9.0. Can someone please let me know if this can be done? And how can this be achieved?
I have not started the cluster installation yet and I'm done with Ambari installation.
Ambari uses pre-defined software stacks.
HDP does not offer any stack with Hadoop 2.9.0
You would therefore need to manually install that version of Hadoop yourself, although you can still manage the servers (but not the Hadoop configuration) using Ambari
In any case, there's little benefit to installing a lower version of the software, plus you won't get Hortonworks support if you do that

How to Install Apache Ambari on MacOS?

Is there any other alternative procedure of installing either from source as described in below link?
https://cwiki.apache.org/confluence/display/AMBARI/Installation+Guide+for+Ambari+2.5.2
or
installing HDP version of Ambari.
Do I have any other alternative source to easy install Ambari eg., like from binary files?
You can build and run Apache Ambari on OSX from source (via the doc link you provided).
There are no binary files that are pre-built and offered by Ambari for OSX.
Even if you get Ambari up and running on OSX you will NOT be able to provision an HDP stack on OSX. This is because the services defined by the HDP stack only support linux or windows. Yum, a package manager utility, is required to install the services (ie. Yarn, Spark, etc.) on linux. Yum is not available on osx.
You would have to write a custom stack for it to work on osx. That would be a good deal of work.
That being said there is nothing stopping you from installing services via source or pre-packaged bins from the various apache repos on OSX. This will allow you to run Hadoop processes locally on OSX but without the niceties provided by Ambari for provisioning, managing, and monitoring.

Orange3-spark addon installed but no widgets inside

I want to install the Orange3-spark addon,
I checked the requirements and it was (Apache Spark, Pandas, and orange3),
Problem Description
I installed it according to the order below, added the spark addon (GUI approach), no error messages, but there were no widgets shown in the Spark ML section.. The Spark ML section is empty,
Installation Steps
Installed Apache Spark 2.1.1 with Hadoop 2.7 on a Windows 10 machine.
Scala 2.11.8 (comes with Spark)
Checked Spark using (spark-shell) in command prompt
Installed Anaconda 4.4.0 Python 3.6 version
Verified that Pandas is installed within Anaconda
Installed Orange version 3.4.4
Installed Spark Addon (GUI way from orange)
Can you please instruct me on what to do ?
I have never used Python before, I know the job of most of the above-mentioned components, however, this is the first time that I install any of the above. So please bear with me and be clear with your comments ;))
I had a similar problem with the network add on in Orange3 and overcame it by doing a terminal install with root access via sudo instead of the canvas gui install. Try a command line install running as administrator via runas.

Install HDP 2.2 on CentOS 7

As I can see HDP 2.2 needs Centos 6.5 as an operation system, probably because Ambari needs Centos 6.5. My question is if anyone has installed it on Centos 7. Is there any hard dependencies that will not allow me to complete the installation successfully?
Ambari 2.2+ can be installed successfully and works fine on CentOS 7. Then you can install HDP 2.0+.
As far as I'm aware there are no hard dependencies, per se. However Ambari itself looks at the operating system version, and if its CentOS 7, it'll stop the install.
In order to work around that you'd need to edit Ambari's source code.
Just consult the official Installation Guide for a relevant Ambari version
https://cwiki.apache.org/confluence/display/AMBARI/Install+Ambari+2.2.0+from+Public+Repositories
It's an up-to-date source of OS compatibility information.
Here you can see that Centos 7 is officially supported.

accumulo, zookeeper hadoop Installation instructions, downloads and versions for CENTOS 6

I would appreciate guidance on accumulo, zookeeper hadoop Installation instructions, downloads and versions for CENTOS 6.
Thanks,
Chris
You can do the installation via cloudera manager version 5. I recently installed accumulo using the same.
Here is the link for Cloudera manager 5.
You can use this Youtube video as reference.

Resources