How to deploy ambari for an existing hadoop cluster - hadoop

As I mention in this title, can I skip the step of install hadoop cluster for that cluster already exist and which in service?

Ambari relies on 'Stack' definitions to describe what services the Hadoop cluster consists of. Hortonworks defined a custom Ambari stack, its called HDP.
You could define your own stack and use any services and respective versions that you wanted. See the ambari wiki for more information about defining stacks and services.
That being said, I don't think it's possible to use your pre-existing installation of Hadoop with Ambari. Ambari is used to provision and manage hadoop clusters. It keeps track of the state of each of its stacks services, and the states of each services components. Since your cluster is already provisioned it would be difficult (maybe impossible) to add it to an Ambari instance.

One of the minimum requierments of installing Ambari is removing the pre-existing installations of tools mentioned here.It is not mentioned to remove any pre-existing hadoop installation.

Related

What's the order to start up hdp services manually?

I face some problems to launch hortonworks services through Ambari by starting all services, So I decide to start those services manually and I'm not sure if there is a order I should respect when starting those services. I've installed almost all sevices that we could find on hortonworks data platform.
To start hortonworks data platform services manually through Ambari, there is a order to respect, the following link displays the list of the most frequent services we can use on HDP :
Ranger
Knox
ZooKeeper
HDFS
YARN
HBase
Hive Metastore
HiveServer2
WebHCat
Oozie
Hue
Storm
Kafka
Atlas
To be precise, Ambari starts services by following the Role command order definition files. These files may be defined per-service, or once for the entire stack.
So you may take a look at role_command_order.json files in your stack. For example, here is the role_command_order.json file for HDP-2.0.6 stack.
If the role_command_order.json file is missing, then it is inherited from some other stack. For example the <extends> tag here means that HDP-2.6 stack extends HDP-2.5 stack. Basically, all HDP-2.x stacks inherit
the role_command_order.json file for HDP-2.0.6 stack.

Apache ambari installation

I'm trying to install ambari server + agents.
I have a doubt regarding ambari.
I tried to install ambari.
It always gets link with hortonwork
My doubt is that I have hadoop cluster of my own in Ubunu 16.0.Will ambari only work with HDP or is it possible to also make it work with custom built clusters?
Or if possible please share me detailed descriptive documentation
It's not clear where you downloaded Ambari from, but it sounds like you used the Hortonworks version of it. Not directly from https://ambari.apache.org
Ambari works with the concept of stacks. Each stack has a set of services and components. HDP is such a stack, but there are others, or you can even define your own, so yes, you can manage your own Hadoop installation components, but that really would be not much different from what Hortonworks already provides.
Besides, the HDP services and components have been tested to work together more throughly than off the shelf Hadoop installation.
If you don't want HDP components, there is also the Apache Bigtop project that provides installation packs for many Hadoop related services
Ambari expects Java and Hadoop to be installed in a certain way. I'm not sure how easy it is to setup for an existing Hadoop install.

Can I setup plain hadoop not HDP with apache ambari

I have seen many tutorials for Ambari, all uses HDP stack from hortonworks so I want to know I can setup plain hadoop cluster without HDP.
Can I setup Hadoop using Ambari without HDP?
To set up a 'plain' Hadoop cluster, you will have to define your own stack definition to work with Ambari.
Documentation on how to setup your own stack and services can be found here here.
Additional Note: There are efforts from ODPi to make management of custom stacks and services through Ambari standardized.

How to intergrate hadoop using ambari without HDP?

I have a hadoop cluster with apache hadoop 2.0.7.
I want to know how to integrate Ambari with the apache hadoop without the HDP(HortonWorks).
Actually, If I use HDP the solution is easy. but , I don't want to use the in my situation.
Do you have an any Idea?
Ambari relies on 'Stack' definitions to describe what services the Hadoop cluster consists of. Hortonworks defined a custom Ambari stack, its called HDP.
You could define your own stack and use any services and respective versions that you wanted. See the ambari wiki for more information about defining stacks and services.
That being said, I don't think it's possible to use your pre-existing installation of Hadoop with Ambari. Ambari is used to provision and manage hadoop clusters. It keeps track of the state of each of its stacks services, and the states of each services components. Since your cluster is already provisioned it would be difficult (maybe impossible) to add it to an Ambari instance.

Existing Cluster monitoring by Hortonworks Ambari

I have a 10 node existing cluster in RHEL 6.6 which was prepared by plain apache Hadoop configuration XMLs. Now I wanted to check the cluster status by Ambari. Would it be possible to install Hortonworks Ambari just to monitor only not to install Hadoop.
No, Ambari must provision the cluster it's monitoring.
Ambari is designed around a Stack concept where each stack consists of several services. A stack definition is what allows Ambari to install, manage and monitor the services in the cluster.
In order for you to use Ambari with the hadoop core that you built you would have to provide your own Ambari stack definition.
Specifically in your case your existing Hadoop installation would not have the necessary alert.json descriptors used by Ambari to provide alerts for any given service.

Resources