I'm have DSE 4.5 cassandra cluster with multi region data center on EC2. So i'm using EC2MultiRegionSnitch which is returning public IP. And I'm trying to create Spark node but logs says "Failed to bind PUBLIC IP:7077. I am sure it is due to EC2MultiRegionSnitch property. I spoke to Amazon guys , as they are not able to help me in binding port with public ip.
Now, I'm not sure which snitch I can use for EC2 multi region data center cluster. (apart from EC2MultiRegionSnitch). So that I'm bind cluster in multi region data center and run Spark.
Can you please suggest ??
The EC2MultiRegionSnitch should be fine for your setup. I suspect the issue is that port 7077 is not opened in your security group in each data center.
Related
I don't see a way to configure the cluster FQDN for On Premise installation.
I create a 6 nodes cluster (each nodes running on a physical server) and I'm only able to contact each node on their own IP instead of contacting the cluster on a "general FQDN". With this model, I'm to be are of which node is up, and which node is down.
Does somebody know how to achieve it, based on the sample configurations files provided with Service Fabric standalone installation package?
You need to add a network load balancer to your infrastructure for that. This will be used to route traffic to healthy nodes.
I am newbie in Ejabberd. I installed two nodes in Ubuntu Amazon EC2 and using Amazon Public DNS ec2-00-00-00-00.compute-1.amazonaws.com.
When I tried to do cluster, I received message with no_ping.
Can someone help me please ?
Newbie w/ etcd/zookeeper type services ...
I'm not quite sure how to handle cluster installation for etcd. Should the service be installed on each client or a group of independent servers? I ask because if I'm on a client, how would I query the cluster? Every tutorial I've read shows a curl command running against localhost.
For etcd cluster installation, you can install the service on independent servers and form a cluster. The cluster information can be queried by logging onto one of the machines and running curl or remotely by specifying the IP address of one of the cluster member node.
For more information on how to set it up, follow this article
I have installed a single node haodoop cluster on using Hortonworks/Ambari on Amazon's ec2 host.
Since I don't want this cluster running 24/7, I stop the instance when done. When I reboot the instance later, I get a new IP address and then ambari no longer is able to start the Hadoop related services.
Is there a way other than completely redeploying to reconfigure the cluster so the services will start?
It looks like the IP address lives in various xml files under /etc, in the postgres database table ambari, and possibly other places I haven't found yet.
I tried updating the xml files and postgres database with updated versions of the ip address, internal and external dns names as I could find them, but to no avail. I have not been able to restart the services.
The reason I am doing this is to possibly save the deployment time and data configuration on hdfs and other project specific setup each time I restart the host.
Any suggestions?
Thanks!
Elastic IP can be used. Also, since you mentioned it being a single node cluster - you can use localhost or private IP.
If you use elastic IP, your UIs will always be on the same public IP. However, if you use private IP or localhost and do not associate your instance with an elastic IP you will have to look for public IP everytime you start the instance and then connect to the web UI using the IP.
Thanks for the help, both Harman and TJ are correct. I haven't used an elastic IP because I might have more than one of these running and a time, and for now at least, I don't mind looking up the public ip address.
Harman's suggestion of using "localhost" as the fqdn when setting up ambari in the first place is a really good idea in retrospect. Unless I go through the whole setup again, that's water under the bridge for me, but I recommend this to others who might read this post.
In my case, I figured this out on my own before coming back to the page. The specific step I took was insanely simple after all, thanks to Occam's Razor.
I added the following line in /etc/hosts:
<new internal IP> <old internal dns name>
and then did
ambari-server restart. from the command line. Then I am able to restart all services after logging into ambari.
I am trying to setup Apache Traffic Server cluster on two Amazon EC2 instances.
I followed the steps from http://docs.trafficserver.apache.org/en/latest/admin/cluster-howto.en.html to set up the cluster. But when I give the following command
traffic_line -r proxy.process.cluster.nodes still gives me value 1. But I am expecting value 2. I want to know whether is it possible to set up Apache Traffic Server cluster on EC2 instances? If possible could anyone let me know what are the other steps that need to be considered apart from the steps mentioned in above link.