cluster installation stuck at "installation in progress" in cloudera manager - hadoop

I have installed Cloudera manager 5.4 version into ubuntu OS. When I try to install cluster using cloudera manager it got stucked "Installation in progress" with no errors.
I could see only two errors from /var/log/cloudera-scm-server/cloudera-scm-server.log
2015-05-12 19:11:42,715 ERROR main:org.hibernate.engine.jdbc.spi.SqlExceptionHelper: ERROR: relation "cm_version" does not exist
Position: 21
2015-05-12 19:16:58,585 ERROR main:com.cloudera.server.web.cmf.cloud.EC2MetadataFetcher: Request to EC2 metadata failed: I/O error: The host did not accept the connection within timeout of 2000 ms; nested exception is org.apache.commons.httpclient.ConnectTimeoutException: The host did not accept the connection within timeout of 2000 ms
can someone pls help

Based on your comments looks to be a DNS/ IP loopback issue.
Update your /etc/hosts file like the following:
I updated the /etc/hosts file to look like this and the inspector worked.
> 127.0.0.1 localhost.localdomain localhost
> xx.xx.xx.xx mc.domain.com mc
Substitute xx.xx with your ip & mc.domain.com with appropriate values for you machines hostname. HTH.

Related

Getting error instalingl wget in RHEL EC2 instance

Getting this message when trying to install wget in RHEL EC2 instance. How to resolve this?
Error: Failed to download metadata for repo 'rhui-client-config-server-8': Cannot prepare internal mirrorlist: Curl error (28): Timeout was reached for https://rhui3.ap-south-1.aws.ce.redhat.com/pulp/mirror/protected/rhui-client-config/rhel/server/8/x86_64/os [Connection timed out after 30000 milliseconds]
If you set up SSH for the EC2 instance, try and SSH into the machine, and try
curl -I https://www.google.com
and if it times out, you perhaps can't connect to the internet.
You need to check if the relevant Security Group has an outbound rule allowing connections to the internet:
https://docs.aws.amazon.com/AWSEC2/latest/UserGuide/ec2-security-groups.html?icmpid=docs_ec2_console#security-group-rules
There are a few useful hints in https://access.redhat.com/discussions/4656371 as well, if the issue is related to your subscription.

clickhouse-client get error "Timeout exceeded while reading from socket"

I'm newbie to clickhouse, I'm trying to create a clickhouse database on my unbuntu 18.04 remote server, I follow instruction to install click house from DEB package in this link: https://clickhouse.tech/docs/en/getting_started/install/#from-sources
after that when I run command clickhouse-client it shows something like this :
root#busmap-api-test:~# clickhouse-client
ClickHouse client version 20.3.5.21 (official build)
Connecting to localhost:9000 as user default.
Code: 209. DB::NetException: Timeout exceeded while reading from socket (127.0.0.1:9000)
Can someone help me to figure out what is the problem and how I can solve it?
Thanks,
Follow these steps to resolve the issue:
check that clickhouse-server-service started
service clickhouse-server status
check the server logs to find the possible reason
cat /var/log/clickhouse-server/clickhouse-server.err.log
if occured the error 'Address already in use':
{} <Error> Application: Net Exception: Address already in use: [::1]:9000
{} <Error> Application: Net Exception: Address already in use: 127.0.0.1:9000
need to switch CH-server to any other port by editing tcp_port-param in /etc/clickhouse-server/config.xml-file:
..
<tcp_port>9032</tcp_port>
..
restart CH-server service:
service clickhouse-server restart
and connect this way
clickhouse-client --port 9032
I actually had this problem too but I got it working with the default port.
The setting should be this way if you want to connect remotely and be able to use the loopback from localhost.
<listen_host>::1</listen_host>
<listen_host>0.0.0.0</listen_host>
This allows the loopback method to work (i.e clickhouse-client no args) on localhost to connect through the IPV6 route, and the remote connection (i.e clickhouse-client -h <hostname>) through the IPV4 connection.
My original problem was that i only used <listen_host>0.0.0.0</listen_host> in my config which meant theclickhouse-client no args would not work on localhost. And I could not get both to work by adding <listen_host>127.0.0.1</listen_host>

Hive problems of connection to port 10000

Currently we are running mapr three node cluster where hive is installed and we use it very frequently for analytics and reporting but due to many connection or some other reason hue(UI panel) show error that
"Could not connect to cm:10000"
and after hs2 stop and start(restart) that start working fine what is issue I am really failed to fix it what should I do and how I can fix it
Could not connect to cm:10000 (code THRIFTTRANSPORT): TTransportException('Could not connect to cm:10000',)
its happening very frequently 1 - 2 times in a day .
Please suggest .
I had the same issue connecting to Port 10000. Below is the error.
WARN jdbc.HiveConnection: Failed to connect to localhost:10000
Could not open connection to the HS2 server. Please check the server URI and if the URI is correct, then ask the administrator to check the server status.
Error: Could not open client transport with JDBC Uri: jdbc:hive2://localhost:10000: java.net.ConnectException: Connection refused (Connection refused) (state=08S01,code=0)
Beeline version 3.1.0.3.1.0.0-78 by Apache Hive
This forum has several posts about this error already. Common suggestions are:
Checking hive.log. I am using HDP but do not know the location of
hive.log. I checked the path "/var/log/hive" but files in that folder are
days old.
I checked the port using "netstat -an |grep 10000". It did
not return anything. But I know HiveServer2 is running from the
Ambari interface. Besides, I am able to connect to HiveServer2 via
embedded mode.

DNS incosistent

I am trying to create a MAPR cluster, nut I have DNS inconsistent warning. I have edited my /etc/hosts file as
10.0.0.10 master.aptus.com
10.0.0.20 slave1.aptus.com
10.0.0.30 slave2.aptus.com
These system contribute to establishing cluster. When i execute :
host 10.0.0.10
the output is:
10.0.0.10.in-addr.arpa has no PTR record
And when I execute :
host master.aptus.com
I get the following output :
master.aptus.com has address 128.199.41.186
I tried to run the installation with warning, but installation fails.
MAPR installation fails
Update DNS entry in domain server
Update the DNS server (AD server) with the new ip address for the cluster node which is being migrated. Once it is updated, all the nodes in the cluster will be able to resolve it.
Check the reverse lookup zone and make sure it is the same using dns tool or ssh
Try it let me know if it helps.

Cassandra: target machine actively refused it

I am trying to run Cassandra (CQL Shell) and I am receiving the following error, I have tried all the google responses to existing questions, nothing has fixed it so far.
Connection error: ('Unable to connect to any servers', {'127.0.0.1': error(10061, "Tried connecting to [('127.0.0.1', 9042)]. Last error: No connection could be made because the target machine actively refused it")})
Before installing Apache Cassandra, JDK must be installed.
Can you make sure the IP address is set correctly on your rpc_address setting in your cassandra.yaml file, on your cassandra server.
Also, you need to make sure port 9042 is open and available for incoming traffic (if your IT department is setting up servers, it is possible this port is blocked, unless otherwise specified...)
Hope it helps.
I also faced the same issue , but may be the below 2 way's can help :
Option 1 :
In my case i haven't started the Cassandra Server and was directly trying to connect to Cassandra.
(a) Firstly start the cassandra server via cmd --> \bin>cassandra.bat -f
and then
(b) Try to connect to it's node --> \bin>cqlsh.bat -u cassandra
Option 2:
Try changing the rpc_address in your cassandra.yaml file to eihter 127.0.0.1 instead of localhost
or to 0.0.0.0 instead of localhost
and then again start the server from new CMD.

Resources