HAWQ service check Fails From Ambari - hadoop

We have small Hadoop-Hawq Cluster. In that, once of slave host service fails When I am executing the check of HAWQ services.
Getting following error From Ambari UI:
**stderr:**
Traceback (most recent call last):
File "/var/lib/ambari-agent/cache/stacks/PHD/3.0/services/HAWQ/package/scripts/service_check.py", line 9, in <module>
HAWQServiceCheck().execute()
File "/usr/lib/python2.6/site-packages/resource_management/libraries/script/script.py", line 123, in execute
method(env)
File "/var/lib/ambari-agent/cache/stacks/PHD/3.0/services/HAWQ/package/scripts/service_check.py", line 6, in service_check
hawq.verify_segments_state(env)
File "/var/lib/ambari-agent/cache/stacks/PHD/3.0/services/HAWQ/package/scripts/hawq.py", line 20, in verify_segments_state
raise Exception("gpstate command returned non-zero result: {0}. Out: {1} Error: {2}".format(retcode, out, err))
Exception: gpstate command returned non-zero result: 255. Out: Error: Permission denied, please try again.
Permission denied, please try again.
Permission denied (publickey,gssapi-keyex,gssapi-with-mic,password).
**stdout:**
(255, '', 'Permission denied, please try again.\r\nPermission denied, please try again.\r\nPermission denied (publickey,gssapi-keyex,gssapi-with-mic,password).\r\n')
Any help on it would be much appreciated ?

HAWQ requires passwordless ssh access to all cluster nodes. Check if the system is configured to allow that, and execute the following hawq command to set up passwordless ssh on the nodes defined in your hostfile:
$ gpssh-exkeys -f hostfile (in version 1.x)
$ hawq ssh-exkeys -f hostfile (in version 2.x)

Related

"[Errno 13] Permission denied " while using dbt

I am new to using dbt. I am trying to set up dbt on my windows 10 machine for local development purposes which later I will send the changes to the production environment. When I run dbt debug. The connection to Bigquery is not successful. If I check the logs am getting the following error
BigQuery adapter: Got an error when attempting to create a bigquery client: '[Errno 13] Permission denied: 'C:\\Users\\path\\to\\my dbt installation''
I have tried to run cmd as administartor as suggested on some other post but am still getting this error.
Please assit.Thanks

greenplum initialization failed

When I tried to initialize Greenplum I got the following error
20180408:23:21:02:017614 gpstop:datanode3:root-[INFO]:-Starting gpstop with args:
20180408:23:21:02:017614 gpstop:datanode3:root-[INFO]:-Gathering information and validating the environment...
20180408:23:21:02:017614 gpstop:datanode3:root-[ERROR]:-gpstop error: postmaster.pid file does not exist. is Greenplum instance already stopped?
also when i tried to check gpstate command i got the following error
20180408:23:21:48:017711 gpstate:datanode3:root-[INFO]:-Starting gpstate with args:
20180408:23:21:48:017711 gpstate:datanode3:root-[INFO]:-local Greenplum Version: 'postgres (Greenplum Database) 5.7.0 build f7c6eb5-oss'
20180408:23:21:48:017711 gpstate:datanode3:root-[CRITICAL]:-gpstate failed. (Reason='could not connect to server: Connection refused
I also did the configuration an add a permission on PostgreSQL.conf, but the same issue
You have pasted the output of gpstop.
gpstop error: postmaster.pid file does not exist. is Greenplum
instance already stopped?
Which means that the database is not running.

Oracle Endeca (Commerce) forge error : Failed to create the directory C:: Permission denied at -e line 1

When trying to run baseline update in Oracle Commerce 11.3.0 I get the following error :
Failed to create the directory C:: Permission denied at -e line 1.
in the mkpath_repository log in endeca Platform services.The component Forge fails without any logs in the directory .\logs\forges\Forge . I have run by both endeca user as well as administrator which are both the owner of [appdir]. How do I fix this?
Thanks in advance.
If you haven't solved this yet, from your error message it sounds like it's trying to create the C directory. When the error message presents itself it should print out the path of the directory it's trying to create.
Example from Oracle Support: "Failed to remove directory D:\Endeca\apps\\config\script.....\data\complete_index_config: Permission denied at -e line 1."
Sounds like you might have a variable configured incorrectly somewhere.

Confirm host fails for Single node Cluster while setting up cluster on Ambari

I am trying to setup Ambari on single node cluster.
Ambari setup was done as root user
I tried all the post related to this , change the permission and did set up as permission
http://docs.hortonworks.com/HDPDocuments/Ambari-2.1.2.1/bk_Installing_HDP_AMB/content/_set_up_password-less_ssh.html
cd ~/.ssh
rm -rf /root/.ssh
ssh-keygen -t dsa
cat /root/.ssh/id_dsa.pub >> /root/.ssh/authorized_keys
cat /root/.ssh/authorized_keys
Copied the the Key from above line in Ambari while setting up cluster Step
ambari-server restart
When I try to Register and Confirm in lInstall Options I get below error
However I am able to do "ssh root#hadoop.maxsjohn.com without giving the password.
==========================
Creating target directory...
==========================
Command start time 2017-03-13 03:35:43
Permission denied (publickey,gssapi-keyex,gssapi-with-mic,password).
SSH command execution finished
host=hadoop.maxsjohn.com, exitcode=255
Command end time 2017-03-13 03:35:43
ERROR: Bootstrap of host hadoop.maxsjohn.com fails because previous action finished with non-zero exit code (255)
ERROR MESSAGE: Permission denied (publickey,gssapi-keyex,gssapi-with-mic,password).
STDOUT:
Permission denied (publickey,gssapi-keyex,gssapi-with-mic,password).[Error Message][1]
So, coming in a year later I got a very similar error but with a multiple host cluster. In case it helps, I found this error happens for the host running Ambari Server when the private key file chosen on the 'Install Options' page in the 'Cluster Install Wizard' is incorrect (in my case I re-created the keys but neglected to update Ambari). From the host OS perspective the passwordless SSH works just fine but Ambari fails to install the host until the corresponding SSH Private Key file is chosen.
I suspect the password cannot be blank. You need to set a password. If this is for your learning, i would suggest take a copy of VM from hortonworks site and use it. You don't have to go through the pain of installing and configuring. Here is the link

Cloudera Host Installation failure: Failed to detect Cloudera Manager Server

I'm trying to put a host on a hadoop cluster using the Cloudera Manager.
The two computers I am using for this are the following:
10.10.10.9 is supposed to be a DataNode and my first host
10.10.10.10 has the Cloudera Manager and will be the NameNode
The manager is having trouble with the "Cluster Installation" part of the "Add Hosts to the Cluster" scenario on the GUI.
I get the following error when the manager tries to detect the Cloudera Manager Server:
BEGIN host -t PTR 10.10.10.10
10.10.10.10.in-addr.arpa domain name pointer stardestroyer.riis.local.
END (0)
using stardestroyer.riis.local as scm server hostname
BEGIN which python
/usr/bin/python
END (0)
BEGIN python -c 'import socket; import sys; s = socket.socket(socket.AF_INET); s.settimeout(5.0); s.connect((sys.argv[1], int(sys.argv[2]))); s.close();' stardestroyer.riis.local 7182
Traceback (most recent call last):
File "<string>", line 1, in <module>
File "/usr/lib/python2.7/socket.py", line 224, in meth
return getattr(self._sock,name)(*args)
socket.gaierror: [Errno -2] Name or service not known
END (1)
could not contact scm server at stardestroyer.riis.local:7182, giving up
waiting for rollback request
I tried to do as Cheloute instructs in the following link, but it didn't seem to fix my issue. I also had a different error than the poster.
Cloudera Manager. Failed to detect Cloudera Manager Server
If the following is used on the commandline, there's no error.
python -c 'import socket; import sys; s = socket.socket(socket.AF_INET); s.settimeout(5.0); s.connect((sys.argv[1], int(sys.argv[2]))); s.close();' 10.10.10.10 7182
I'm not really sure how to fix this in the Cloudera Manager, though.
I found that I could have my system administrator delete my reverse DNS entry and the manager ended up working. I don't know if there is a better solution (there likely is), but this is the one I came up with.

Resources