We are in process of building active-passive cluster via DRBD installed in Centos-7.4 which running kernel-3.10.0-862.el7. While creating cluster with drbadm is unable to create a volume and giving below error. Can you please help me out.
open(/dev/vdb) failed: Invalid argument
could not open with O_DIRECT, retrying without
'/dev/vdb' is not a block device!
open(/dev/vdb) failed: Invalid argument
could not open with O_DIRECT, retrying without
'/dev/vdb' is not a block device!
Command 'drbdmeta 0 v08 /dev/vdb internal create-md' terminated with exit code 20
Related
I'm setting up Azure SQL Database on the local machine (Windows 11) using Azure Data Studio.
I followed the below article to create an Azure SQL Edge instance:
https://learn.microsoft.com/en-us/azure/azure-sql/database/local-dev-experience-quickstart?view=azuresql
And after publishing (i.e after step 11 in above article) I'm getting the below error logs:
Waiting for 2 seconds before another attempt for operation 'Validating the docker container'
Running operation 'Validating the docker container' Attempt 0 of 10
> docker ps -q -a --filter label=source=sqldbproject-choicemls -q
stdout: 142c44a8b420
stdout:
>>> docker ps -q -a --filter label=source=sqldbproject-choicemls -q … exited with code: 0
Operation 'Validating the docker container' completed successfully. Result: 142c44a8b420
Docker created id: '142c44a8b420
'
Waiting for 10 seconds before another attempt for operation 'Connecting to SQL Server'
Running operation 'Connecting to SQL Server' Attempt 0 of 3
Operation 'Connecting to SQL Server' failed. Re-trying... Current Result: undefined. Error: 'Connection failed error: 'A connection was successfully established with the server, but then an error occurred during the pre-login handshake. (provider: TCP Provider, error: 0 - An existing connection was forcibly closed by the remote host.)''
Waiting for 10 seconds before another attempt for operation 'Connecting to SQL Server'
Running operation 'Connecting to SQL Server' Attempt 1 of 3
Operation 'Connecting to SQL Server' failed. Re-trying... Current Result: undefined. Error: 'Connection failed error: 'A connection was successfully established with the server, but then an error occurred during the pre-login handshake. (provider: TCP Provider, error: 0 - An existing connection was forcibly closed by the remote host.)''
Waiting for 10 seconds before another attempt for operation 'Connecting to SQL Server'
Running operation 'Connecting to SQL Server' Attempt 2 of 3
Operation 'Connecting to SQL Server' failed. Re-trying... Current Result: undefined. Error: 'Connection failed error: 'A connection was successfully established with the server, but then an error occurred during the pre-login handshake. (provider: TCP Provider, error: 0 - An existing connection was forcibly closed by the remote host.)''
Please give suggestions to solve this issue.
Thanks,
Saurabh
When I tried to initialize Greenplum I got the following error
20180408:23:21:02:017614 gpstop:datanode3:root-[INFO]:-Starting gpstop with args:
20180408:23:21:02:017614 gpstop:datanode3:root-[INFO]:-Gathering information and validating the environment...
20180408:23:21:02:017614 gpstop:datanode3:root-[ERROR]:-gpstop error: postmaster.pid file does not exist. is Greenplum instance already stopped?
also when i tried to check gpstate command i got the following error
20180408:23:21:48:017711 gpstate:datanode3:root-[INFO]:-Starting gpstate with args:
20180408:23:21:48:017711 gpstate:datanode3:root-[INFO]:-local Greenplum Version: 'postgres (Greenplum Database) 5.7.0 build f7c6eb5-oss'
20180408:23:21:48:017711 gpstate:datanode3:root-[CRITICAL]:-gpstate failed. (Reason='could not connect to server: Connection refused
I also did the configuration an add a permission on PostgreSQL.conf, but the same issue
You have pasted the output of gpstop.
gpstop error: postmaster.pid file does not exist. is Greenplum
instance already stopped?
Which means that the database is not running.
I have two machines both with MS MPI 7.1 installed, one called SERVER and one called COMPUTE.
The machines are set up on LAN in a simple windows workgroup (No DA), and both have an account with the same name and password.
Both are running the MSMPILaunchSvc service.
Both machines can execute MPI jobs locally, verified by testing with the hostname command
SERVER> mpiexec -hosts 1 SERVER 1 hostname
SERVER
or
COMPUTE> mpiexec -hosts 1 COMPUTE 1 hostname
COMPUTE
in a terminal on the machines themselves.
I have disabled the firewall on both machines to make things easier.
My problem is I can not get MPI to run jobs from SERVER on a remote host:
1: SERVER with MSMPILaunchSvc -> COMPUTE with MSMPILaunchSvc
SERVER> mpiexec -hosts 1 COMPUTE 1 hostname -pwd
ERROR: Failed RpcCliCreateContext error 1722
Aborting: mpiexec on SERVER is unable to connect to the smpd service on COMPUTE:8677
Other MPI error, error stack:
connect failed - The RPC server is unavailable. (errno 1722)
What's even more frustrating here is that only sometimes I get prompted to enter a password. It suggests SERVER\Maarten as the user for COMPUTE, the account I am already logged in as on SERVER and shouldn't exist on COMPUTE (should be COMPUTE\Maarten then?). Nonetheless it also fails:
SERVER>mpiexec -hosts 1 COMPUTE 1 hostname.exe -pwd
Enter Password for SERVER\Maarten:
Save Credentials[y|n]? n
ERROR: Failed to connect to SMPD Manager Instance error 1726
Aborting: mpiexec on SERVER is unable to connect to the
smpd manager on COMPUTE:50915 error 1726
2: COMPUTE with MSMPILaunchSvc -> SERVER with MSMPILaunchSvc
COMPUTE> mpiexec -hosts 1 SERVER 1 hostname -pwd
ERROR: Failed RpcCliCreateContext error 5
Aborting: mpiexec on COMPUTE is unable to connect to the smpd service on SERVER:8677
Other MPI error, error stack:
connect failed - Access is denied. (errno 5)
3: COMPUTE with MSMPILaunchSvc -> SERVER with smpd daemon
Aborting: mpiexec on COMPUTE is unable to connect to the smpd service on SERVER:8677
Other MPI error, error stack:
connect failed - Access is denied. (errno 5)
4: SERVER with MSMPILaunchSvc -> COMPUTE with smpd daemon
ERROR: Failed to connect to SMPD Manager Instance error 1726
Aborting: mpiexec on SERVER is unable to connect to the smpd manager on
COMPUTE:51022 error 1726
Update:
Trying with smpd daemon on both nodes I get this error:
[-1:9796] Authentication completed. Successfully obtained Context for Client.
[-1:9796] version check complete, using PMP version 3.
[-1:9796] create manager process (using smpd daemon credentials)
[-1:9796] smpd reading the port string from the manager
[-1:9848] Launching smpd manager instance.
[-1:9848] created set for manager listener, 376
[-1:9848] smpd manager listening on port 51149
[-1:9796] closing the pipe to the manager
[-1:9848] Authentication completed. Successfully obtained Context for Client.
[-1:9848] Authorization completed.
[-1:9848] version check complete, using PMP version 3.
[-1:9848] Received session header from parent id=1, parent=0, level=0
[01:9848] Connecting back to parent using host SERVER and endpoint 17979
[01:9848] Previous attempt failed with error 5, trying to authenticate without Kerberos
[01:9848] Failed to connect back to parent error 5.
[01:9848] ERROR: Failed to connect back to parent 'ncacn_ip_tcp:SERVER:17979' error 5
[01:9848] smpd manager successfully stopped listening.
[01:9848] SMPD exiting with error code 4294967293.
and on the host:
[-1:12264] Launching SMPD service.
[-1:12264] smpd listening on port 8677
[-1:12264] Authentication completed. Successfully obtained Context for Client.
[-1:12264] version check complete, using PMP version 3.
[-1:12264] create manager process (using smpd daemon credentials)
[-1:12264] smpd reading the port string from the manager
[-1:16668] Launching smpd manager instance.
[-1:16668] created set for manager listener, 364
[-1:16668] smpd manager listening on port 18033
[-1:12264] closing the pipe to the manager
[-1:16668] Authentication completed. Successfully obtained Context for Client.
[-1:16668] Authorization completed.
[-1:16668] version check complete, using PMP version 3.
[-1:16668] Received session header from parent id=1, parent=0, level=0
[01:16668] Connecting back to parent using host SERVER and endpoint 18031
[01:16668] Authentication completed. Successfully obtained Context for Client.
[01:16668] Authorization completed.
[01:16668] handling command SMPD_CONNECT src=0
[01:16668] now connecting to COMPUTE
[01:16668] 1 -> 2 : returning SMPD_CONTEXT_LEFT_CHILD
[01:16668] using spn msmpi/COMPUTE to contact server
[01:16668] SERVER posting a re-connect to COMPUTE:51161 in left child context.
[01:16668] ERROR: Failed to connect to SMPD Manager Instance error 1726
[01:16668] sending abort command to parent context.
[01:16668] posting command SMPD_ABORT to parent, src=1, dest=0.
[01:16668] ERROR: smpd running on SERVER is unable to connect to smpd service on COMPUTE:8677
[01:16668] Handling cmd=SMPD_ABORT result
[01:16668] cmd=SMPD_ABORT result will be handled locally
[01:16668] parent terminated unexpectedly - initiating cleaning up.
[01:16668] no child processes to kill - exiting with error code -1
I found after trial and error that these and other unspecific errors come up when trying to run MS MPI with different configurations (in my case a mix of HPC Cluster 2008 and HPC Cluster 2012 with MSMPI).
The solution was to downgrade all nodes to Windows Server 2008 R2 with HPC Cluster 2008. Because I dont use AD, I had to fall back to using the SMPD daemon and add firewall rules for it (skipping the cluster management tools alltogether).
Running the command fleetctl load registry#1.service registry-presence#1.service I get the following output:
2015/05/08 10:25:26 WARN fleetctl.go:772: Error retrieving Unit(registry#1.service) from Registry: Get http://domain-sock/fleet/v1/units/registry%401.service?alt=json: forwarding request denied
2015/05/08 10:30:31 WARN fleetctl.go:772: Error retrieving Unit(registry-presence#1.service) from Registry: Get http://domain-sock/fleet/v1/units/registry-presence%401.service?alt=json: forwarding request denied
2015/05/08 10:36:14 WARN fleetctl.go:772: Error retrieving Unit(registry#1.service) from Registry: Get http://domain-sock/fleet/v1/units/registry%401.service?alt=json: ssh: rejected: administratively prohibited (open failed)
2015/05/08 10:42:44 WARN fleetctl.go:772: Error retrieving Unit(registry-presence#1.service) from Registry: Get http://domain-sock/fleet/v1/units/registry-presence%401.service?alt=json: ssh: rejected: administratively prohibited (open failed)
2015/05/08 10:54:46 WARN fleetctl.go:772: Error retrieving Unit(registry#1.service) from Registry: Get http://domain-sock/fleet/v1/units/registry%401.service?alt=json: ssh: rejected: administratively prohibited (open failed)
2015/05/08 10:57:51 WARN fleetctl.go:772: Error retrieving Unit(registry-presence#1.service) from Registry: Get http://domain-sock/fleet/v1/units/registry-presence%401.service?alt=json: ssh: rejected: administratively prohibited (open failed)
2015/05/08 10:58:12 WARN fleetctl.go:772: Error retrieving Unit(registry#1.service) from Registry: Get http://domain-sock/fleet/v1/units/registry%401.service?alt=json: ssh: rejected: administratively prohibited (open failed)
2015/05/08 11:02:43 WARN fleetctl.go:772: Error retrieving Unit(registry-presence#1.service) from Registry: Get http://domain-sock/fleet/v1/units/registry-presence%401.service?alt=json: ssh: rejected: administratively prohibited (open failed)
This just repeats, I've left the command running for over 30 minutes. When I press CTRL-C and run fleetctl list-unit-files I see the following:
UNIT HASH DSTATE STATE TARGET
registry-presence#1.service f54aa0d loaded inactive 0d8d13be.../172.17.8.101
registry#1.service d233714 loaded inactive 0d8d13be.../172.17.8.101
And the output of fleetctl list-units is:
UNIT MACHINE ACTIVE SUB
If I run the load command with -block-attempts=2 it gives the same errors but completes and the output of fleetctl list-unit-files is:
UNIT HASH DSTATE STATE TARGET
registry-presence#1.service f54aa0d loaded loaded 0d8d13be.../172.17.8.101
registry#1.service d233714 loaded loaded 0d8d13be.../172.17.8.101
And the output of fleetctl list-units is:
UNIT MACHINE ACTIVE SUB
registry-presence#1.service 0d8d13be.../172.17.8.101 inactive dead
registry#1.service 0d8d13be.../172.17.8.101 inactive dead
I'm wondering what the WARN logs trying to tell me, what registry is it talking about?
That happens to me also when running fleetctl from my local machine to a distant cluster.
The registry it's talking about is the fleet registry, the units that have been submitted and/or loaded in with fleetctl submit or fleetctl load.
What I usually do in those cases is connecting to one of the nodes, and run my fleetctl start command from there. Then you can run them from your local machine without any more problems.
What I suspect is that for some reason the unit file is not loaded in the whole cluster when you run that command.
I'm having problems installing composer. The error being displayed is:
Connection Error [ERR_CONNECTION]: Unable to connect to getcomposer.org
Proxy http####://10.50.7.154:3128### [from Internet Settings] failed with errors:
Failed to open stream: A connection attempt failed because the connected party did not properly respond after a period of time, or established connection failed because connected host has failed to respond.
Request to https://getcomposer.org/installer failed with errors:
Failed to open stream: A connection attempt failed because the connected party did not properly respond after a period of time, or established connection failed because connected host has failed to respond.
That's a proxy problem, right there. Please step away from the building you are in and find a regular coffeeshop with regular broadband WiFi.
Hope this helps.