After creating a new Queue-Manager using the MQ Explorer, it's fails to be started properly and provide the following message:
Command: "C:\Program Files (x86)\IBM\WebSphere MQ\bin\crtmqm" -sa test_qm
WebSphere MQ queue manager created.
Directory 'C:\Program Files (x86)\IBM\WebSphere MQ\qmgrs\test_qm'
created.
The queue manager is associated with installation 'WMQ75Install'.
exitvalue = 2059
I couldn't figure out how to solve it from the logs and tried to start it manually from the MQ Explorer and from the command-line shell as well, but without any success - it just not started.
HERE IS MY AMQERR01.LOG TEXT:
21/01/2015 14:18:46 - Process(7960.3) User(johnsmith) Program(amqzmuc0.exe)
Host(NY0035546) Installation(WMQ75Install)
VRMF(7.5.0.2) QMgr(test_qm)
AMQ6287: WebSphere MQ V7.5.0.2 (p750-002-131001_DE).
EXPLANATION:
WebSphere MQ system information:
Host Info :- Windows 7 Enterprise x64 Edition, Build 7601: SP1 (MQ
Windows 32-bit)
Installation :- C:\Program Files (x86)\IBM\WebSphere MQ (WMQ75Install)
Version :- 7.5.0.2 (p750-002-131001_DE)
ACTION:
None.
21/01/2015 14:18:46 - Process(7960.3) User(johnsmith) Program(amqzmuc0.exe)
Host(NY0035546) Installation(WMQ75Install)
VRMF(7.5.0.2) QMgr(test_qm)
AMQ5051: The queue manager task 'LOGGER-IO' has started.
EXPLANATION:
The critical utility task manager has started the LOGGER-IO task. This task has
now started 1 times.
ACTION:
None.
21/01/2015 14:18:46 - Process(7960.1) User(johnsmith) Program(amqzmuc0.exe)
Host(NY0035546) Installation(WMQ75Install)
VRMF(7.5.0.2) QMgr(test_qm)
AMQ5041: The queue manager task 'LOGGER-IO' has ended.
EXPLANATION:
The queue manager task LOGGER-IO has ended.
ACTION:
None.
21/01/2015 14:18:49 - Process(7528.3) User(johnsmith) Program(amqzmuc0.exe)
Host(NY0035546) Installation(WMQ75Install)
VRMF(7.5.0.2) QMgr(test_qm)
AMQ5051: The queue manager task 'LOGGER-IO' has started.
EXPLANATION:
The critical utility task manager has started the LOGGER-IO task. This task has
now started 1 times.
ACTION:
None.
21/01/2015 14:18:49 - Process(9760.1) User(johnsmith) Program(amqzxma0.exe)
Host(NY0035546) Installation(WMQ75Install)
VRMF(7.5.0.2) QMgr(test_qm)
AMQ7229: 4 log records accessed on queue manager 'test_qm' during the log
replay phase.
EXPLANATION:
4 log records have been accessed so far on queue manager test_qm during the log
replay phase in order to bring the queue manager back to a previously known
state.
ACTION:
None.
21/01/2015 14:18:49 - Process(9760.1) User(johnsmith) Program(amqzxma0.exe)
Host(NY0035546) Installation(WMQ75Install)
VRMF(7.5.0.2) QMgr(test_qm)
AMQ7230: Log replay for queue manager 'test_qm' complete.
EXPLANATION:
The log replay phase of the queue manager restart process has been completed
for queue manager test_qm.
ACTION:
None.
21/01/2015 14:18:49 - Process(9760.1) User(johnsmith) Program(amqzxma0.exe)
Host(NY0035546) Installation(WMQ75Install)
VRMF(7.5.0.2) QMgr(test_qm)
AMQ7231: 0 log records accessed on queue manager 'test_qm' during the recovery
phase.
EXPLANATION:
0 log records have been accessed so far on queue manager test_qm during the
recovery phase of the transactions manager state.
ACTION:
None.
21/01/2015 14:18:49 - Process(9760.1) User(johnsmith) Program(amqzxma0.exe)
Host(NY0035546) Installation(WMQ75Install)
VRMF(7.5.0.2) QMgr(test_qm)
AMQ7232: Transaction manager state recovered for queue manager 'test_qm'.
EXPLANATION:
The state of transactions at the time the queue manager ended has been
recovered for queue manager test_qm.
ACTION:
None.
21/01/2015 14:18:49 - Process(9760.1) User(johnsmith) Program(amqzxma0.exe)
Host(NY0035546) Installation(WMQ75Install)
VRMF(7.5.0.2) QMgr(test_qm)
AMQ7233: 0 out of 0 in-flight transactions resolved for queue manager
'test_qm'.
EXPLANATION:
0 transactions out of 0 in-flight at the time queue manager test_qm ended have
been resolved.
ACTION:
None.
21/01/2015 14:18:49 - Process(7528.4) User(johnsmith) Program(amqzmuc0.exe)
Host(NY0035546) Installation(WMQ75Install)
VRMF(7.5.0.2) QMgr(test_qm)
AMQ5051: The queue manager task 'CHECKPOINT' has started.
EXPLANATION:
The critical utility task manager has started the CHECKPOINT task. This task
has now started 1 times.
ACTION:
None.
21/01/2015 14:18:51 - Process(9796.3) User(johnsmith) Program(amqzmur0.exe)
Host(NY0035546) Installation(WMQ75Install)
VRMF(7.5.0.2) QMgr(test_qm)
AMQ5037: The queue manager task 'ERROR-LOG' has started.
EXPLANATION:
The restartable utility task manager has started the ERROR-LOG task. This task
has now started 1 times.
ACTION:
None.
21/01/2015 14:18:51 - Process(9796.4) User(johnsmith) Program(amqzmur0.exe)
Host(NY0035546) Installation(WMQ75Install)
VRMF(7.5.0.2) QMgr(test_qm)
AMQ5037: The queue manager task 'APP-SIGNAL' has started.
EXPLANATION:
The restartable utility task manager has started the APP-SIGNAL task. This task
has now started 1 times.
ACTION:
None.
21/01/2015 14:18:51 - Process(9796.5) User(johnsmith) Program(amqzmur0.exe)
Host(NY0035546) Installation(WMQ75Install)
VRMF(7.5.0.2) QMgr(test_qm)
AMQ5037: The queue manager task 'APP-SIGNAL' has started.
EXPLANATION:
The restartable utility task manager has started the APP-SIGNAL task. This task
has now started 2 times.
ACTION:
None.
21/01/2015 14:18:51 - Process(9796.7) User(johnsmith) Program(amqzmur0.exe)
Host(NY0035546) Installation(WMQ75Install)
VRMF(7.5.0.2) QMgr(test_qm)
AMQ5037: The queue manager task 'APP-SIGNAL' has started.
EXPLANATION:
The restartable utility task manager has started the APP-SIGNAL task. This task
has now started 4 times.
ACTION:
None.
21/01/2015 14:18:51 - Process(9796.6) User(johnsmith) Program(amqzmur0.exe)
Host(NY0035546) Installation(WMQ75Install)
VRMF(7.5.0.2) QMgr(test_qm)
AMQ5037: The queue manager task 'APP-SIGNAL' has started.
EXPLANATION:
The restartable utility task manager has started the APP-SIGNAL task. This task
has now started 3 times.
ACTION:
None.
21/01/2015 14:18:52 - Process(10328.1) User(johnsmith) Program(amqzfuma.exe)
Host(NY0035546) Installation(WMQ75Install)
VRMF(7.5.0.2) QMgr(test_qm)
AMQ8077: Entity 'johnsmith#intranet' has insufficient authority to access object
'test_qm'.
EXPLANATION:
The specified entity is not authorized to access the required object. The
following requested permissions are unauthorized: connect/system
ACTION:
Ensure that the correct level of authority has been set for this entity against
the required object, or ensure that the entity is a member of a privileged
group.
----- amqzfubn.c : 515 --------------------------------------------------------
21/01/2015 14:18:52 - Process(9760.1) User(johnsmith) Program(amqzxma0.exe)
Host(NY0035546) Installation(WMQ75Install)
VRMF(7.5.0.2) QMgr(test_qm)
AMQ5525: The WebSphere MQ Object Authority Manager has failed.
EXPLANATION:
The Object Authority Manager has failed to complete an MQ request.
ACTION:
Check the queue manager error logs for messages explaining the failure and try
to correct the problem accordingly.
----- amqzxma0.c : 3825 -------------------------------------------------------
21/01/2015 14:18:52 - Process(9760.1) User(johnsmith) Program(amqzxma0.exe)
Host(NY0035546) Installation(WMQ75Install)
VRMF(7.5.0.2) QMgr(test_qm)
AMQ8003: WebSphere MQ queue manager 'test_qm' started using V7.5.0.2.
EXPLANATION:
WebSphere MQ queue manager 'test_qm' started using V7.5.0.2.
ACTION:
None.
21/01/2015 14:18:52 - Process(9760.1) User(johnsmith) Program(amqzxma0.exe)
Host(NY0035546) Installation(WMQ75Install)
VRMF(7.5.0.2) QMgr(test_qm)
AMQ5008: An essential WebSphere MQ process 10328 (amqzfuma.exe) cannot be found
and is assumed to be terminated.
EXPLANATION:
1) A user has inadvertently terminated the process. 2) The system is low on
resources. Some operating systems terminate processes to free resources. If
your system is low on resources, it is possible it has terminated the process
so that a new process can be created.
ACTION:
WebSphere MQ will stop all MQ processes. Inform your systems administrator.
When the problem is rectified WebSphere MQ can be restarted.
----- amqzxmb0.c : 9956 -------------------------------------------------------
21/01/2015 14:18:52 - Process(9760.1) User(johnsmith) Program(amqzxma0.exe)
Host(NY0035546) Installation(WMQ75Install)
VRMF(7.5.0.2) QMgr(test_qm)
AMQ5050: An essential WebSphere MQ process 9188 (zllPUBSUB) cannot be found and
is assumed to be terminated.
EXPLANATION:
1) A user has inadvertently terminated the process. 2) The system is low on
resources. Some operating systems terminate processes to free resources. If
your system is low on resources, it is possible it has terminated the process
so that a new process can be created. 3) MQ has encountered an unexpected
error. Check for possible errors reported in the MQ error logs and for any
FFSTs that have been generated.
ACTION:
WebSphere MQ will attempt to restart the terminated process.
----- amqzxmb0.c : 9679 -------------------------------------------------------
21/01/2015 14:18:53 - Process(9760.1) User(johnsmith) Program(amqzxma0.exe)
Host(NY0035546) Installation(WMQ75Install)
VRMF(7.5.0.2) QMgr(test_qm)
AMQ8004: WebSphere MQ queue manager 'test_qm' ended.
EXPLANATION:
WebSphere MQ queue manager 'test_qm' ended.
ACTION:
None.
At a guess, is johnsmith a domain id? Is this a domain workstation (what is 'johnsmith#intranet')? My suspicion is you have a machine in a domain but you have not configured MQ to run on a machine in a domain. I think the issue is that at startup MQ is trying to determine what groups the userid 'johnsmith#intranet' is a member of, and hence fails. Its possible, adding johnsmith to the local mqm group may get you futher although my suspicion is you need to do the domain configuration. See another answer for details of what to do:
Issue with permission grant to domain users in IBM web sphere queue manager
First, make sure your UserID is part of the 'mqm' group and second reboot your PC (wonderful things happen when you reboot Windows!).
Related
I followed the article - https://github.com/nuviosoftware/custom-ibm-mq and deployed the container using the below commands in my windows laptop using docker for desktop
docker build -t nuviosoftware-mq-local .
docker volume create qm1data
and executed
docker run --env LICENSE=accept --env MQ_QMGR_NAME=QM1 --volume qm1data:/mnt/mqm --publish 1414:1414 --publish 9443:9443 --env MQ_APP_PASSWORD=passw0rd nuviosoftware-mq-local
below is the container log
WARNING: The requested image's platform (linux/amd64) does not match the detected host platform (windows/amd64) and no specific platform was requested
2021-07-15T08:58:20.659Z CPU architecture: amd64
2021-07-15T08:58:20.659Z Linux kernel version: 4.19.27-linuxkit
2021-07-15T08:58:20.661Z Base image: Red Hat Enterprise Linux 8.3 (Ootpa)
2021-07-15T08:58:20.661Z Running as user ID 0 with primary group 0
2021-07-15T08:58:20.661Z Capabilities (bounding set): chown,dac_override,fowner,fsetid,kill,setgid,setuid,setpcap,net_bind_service,net_raw,sys_chroot,mknod,audit_write,setfcap
2021-07-15T08:58:20.661Z seccomp enforcing mode: disabled
2021-07-15T08:58:20.661Z Process security attributes: none
2021-07-15T08:58:20.661Z Detected '9p' volume mounted to /mnt/mqm
2021-07-15T08:58:25.540Z Using queue manager name: QM1
2021-07-15T08:58:27.316Z Created directory structure under /var/mqm
2021-07-15T08:58:27.316Z Image created: 2021-03-12T09:57:54+00:00
2021-07-15T08:58:27.316Z Image tag: ibm-mqadvanced-server-dev:9.2.2.0-r1.20210312094918.14987c1-amd64
2021-07-15T08:58:27.794Z MQ version: 9.2.2.0
2021-07-15T08:58:27.794Z MQ level: p922-L210310.DE
2021-07-15T08:58:27.794Z MQ license: Developer
2021-07-15T08:58:29.405Z Creating queue manager QM1
2021-07-15T08:58:29.405Z Starting web server
2021-07-15T08:58:29.652Z Detected existing queue manager QM1
2021-07-15T08:58:29.675Z Error 126 starting web server: AMQ5061E: Program 'strmqweb' cannot be run with root privileges.
2021-07-15T08:58:29.675Z Error starting web server: exit status 126
2021-07-15T08:58:29.982Z Removing existing ServiceComponent configuration
2021-07-15T08:58:29.991Z Starting queue manager
2021-07-15T08:58:31.031Z AMQ5657W: The system resource RLIMIT_NOFILE is set at an unusually low level for IBM MQ. [ArithInsert1(1024), ArithInsert2(10240), CommentInsert1(RLIMIT_NOFILE)]
2021-07-15T08:58:31.216Z AMQ5657W: The system resource RLIMIT_NPROC is set at an unusually low level for IBM MQ. [ArithInsert1(3844), ArithInsert2(4096), CommentInsert1(RLIMIT_NPROC)]
2021-07-15T08:58:31.962Z AMQ5775I: Successfully applied automatic configuration INI definitions. [CommentInsert1(INI)]
2021-07-15T08:58:36.368Z Initializing MQ Advanced for Developers custom authentication service
2021-07-15T08:58:39.939Z AMQ5051I: The queue manager task 'LOGGER-IO' has started. [ArithInsert2(1), CommentInsert1(LOGGER-IO)]
2021-07-15T08:58:42.052Z AMQ7229I: 6 log records accessed on queue manager 'QM1' during the log replay phase. [ArithInsert1(6), CommentInsert1(QM1)]
2021-07-15T08:58:42.104Z AMQ7230I: Log replay for queue manager 'QM1' complete. [ArithInsert1(6), CommentInsert1(QM1)]
2021-07-15T08:58:42.244Z AMQ5051I: The queue manager task 'CHECKPOINT' has started. [ArithInsert2(1), CommentInsert1(CHECKPOINT)]
2021-07-15T08:58:42.359Z AMQ7231I: 0 log records accessed on queue manager 'QM1' during the recovery phase. [CommentInsert1(QM1)]
2021-07-15T08:58:42.405Z AMQ7232I: Transaction manager state recovered for queue manager 'QM1'. [CommentInsert1(QM1)]
2021-07-15T08:58:42.542Z AMQ7233I: 0 out of 0 in-flight transactions resolved for queue manager 'QM1'. [CommentInsert1(QM1)]
2021-07-15T08:58:43.559Z AMQ5037I: The queue manager task 'APP-SIGNAL' has started. [ArithInsert2(1), CommentInsert1(APP-SIGNAL)]
2021-07-15T08:58:43.603Z AMQ5037I: The queue manager task 'APP-SIGNAL' has started. [ArithInsert2(2), CommentInsert1(APP-SIGNAL)]
2021-07-15T08:58:43.643Z AMQ5037I: The queue manager task 'APP-SIGNAL' has started. [ArithInsert2(3), CommentInsert1(APP-SIGNAL)]
2021-07-15T08:58:43.689Z AMQ5037I: The queue manager task 'APP-SIGNAL' has started. [ArithInsert2(4), CommentInsert1(APP-SIGNAL)]
2021-07-15T08:58:43.731Z AMQ5037I: The queue manager task 'APP-SIGNAL' has started. [ArithInsert2(5), CommentInsert1(APP-SIGNAL)]
2021-07-15T08:58:43.793Z AMQ5037I: The queue manager task 'ERROR-LOG' has started. [ArithInsert2(1), CommentInsert1(ERROR-LOG)]
2021-07-15T08:58:43.822Z AMQ5037I: The queue manager task 'APP-SIGNAL' has started. [ArithInsert2(6), CommentInsert1(APP-SIGNAL)]
2021-07-15T08:58:43.849Z AMQ5037I: The queue manager task 'APP-SIGNAL' has started. [ArithInsert2(7), CommentInsert1(APP-SIGNAL)]
2021-07-15T08:58:43.896Z AMQ5037I: The queue manager task 'APP-SIGNAL' has started. [ArithInsert2(8), CommentInsert1(APP-SIGNAL)]
2021-07-15T08:58:45.037Z AMQ8003I: IBM MQ queue manager 'QM1' started using V9.2.2.0. [CommentInsert1(9.2.2.0), CommentInsert3(QM1)]
2021-07-15T08:58:45.551Z AMQ5051I: The queue manager task 'DUR-SUBS-MGR' has started. [ArithInsert2(1), CommentInsert1(DUR-SUBS-MGR)]
2021-07-15T08:58:45.559Z AMQ9410I: Repository manager started.
2021-07-15T08:58:45.643Z AMQ5051I: The queue manager task 'TOPIC-TREE' has started. [ArithInsert2(1), CommentInsert1(TOPIC-TREE)]
2021-07-15T08:58:45.689Z AMQ5024I: The command server has started. ProcessId(135). [ArithInsert1(135), CommentInsert1(SYSTEM.CMDSERVER.1)]
2021-07-15T08:58:45.758Z AMQ5022I: The channel initiator has started. ProcessId(137). [ArithInsert1(137), CommentInsert1(SYSTEM.CHANNEL.INITQ)]
2021-07-15T08:58:45.802Z AMQ5051I: The queue manager task 'AUTOCONFIG' has started. [ArithInsert2(1), CommentInsert1(AUTOCONFIG)]
2021-07-15T08:58:45.838Z AMQ5051I: The queue manager task 'IQM-COMMS-MANAGER' has started. [ArithInsert2(1), CommentInsert1(IQM-COMMS-MANAGER)]
2021-07-15T08:58:46.262Z AMQ8942I: Starting to process automatic MQSC configuration script.
2021-07-15T08:58:46.291Z AMQ8024I: IBM MQ channel initiator started. [CommentInsert1(SYSTEM.CHANNEL.INITQ)]
2021-07-15T08:58:59.758Z AMQ8939I: Automatic MQSC configuration script has completed, and contained 24 command(s), of which 0 had errors. [ArithInsert1(24), CommentInsert1(0)]
2021-07-15T08:58:59.809Z AMQ5051I: The queue manager task 'ACTVTRC' has started. [ArithInsert2(1), CommentInsert1(ACTVTRC)]
2021-07-15T08:58:59.818Z AMQ5037I: The queue manager task 'DEFERRED_DELIVERY' has started. [ArithInsert2(1), CommentInsert1(DEFERRED_DELIVERY)]
2021-07-15T08:58:59.831Z AMQ5052I: The queue manager task 'QPUBSUB-CTRLR' has started. [ArithInsert2(1), CommentInsert1(QPUBSUB-CTRLR)]
2021-07-15T08:58:59.840Z AMQ5051I: The queue manager task 'EXPIRER' has started. [ArithInsert2(1), CommentInsert1(EXPIRER)]
2021-07-15T08:58:59.846Z AMQ5037I: The queue manager task 'DEFERRED-MSG' has started. [ArithInsert2(1), CommentInsert1(DEFERRED-MSG)]
2021-07-15T08:58:59.859Z AMQ5052I: The queue manager task 'QPUBSUB-SUBPT-NLCACHE' has started. [ArithInsert2(1), CommentInsert1(QPUBSUB-SUBPT-NLCACHE)]
2021-07-15T08:58:59.867Z AMQ5051I: The queue manager task 'Q-DELETION' has started. [ArithInsert2(1), CommentInsert1(Q-DELETION)]
2021-07-15T08:58:59.879Z AMQ5052I: The queue manager task 'QPUBSUB-QUEUE-NLCACHE' has started. [ArithInsert2(1), CommentInsert1(QPUBSUB-QUEUE-NLCACHE)]
2021-07-15T08:58:59.882Z AMQ5037I: The queue manager task 'STATISTICS' has started. [ArithInsert2(1), CommentInsert1(STATISTICS)]
2021-07-15T08:58:59.892Z AMQ5051I: The queue manager task 'ASYNCQ' has started. [ArithInsert2(1), CommentInsert1(ASYNCQ)]
2021-07-15T08:58:59.915Z AMQ5051I: The queue manager task 'RESOURCE_MONITOR' has started. [ArithInsert2(1), CommentInsert1(RESOURCE_MONITOR)]
2021-07-15T08:58:59.918Z AMQ5037I: The queue manager task 'MARKINTSCAN' has started. [ArithInsert2(1), CommentInsert1(MARKINTSCAN)]
2021-07-15T08:58:59.930Z AMQ5052I: The queue manager task 'PUBSUB-DAEMON' has started. [ArithInsert2(1), CommentInsert1(PUBSUB-DAEMON)]
2021-07-15T08:58:59.930Z AMQ5975I: 'IBM MQ Distributed Pub/Sub Controller' has started. [CommentInsert1(IBM MQ Distributed Pub/Sub Controller)]
2021-07-15T08:58:59.966Z AMQ5026I: The listener 'SYSTEM.LISTENER.TCP.1' has started. ProcessId(174). [ArithInsert1(174), CommentInsert1(SYSTEM.LISTENER.TCP.1)]
2021-07-15T08:58:59.984Z AMQ5051I: The queue manager task 'PRESERVED-Q' has started. [ArithInsert2(1), CommentInsert1(PRESERVED-Q)]
2021-07-15T08:59:00.011Z AMQ5051I: The queue manager task 'MULTICAST' has started. [ArithInsert2(1), CommentInsert1(MULTICAST)]
2021-07-15T08:59:00.142Z Started queue manager
2021-07-15T08:59:00.142Z Metrics are disabled
2021-07-15T08:59:00.225Z AMQ5975I: 'IBM MQ Distributed Pub/Sub Command Task' has started. [CommentInsert1(IBM MQ Distributed Pub/Sub Command Task)]
2021-07-15T08:59:00.279Z AMQ5975I: 'IBM MQ Distributed Pub/Sub Publish Task' has started. [CommentInsert1(IBM MQ Distributed Pub/Sub Publish Task)]
2021-07-15T08:59:00.294Z AMQ5975I: 'IBM MQ Distributed Pub/Sub Fan Out Task' has started. [CommentInsert1(IBM MQ Distributed Pub/Sub Fan Out Task)]
2021-07-15T08:59:00.498Z AMQ5806I: Queued Publish/Subscribe Daemon started for queue manager QM1. [CommentInsert1(QM1)]
2021-07-15T08:59:07.843Z AMQ5041I: The queue manager task 'AUTOCONFIG' has ended. [CommentInsert1(AUTOCONFIG)]
however I could access the web management console
as well as I could not connect with the MQ from the client application
2021-07-15 14:31:40.922 ERROR 4884 --- [enerContainer-1] o.s.j.l.DefaultMessageListenerContainer : Could not refresh JMS Connection for destination 'ORDER.RESPONSE' - retrying using FixedBackOff{interval=5000, currentAttempts=0, maxAttempts=unlimited}. Cause: JMSWMQ2013: The security authentication was not valid that was supplied for QueueManager 'QM1' with connection mode 'Client' and host name 'localhost(1414)'.; nested exception is com.ibm.mq.MQException: JMSCMQ0001: IBM MQ call failed with compcode '2' ('MQCC_FAILED') reason '2035' ('MQRC_NOT_AUTHORIZED').
2021-07-15 14:31:47.368 ERROR 4884 --- [enerContainer-1] o.s.j.l.DefaultMessageListenerContainer : Could not refresh JMS Connection for destination 'ORDER.RESPONSE' - retrying using FixedBackOff{interval=5000, currentAttempts=1, maxAttempts=unlimited}. Cause: JMSWMQ2013: The security authentication was not valid that was supplied for QueueManager 'QM1' with connection mode 'Client' and host name 'localhost(1414)'.; nested exception is com.ibm.mq.MQException: JMSCMQ0001: IBM MQ call failed with compcode '2' ('MQCC_FAILED') reason '2035' ('MQRC_NOT_AUTHORIZED')
try to access it through with https:
https://localhost:9443/ibmmq/console
I was using the windows mode and switched to Linux mod, it worked.
I have AMQ Artemis cluster, shared-store HA (master-slave), 2.17.0.
I noticed that all my clusters (active servers only) that are idle (no one is using them) using from 10% to 20% of CPU, except one, which is using around 1% (totally normal). I started investigating...
Long story short - only one cluster has a completely normal CPU usage. The only difference I've managed to find that if I connect to that normal cluster's master node and attempt telnet slave 61616 - it will show as connected. If I do the same in any other cluster (that has high CPU usage) - it will show as rejected.
In order to better understand what is happening, I enabled DEBUG logs in instance/etc/logging.properties. Here is what master node is spamming:
2021-05-07 13:54:31,857 DEBUG [org.apache.activemq.artemis.core.client.impl.ClientSessionFactoryImpl] Backup is not active, trying original connection configuration now.
2021-05-07 13:54:32,357 DEBUG [org.apache.activemq.artemis.core.client.impl.ClientSessionFactoryImpl] Trying reconnection attempt 0/1
2021-05-07 13:54:32,357 DEBUG [org.apache.activemq.artemis.core.client.impl.ClientSessionFactoryImpl] Trying to connect with connectorFactory = org.apache.activemq.artemis.core.remoting.impl.netty$NettyConnectorFactory#6cf71172, connectorConfig=TransportConfiguration(name=slave-connector, factory=org-apache-activemq-artemis-core-remoting-impl-netty-NettyConnectorFactory) ?trustStorePassword=****&port=61616&keyStorePassword=****&sslEnabled=true&host=slave-com&trustStorePath=/path/to/ssl/truststore-jks&keyStorePath=/path/to/ssl/keystore-jks
2021-05-07 13:54:32,357 DEBUG [org.apache.activemq.artemis.core.remoting.impl.netty.NettyConnector] Connector NettyConnector [host=slave.com, port=61616, httpEnabled=false$ httpUpgradeEnabled=false, useServlet=false, servletPath=/messaging/ActiveMQServlet, sslEnabled=true, useNio=true] using native epoll
2021-05-07 13:54:32,357 DEBUG [org.apache.activemq.artemis.core.client] AMQ211002: Started EPOLL Netty Connector version 4.1.51.Final to slave.com:61616
2021-05-07 13:54:32,358 DEBUG [org.apache.activemq.artemis.core.remoting.impl.netty.NettyConnector] Remote destination: slave.com/123.123.123.123:61616
2021-05-07 13:54:32,358 DEBUG [org.apache.activemq.artemis.spi.core.remoting.ssl.SSLContextFactory] Creating SSL context with configuration
trustStorePassword=****
port=61616
keyStorePassword=****
sslEnabled=true
host=slave.com
trustStorePath=/path/to/ssl/truststore.jks
keyStorePath=/path/to/ssl/keystore.jks
2021-05-07 13:54:32,448 DEBUG [org.apache.activemq.artemis.core.remoting.impl.netty.NettyConnector] Added ActiveMQClientChannelHandler to Channel with id = 77c078c2
2021-05-07 13:54:32,448 DEBUG [org.apache.activemq.artemis.core.client.impl.ClientSessionFactoryImpl] Connector towards NettyConnector [host=slave.com, port=61616, httpEnabled=false, httpUpgradeEnabled=false, useServlet=false, servletPath=/messaging/ActiveMQServlet, sslEnabled=true, useNio=true] failed
This is what slave is spamming:
2021-05-07 14:06:53,177 DEBUG [org.apache.activemq.artemis.core.server.impl.FileLockNodeManager] trying to lock position: 1
2021-05-07 14:06:53,178 DEBUG [org.apache.activemq.artemis.core.server.impl.FileLockNodeManager] failed to lock position: 1
If I attempt to telnet from master node to slave node (same if I do it from slave to slave):
[root#master]# telnet slave.com 61616
Trying 123.123.123.123...
telnet: connect to address 123.123.123.123: Connection refused
However if I attempt the same telnet in that the only working cluster, I can successfully "connect" from master to slave...
Here is what I suspect:
Master acquires lock in instance/data/journal/server.lock
Master keeps trying to connect to slave server
Slave unable to start, because it cannot acquire the same server.lock on shared storage.
Master uses high CPU because of such hard-trying to connect to slave, which is not running.
What am I doing wrong?
EDIT: This is how my NFS mounts look like (taken from mount command):
some_server:/some_dir on /path/to/artemis/instance/data type nfs4 (rw,relatime,sync,vers=4.1,rsize=65536,wsize=65536,namlen=255,acregmin=0,acregmax=0,acdirmin=0,acdirmax=0,soft,noac,proto=tcp,timeo=50,retrans=1,sec=sys,clientaddr=123.123.123.123,local_lock=none,addr=123.123.123.123)
Turns out issue was in broker.xml configuration. In static-connectors I somehow decided to list only a "non-current server" (e.g. I have srv0 and srv1 - in srv0 I only added connector of srv1 and vice versa).
What it used to be (on 1st master node):
<cluster-connections>
<cluster-connection name="abc">
<connector-ref>srv0-connector</connector-ref>
<message-load-balancing>ON_DEMAND</message-load-balancing>
<max-hops>1</max-hops>
<static-connectors>
<connector-ref>srv1-connector</connector-ref>
</static-connectors>
</cluster-connection>
</cluster-connections>
How it is now (on 1st master node):
<cluster-connections>
<cluster-connection name="abc">
<connector-ref>srv0-connector</connector-ref>
<message-load-balancing>ON_DEMAND</message-load-balancing>
<max-hops>1</max-hops>
<static-connectors>
<connector-ref>srv0-connector</connector-ref>
<connector-ref>srv1-connector</connector-ref>
</static-connectors>
</cluster-connection>
</cluster-connections>
After listing all cluster's nodes, the CPU normalized and it's not only ~1% on active node. The issue is totally not related AMQ Artemis connections spamming or file locks.
We are using MQ 9.0.0.1 version on linux machine and it is active/passive configuration.
We have faced issue in production server issue is "queue manager ended unexpectedly" around 10:30 AM IST, we have taken queue manager restart manually post restart everything is normal issue got resolved.
From the Queue manager logs we have observed below error stacks. Request to all,
please provide your analysis from the below queue manager logs why the queue manager ended unexpectedly?
-------------------------------------------------------------------------------
03/07/2019 07:21:10 AM - Process(5082.15855) User(tsg) Program(amqzlaa0)
Host(xxxxxxx) Installation(Installation1)
VRMF(9.0.0.1) QMgr(QMGR_NEFT)
AMQ7234: 20000 messages from queue 'NEFT_SMS_INQUIRY' loaded on queue manager
'QMGR_NEFT'.
EXPLANATION:
20000 messages from queue NEFT_SMS_INQUIRY have been loaded on queue manager
QMGR_NEFT.
ACTION:
None.
-------------------------------------------------------------------------------
03/07/2019 10:30:15 AM - Process(5003.1) User(tsg) Program(amqzxma0)
Host(xxxxxxx) Installation(Installation1)
VRMF(9.0.0.1) QMgr(QMGR_NEFT)
AMQ5008: An essential IBM MQ process 5016 (zllCRIT) cannot be found and is
assumed to be terminated.
EXPLANATION:
1) A user has inadvertently terminated the process. 2) The system is low on
resources. Some operating systems terminate processes to free resources. If
your system is low on resources, it is possible it has terminated the process
so that a new process can be created.
ACTION:
IBM MQ will stop all MQ processes. Inform your systems administrator. When
the problem is rectified IBM MQ can be restarted.
----- amqzxmb0.c : 10095 ------------------------------------------------------
03/07/2019 10:30:15 AM - Process(5080.1) User(tsg) Program(amqpcsea)
Host(xxxxxxx) Installation(Installation1)
VRMF(9.0.0.1) QMgr(QMGR_NEFT)
AMQ8506: Command server MQGET failed with reason code 2009.
EXPLANATION:
An MQGET request by the command server, for the IBM MQ queue
SYSTEM.ADMIN.COMMAND.QUEUE , failed with reason code 2009.
ACTION:
None.
----- amqphrea.c : 86 ---------------------------------------------------------
03/07/2019 10:30:15 AM - Process(5069.1) User(tsg) Program(amqzmgr0)
Host(xxxxxxx) Installation(Installation1)
VRMF(9.0.0.1) QMgr(QMGR_NEFT)
AMQ5025: The command server has ended. ProcessId(5080).
EXPLANATION:
The command server process has ended.
ACTION:
None.
-------------------------------------------------------------------------------
03/07/2019 10:30:15 AM - Process(5051.1) User(tsg) Program(amqrrmfa)
Host(xxxxxxx) Installation(Installation1)
VRMF(9.0.0.1) QMgr(QMGR_NEFT)
AMQ9542: Queue manager is ending.
EXPLANATION:
The program will end because the queue manager is quiescing.
ACTION:
None.
----- amqrrmfa.c : 3011 -------------------------------------------------------
03/07/2019 10:30:15 AM - Process(5051.1) User(tsg) Program(amqrrmfa)
Host(xxxxxxx) Installation(Installation1)
VRMF(9.0.0.1) QMgr(QMGR_NEFT)
AMQ9411: Repository manager ended normally.
EXPLANATION:
The repository manager ended normally.
ACTION:
None.
-------------------------------------------------------------------------------
03/07/2019 10:30:15 AM - Process(5047.9) User(tsg) Program(amqzmuf0)
Host(xxxxxxx) Installation(Installation1)
VRMF(9.0.0.1) QMgr(QMGR_NEFT)
AMQ5976: 'IBM MQ Distributed Pub/Sub Command Task' has ended.
EXPLANATION:
'IBM MQ Distributed Pub/Sub Command Task' has ended.
ACTION:
-------------------------------------------------------------------------------
03/07/2019 10:30:15 AM - Process(5047.10) User(tsg) Program(amqzmuf0)
Host(xxxxxxx) Installation(Installation1)
VRMF(9.0.0.1) QMgr(QMGR_NEFT)
AMQ5976: 'IBM MQ Distributed Pub/Sub Publish Task' has ended.
EXPLANATION:
'IBM MQ Distributed Pub/Sub Publish Task' has ended.
ACTION:
-------------------------------------------------------------------------------
03/07/2019 10:30:15 AM - Process(5047.8) User(tsg) Program(amqzmuf0)
Host(xxxxxxx) Installation(Installation1)
VRMF(9.0.0.1) QMgr(QMGR_NEFT)
AMQ5976: 'IBM MQ Distributed Pub/Sub Fan Out Task' has ended.
EXPLANATION:
'IBM MQ Distributed Pub/Sub Fan Out Task' has ended.
ACTION:
-------------------------------------------------------------------------------
03/07/2019 10:30:15 AM - Process(5047.7) User(tsg) Program(amqzmuf0)
Host(xxxxxxx) Installation(Installation1)
VRMF(9.0.0.1) QMgr(QMGR_NEFT)
AMQ5976: 'IBM MQ Distributed Pub/Sub Controller' has ended.
EXPLANATION:
'IBM MQ Distributed Pub/Sub Controller' has ended.
ACTION:
-------------------------------------------------------------------------------
03/07/2019 10:30:15 AM - Process(31712.199) User(tsg) Program(amqrmppa)
Host(xxxxxxx) Installation(Installation1)
VRMF(9.0.0.1) QMgr(QMGR_NEFT)
AMQ9508: Program cannot connect to the queue manager.
EXPLANATION:
The connection attempt to queue manager 'QMGR_NEFT' failed with reason code
2059.
ACTION:
Ensure that the queue manager is available and operational.
----- cmqxrmsa.c : 6146 -------------------------------------------------------
03/07/2019 10:30:15 AM - Process(31712.199) User(tsg) Program(amqrmppa)
Host(xxxxxxx) Installation(Installation1)
VRMF(9.0.0.1) QMgr(QMGR_NEFT)
AMQ9772: MQCTL failed with MQRC=2009.
EXPLANATION:
The indicated IBM MQ API call failed for the specified reason code.
ACTION:
Refer to the Application Programming Reference manual for information about
Reason Code 2009.
----- cmqxrstf.c : 2663 -------------------------------------------------------
03/07/2019 10:30:15 AM - Process(31712.199) User(tsg) Program(amqrmppa)
Host(xxxxxxx) Installation(Installation1)
VRMF(9.0.0.1) QMgr(QMGR_NEFT)
AMQ9999: Channel 'GPSSVRCONN' to host 'XXXXXX' ended abnormally.
EXPLANATION:
The channel program running under process ID 31712 for channel 'GPSSVRCONN'
ended abnormally. The host name is 'XXXXXX'; in some cases the host name
cannot be determined and so is shown as '????'.
ACTION:
Look at previous error messages for the channel program in the error logs to
determine the cause of the failure. Note that this message can be excluded
completely or suppressed by tuning the "ExcludeMessage" or "SuppressMessage"
attributes under the "QMErrorLog" stanza in qm.ini. Further information can be
found in the System Administration Guide.
----- amqrmrsa.c : 938 --------------------------------------------------------
03/07/2019 10:30:15 AM - Process(5081.5) User(tsg) Program(runmqlsr)
Host(xxxxxxx) Installation(Installation1)
VRMF(9.0.0.1) QMgr(QMGR_NEFT)
AMQ9508: Program cannot connect to the queue manager.
EXPLANATION:
The connection attempt to queue manager 'QMGR_NEFT' failed with reason code
2059.
ACTION:
Ensure that the queue manager is available and operational.
----- cmqxrmsa.c : 471 --------------------------------------------------------
03/07/2019 10:30:15 AM - Process(5081.5) User(tsg) Program(runmqlsr)
Host(xxxxxxx) Installation(Installation1)
VRMF(9.0.0.1) QMgr(QMGR_NEFT)
AMQ9999: Channel 'GPSSVRCONN' to host 'XXXXXX' ended abnormally.
EXPLANATION:
The channel program running under process ID 5081 for channel 'GPSSVRCONN''
ended abnormally. The host name is 'XXXXXX'; in some cases the host name
cannot be determined and so is shown as '????'.
ACTION:
Look at previous error messages for the channel program in the error logs to
determine the cause of the failure. Note that this message can be excluded
completely or suppressed by tuning the "ExcludeMessage" or "SuppressMessage"
attributes under the "QMErrorLog" stanza in qm.ini. Further information can be
found in the System Administration Guide.
----- amqrmrsa.c : 938 --------------------------------------------------------
03/07/2019 10:30:17 AM - Process(5078.1) User(tsg) Program(runmqchi)
Host(xxxxxxx) Installation(Installation1)
VRMF(9.0.0.1) QMgr(QMGR_NEFT)
AMQ9510: Messages cannot be retrieved from a queue.
EXPLANATION:
The attempt to get messages from queue 'SYSTEM.CHANNEL.INITQ' on queue manager
'QMGR_NEFT' failed with reason code 2009.
ACTION:
If the reason code indicates a conversion problem, for example
MQRC_SOURCE_CCSID_ERROR, remove the message(s) from the queue. Otherwise,
ensure that the required queue is available and operational.
----- amqrimna.c : 1085 -------------------------------------------------------
03/07/2019 10:30:18 AM - Process(5069.1) User(tsg) Program(amqzmgr0)
Host(xxxxxxx) Installation(Installation1)
VRMF(9.0.0.1) QMgr(QMGR_NEFT)
AMQ5023: The channel initiator has ended. ProcessId(5078).
EXPLANATION:
The channel initiator process has ended.
ACTION:
None.
-------------------------------------------------------------------------------
03/07/2019 10:30:24 AM - Process(31712.192) User(tsg) Program(amqrmppa)
Host(xxxxxxx) Installation(Installation1)
VRMF(9.0.0.1) QMgr(QMGR_NEFT)
AMQ9540: Commit failed.
EXPLANATION:
The program ended because return code 2009 was received when an attempt was
made to commit change to the resource managers. The commit ID was
'AMQRMRSASFMS.TO.IPAYNEFT XXXXXX E怜͇'.
ACTION:
Tell the systems administrator.
----- amqrmrca.c : 2977 -------------------------------------------------------
03/07/2019 10:30:24 AM - Process(31712.192) User(tsg) Program(amqrmppa)
Host(xxxxxxx) Installation(Installation1)
VRMF(9.0.0.1) QMgr(QMGR_NEFT)
AMQ9999: Channel 'SFMS.TO.IPAYNEFT' to host 'XXXXXX' ended abnormally.
EXPLANATION:
The channel program running under process ID 31712 for channel
'SFMS.TO.IPAYNEFT' ended abnormally. The host name is 'XXXXXX'; in some
cases the host name cannot be determined and so is shown as '????'.
ACTION:
Look at previous error messages for the channel program in the error logs to
determine the cause of the failure. Note that this message can be excluded
completely or suppressed by tuning the "ExcludeMessage" or "SuppressMessage"
attributes under the "QMErrorLog" stanza in qm.ini. Further information can be
found in the System Administration Guide.
----- amqrmrsa.c : 938 --------------------------------------------------------
03/07/2019 10:30:24 AM - Process(5071.1) User(tsg) Program(amqfqpub)
Host(xxxxxxx) Installation(Installation1)
VRMF(9.0.0.1) QMgr(QMGR_NEFT)
AMQ5807: Queued Publish/Subscribe Daemon for queue manager QMGR_NEFT ended.
EXPLANATION:
The Queued Publish/Subscribe Daemon on queue manager QMGR_NEFT has ended.
ACTION:
None.
-------------------------------------------------------------------------------
03/07/2019 10:30:24 AM - Process(5003.1) User(tsg) Program(amqzxma0)
Host(xxxxxxx) Installation(Installation1)
VRMF(9.0.0.1) QMgr(QMGR_NEFT)
AMQ5050: An essential IBM MQ process 5071 (amqfqpub) cannot be found and is
assumed to be terminated.
EXPLANATION:
1) A user has inadvertently terminated the process. 2) The system is low on
resources. Some operating systems terminate processes to free resources. If
your system is low on resources, it is possible it has terminated the process
so that a new process can be created. 3) MQ has encountered an unexpected
error. Check for possible errors reported in the MQ error logs and for any
FFSTs that have been generated.
ACTION:
IBM MQ will attempt to restart the terminated process.
----- amqzxmb0.c : 10095 ------------------------------------------------------
03/07/2019 10:30:24 AM - Process(5003.1) User(tsg) Program(amqzxma0)
Host(xxxxxxx) Installation(Installation1)
VRMF(9.0.0.1) QMgr(QMGR_NEFT)
AMQ8004: IBM MQ queue manager 'QMGR_NEFT' ended.
EXPLANATION:
IBM MQ queue manager 'QMGR_NEFT' ended.
ACTION:
None.
-------------------------------------------------------------------------------
03/07/2019 10:30:25 AM - Process(5069.1) User(tsg) Program(amqzmgr0)
Host(xxxxxxx) Installation(Installation1)
VRMF(9.0.0.1) QMgr(QMGR_NEFT)
AMQ5027: The listener 'LISTENER.TCP' has ended. ProcessId(5081).
EXPLANATION:
The listener process has ended.
ACTION:
None.
-------------------------------------------------------------------------------
I'm trying to set up a RocketMQ cluster, with a single name server, 1 master and 2 slaves. But, I'm running into some problems.
The version I'm running is downloaded from github/rocketmq-all-4.1.0-incubating.zip.
The brokers are run using mqbroker -c broker.conf, where broker.conf
differs for master and slave. For the master I have:
listenPort=10911
brokerName=mybroker
brokerClusterName=mybrokercluster
brokerId=0
deleteWhen=04
fileReservedTime=48
brokerRole=SYNC_MASTER
flushDiskType=ASYNC_FLUSH
And for slaves:
listenPort=10911
brokerName=mybroker
brokerClusterName=mybrokercluster
brokerId=1
deleteWhen=04
fileReservedTime=48
brokerRole=SLAVE
flushDiskType=ASYNC_FLUSH
The second slave has brokerId=2.
Brokers start up fine, some parts of the logs for a slave:
2017-10-02 20:31:35 INFO main - brokerRole=ASYNC_MASTER
2017-10-02 20:31:35 INFO main - flushDiskType=ASYNC_FLUSH
(...)
2017-10-02 20:31:35 INFO main - Replace, key: brokerId, value: 0 -> 1
2017-10-02 20:31:35 INFO main - Replace, key: brokerRole, value:
ASYNC_MASTER -> SLAVE
(...)
2017-10-02 20:31:37 INFO main - Set user specified name server address:
172.22.1.38:9876
2017-10-02 20:31:37 INFO ShutdownHook - Shutdown hook was invoked, 1
2017-10-02 20:31:37 INFO ShutdownHook - shutdown thread
PullRequestHoldService interrupt false
2017-10-02 20:31:37 INFO ShutdownHook - join thread PullRequestHoldService
eclipse time(ms) 0 90000
2017-10-02 20:31:37 WARN ShutdownHook - unregisterBroker Exception,
172.22.1.38:9876
org.apache.rocketmq.remoting.exception.RemotingConnectException: connect to
<172.22.1.38:9876> failed
at
org.apache.rocketmq.remoting.netty.NettyRemotingClient.invokeSync(NettyRemotingClient.java:359)
~[rocketmq-remoting-4.1.0-incubating.jar:4.1.0-incubating]
at
org.apache.rocketmq.broker.out.BrokerOuterAPI.unregisterBroker(BrokerOuterAPI.java:221)
~[rocketmq-broker-4.1.0-incubating.jar:4.1.0-incubating]
at
org.apache.rocketmq.broker.out.BrokerOuterAPI.unregisterBrokerAll(BrokerOuterAPI.java:198)
~[rocketmq-broker-4.1.0-incubating.jar:4.1.0-incubating]
at
org.apache.rocketmq.broker.BrokerController.unregisterBrokerAll(BrokerController.java:623)
[rocketmq-broker-4.1.0-incubating.jar:4.1.0-incubating]
at
org.apache.rocketmq.broker.BrokerController.shutdown(BrokerController.java:589)
[rocketmq-broker-4.1.0-incubating.jar:4.1.0-incubating]
at org.apache.rocketmq.broker.BrokerStartup$1.run(BrokerStartup.java:218)
[rocketmq-broker-4.1.0-incubating.jar:4.1.0-incubating]
at java.lang.Thread.run(Thread.java:748) [na:1.8.0_141]
2017-10-02 20:31:37 INFO ShutdownHook - Shutdown hook over, consuming total
time(ms): 25
2017-10-02 20:31:45 INFO BrokerControllerScheduledThread1 - dispatch behind
commit log 0 bytes
2017-10-02 20:31:45 INFO BrokerControllerScheduledThread1 - Slave fall
behind master: 0 bytes
2017-10-02 20:31:45 INFO BrokerControllerScheduledThread1 - register broker
to name server 172.22.1.38:9876 OK
2017-10-02 20:32:15 INFO BrokerControllerScheduledThread1 - register broker
to name server 172.22.1.38:9876 OK
As I suspect the broker is trying to connect to the name server, which
isn't running initially, so it retries and eventually succeeds?
However, later when trying clusterList I only see one broker listed, which happens to be a slave (172.22.1.17) and has brokerId=2 in the configuration (although here it's listed as 0):
$ ./mqadmin clusterList -n 172.22.1.38:9876
#Cluster Name #Broker Name #BID #Addr
#Version #InTPS(LOAD) #OutTPS(LOAD) #PCWait(ms) #Hour
#SPACE
mybrokercluster mybroker 0 172.22.1.17:10911
V4_1_0_SNAPSHOT 0.00(0,0ms) 0.00(0,0ms) 0
418597.80 -1.0000
Moreover, when sending messages to the master, I get SLAVE_NOT_AVAILABLE.
Why is that? Are the brokers configured properly? If so, wy does
clusterList report them incorrectly?
you should change slave port,as you know 10911 has been used by anthoer process(master node),slave should be use different tcp port(eg.10921/10931 and so on)
tips:my cluster deploy on one machine,so i changed tcp port and startup successful,if you master&slave deploy on different machine and startup failed,you should visit rocketmq error log for more information.
notice:one master which have more than one slave,brokerId should be different
I got a cluster of 1 master node and 2 slaves and I'm trying to compile my application with mesos.
Basically, here is the command that I use:
mesos-execute --name=alc1 --command="ccmake -j myapp" --master=10.11.12.13:5050
Offers are made from the slave but this compilation task keeps failing.
[root#master-node ~]# mesos-execute --name=alc1 --command="ccmake -j myapp" --master=10.11.12.13:5050
I0511 22:26:11.623016 11560 sched.cpp:222] Version: 0.28.0
I0511 22:26:11.625602 11564 sched.cpp:326] New master detected at master#10.11.12.13:5050
I0511 22:26:11.625952 11564 sched.cpp:336] No credentials provided. Attempting to register without authentication
I0511 22:26:11.627279 11564 sched.cpp:703] Framework registered with 70582e35-5d6e-4915-a919-cae61c904fd9-0139
Framework registered with 70582e35-5d6e-4915-a919-cae61c904fd9-0139
task alc1 submitted to slave 70582e35-5d6e-4915-a919-cae61c904fd9-S2
Received status update TASK_RUNNING for task alc1
Received status update TASK_FAILED for task alc1
I0511 22:26:11.759610 11567 sched.cpp:1903] Asked to stop the driver
I0511 22:26:11.759639 11567 sched.cpp:1143] Stopping framework '70582e35-5d6e-4915-a919-cae61c904fd9-0139'
On the sandbox slave node, here is the stderr logs:
I0511 22:26:13.781070 5037 exec.cpp:143] Version: 0.28.0
I0511 22:26:13.785001 5040 exec.cpp:217] Executor registered on slave 70582e35-5d6e-4915-a919-cae61c904fd9-S2
sh: ccmake: command not found
I0511 22:26:13.892653 5042 exec.cpp:390] Executor asked to shutdown
Just to mentionned that commands like this work fine and get me the expected results:
[root#master-node ~]# mesos-execute --name=alc1 --command="find / -name a" --master=10.11.12.13:5050
I0511 22:26:03.733172 11550 sched.cpp:222] Version: 0.28.0
I0511 22:26:03.736112 11554 sched.cpp:326] New master detected at master#10.11.12.13:5050
I0511 22:26:03.736383 11554 sched.cpp:336] No credentials provided. Attempting to register without authentication
I0511 22:26:03.737730 11554 sched.cpp:703] Framework registered with 70582e35-5d6e-4915-a919-cae61c904fd9-0138
Framework registered with 70582e35-5d6e-4915-a919-cae61c904fd9-0138
task alc1 submitted to slave 70582e35-5d6e-4915-a919-cae61c904fd9-S2
Received status update TASK_RUNNING for task alc1
Received status update TASK_FINISHED for task alc1
I0511 22:26:04.184813 11553 sched.cpp:1903] Asked to stop the driver
I0511 22:26:04.184844 11553 sched.cpp:1143] Stopping framework '70582e35-5d6e-4915-a919-cae61c904fd9-0138'
I don't really get what is needed for even troubleshot this issue.