OUT_ONLY Issue on Multi-Tenant Mode (WSO2 ESB 5.0 BETA) - multi-tenant

I'm experiencing the following issue when in Multi-Tenant mode on the WSO2 ESB 5.0 Beta version. This occurs when I send to an endpoint(which is a webservice running on WSO2 AS 5.3). I've tried changing the method on the Webservice, both to return a value and to return nothing(i.e. a void method) with no luck. Everything seems to process fine, except the following error seems to appear on my log file, exactly 3 minutes, after the message was sent.
Issue similar to
[Errors using input-only web service (OUT_ONLY from ESB) ][1]
I also think this may be related to bug 'ESBJAVA-3989' - "Exception thrown when OUT_ONLY is set and a response is received for an API in Tenant Mode", however I see no anticipated fix for it. My rationale of moving to ESB 5.0 was to maximize the Multi-tenant features, which had few bugs in the previous versions. However, this is proving to be a stumbling block.
Any ideas ?
TID: [-1] [] [2016-07-21 15:04:06,810] ERROR {org.apache.axis2.transport.base.threads.NativeWorkerPool} - Uncaught exception {org.apache.axis2.transport.base.threads.NativeWorkerPool}
java.lang.UnsupportedOperationException: Not yet implemented
at org.apache.axis2.description.OutOnlyAxisOperation.getMessage(OutOnlyAxisOperation.java:124)
at org.wso2.carbon.core.multitenancy.MultitenantMessageReceiver.processResponse(MultitenantMessageReceiver.java:140)
at org.wso2.carbon.core.multitenancy.MultitenantMessageReceiver.receive(MultitenantMessageReceiver.java:82)
at org.apache.axis2.engine.AxisEngine.receive(AxisEngine.java:180)
at org.apache.synapse.transport.passthru.ClientWorker.run(ClientWorker.java:255)
at org.apache.axis2.transport.base.threads.NativeWorkerPool$1.run(NativeWorkerPool.java:172)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
at java.lang.Thread.run(Thread.java:745)

It is a bug identified in ESB tenant mode and reported in jira [1]. Still team did not fixed that and will do according to the priority.
[1] https://wso2.org/jira/browse/ESBJAVA-3989

Related

Problem with Queue browsing when upgrading from 5.15.14 to 5.16.2

Has anyone also noticed an issue with queue browsing with 5.16.2?
It happens when using an XA connection and starting a session without a transaction context.
This always worked fine but no longer does. Instead one can see a exception, but only when debug logging is turned on:
javax.jms.JMSException: Session's XAResource has not been enlisted in a distributed transaction.
at org.apache.activemq.ActiveMQXASession.doStartTransaction(ActiveMQXASession.java:101) ~[?:?]
This was changed a few years back in 5.16.0 via AMQ-2659. You need to set xaAckMode=1 on your URL now, e.g.:
tcp://localhost:61616?jms.xaAckMode=1

WebRTC app not working on firefox

I have a very tremendous issue in my app. App is using webrtc to create video connection between two people.
Currently the app it's in the test phase.
Everything is working fine on Chrome, but on Firefox there is a strange issue.
When the second Peer connects I receive this error:
Error adding ice candidate for pcInvalidStateError: setRemoteDescription needs to called before addIceCandidate
I know that the error message seems to be clear, but how it's possible that on Chrome this error does not exist?
I mean, maybe there is a bigger issue, not completely depending on this error message.
Do you have any ideas or solutions to this?
A part from WebRTC doc: (See Deprecated Exception section)
Deprecated exceptions
When using the deprecated callback-based version of
setRemoteDescription(), the following exceptions may occur:
InvalidStateError The connection's signalingState is "closed",
indicating that the connection is not currently open, so negotiation
cannot take place.
You should check you are not using deprecated callback version of this function. Also you should keep an eye on signalingState of the peer connection.
Hope it helps!

SignalR Intermittent Failure to call client method

I am currently using SignalR 2.0 and MVC 5.0 for my project. All the signalR update works fine except when I have updated the web.config file which causes the application pool to recycle. This is the starting point I notice SignalR behaving inconsistently. Sometimes I need to perform an action multiple times before the SignalR update can be propagated to the client. I have enabled the SignalR logging and I can't see anything displayed in the log whenever it failed to propagate to the client.
Do we need to do something in the hub, like reconnect or anything when the app pool got recycled?
At the moment, i didnt do anything in terms of coding and simply refresh the page. But it gives me that problems with SignalR. And the only way to fix that intermittent problem completely is to reboot the server.
Have anyone experienced this kind of behavior in SignalR?
Thanks in advance.

Playframework-2.0 NettyDispatcher 500 Error on Heroku

I started getting 500 Internal Server Error(ISE) on Heroku, tried enabling debug level logs. The error is not consistent, it occurs for some request, for others it goes thru properly. When there is an ISE there is not even a single log line(even with debug on) in web dynos... I am able to see that the 500 response is given for the request.
From the client side am seeing the following
This exception has been logged with id 6gimbegj7
The above line tells that it should a response from Play! when it is running in production mode.
Attached New Relic monitoring plugin where it says the Exception occurred on NettyDispatcher (Netty IO server), not further info.
Any idea what could be the issue?
We have identified the problem here,
In play! we have a Global.java on the root of the packages, which has the methods like
onRequest(), onRequestRoute()
If there is any error at this place, and if we do not handle the exceptions, then Netty directly throws an 500 response to the request. This can happen from any part of the application. We did catch the exceptions in other part of the application but missed this one.
Mimicking the heoku env with the help of Foreman tool helped us to narrow down on the cause. As such setting the playframework log level to DEBUG did not throw any logs/stacktrace, Not sure why.

Validation of viewstate MAC failed

We've recently moved all our sites from one server (IIS6) to a new server (IIS7) which has also involved changing all the DNS records. However, on some of the sites, when you make a post-back on certain pages, you get the following error:
Validation of viewstate MAC failed. If this application is hosted by a Web Farm or cluster, ensure that <machineKey> configuration specifies the same validationKey and validation algorithm. AutoGenerate cannot be used in a cluster.
Description: An unhandled exception occurred during the execution of the current web request. Please review the stack trace for more information about the error and where it originated in the code.
Exception Details: System.Web.HttpException: Validation of viewstate MAC failed. If this application is hosted by a Web Farm or cluster, ensure that <machineKey> configuration specifies the same validationKey and validation algorithm. AutoGenerate cannot be used in a cluster.
At first I thought this may be a propogation issue with the site (Ie. its trying to post-back to the old server). However, the DNS records were changed over 2 weeks ago, so surely this cannot be the case.
Why would this error be occuring?
It might be worth noting, this error never occurred on the old server.
This may be a bug related to asp.net-2.0. Which version are you using - the bug is fixed in SP2.

Resources