Exception while using Windows Azure Caching : No such host is known - caching

I am trying to get started with Azure and am trying to use the Caching feature. I created a cloud service project and added a Cache worker role and a web role. I installed "Windows Azure Caching" nuget into projects for both the roles and added the name of cache worker role as identifier in DataCacheClients element in web.config of the web role.
I added the following code into the web role:
DataCacheFactory cf = new DataCacheFactory();
DataCache c = cf.GetDefaultCache();
When I try to run this locally on the emulator, I get the following exception:
ErrorCode<ERRCA0017>:SubStatus<ES0006>:There is a temporary failure.
Please retry later. (One or more specified cache servers are unavailable,
which could be caused by busy network or servers. For on-premises cache clusters,
also verify the following conditions. Ensure that security permission has been granted
for this client account, and check that the AppFabric Caching Service is allowed through
the firewall on all cache hosts. Also the MaxBufferSize on the server must be greater
than or equal to the serialized object size sent from the client.).
Additional Information : The client was trying to communicate with the server: net.tcp://MvcWebRole1:24233.
Inner Exception : No such host is known
Can you please tell me what I am missing here?
Azure SDK used : v2.0

Timing of your question couldn't be better. We also faced exactly the same issue and were scrathing our head as to what the problem could be. We had one project where everything worked perfectly fine and in one we were getting the same error. Based on our research, we have identified the problem with the Nuget package for caching. It seems a new version (2.1.0.0) was released yesterday and we found that if we install that package, we get this error. Can you check the package version in your case? The documentation states that this new version can only be used with the latest SDK (2.1) released today.
One solution would be to uninstall version 2.1.0.0 and install version 2.0.0.0. To install version 2.0.0.0, open Package Manager Console (View --> Other Windows --> Package Manager Console) and type following command there:
Install-Package Microsoft.WindowsAzure.Caching -Version 2.0.0.0
This fixed our problem. Hopefully it should fix yours too.

Here is a link to the Windows Azure Cloud Integration Engineering blog on how to deal with this same issue. They recommend upgrading to Azure SDK v 2.1 or rolling back as the accepted answer states.
http://blogs.msdn.com/b/cie/archive/2013/08/08/windows-azure-caching-2-1-0-0-no-such-host-is-known.aspx

This exception could also occur under compute emulator if for the role in the client library configuration there is no cache configured. In my case this happened on purpose, since the cache emulator has some problems that can slow down the test and debugging on the service.
In previous version of Windows Azure Caching, in this scenario the construction of DataCacheFactory would fail with an exception (handled by my code); with Windows Azure Caching 2.1 (and Azure SDK 2.1) in this same scenario the code would consider the role name as a server address and thus on DataCache construction would try to communicate with the non existing cache -- this leads to a 3 minute wait and the exception reported in the question.
I have changed my code to detect the new situation for this scenario -- you can find more detail in this SO question.

Related

Self-Hosted Integration Runtime times out after ADF pipeline loads a few tables

I have recently installed Integration Runtime for a local server which hosts an Access DB. The idea is to pull data from it and store in in Azure SQL DB. I have done the following:
Integration Runtime Services Installed on local machine hosting the MS Access DB and connected to it using ODBC
Created linked services in ADF to connect to the DB
Created Datasets for source and destination DB for each table required. One for the source msaccess and one for the target in azure sql db
Created a pipeline to copy the data from the source and sink into the asql db mentioned in step #3
Basically, all the connections work however when I trigger my pipeline to load around 10 of these tables, it runs and loads the first two and then fails afterwards by timing out. I must restart the Integration Runtime everytime to get it back up and running otherwise I can no longer query the tables.
To mitigate this I figured there was too much traffic and the server needed to rest between calls so I added wait timers in between each step of the pipelines but not much success. It did help a bit but that might be coincidence.
The error log in the monitor spits out at the failed step is :
Error: 2200
ErrorCode=UserErrorFailedToConnectOdbcSource,'Type=Microsoft.DataTransfer.Common.Shared.HybridDeliveryException,Message=,Source=Microsoft.DataTransfer.Runtime.GenericOdbcConnectors,''Type=System.Data.Odbc.OdbcException,Message=,Source=,'
btw the integration runtime install service running is 5.12.7984.1 and the version of access installed is through office 365 x64. The exact MS access driver is 16.14430.20006. The OS is Windows Server 2019
I am getting the exact same error. To start, I did check the event viewer logs and saw some errors to do with access. So I gave the user running the IR more access to the registry keys/general log on as a service rights. This helped a little, but I am still stuck with the same problem.
When copying from an Access DB located on the SHIR itself to flat files in lake storage, I encountered the same error.
Removing Office 365 from the machine, then re-installing the Access runtime solved the problem.
This answer is from #ezaidi comments above.

Blob Trigger not working when published VS2017 .NET Core 2.0 Azure Function v2

When I create a Azure Function with VS2017 in .NET Core 2.0, running locally, my blob trigger works fine. But when I publish to Azure, the trigger does not fire and no logs are written to the Azure Portal console.
My Function code.
My local.settings.json
After publishing, I go to the Azure portal and see this error: "The binding type(s) 'blobTrigger' are not registered.
I then attempt to install the Microsoft.Azure.WebJobs.Extensions.Storage but it fails. (I got here by trying to create a new trigger on the portal.)
If I keep trying, eventually it seems to install. But my published trigger doesn't respond to any file drops in my Blob Storage. The Function App's application setting, FUNCTIONS_EXTENSION_VERSION = beta and the run time is 2.x.
Has anyone gotten this to work?
There is a breaking change for Azure Storage bindings for 2.x runtime:
The Azure Storage bindings (blob, queues and tables triggers and bindings) were previously part of the runtime, but moving forward, will need to be explicitly installed. This change was made to allow Azure Functions to support multiple versions of the Azure Storage SDK
To temporarily avoid impact, you can pin your Function App to the previous version of the runtime by updating the app setting FUNCTIONS_EXTENSION_VERSION to 2.0.11961-alpha
https://github.com/Azure/app-service-announcements/issues/129

Unable to deploy project using dedicated window azure caching (preview) in production environment

I made a sample application using windows azure dedicated caching (preview).
The sample application runs perfectly in emulated environment but does not get properly deployed in production environment: The project and the cache instances always shows that "waiting for role to startup". I am not able to understand whether it is a configuration issue or something else.
Which deployment mode are you using? Colocated or Dedicated?
If it is colocated can you tell me what value of Cache Size (%) have you set? And which vm type are you deploying?
Also, please check whether you have provided storage account connection string in "Caching Tab" under "storage account credentials to use for maintaining the cache cluster's runtime state" section.
In the Cache Worker Role (the role where you have caching service enabled) if you go to "Caching Tab" you will see "storage account credentials to use for maintaining the cache cluster's runtime state" section.
For Emulated environment this value is "UseDevelopmentStorage=true", for deploying to cloud you need to replace this setting with a valid cloud storage account. Please follow detailed instructions for proper configuration at http://msdn.microsoft.com/en-us/library/windowsazure/jj131263.aspx
Further if you are still facing the issues, enabling remote desktop and checking "Windows->Application" and/or "Microsoft->Windows-Application Server-System Services/Admin" channel would also help.

How I can update the new code to my site in windows azure cloud?

I am new windows azure user. I have gotten selected for 90 days trial account and I am able to upload my ASP.NET MVC3 application to my account. My site is also running now. After I did publish my site, I added more model, views and controller to my proramme. Now I can not find a way to update my application. I can again publish my application but update option is not there. I want to update my new code only but the package option is creating full application. How I can update the new code to my site in windows azure cloud?
[Changed spelling]
With Windows Azure you can publish/update an application following ways:
Log into you Windows Azure account. Select you hosted server name and at the top panel you will see "Upgrade" option, when you will use this option you will be given a chance to select your CSPKG and CSCFG file from local file system or from Windows Azure storage. Once you selected new or updated CSPKG, your current running service will bee upgraded.
You can also use Windows Azure PowerShell Cmdlets to upgrade your current running hosted service using "Update-Deployment" command:
2.1 http://wappowershell.codeplex.com/
You can other 3rd party applications created using Windows Azure Service Management API to upgrade/manage your current running hosted service.
3.1 http://wapmmc.codeplex.com/
3.2 http://www.cerebrata.com/Products/CloudStorageStudio/Default.aspx
Note: With Visual Studio if you again publish your application, it will delete the current running hosted service and then create the new on so for update it is not the good one.
Finally based on your question about partial update, that is not supported. Even when you make a single line change in your code the deployment will be considered a full deployment even when the action is "update/upgrade". There is no diff package deployment so evertime you update your Windows Azure application, you will use the newly created CSPKG file and upgrade your hosted application.
Regarding partial update: If you have multiple Roles, you may choose to upgrade a single role (so that would be a partial update of the deployment). For a given Role, all code is redeployed. If you're running more than one instance, the update will be rolled out across groups of instances, not all instances at once.
For updates such as static content: if you move these into blob storage (a great place for css, jquery, images, etc.), then you may update this content by simply uploading new items to blob storage individually. These updates don't require any code to be rebuilt or redeployed.
If you're in dev mode (e.g. non-production), you may enable Web Deploy, which then allows very fast updates of your app to the running instance. This only works in single-instance mode, and it's great when doing frequent code+test cycles.

How to actually configure debugging in CFBuilder

I have ColdFusion Builder 2.0.0 installed and I am trying to look at the much vaunted step debugging. However, I cannot seem to get it to work as I don't have my site / JRun install setup in the naive way the examples show.
I am using version 9,0,1,274733 of ColdFusion and my configuration is as follows:-
Installed as multi-server version with Jrun here:- c:\Apps\JRun4
application files are here:- d:\websites\my.website.com
web root is here d:\websites\my.website.com\www
core library of CFCs is here d:\websites\frameworks\core which is mapped in CF as core
I have read this watched this http://help.adobe.com/en_US/ColdFusionBuilder/Using/WS0ef8c004658c1089-31c11ef1121cdfd6aa0-7fff.html and this http://forta.com/blog/index.cfm/2007/5/30/CF8-Debugger-Getting-Started and watched this https://experts.adobeconnect.com/_a204547676/p33029638/?launcher=false&fcsContent=true&pbMode=normal but I get stuck at the point after you have configured RDS and you are setting up the server for your project.
Now I am pretty sure the above is correct, when I move to the next page in the wizard I get the following:-
Now I as I understand it my Server Home should be c:\Apps\JRun4 and my Document root should be d:\websites\my.website.com
This all looks like it is going to be fine until you actually try and debug when I get
followed by
I can confirm that the server is running and RDS is enabled as in the RDS Dataview I can see all my databases.
Any help would be gratefully received as this is very frustrating and the documentation is very lacking.
There is a video tutorial as well that you may want to check and see if that helps. http://blogs.adobe.com/anand/2011/01/learn-how-to-debug-coldfusion-applications-using-coldfusion-builder-2.html
You need to specify the RDS username/password and the "application server name". If you are using the base instance that was installed when you setup the multiserver install of CF that is "cfusion", otherwise its the name of the instance you are using.
The RDS username is most likely "admin" unless you setup custom users for RDS. The password is the RDS password you specified when you installed CF.

Resources