How I can update the new code to my site in windows azure cloud? - asp.net-mvc-3

I am new windows azure user. I have gotten selected for 90 days trial account and I am able to upload my ASP.NET MVC3 application to my account. My site is also running now. After I did publish my site, I added more model, views and controller to my proramme. Now I can not find a way to update my application. I can again publish my application but update option is not there. I want to update my new code only but the package option is creating full application. How I can update the new code to my site in windows azure cloud?
[Changed spelling]

With Windows Azure you can publish/update an application following ways:
Log into you Windows Azure account. Select you hosted server name and at the top panel you will see "Upgrade" option, when you will use this option you will be given a chance to select your CSPKG and CSCFG file from local file system or from Windows Azure storage. Once you selected new or updated CSPKG, your current running service will bee upgraded.
You can also use Windows Azure PowerShell Cmdlets to upgrade your current running hosted service using "Update-Deployment" command:
2.1 http://wappowershell.codeplex.com/
You can other 3rd party applications created using Windows Azure Service Management API to upgrade/manage your current running hosted service.
3.1 http://wapmmc.codeplex.com/
3.2 http://www.cerebrata.com/Products/CloudStorageStudio/Default.aspx
Note: With Visual Studio if you again publish your application, it will delete the current running hosted service and then create the new on so for update it is not the good one.
Finally based on your question about partial update, that is not supported. Even when you make a single line change in your code the deployment will be considered a full deployment even when the action is "update/upgrade". There is no diff package deployment so evertime you update your Windows Azure application, you will use the newly created CSPKG file and upgrade your hosted application.

Regarding partial update: If you have multiple Roles, you may choose to upgrade a single role (so that would be a partial update of the deployment). For a given Role, all code is redeployed. If you're running more than one instance, the update will be rolled out across groups of instances, not all instances at once.
For updates such as static content: if you move these into blob storage (a great place for css, jquery, images, etc.), then you may update this content by simply uploading new items to blob storage individually. These updates don't require any code to be rebuilt or redeployed.
If you're in dev mode (e.g. non-production), you may enable Web Deploy, which then allows very fast updates of your app to the running instance. This only works in single-instance mode, and it's great when doing frequent code+test cycles.

Related

Files not updating on google cloud platform instance

i am using google cloud platform and i have created an instance from compute engine. i installed apache server and then fresh laravel installation using ssh. All my laravel files exist on this path
var/www/html
but now when i am trying to edit any file its not reflecting. When i access my site using this link
https://project-id.appspot.com/
it only display a fresh laravel installation. not reflecting new changes.
I am using filezilla to update files.
The link you have provided is the URL format used by App Engine applications, not a Compute Engine instance. I believe you may be confusing the two.
To view the changes you have done to the files in a Compute Engine instance, you have to access the external IP of said instance, just as you would with a regular machine or VM.
Therefore, navigate to the Compute Engine section of the Cloud Console and look for the external IP of the Compute Engine instance where you have installed Laravel.

Blob Trigger not working when published VS2017 .NET Core 2.0 Azure Function v2

When I create a Azure Function with VS2017 in .NET Core 2.0, running locally, my blob trigger works fine. But when I publish to Azure, the trigger does not fire and no logs are written to the Azure Portal console.
My Function code.
My local.settings.json
After publishing, I go to the Azure portal and see this error: "The binding type(s) 'blobTrigger' are not registered.
I then attempt to install the Microsoft.Azure.WebJobs.Extensions.Storage but it fails. (I got here by trying to create a new trigger on the portal.)
If I keep trying, eventually it seems to install. But my published trigger doesn't respond to any file drops in my Blob Storage. The Function App's application setting, FUNCTIONS_EXTENSION_VERSION = beta and the run time is 2.x.
Has anyone gotten this to work?
There is a breaking change for Azure Storage bindings for 2.x runtime:
The Azure Storage bindings (blob, queues and tables triggers and bindings) were previously part of the runtime, but moving forward, will need to be explicitly installed. This change was made to allow Azure Functions to support multiple versions of the Azure Storage SDK
To temporarily avoid impact, you can pin your Function App to the previous version of the runtime by updating the app setting FUNCTIONS_EXTENSION_VERSION to 2.0.11961-alpha
https://github.com/Azure/app-service-announcements/issues/129

MVC 3, IIS 8.5 deployment = 404 - File or directory not found

I have an existing deployed ASP.NET MVC3 application using Entity Framework 4.1 and .NET 4 that I need to copy onto a new server. Problem is I have little or no documentation about how the deployed environment was configured. On the up side I do administrator access to the deployment environment and full access to the source code to the application.
So starting from bare metal I have:
Fresh install of Windows Server 2012, SQL Server 2014 Enterprise
Installed MVC 3 and Tools via Web Platform Installer
Copied entire website from existing deployment to its own folder under C:\inetpub
Added website under IIS and started it with the App Pool configured as
Browse to website and it returns the default.htm page in the root folder. This is not expected behaviour as it should serve up the default controller view.
Check to see that I can access a html file elsewhere within the website eg. http://myhost.com:443/SomeDir/test.htm and it serves that up
See if it will hit a route I know exists eg. http://myhost.com:443/Auth/NewLogin and I get
Various readings indicate that occasionally this is caused by "UrlRoutingModule-4.0" not being installed in modules. It wasn't so I added same server wide and ensure it was present for the website (Made no difference):
There is also mention of ensuring runAllManagedModulesForAllRequests="true" was present in the web.config which it is
UPDATE
.NET is installed:
So before I post a "Failed Request Trace" can anyone give me any suggestions on how I proceed to get MVC working in this environment?
I have read may SO posts and Googled but nothing seems to work.

Copying a TFS 2010 instance using Team Project Collection Cloning?

We are looking to create a test TFS 2010 server based on our live instance.
One method which has been suggested is to clone the Team Project Collection (TPC) onto to another server - as detailed in this existing answer but I think there are a few additional steps?
In order to get the cloned TPC's GUID reset, I take it we would have to first reattach the cloned TPC in the admin console on the original server then detach, move and reattach on to test Server/TFS instance.
We are not running Sharepoint/WSS but would there be additional config work required on the test server with SSRS - in order for new projects to be created against the cloned TPC?
Are there additional using diffrent AD accounts for services or can all of that be resolved within the admin console on the new server?
Both servers will running on VMWare and on the same domain but different AD accounts would be used on the two servers to help prevent any unwanted interactions between the TFS instances.
I will recommended convert your TFS to virtual environment P2V using SCVMM, see this article,
http://mohamedradwan.wordpress.com/2011/06/23/converting-my-physical-domain-controller-to-a-virtual-machine-p2v/

WindowsAzure: Remote Desktop, what can I do?

I'm connected to one of my WebRole instances. But what can I do now? I startet the Server Manager and navigated to Roles -> WebServer -> RD00123..., but what I see seems not to be the active instance. When I click on "Browse WebSite" I don't see the deployed site even I can't see or edit for instance the configured rewrite rules (defined in the web.config).
It's also funny that the inetpub is on systemdrive D (d:\inetpub\wwwroot) but the deployed (editable) files are on E (e:\approot). I'm wondering how this comes togehter.
Ok the things I want to know are:
How can I launch the unique instance version of the web role (I can't find an IP or port)?
How can I assume temporary debug changes in the web.config?
How can I restore the deployed version of the instance?
1) I don't believe you can. This is managed by the Azure Fabric controller. The most you can do is force an individual instance (the one you've RDP'd into) to reboot/restart.
2) Yes. But they won't be reverted unless the VM is recycled, so be prepared to back them out yourself. This is a good use of RDP
3) Not directly. Again, this is controlled by the Azure Fabric controller. But you could put the original bits you had back into the VM manually.
I'm not sure I understand the questions regarding server manager, but I can answer the question about where things live. The "base" Windows Azure VM image is the root drive where you see INetPub at. The application is deployed to a seperate VHD (thus its presence on another drive). If you request local storage, that will come from yet another virtual drive. This is just how the Azure Fabric manages the application deployments. Properly constructed, the roles of course won't care where things live.

Resources