Complete Azure deployment takes a while for a reason, but here's what I do.
I stop the deployment, then while it is in "stopped" state I click "upgrade" and choose the service package. The deployment gets into "upgrading" state and it takes up to three minutes to complete and after that it gets into "stopped" state. I have to click "start" and wait for several more minutes to start the deployment.
I don't get it. The deployment doesn't need to be stopped prior to upgrading - it is already stopped. The deployment isn't started after upgrading. What is taking three minutes then?
Why are you stopping the deployment? The point of the upgrade is to get zero-downtime. You should simply choose upgrade and either do a manual or automatic rolling upgrade across your upgrade domains. Only new deployments require a stop/delete.
I think the reason is because the automated upgrade is careful to make sure that you could start the Azure service again at any time during the upgrade - so it still takes its time upgrading the individual upgrade zones separately.
Related
Running an EC2 instance with Server 2012. Windows is set to Automatically Install Updates. I am not sure what update is causing it, but it was installed after December 1. When this update reboots the server, the server goes into a boot loop where it comes back up for 20 seconds, then goes back down again. I can't RDP to it given the short window. Restoring the root drive back to the December 1 version works, but with auto-updates installed, I am afraid it will go down again. I am disabling auto updates for now.
Here is a screen shot of the 12/1 drive version and "Check for Updates" ran. There are 2 "important" which I assume are the only ones that are auto-installed, and 3 option.
Important: https://imgur.com/a/Nw8cmDY
Optional: https://imgur.com/a/jK9P3V5
I don't think any of these have a known issue of causing the server to be stuck in a boot loop. If you know which updates were installed at that date, then why not remove those updates and see if it corrects your issue?
What events are triggering in the event viewer when the machine goes down?
We're running a chain of updates and installs via a Task Sequence, SCCM 2012 R2 and a mix of 7 and 10 labs. Some of the stations reported success in one deployment I made, others are reporting success in a deployment my coworker made (we were both testing at the same time and didn't communicate well enough).
While the machines have been removed from the deployment, Deployment Status still reports them in one deployment or the other. We'd like to remove the stations that Errored in one group, as they're reporting as Success in the other.
Is there any way to remove records from a Deployment Status for a collection in Monitoring?
I've been unsuccessful in my searches, won't be surprised if the answer is no, and sorry if this is a duplicate, I searched before asking.
It is possible to clear out status messages from sccm for a given period but not for specific systems that i know of. SCCM probably saves these in the database, You could try to delete them from there.
Has anyone else encountered this problem:
Every month I apply windows updates to servers using SCCM Software Update Groups. Some servers are considered lower priority so I push the updates as required to the server and expect the updates to install and the server to reboot if necessary during its assigned maintenance window only to find out that the some of the updates are failing. With experience, I have found this is because the system is waiting for a reboot. I would expect that SCCM would know that there is a pending reboot and reboot the server during the maintenance window to finish applying the updates but it does not. It seems as though these are "pending reboots" that SCCM cannot detect.
As a result, this requires manual intervention each month on a dozen or more servers that have to be manually rebooted in the middle of the night so as to not interrupt production.
One of the biggest culprits to this issue is the monthly Malicious Software Removal Tool. It always seems to fail to apply then works after a reboot.
The Computer Restart related setting can be configured in the "Client Setting" node on your console. No matter if you determine to use the Default client setting or the custom client settings, you should make sure that that the value for the restart temporary notification interval and the value for the final countdown interval are shorter in duration than the shortest maintenance window that is applied to the computer (the default values are 90 and 15 mins). This is important for the deployments which require a reboot completed on your clients.
Additionally, you can examine the logs on the client-side as below:
Update deployments related logs: UpdatesDeployment.log, WindowsUpdate.log
Reboot & Maintenance related logs:RebootCoordinator.log,ServiceWindowManager.log
More details about how to track the Update deployment process in ConfigMgr can be found here.
I've created a bot on my local server that will build periodically every day. When I do a manual integration (Integrate now) it works fine. Everything builds.
However, when the bot should build on a specific time it won't start integrating.
When I view the bot in the browser it gets stuck on:
the bot has been queued and will integrate shortly
and nothing happens. I've tried hourly and daily.
Even tried restarting the computer and server. Any ideas?
After a bit of tinkering back and forth I now managed to fix the problem. It was the easiest solution ever. Just delete old bots and create new ones. I thought I did this, but maybe I didn't.
Anyway it now builds as it should.
First off, I'm very new to Azure.
I've successfully deployed an ASP.NET MVC 3 web application to Azure, using a web role. The app uses Entity Framework and SQL Azure.
Recently I've done some changes (some including adding appsettings), and tried to upgrade the application. When upgrading, it took quite a long time, before Aborting. I've always deployed through the management portal Silverlight application at http://windows.azure.com.
When trying again to no avail, I setup remote desktop and deployed again. The remote desktop session was extremely slow, and it turned out to be because WaIISHost was putting the CPU to 100%.
The IIS Manager shows that the application is deployed and 'started', however I cannot navigate to the site in the VM, and the deployment constantly seems to be trying to update without success and eventually aborting and retrying, (as I write this, it's currently Busy and Waiting for role to start...).
Does anyone have any ideas as to what the problem could be?
I believe all the right dependencies are set to copy local, which is a possible problem. It is extremely hard to debug this issue, as the remote desktop session hangs so often due to the 100% CPU utilization, and the recycling/restarting/reupdating of the web role from time to time.
Thanks,
James
P.S. Hope some of that made at least some sense...
I doubt that there's something doing in your WebRole.OnStart and/or Run, which caused the WaIISHost uses 100% CPU. Can you remove all codes from the WebRole.OnStart and/or Run and try again.
And it might be helpful to turn on the IntelliTrace when deploying, so that you can download the trace and find out any exceptions occurred when your application started, even before the website started.