Never ending deployment on Azure ACS Kubernetes - azure-container-service

Any idea to have a running kubernetes on Azure?
I try the deploy with the Azure UI and after 49 minutes 6 secondes the deployment is still running.
Thanks.

The deployment should take 5 to 10 minutes. There are three reasons why the deployment may take longer:
Service principal is not setup correctly: https://github.com/Azure/acs-engine/blob/master/docs/serviceprincipal.md
There is a platform issue with the VM or extension.
If you are using windows, the image may not be provisioning.
Try the deployment again, and if it falls in reason #2 or #3, please submit an incident through the portal. If you suspect different than #1-3, please submit an issue through https://github.com/azure/acs/issues.

Related

Azure App Service getting "503 Service Unavailable" after deploying through DevOps pipeline

I have asp.net core web app that has been running fine in an Azure app service for the past couple of years. I have regularly published new versions of the app using the "Publish" feature in Visual Studio 2019, and have never received the 503 error while the app service is running. I recently set up a multi-stage release pipeline for the app in Azure DevOps. The first stage of the pipeline deploys the build to an Azure app service that we use as a test environment. The second stage has an approval condition set up on it and after the approval is completed, it deploys the build to the app service in our production environment.
This is my first deployment using the pipeline. For the past couple weeks, I've deployed builds through the first stage of the pipeline and tested the app service with no problems. When we approved the build for Stage 2, the pipeline completed successfully (with no errors in the logs) and the build was deployed to our production environment. However, when I try to access the app service I receive a "503 Service Unavailable" error.
I have tried restarting the app service (several times), and compared the configurations between our test and production environments, but did not find any significant differences between the them. I compared the pipeline stages and the only difference I saw was the app services they deployed to.
Unable to resolve the issue, I ended up setting the "WEBSITE_RUN_FROM_PACKAGE" setting to "0" and published the app from Visual Studio, instead of using the pipeline. This was done only as a temporary work-around, so our users would not experience an extended outage while I work to resolve the pipeline issue.
So, now I need to figure out what is causing the 503 error when I deploy through the pipeline. If anyone has some suggestions on how I can fix the issue, or point me in the direction of what to look for, I would greatly appreciate the help.

IIS Restart Web API in ordered

When IIS Server is the outage, the server is restarted and all applications and web APIs are restarted as well, the problem I'm facing that some Web API depends on BUS. Could we let site/app wait until BUS ready before starting from IIS without touching the application code? In Docker, we can use the WAIT command or other third parties to wait until service is available before starting a container.
The Web API we are built on .Net Core 3.1
Any help is appreciated.
During the startup could you use something like
https://github.com/App-vNext/Polly to enter a period of retry
(https://github.com/App-vNext/Polly#retry). This would allow you to
retry the call to the API until it's hopefully it's available.
Use a windows service to monitor heartbeats of the applications and trigger application pool restarts on applications not working correctly. This should help you get into a running state.
Ultimately I'd try and remove this dependency, if you could give a little more information around the webapi requirements I'd be happy to suggest more ideas.

Automatic cluster setup and app deployment on GCE Kubernetes

We are looking for a solid, declarative (yaml), based proceedure to automate the setup of our Kubernetes cluster and application deployments on Google Container Engine.
As our last resort in a serious failure we want to be able to:
Create a new GCE cluster
Execute all our deployments to their latest versions
Execute all the steps in the correct order
What are the solutions people are currently using. Doing this manually takes us about an hour and is error prone. Really it could take 15-20 mins if automated.
You should take a look at Google Cloud Deployment Manager. It "automates the creation and management of your Google Cloud Platform resources for you" meaning that it can create a Google Container Engine cluster as well as create your deployments.
Looking through the GKE deployment manager example should help get you started.

Azure diagnostics with WCF hosted in azure

Could someone please confirm whether azure diagnostics is possible for WCF hosted in azure?
I followed this guide: http://blogs.msdn.com/b/sajid/archive/2010/04/30/basic-azure-diagnostics.aspx
After doing Trace.WriteLine("Information","testing").
I was expecting a WADTable on azure storage, but not appearing.
Thanks
What was your transfer period, filter level, and how long did you wait to see it appear? Do you have the AzureDiagnostics trace listener in your config file (or added programmatically). Nothing willl appear if you are not using the listener.
Diagnostics by default are aggregated locally and will never appear in your storage account unless you actually initiate a transfer with the correct filter level (Undefined will do it all). There are billing reasons why that is the case, btw (it costs you money in Tx and storage).
This blog post is about 18 months old and there have been some breaking changes for Windows Azure Diagnostics since then from SDK perspective. Please refer to the latest SDK documentation or these blog posts:
http://convective.wordpress.com/2010/12/01/configuration-changes-to-windows-azure-diagnostics-in-azure-sdk-v1-3/
http://blog.bareweb.eu/2011/03/implementing-azure-diagnostics-with-sdk-v1-4/

Azure Deployment Problem: One instance of an Azure Role starts, the other is "waiting for host" indefinitely - what can cause this?

I have a fairly simple WCF Role on Azure that I am trying to deploy two instances of - the role is fairly-well united tested and I've been able to run it successfully on the local emulator with no readily apparent issues.
The role has a couple of startup tasks that run in the background as the role starts, namely the installation of a pair of Windows Services that run asynchronously in the background doing data processing.
When I deploy the Azure, the first role instance boots normally and quickly, with the Windows Services successfully installed and running (I RDPed in to verify.) The second instance permanently hangs in a "Waiting for Host" state. I've tried rebooting that role instance individually and it doesn't appear to fix the problem.
I've also tried redeploying the entire package to Azure with the same results - first role instance starts fine, the second hangs.
What can cause this problem? Where should I look to try to fix the issue?
So ultimately I resolved this issue by turning to Windows Azure support - turns out I was being consistently deployed onto a "bad node" which I didn't think was possible given how the Fabric Controller works.
Nonetheless, if you run into this issue - Azure support is your best bet.