This repo contains the implementation that backs the Deploy microservices with Azure Container Apps article in the Azure architecture center. It's encouraged that you read through that guidance before proceeding here as it gives you an overview of the scenario and architecture; this repo is its deployment guide.
Fabrikam inc has created a new operations team, and under its organization is a brownfield application called Drone Delivery. This application been running for a while in Azure Kubernetes Service (AKS), and while they are obtaining the benefits of containers to run microservices and Kubernetes to host them, it has been discovered that they are not making use of any of the advance features of AKS like custom service mesh or autoscaling among others.
The team has detected an opportunity to simplify and be more efficient at the DevOps level, and this is why they are now looking into Azure Container Apps to evaluate hosting Fabrikam Drone Delivery. This will allow them to publish and run containerized microservices at scale, faster than before, reducing the complexity, saving resources by using scale-to-zero, built-in autoscaling capability, and without losing all the container advantages they love.
Azure Container Apps is a fully managed environment that enables you to run microservices and containerized applications on a serverless platform. Technically speaking, it's an app-centric abstraction on top of AKS, with native features such as KEDA and Dapr integrated.
This repository guides you through the process of running a single workload composed of multiple microservices in Azure Container Apps. In this example scenario, the Fabrikam Drone Delivery app that was previously running in Azure Kubernetes Services will be run in a newly created Azure Container App environment. This application platform is optimized for running applications that span multiple microservices. This example will make some containers internet-facing via an HTTPS ingress, and internally accessible thanks to its built-in DNS-based service discovery capability. Additionally, it will manage their secrets in a secure manner and authenticate against Azure Key Vault resources using managed identities.
For more information on how the Container Apps features are being used in this reference implementation, please take a look below:
- HTTPS ingress, this allows to expose the Ingestion service to internet.
- Internal service discovery, Delivery, DroneScheduler and Package services must be internally reachable by Workflow service
- Use user-assigned identities when authenticating into Azure KeyVault from Delivery and DroneScheduler services
- Securely manage secrets for Package, Ingestion and Workflow services
- Run containers from any registry, the Fabrikam Drone Delivery uses Azure Container Registry (ACR) to publish its Docker images
- Use Revisions in Azure Container Apps to safely deploy updates, where appropriate. ❗ Workflow Service is a message consumer app, so it needs to be deployed in single revision mode, otherwise an old version could still process a message if it happens to be the one that retrieves it first.
- Use ARM templates to deploy my application, there is no need for another layer of indirection like Helm charts. All the Drone Delivery containers are part of the ARM templates
- Logs, see the container logs directly in Log Analytics without configuring any provider from code or Azure service.
Following the steps below will result in the creation of the following Azure resources that will be used throughout this example scenario.
Object | Purpose |
---|---|
An Azure Container App Environment | This is the managed Container App environment where Container Apps are deployed |
Five Azure Container Apps | These are the Azure resources that represents the five Fabrikam microservices in the Azure Container App environment |
An Azure Container Registry | This is the private container registry where all Fabrikam workload images are uploaded and later pulled from the different Azure Container Apps |
An Azure Log Analytics Workspace | This is where all the Container Apps logs are sent, along with Azure Diagnostics on all services |
An Azure Application Insights instance | All services are sending trace information to a shared Azure Application Insights instance |
Two Azure Cosmos DB instances | Delivery and Package services have dependencies on Azure Cosmos DB |
An Azure Redis Cache instance | Delivery service uses Azure Redis cache to keep track of inflight deliveries |
An Azure Service Bus | Ingestion and Workflow services communicate using Azure Service Bus queues |
Five Azure User Managed Identities | These are going to give Read and List secrets permissions over Azure Key Vault to the microservices. |
Five Azure Key Vault instances | Secrets are saved into Azure Key Vault instances. |
-
An Azure subscription in which you have at least Contributor access to. You can open an account for free.
-
Latest version of Azure CLI installed or you can perform this from Azure Cloud Shell by clicking below.
-
The required resource providers registered.
Microsoft.App
Microsoft.Cache
Microsoft.DocumentDB
Microsoft.KeyVault
Microsoft.ServiceBus
-
Set environment variables.
export LOCATION=eastus2
-
Clone this repository.
git clone --recurse-submodules https://github.com/mspnp/container-apps-fabrikam-dronedelivery.git
💡 The steps shown here and elsewhere in the reference implementation use Bash shell commands. On Windows, you can install Windows Subsystem for Linux.
-
Navigate to the container-apps-fabrikam-dronedelivery folder
cd ./container-apps-fabrikam-dronedelivery
-
Log into Azure from the CLI.
az login
-
Create a resource group for your deployment.
export PREREQS_DEPLOYMENT_NAME=workload-stamp-prereqs az deployment sub create --name $PREREQS_DEPLOYMENT_NAME --location ${LOCATION} --template-file ./workload/workload-stamp-prereqs.bicep --parameters resourceGroupLocation=${LOCATION}
-
Get the user identities.
export DELIVERY_PRINCIPAL_ID=$(az identity show -g rg-far2-capp-shipping-dronedelivery-${LOCATION} -n uid-delivery --query principalId -o tsv) && \ export DRONESCHEDULER_PRINCIPAL_ID=$(az identity show -g rg-far2-capp-shipping-dronedelivery-${LOCATION} -n uid-dronescheduler --query principalId -o tsv) && \ export WORKFLOW_PRINCIPAL_ID=$(az identity show -g rg-far2-capp-shipping-dronedelivery-${LOCATION} -n uid-workflow --query principalId -o tsv) && \ export PACKAGE_ID_PRINCIPAL_ID=$(az identity show -g rg-far2-capp-shipping-dronedelivery-${LOCATION} -n uid-package --query principalId -o tsv) && \ export INGESTION_ID_PRINCIPAL_ID=$(az identity show -g rg-far2-capp-shipping-dronedelivery-${LOCATION} -n uid-ingestion --query principalId -o tsv) # Wait for Microsoft Entra ID propagation until az ad sp show --id $DELIVERY_PRINCIPAL_ID &> /dev/null ; do echo "Waiting for Microsoft Entra ID propagation" && sleep 5; done until az ad sp show --id $DRONESCHEDULER_PRINCIPAL_ID &> /dev/null ; do echo "Waiting for Microsoft Entra ID propagation" && sleep 5; done until az ad sp show --id $WORKFLOW_PRINCIPAL_ID &> /dev/null ; do echo "Waiting for Microsoft Entra ID propagation" && sleep 5; done until az ad sp show --id $PACKAGE_ID_PRINCIPAL_ID &> /dev/null ; do echo "Waiting for Microsoft Entra ID propagation" && sleep 5; done until az ad sp show --id $INGESTION_ID_PRINCIPAL_ID &> /dev/null ; do echo "Waiting for Microsoft Entra ID propagation" && sleep 5; done ``
-
Deploy all the dependencies of the various microservices that comprise the workload.
None of these resources are for the application platform hosting the workload, but instead are tied directly to the drone delivery workload. For example, the per-microservice Key Vault, the per-microservice data stores, the message queue, logging sinks, etc. These same resources would exist no matter if the application platform was Azure Container Apps, Kubernetes, or App Service.
# [This takes about 18 minutes.] az deployment group create -n workload-stamp -g rg-shipping-dronedelivery-${LOCATION} -f ./workload/workload-stamp.bicep -p droneSchedulerPrincipalId=$DRONESCHEDULER_PRINCIPAL_ID -p workflowPrincipalId=$WORKFLOW_PRINCIPAL_ID -p deliveryPrincipalId=$DELIVERY_PRINCIPAL_ID -p ingestionPrincipalId=$INGESTION_ID_PRINCIPAL_ID -p packagePrincipalId=$PACKAGE_ID_PRINCIPAL_ID
-
Deleting Kubernetes Enricher dependency (Optional) The Kubernetes Enricher is not needed on C# implementations when it is deployed on Container Apps, because it is a cluster abstracting solution. It will work anyway if we keep it.
sed -i '/"KubernetesEnricher": "true",/d' ./workload/src/shipping/delivery/Fabrikam.DroneDelivery.DeliveryService/appsettings.json sed -i '/"KubernetesEnricher": "true",/d' ./workload/src/shipping/workflow/Fabrikam.Workflow.Service/appsettings.json sed -i '/"KubernetesEnricher": "true",/d' ./workload/src/shipping/dronescheduler/Fabrikam.DroneDelivery.DroneSchedulerService/appsettings.json
-
Build, tag, and host the five microservice container images in ACR.
ACR_NAME=$(az deployment group show -g rg-shipping-dronedelivery-${LOCATION} -n workload-stamp --query properties.outputs.acrName.value -o tsv) ACR_SERVER=$(az acr show -n $ACR_NAME --query loginServer -o tsv) # [This takes about 10 minutes.] az acr build -r $ACR_NAME -t $ACR_SERVER/shipping/delivery:0.1.0 ./workload/src/shipping/delivery/. az acr build -r $ACR_NAME -t $ACR_SERVER/shipping/ingestion:0.1.0 ./workload/src/shipping/ingestion/. az acr build -r $ACR_NAME -t $ACR_SERVER/shipping/workflow:0.1.0 ./workload/src/shipping/workflow/. az acr build -r $ACR_NAME -f ./workload/src/shipping/dronescheduler/Dockerfile -t $ACR_SERVER/shipping/dronescheduler:0.1.0 ./workload/src/shipping/. az acr build -r $ACR_NAME -t $ACR_SERVER/shipping/package:0.1.0 ./workload/src/shipping/package/.
-
Get the microservices configuration details from the workload resource deployment.
AI_NAME=$(az deployment group show -g rg-shipping-dronedelivery-${LOCATION} -n workload-stamp --query properties.outputs.appInsightsName.value -o tsv) AI_KEY=$(az monitor app-insights component show -g rg-shipping-dronedelivery-${LOCATION} -a $AI_NAME --query instrumentationKey -o tsv) AI_ID=$(az monitor app-insights component show -g rg-shipping-dronedelivery-${LOCATION} -a $AI_NAME --query appId -o tsv) ACR_ID=$(az deployment group show -g rg-shipping-dronedelivery-${LOCATION} -n workload-stamp --query properties.outputs.acrId.value -o tsv) LA_WORKSPACE_ID=$(az deployment group show -g rg-shipping-dronedelivery-${LOCATION} -n workload-stamp --query properties.outputs.laWorkspace.value -o tsv) echo -e "\nCommon Config:\nAI_NAME=${AI_NAME}\nAI_KEY=${AI_KEY}\nAI_ID=${AI_ID}\nACR_ID=${ACR_ID}\nLA_WORKSPACE_ID=${LA_WORKSPACE_ID}\n" # Delivery DELIVERY_COSMOSDB_NAME=$(az deployment group show -g rg-shipping-dronedelivery-${LOCATION} -n workload-stamp --query properties.outputs.deliveryCosmosDbName.value -o tsv) DELIVERY_DATABASE_NAME="${DELIVERY_COSMOSDB_NAME}-db" DELIVERY_COLLECTION_NAME="${DELIVERY_COSMOSDB_NAME}-col" DELIVERY_COSMOSDB_ENDPOINT=$(az cosmosdb show -g rg-shipping-dronedelivery-${LOCATION} -n $DELIVERY_COSMOSDB_NAME --query documentEndpoint -o tsv) DELIVERY_REDIS_NAME=$(az deployment group show -g rg-shipping-dronedelivery-${LOCATION} -n workload-stamp --query properties.outputs.deliveryRedisName.value -o tsv) DELIVERY_REDIS_ENDPOINT=$(az redis show -g rg-shipping-dronedelivery-${LOCATION} -n $DELIVERY_REDIS_NAME --query hostName -o tsv) DELIVERY_KEYVAULT_URI=$(az deployment group show -g rg-shipping-dronedelivery-${LOCATION} -n workload-stamp --query properties.outputs.deliveryKeyVaultUri.value -o tsv) echo -e "\nDelivery Config:\nDELIVERY_COSMOSDB_NAME=${DELIVERY_COSMOSDB_NAME}\nDELIVERY_DATABASE_NAME=${DELIVERY_DATABASE_NAME}\nDELIVERY_COLLECTION_NAME=${DELIVERY_COLLECTION_NAME}\nDELIVERY_COSMOSDB_ENDPOINT=${DELIVERY_COSMOSDB_ENDPOINT}\nDELIVERY_REDIS_NAME=${DELIVERY_REDIS_NAME}\nDELIVERY_REDIS_ENDPOINT=${DELIVERY_REDIS_ENDPOINT}\nDELIVERY_KEYVAULT_URI=${DELIVERY_KEYVAULT_URI}\n" # Drone scheduler DRONESCHEDULER_COSMOSDB_NAME=$(az deployment group show -g rg-shipping-dronedelivery-${LOCATION} -n workload-stamp --query properties.outputs.droneSchedulerCosmosDbName.value -o tsv) DRONESCHEDULER_COSMOSDB_ENDPOINT=$(az cosmosdb show -g rg-shipping-dronedelivery-${LOCATION} -n $DRONESCHEDULER_COSMOSDB_NAME --query documentEndpoint -o tsv) DRONESCHEDULER_KEYVAULT_URI=$(az deployment group show -g rg-shipping-dronedelivery-${LOCATION} -n workload-stamp --query properties.outputs.droneSchedulerKeyVaultUri.value -o tsv) echo -e "\nScheduler Config:\nDRONESCHEDULER_COSMOSDB_NAME=${DRONESCHEDULER_COSMOSDB_NAME}\nDRONESCHEDULER_COSMOSDB_ENDPOINT=${DRONESCHEDULER_COSMOSDB_ENDPOINT}\nDRONESCHEDULER_KEYVAULT_URI=${DRONESCHEDULER_KEYVAULT_URI}\n" # Workflow WORKFLOW_NAMESPACE_NAME=$(az deployment group show -g rg-shipping-dronedelivery-${LOCATION} -n workload-stamp --query properties.outputs.ingestionQueueNamespace.value -o tsv) WORKFLOW_NAMESPACE_ENDPOINT=$(az servicebus namespace show -g rg-shipping-dronedelivery-${LOCATION} -n $WORKFLOW_NAMESPACE_NAME --query serviceBusEndpoint -o tsv) WORKFLOW_NAMESPACE_SAS_NAME=$(az deployment group show -g rg-shipping-dronedelivery-${LOCATION} -n workload-stamp --query properties.outputs.workflowServiceAccessKeyName.value -o tsv) WORKFLOW_NAMESPACE_SAS_KEY=$(az servicebus namespace authorization-rule keys list -g rg-shipping-dronedelivery-${LOCATION} --namespace-name $WORKFLOW_NAMESPACE_NAME -n $WORKFLOW_NAMESPACE_SAS_NAME --query primaryKey -o tsv) WORKFLOW_QUEUE_NAME=$(az deployment group show -g rg-shipping-dronedelivery-${LOCATION} -n workload-stamp --query properties.outputs.ingestionQueueName.value -o tsv) echo -e "\nWorkflow Config:\nWORKFLOW_NAMESPACE_NAME=${WORKFLOW_NAMESPACE_NAME}\nWORKFLOW_NAMESPACE_ENDPOINT=${WORKFLOW_NAMESPACE_ENDPOINT}\nWORKFLOW_NAMESPACE_SAS_NAME=${WORKFLOW_NAMESPACE_SAS_NAME}\nWORKFLOW_NAMESPACE_SAS_KEY=${WORKFLOW_NAMESPACE_SAS_KEY}\nWORKFLOW_QUEUE_NAME=${WORKFLOW_QUEUE_NAME}\n" # Package PACKAGE_MONGODB_NAME=$(az deployment group show -g rg-shipping-dronedelivery-${LOCATION} -n workload-stamp --query properties.outputs.packageMongoDbName.value -o tsv) PACKAGE_MONGODB_CONNNECTIONSTRING=$(az cosmosdb keys list --type connection-strings -g rg-shipping-dronedelivery-${LOCATION} --name $PACKAGE_MONGODB_NAME --query "connectionStrings[0].connectionString" -o tsv | sed 's/==/%3D%3D/g') echo -e "\nPackage Config:\nPACKAGE_MONGODB_NAME=${PACKAGE_MONGODB_NAME}\nPACKAGE_MONGODB_CONNNECTIONSTRING=${PACKAGE_MONGODB_CONNNECTIONSTRING}\n" # Ingestion INGESTION_NAMESPACE_NAME=$(az deployment group show -g rg-shipping-dronedelivery-${LOCATION} -n workload-stamp --query properties.outputs.ingestionQueueNamespace.value -o tsv) INGESTION_NAMESPACE_SAS_NAME=$(az deployment group show -g rg-shipping-dronedelivery-${LOCATION} -n workload-stamp --query properties.outputs.ingestionServiceAccessKeyName.value -o tsv) INGESTION_NAMESPACE_SAS_KEY=$(az servicebus namespace authorization-rule keys list -g rg-shipping-dronedelivery-${LOCATION} --namespace-name $INGESTION_NAMESPACE_NAME -n $INGESTION_NAMESPACE_SAS_NAME --query primaryKey -o tsv) INGESTION_QUEUE_NAME=$(az deployment group show -g rg-shipping-dronedelivery-${LOCATION} -n workload-stamp --query properties.outputs.ingestionQueueName.value -o tsv) echo -e "\nIngestion Config:\nINGESTION_NAMESPACE_NAME=${INGESTION_NAMESPACE_NAME}\nINGESTION_NAMESPACE_SAS_NAME=${INGESTION_NAMESPACE_SAS_NAME}\nINGESTION_NAMESPACE_SAS_KEY=${INGESTION_NAMESPACE_SAS_KEY}\nINGESTION_QUEUE_NAME=${INGESTION_QUEUE_NAME}"
If any of the config values were empty above, please stop and troubleshoot before proceeding.
-
Deploy the Container Apps ARM template.
This deploys the Azure Container Apps Environment and each microservice.
# [This takes about four minutes.] az deployment group create -f main.bicep -g rg-shipping-dronedelivery-${LOCATION} -p \ logAnalyticsResourceId=$LA_WORKSPACE_ID \ applicationInsightsInstrumentationKey=$AI_KEY \ containerRegistryResourceId=$ACR_ID \ deliveryCosmosdbDatabaseName=$DELIVERY_DATABASE_NAME \ deliveryCosmosdbCollectionName=$DELIVERY_COLLECTION_NAME \ deliveryCosmosdbEndpoint=$DELIVERY_COSMOSDB_ENDPOINT \ deliveryRedisEndpoint=$DELIVERY_REDIS_ENDPOINT \ deliveryKeyVaultUri=$DELIVERY_KEYVAULT_URI \ droneSchedulerCosmosdbEndpoint=$DRONESCHEDULER_COSMOSDB_ENDPOINT \ droneSchedulerKeyVaultUri=$DRONESCHEDULER_KEYVAULT_URI \ wokflowNamespaceEndpoint=$WORKFLOW_NAMESPACE_ENDPOINT \ workflowNamespaceSASName=$WORKFLOW_NAMESPACE_SAS_NAME \ workflowNamespaceSASKey=$WORKFLOW_NAMESPACE_SAS_KEY \ workflowQueueName=$WORKFLOW_QUEUE_NAME \ packageMongodbConnectionString=$PACKAGE_MONGODB_CONNNECTIONSTRING \ ingestionNamespaceName=$INGESTION_NAMESPACE_NAME \ ingestionNamespaceSASName=$INGESTION_NAMESPACE_SAS_NAME \ ingestionNamespaceSASKey=$INGESTION_NAMESPACE_SAS_KEY \ ingestionQueueName=$INGESTION_QUEUE_NAME
Now that you have deployed your Container Apps Environment, you can validate its functionality. This section will help you to validate the workload is exposed through a Container Apps HTTP ingress flow and responding to HTTP requests correctly.
-
Get the Ingestion service's FQDN
📖 The app team conducts a final acceptance test to ensure that traffic is flowing end-to-end as expected. To do so, an HTTP request is submitted against the ingestion external ingress.
INGESTION_FQDN=$(az deployment group show -g rg-shipping-dronedelivery-${LOCATION} -n main --query properties.outputs.ingestionFqdn.value -o tsv)
-
Create a delivery request using your microservices hosted on ACA.
This calls the only Internet-exposed service. This kicks off the five microservices to perform the request.
curl -X POST "https://${INGESTION_FQDN}/api/deliveryrequests" --header 'Content-Type: application/json' --header 'Accept: application/json' -d '{ "confirmationRequired": "None", "deadline": "", "dropOffLocation": "drop off", "expedited": true, "ownerId": "myowner", "packageInfo": { "packageId": "mypackage", "size": "Small", "tag": "mytag", "weight": 10 }, "pickupLocation": "mypickup", "pickupTime": "'$(date -u +%FT%TZ)'" }'
The response to the request printed in your terminal should look similar to the one shown below:
{"deliveryId":"5453d09a-a826-436f-8e7d-4ff706367b04","ownerId":"myowner","pickupLocation":"mypickup","pickupTime":"2023-05-14T20:00:00.000+0000","deadline":"","expedited":true,"confirmationRequired":"None","packageInfo":{"packageId":"mypackage","size":"Small","weight":10.0,"tag":"mytag"},"dropOffLocation":"drop off"}
-
Query Application Insights to ensure your request has been ingested by the underlaying services.
⏱️ It might take ten minutes for the full query results to be available.
az monitor app-insights query --app $AI_ID --analytics-query 'requests | summarize count_=sum(itemCount) by operation_Name | order by operation_Name | project strcat(operation_Name," (", count_, ")")' --query tables[0].rows[] -o table
The following output demonstrates the type of response to expect from the Application Insights query.
Result -------------------------------------------------- PUT DroneDeliveries/Put [id] (1) PUT Deliveries/Put [id] (1) PUT /api/packages/mypackage (1) POST IngestionController/scheduleDeliveryAsync (1) GET /api/packages/mypackage (1)
📖 Above result demonstrates that the HTTP request, initiated from the client, has been ingested by
IngestionController/scheduleDeliveryAsync
to be later consumed by the Workflow background service to be sent toDeliveries/Put
,/api/packages/mypackage
, andDroneDeliveries/Put
endpoints respectively. Them all are microservices running within Azure Container Apps.
If you need a restart a revision with Provision Status Failed
or for another reason you can use az cli:
az containerapp revision restart -g rg-shipping-dronedelivery --app <containerapp-name> -n <containerapp-revision-name>
-
Delete the resource group that contains all the resources
az group delete -n rg-shipping-dronedelivery-${LOCATION} -y
-
Purge deleted Key Vaults related to this deployment.
az keyvault list-deleted --query [].name -o tsv # Per Key Vault that was related to this deployment (should be five of them) az keyvault purge -n <name>
The team has been able to migrate and run Fabrikam Drone Delivery on top of Azure Container Apps. They are now laying out a new migration and modernization plan that will include:
Please see our contributor guide.
This project has adopted the Microsoft Open Source Code of Conduct. For more information see the Code of Conduct FAQ or contact opencode@microsoft.com with any additional questions or comments.