How to Set Up Azure Kubernetes Service for n8n Workflow Automation


This is Part 2 of the "Building a Production-Ready n8n Workflow Automation Platform on Azure Kubernetes Service" series. View the complete series here.
Setting Up the Foundation
Creating the AKS Cluster
The first step in our implementation is setting up the Azure Kubernetes Service (AKS) cluster. This forms the foundation of our entire deployment.
Resource Planning
Before creating the cluster, we determined our resource requirements:
- Node Count: 2 nodes for basic high availability
- VM Size: D2s v3 (2 vCPUs, 8GB RAM) for good performance
- Region: East US (chosen for proximity to our users)
- Kubernetes Version: 1.25.5 (stable version with good feature support)
Creating the Cluster with Azure CLI
We used Azure CLI for cluster creation to make the process reproducible:
# Create resource group
az group create --name n8n-aks-rg --location eastus
# Create AKS cluster
az aks create \
--resource-group n8n-aks-rg \
--name n8n-cluster \
--node-count 2 \
--node-vm-size Standard_D2s_v3 \
--kubernetes-version 1.25.5 \
--enable-managed-identity \
--generate-ssh-keys
This command creates a basic AKS cluster with managed identity for simplified authentication. The SSH keys are automatically generated for node access if needed.
Connecting to the Cluster
After cluster creation, we configured kubectl
to connect to our new cluster:
# Get credentials
az aks get-credentials --resource-group n8n-aks-rg --name n8n-cluster
# Verify connection
kubectl get nodes
The output confirmed our two nodes were running:
NAME STATUS ROLES AGE VERSION
aks-nodepool1-12345678-vmss000000 Ready agent 3m v1.25.5
aks-nodepool1-12345678-vmss000001 Ready agent 3m v1.25.5
Namespace Organization
We created a dedicated namespace for our n8n deployment to isolate it from other applications that might run on the same cluster:
# Create namespace
kubectl create namespace n8n
# Set as default namespace for this context
kubectl config set-context --current --namespace=n8n
Using a dedicated namespace provides several benefits:
- Resource isolation
- Simplified RBAC (Role-Based Access Control)
- Clear resource organization
- Ability to set resource quotas per namespace
Network Architecture
Network Considerations
In our design, we addressed several network requirements:
- External Access: The n8n UI must be accessible externally via HTTPS
- Inter-Service Communication: Components need to communicate within the cluster
- Security: Network policies to restrict unnecessary communication
Implementing the Ingress Controller
For external access, we installed the NGINX Ingress Controller:
# Add the Helm repository
helm repo add ingress-nginx https://kubernetes.github.io/ingress-nginx
helm repo update
# Install NGINX Ingress Controller
helm install nginx-ingress ingress-nginx/ingress-nginx \
--namespace default \
--set controller.replicaCount=2 \
--set controller.nodeSelector."kubernetes\.io/os"=linux \
--set defaultBackend.nodeSelector."kubernetes\.io/os"=linux
After installation, we verified the ingress controller's external IP:
kubectl get service nginx-ingress-ingress-nginx-controller
This returned our external IP (74.179.239.172) which we later used to configure DNS.
Storage Classes and Persistent Volumes
Storage Architecture
For our n8n deployment, we needed persistent storage for:
- PostgreSQL database
- Redis data (if needed for persistence)
AKS provides default storage classes that use Azure Disk or Azure File. We used the default storage class (managed-premium
) which creates Azure Premium Managed Disks.
Verifying Storage Classes
We checked available storage classes with:
kubectl get storageclass
The output confirmed the default storage class:
NAME PROVISIONER RECLAIMPOLICY VOLUMEBINDINGMODE ALLOWVOLUMEEXPANSION AGE
managed-premium (default) disk.csi.azure.com Delete Immediate true 10m
managed-csi disk.csi.azure.com Delete Immediate true 10m
Creating Persistent Volume Claims
For PostgreSQL, we created a Persistent Volume Claim (PVC) to ensure data persistence:
apiVersion: v1
kind: PersistentVolumeClaim
metadata:
name: postgres-data-claim
namespace: n8n
spec:
accessModes:
- ReadWriteOnce
resources:
requests:
storage: 64Gi
storageClassName: managed-premium
We saved this as postgres-pvc.yaml
and applied it:
kubectl apply -f postgres-pvc.yaml
This PVC would be used by the PostgreSQL StatefulSet to store database files.
Deployment Process
The overall deployment process follows this sequence:
flowchart TB
start([Start Deployment]) --> prereq[Prerequisites Check]
prereq --> aks[Create AKS Cluster]
aks --> namespace[Create n8n Namespace]
namespace --> storage[Configure Storage]
storage --> secrets[Apply Secrets]
secrets --> dbPath[Database Path]
secrets --> redisPath[Redis Path]
secrets --> n8nPath[n8n Path]
dbPath --> postgres[Deploy PostgreSQL]
redisPath --> redis[Deploy Redis]
postgres --> dbInit[Initialize Database]
redis --> queueInit[Initialize Queue]
dbInit --> n8n[Deploy n8n Main]
queueInit --> n8n
n8n --> workers[Deploy n8n Workers]
n8n --> ingress[Configure Ingress]
ingress --> certmgr[Install Cert-Manager]
certmgr --> issuer[Configure ClusterIssuer]
issuer --> cert[Obtain SSL Certificate]
cert --> validation[Validation Tests]
workers --> validation
validation --> complete([Deployment Complete])
classDef setup fill:#f9f,stroke:#333,stroke-width:1px
classDef deployment fill:#bbf,stroke:#333,stroke-width:1px
classDef config fill:#bfb,stroke:#333,stroke-width:1px
classDef validation fill:#fbf,stroke:#333,stroke-width:1px
class start,prereq,aks,namespace setup
class postgres,redis,n8n,workers deployment
class storage,secrets,ingress,certmgr,issuer,cert config
class dbInit,queueInit,validation validation
This workflow ensures that dependencies are deployed in the correct order, with each component building upon the previous ones. In the next section, we'll set up the foundational data layer with PostgreSQL and Redis.
Conclusion
With our AKS cluster provisioned and configured, we now have a solid foundation for our n8n deployment. We've set up proper namespaces, configured networking, and prepared our persistent storage requirements.
In the next article, we'll implement the data layer by deploying PostgreSQL and Redis with proper security configurations and persistence. [Continue to Part 3: Data Layer Implementation]
What challenges have you faced when setting up Kubernetes clusters for stateful applications? Share your experiences in the comments!
Subscribe to my newsletter
Read articles from Nikhil Mishra directly inside your inbox. Subscribe to the newsletter, and don't miss out.
Written by

Nikhil Mishra
Nikhil Mishra
I am a student studying in Mumbai University, learning DevOps, looking for opportunities to learn more things by gaining experience at prestigious institutions