WanderLust is a simple MERN travel blog website ✈ This project is aimed to help people to contribute in open source, upskill in react and also master git.
In this demo, we will see how to deploy an end to end three tier MERN stack application on EKS cluster.
- GitHub (Code)
- Docker (Containerization)
- Jenkins (CI)
- OWASP (Dependency check)
- SonarQube (Quality)
- Trivy (Filesystem Scan)
- ArgoCD (CD)
- Redis (Caching)
- AWS EKS (Kubernetes)
- Helm (Monitoring using grafana and prometheus)
Important
Below table helps you to navigate to the particular tool installation section fast.
Tech stack | Installation |
---|---|
Jenkins Master | Install and configure Jenkins |
eksctl | Install eksctl |
Argocd | Install and configure ArgoCD |
Jenkins-Worker Setup | Install and configure Jenkins Worker Node |
OWASP setup | Install and configure OWASP |
SonarQube | Install and configure SonarQube |
Email Notification Setup | Email notification setup |
Monitoring | Prometheus and grafana setup using helm charts |
Clean Up | Clean up |
- Root user access
sudo su
Note
This project will be implemented on North California region (us-west-1).
- Create 1 Master machine on AWS with 2CPU, 8GB of RAM (t2.large) and 29 GB of storage and install Docker on it.
- Open the below ports in security group of master machine and also attach same security group to Jenkins worker node (We will create worker node shortly)
Note
We are creating this master machine because we will configure Jenkins master, eksctl, EKS cluster creation from here.
Install & Configure Docker by using below command, "NewGrp docker" will refresh the group config hence no need to restart the EC2 machine.
apt-get update
apt-get install docker.io -y
usermod -aG docker ubuntu && newgrp docker
- Install and configure Jenkins (Master machine)
sudo apt update -y
sudo apt install fontconfig openjdk-17-jre -y
sudo wget -O /usr/share/keyrings/jenkins-keyring.asc \
https://pkg.jenkins.io/debian-stable/jenkins.io-2023.key
echo "deb [signed-by=/usr/share/keyrings/jenkins-keyring.asc]" \
https://pkg.jenkins.io/debian-stable binary/ | sudo tee \
/etc/apt/sources.list.d/jenkins.list > /dev/null
sudo apt-get update -y
sudo apt-get install jenkins -y
- Now, access Jenkins Master on the browser on port 8080 and configure it.
-
Create EKS Cluster on AWS (Master machine)
- IAM user with access keys and secret access keys
- AWSCLI should be configured (Setup AWSCLI)
curl "https://awscli.amazonaws.com/awscli-exe-linux-x86_64.zip" -o "awscliv2.zip" sudo apt install unzip unzip awscliv2.zip sudo ./aws/install aws configure
- Install kubectl (Master machine)(Setup kubectl )
curl -o kubectl https://amazon-eks.s3.us-west-2.amazonaws.com/1.19.6/2021-01-05/bin/linux/amd64/kubectl chmod +x ./kubectl sudo mv ./kubectl /usr/local/bin kubectl version --short --client
- Install eksctl (Master machine) (Setup eksctl)
curl --silent --location "https://github.com/weaveworks/eksctl/releases/latest/download/eksctl_$(uname -s)_amd64.tar.gz" | tar xz -C /tmp sudo mv /tmp/eksctl /usr/local/bin eksctl version
- Create EKS Cluster (Master machine)
eksctl create cluster --name=wanderlust \ --region=us-west-1 \ --version=1.30 \ --without-nodegroup
- Associate IAM OIDC Provider (Master machine)
eksctl utils associate-iam-oidc-provider \ --region us-west-1 \ --cluster wanderlust \ --approve
- Create Nodegroup (Master machine)
eksctl create nodegroup --cluster=wanderlust \ --region=us-west-1 \ --name=wanderlust \ --node-type=t2.large \ --nodes=2 \ --nodes-min=2 \ --nodes-max=2 \ --node-volume-size=29 \ --ssh-access \ --ssh-public-key=eks-nodegroup-key
Note
Make sure the ssh-public-key "eks-nodegroup-key is available in your aws account"
- Setting up jenkins worker node
- Create a new EC2 instance (Jenkins Worker) with 2CPU, 8GB of RAM (t2.large) and 29 GB of storage and install java on it
sudo apt update -y sudo apt install fontconfig openjdk-17-jre -y
-
Create an IAM role with administrator access attach it to the jenkins worker node Select Jenkins worker node EC2 instance --> Actions --> Security --> Modify IAM role
-
Configure AWSCLI (Setup AWSCLI)
sudo su
curl "https://awscli.amazonaws.com/awscli-exe-linux-x86_64.zip" -o "awscliv2.zip" sudo apt install unzip unzip awscliv2.zip sudo ./aws/install aws configure
- generate ssh keys (Master machine) to setup jenkins master-slave
ssh-keygen
- Now move to directory where your ssh keys are generated and copy the content of public key and paste to authorized_keys file of the Jenkins worker node.
- Now, go to the jenkins master and navigate to Manage jenkins --> Nodes, and click on Add node
- name: Node
- type: permanent agent
- Number of executors: 2
- Remote root directory
- Labels: Node
- Usage: Only build jobs with label expressions matching this node
- Launch method: Via ssh
- Host: <public-ip-worker-jenkins>
- Credentials: Add --> Kind: ssh username with private key --> ID: Worker --> Description: Worker --> Username: root --> Private key: Enter directly --> Add Private key
- Host Key Verification Strategy: Non verifying Verification Strategy
- Availability: Keep this agent online as much as possible
- Install docker (Jenkins Worker)
apt install docker.io -y
usermod -aG docker ubuntu && newgrp docker
- Install and configure SonarQube (Master machine)
docker run -itd --name SonarQube-Server -p 9000:9000 sonarqube:lts-community
- Install Trivy (Jenkins Worker)
sudo apt-get install wget apt-transport-https gnupg lsb-release -y
wget -qO - https://aquasecurity.github.io/trivy-repo/deb/public.key | sudo apt-key add -
echo deb https://aquasecurity.github.io/trivy-repo/deb $(lsb_release -sc) main | sudo tee -a /etc/apt/sources.list.d/trivy.list
sudo apt-get update -y
sudo apt-get install trivy -y
- Install and Configure ArgoCD (Master Machine)
- Create argocd namespace
kubectl create namespace argocd
- Apply argocd manifest
kubectl apply -n argocd -f https://raw.githubusercontent.com/argoproj/argo-cd/stable/manifests/install.yaml
- Make sure all pods are running in argocd namespace
watch kubectl get pods -n argocd
- Install argocd CLI
curl --silent --location -o /usr/local/bin/argocd https://github.com/argoproj/argo-cd/releases/download/v2.4.7/argocd-linux-amd64
- Provide executable permission
chmod +x /usr/local/bin/argocd
- Check argocd services
kubectl get svc -n argocd
- Change argocd server's service from ClusterIP to NodePort
kubectl patch svc argocd-server -n argocd -p '{"spec": {"type": "NodePort"}}'
- Confirm service is patched or not
kubectl get svc -n argocd
- Check the port where ArgoCD server is running and expose it on security groups of a worker node
- Access it on browser, click on advance and proceed with
<public-ip-worker>:<port>
- Fetch the initial password of argocd server
kubectl -n argocd get secret argocd-initial-admin-secret -o jsonpath="{.data.password}" | base64 -d; echo
- Username: admin
- Now, go to User Info and update your argocd password
- Go to your Jenkins Master EC2 instance and allow 465 port number for SMTPS
- Now, we need to generate an application password from our gmail account to authenticate with jenkins
- Open gmail and go to Manage your Google Account --> Security
Important
Make sure 2 step verification must be on
- Once, app password is create and go back to jenkins Manage Jenkins --> Credentials to add username and password for email notification
- Scroll down and search for E-mail Notification and setup email notification
Important
Enter your gmail password which we copied recently in password field E-mail Notification --> Advance
- Go to Jenkins Master and click on Manage Jenkins --> Plugins --> Available plugins install the below plugins:
- OWASP
- SonarQube Scanner
- Docker
- Pipeline: Stage View
-
Configure OWASP, move to Manage Jenkins --> Plugins --> Available plugins (Jenkins Worker)
-
After OWASP plugin is installed, Now move to Manage jenkins --> Tools (Jenkins Worker)
- Login to SonarQube server and create the credentials for jenkins to integrate with SonarQube
- Go to Manage Jenkins --> credentials and add Github credentials to push updated code from the pipeline:
Note
While adding github credentials add Personal Access Token in the password field.
- Now, go to github repository and under Automations directory update the instance-id field on both the updatefrontendnew.sh updatebackendnew.sh with the k8s worker's instance id
- Navigate to Manage Jenkins --> credentials and add credentials for docker login to push docker image:
- Provide permission to docker socket so that docker build and push command do not fail (Jenkins Worker)
chmod 777 /var/run/docker.sock
- Go to Master Machine and add our own eks cluster to argocd for application deployment using cli
- Login to argoCD from CLI
argocd login 52.53.156.187:32738 --username admin
Tip
52.53.156.187:32738 --> This should be your argocd url
- Check how many clusters are available in argocd
argocd cluster list
- Get your cluster name
kubectl config get-contexts
- Add your cluster to argocd
argocd cluster add [email protected] --name wanderlust-eks-cluster
Tip
[email protected] --> This should be your EKS Cluster Name.
Note
Connection should be successful
- Now, go to Applications and click on New App
Important
Make sure to click on the Auto-Create Namespace option while creating argocd application
- Congratulations, your application is deployed on AWS EKS Cluster
- Open port 31000 and 31100 on worker node and Access it on browser
<worker-public-ip>:31000
How to monitor EKS cluster, kubernetes components and workloads using prometheus and grafana via HELM (On Master machine)
-
Install Helm Chart
curl -fsSL -o get_helm.sh https://raw.githubusercontent.com/helm/helm/main/scripts/get-helm-3
chmod 700 get_helm.sh
./get_helm.sh
- Add Helm Stable Charts for Your Local Client
helm repo add stable https://charts.helm.sh/stable
- Add Prometheus Helm Repository
helm repo add prometheus-community https://prometheus-community.github.io/helm-charts
- Create Prometheus Namespace
kubectl create namespace prometheus
kubectl get ns
- Install Prometheus using Helm
helm install stable prometheus-community/kube-prometheus-stack -n prometheus
- Verify prometheus installation
kubectl get pods -n prometheus
- Check the services file (svc) of the Prometheus
kubectl get svc -n prometheus
- Expose Prometheus and Grafana to the external world through Node Port
Important
change it from Cluster IP to NodePort after changing make sure you save the file and open the assigned nodeport to the service.
kubectl edit svc stable-kube-prometheus-sta-prometheus -n prometheus
- Verify service
kubectl get svc -n prometheus
- Now,let’s change the SVC file of the Grafana and expose it to the outer world
kubectl edit svc stable-grafana -n prometheus
- Check grafana service
kubectl get svc -n prometheus
- Get a password for grafana
kubectl get secret --namespace prometheus stable-grafana -o jsonpath="{.data.admin-password}" | base64 --decode ; echo
Note
Username: admin
- Delete eks cluster
eksctl delete cluster --name=wanderlust --region=us-west-1