×

Create your first deployment on a Kubernetes Cluster using Nginx - Step by step guide ?

This article covers the complete steps to create your first NginX Deployment on Kubernetes. In fact, Deployments are managed by the Kubernetes Deployment controller. Deployments use a Pod template, which contains a specification for its Pods. The Pod specification determines how each Pod should look like: what applications should run inside its containers, which volumes the Pods should mount, its labels, and more. When a Deployment's Pod template is changed, new Pods are automatically created one at a time.

Nginx Ingress with Cert-manager on DigitalOcean

This article covers how to set up Nginx ingress on DigitalOcean Kubernetes with cert-manager. Popular Ingress Controllers include Nginx, Contour, HAProxy, and Traefik. Ingresses provide a more efficient and flexible alternative to setting up multiple LoadBalancer services, each of which uses its own dedicated Load Balancer.
Here, you will learn how to set up an Nginx Ingress to load balance and route external requests to backend Services inside of your Kubernetes cluster.

You also secured the Ingress by installing the cert-manager certificate provisioner and setting up a Let's Encrypt certificate for two host paths.
Most Ingress Controllers use only one global Load Balancer for all Ingresses, which is more efficient than creating a Load Balancer per every Service you wish to expose.

Helm is a package manager for managing Kubernetes. Using Helm Charts with your Kubernetes provides configurability and lifecycle management to update, rollback, and delete a Kubernetes application.

Once you’ve set up the Ingress, you’ll install Cert Manager to your cluster to be able to automatically provision Let’s Encrypt TLS certificates to secure your Ingresses.

Nginx ingress on DigitalOcean kubernetes using Helm

This article covers how to set up Nginx ingress on DigitalOcean Kubernetes using Helm. To implement this successfully, an Ingress Controller must be present; its role is to implement the rules by accepting traffic (most likely via a Load Balancer) and routing it to the appropriate Services.
Most Ingress Controllers use only one global Load Balancer for all Ingresses, which is more efficient than creating a Load Balancer per every Service you wish to expose.

To install the Nginx Ingress Controller to your cluster, run the following command:
$ helm install nginx-ingress stable/nginx-ingress --set controller.publishService.enabled=true
This command installs the Nginx Ingress Controller from the stable charts repository, names the Helm release nginx-ingress, and sets the publishService parameter to true.

Django Application with Kubernetes

This article covers deploying a Scalable and Secure Django Application. Kubernetes is a powerful open-source container orchestrator.
Kubernetes is a powerful open-source container orchestrator that automates the deployment, scaling and management of containerized applications.
Kubernetes objects like ConfigMaps and Secrets allow you to centralize and decouple configuration from your containers, while controllers like Deployments automatically restart failed containers and enable quick scaling of container replicas.
TLS encryption is enabled with an Ingress object and the ingress-nginx open-source Ingress Controller.
The cert-manager Kubernetes add-on renews and issues certificates using the free Let’s Encrypt certificate authority.

Distributed Tracing with Jaeger on Kubernetes

This article will guide you on implementing Distributed Tracing with Jaeger on Kubernetes where you will be able to find the cause of our irregular response time. 

The Jaeger agent is a network daemon that listens for spans sent over UDP, which it batches and sends to the collector. It is designed to be deployed to all hosts as an infrastructure component. 

The agent abstracts the routing and discovery of the collectors away from the client.

1. Application tracing allows you to log information when a program is executing. 

2. Tracing is used to help you identify issues with your application (performance of the app or simply debugging an issue) instead of going through stack traces.

3. At a high level, log monitoring helps to speed up the process of identifying specific exceptions, when these exceptions occur and the frequency at which they occur. 4. Additionally, it provides developers and support personnel with a greater level of visibility into the systems and applications being monitored.

5. Insufficient logging and monitoring vulnerability occur when the security-critical event is not logged off properly, and the system is not monitored. 

6. Lack of such functionalities can make malicious activities harder to detect and in turn affects the incident handling process.

Deploying PHP Guestbook application with Redis

This guide will guide you on how to Deploy #PHP Guestbook application with Redis. An effective way of Deploying PHP Guestbook application with Redis is with the use of the Kubernetes cluster.

Redis is a powerful tool for data storage and #caching. 

Redis Cluster extends the functionality by offering sharding and correlated performance benefits, linear scaling, and higher availability because of how Redis stores data. 

The data is automatically split among multiple nodes, which allows operations to continue, even when a subset of the nodes are experiencing failures or are unable to communicate with the rest of the cluster.


Benefits of using #Redis:

1. It is incredibly fast. It is written in ANSI C and runs on POSIX systems such as Linux, Mac OS X, and Solaris.

2. Redis is often ranked the most popular key/value database and the most popular NoSQL database used with containers.

3. Its caching solution reduces the number of calls to a cloud database backend.

4. It can be accessed by applications through its client API library.

5. Redis is supported by all of the popular programming languages.

6. It is open source and stable.


GlusterFS vs Ceph performance

This article will guide you on the differences between GlusterFS vs Ceph Performance. The decision on transfer sizes could itself account for Ceph running faster or slower than Gluster. 

Deciding whether to use Ceph vs. Gluster depends on numerous factors, but either can provide extendable and stable storage of your data.

Ceph and Gluster are both systems used for managing distributed storage. The goal is high performance, massive storage, and compatibility.

Ceph provides a POSIX-compliant network file system (#CephFS) that aims for high performance, large data storage, and maximum compatibility. 

GlusterFS and Ceph are two distributed persistent storage systems. #GlusterFS is at its core a network filesystem. 

Ceph is at the core an object store.

Deploy PHP application with Kubernetes on Ubuntu

This article will guide you on steps to deploy #PHP application with Kubernetes on #Ubuntu. 

Kubernetes, at its basic level, is a system for running and coordinating containerized applications across a cluster of machines. 

It is a platform designed to completely manage the life cycle of containerized applications and services using methods that provide predictability, scalability, and high availability.

#Kubernetes really shines when your #application consists of multiple services running in different containers.

Kubernetes, also referred to as K8s, is an open source system used to manage Linux Containers across private, public and hybrid cloud environments. 

In other words, Kubernetes can be used to manage microservice architectures and is deployable on most cloud providers.

Deploy Zimbra Collaboration using docker

This article will guide you on the steps to run #Zimbra on #Docker #Container. Docker enables apps to quickly assemble from components.

Easy way to fix connection timed out error during http-01 challenge propagation in Kubernetes

Best fix to "Connection timed out" error during http-01 challenge propagation.

'kubectl get' Command - Explained with Examples

This article covers kubectl Commands with Examples. In fact, Kubectl is a set of commands for controlling Kubernetes clusters. Every Kubernetes command has an API endpoint, and kubectl’s primary purpose is to carry out HTTP requests to the API.

While it is possible to issue HTTP requests yourself (e.g., using curl), kubectl is designed to make this process more comfortable and straightforward.


How to delete Resources using kubectl command ?

To remove resources from a file or stdin, use the kubectl delete command.

1. Remove a pod using the name and type listed in pod.yaml:

$ kubectl delete -f pod.yaml

2. Remove all pods and services with a specific label:

$ kubectl delete pods,services -l [label-key]=[label-value]

3. Remove all pods (including uninitialized pods):

$ kubectl delete pods --all

Difference Between LXC, LXD and Docker Containers

This article covers the differences between Docker containers, LXD, and LXC. In fact, You can use any of these containers as per your working environment and the requirements as each container technology has its own advantage and the limitations. 

Install Kubelet on Linux Mint 20 - Step by step guide ?

This article covers how to install Kubelet on a Linux Mint 20 system via the Snap Store. In fact, With the help of this amazing technology, we can easily manage the Kubernetes nodes. 

Install KDE Plasma on Ubuntu 20.04 LTS - Step by step guide ?

This article covers the process of installing the KDE Plasma desktop environment in Ubuntu 20.04 LTS Focal Fossa system. In fact, The name KDE is derived from "K Desktop Environment". KDE Plasma Desktop Environment provides Linux users on various distributions an alternative graphical interface to customize their desktop environment and applications for everyday use enhancement.

Install Juju on Ubuntu 20.04 - Step by step process ?

This article covers the Installation procedure of Juju client on the Linux system Ubuntu 20.04. In fact, Juju is a service-modeling tool for Ubuntu. Connect it to any cloud service, model your application, and press deploy—done. Juju takes care of lower-level configuration, deployments, and scaling, and even monitors your services.


Juju Components includes:

  • Charmed Operator Lifecycle Manager – This is an Orchestration and management system.
  • Charmed Operator Software Development Kit – This is a set of tools for creating and packaging operators.
  • Juju as a Service(jaas) – Managed juju service by canonical for enterprise customers.
  • Charmhub – This is a repository hosting re-usable charms

Install Juju on Linux Mint 20 - Step by step procedure ?

This article covers how you can install Juju on any Linux distribution. In fact, the Juju client is a single binary executable that is published and supported on multiple operating systems and CPU architectures. 

To Install Juju on Linux from snap, run the below command:

$ sudo snap install juju --classic


To Install Juju on macOS, run the below command:

$ brew install juju

Install Kubectl on Linux Mint 20 - Best Method ?

This article covers how you can install Kubectl on your Linux system and use it to operate your Kubernetes clusters if you are fond of using the command-line interfaces. Infact, Kubectl tool is very efficient in terms of cluster management and application deployment. 

Kubectl is a command-line tool for Kubernetes. It allows us to execute Kubernetes operations via the API. We can use Kubectl to deploy apps, check logs as well as manage all the other resources of the cluster.

Kubernetes uses an HTTP-based REST API which is the actual Kubernetes user interface employed to manage it. This means that every Kubernetes operation is represented as an API endpoint and can be carried out based on an HTTP-request sent to the endpoint. 

Install Kong API Gateway on Ubuntu 20.04 - Step by Step Process ?

This article covers how to install the KONG API Gateway on Ubuntu 20.04. These Gateways provides services like Authentication, rate-limiting, analysis, logging etc., to your API endpoints.

To Start kong API, run the following commands:

$ kong migrations up #runs the database migrations
$ sudo kong start
$ curl -i http://localhost:8001

To Stop and see status of kong API, run the following commands:

$ sudo kong stop #stop kong
$ sudo kong health #kong status

Install Minikube on Ubuntu 20.04 - Best Method ?

This article covers how to install Minikube on your Ubuntu 20.04 machine. Minikube is an open source tool that allows you to set up a single-node Kubernetes cluster on your local machine. The cluster is run inside a virtual machine and includes Docker, allowing you to run containers inside the node. 

Now, you can use the single-node Kubernetes with minikube to learn one of the most famous containers orchestration tools in the Cloud Native era. 


To enable and access kubernetes dashboard , run the command:

$ minikube dashboard

This will open the Kubernetes dashboard in the web browser.


To stop the minikube, run:

$ minikube stop

To delete the minikube, run:

$ minikube delete

To Start the minikube, run:

$ minikube start

Install Harbor Image Registry on Ubuntu 20.04 - Best Method ?

This article covers how to install harbor on Ubuntu. The process is the same for all the other distribution of Linux. Harbor is a component of vSphere with Kubernetes. Harbor provides an enterprise-class registry service. 

Harbor is deployed in a dedicated system namespace on the Supervisor Cluster and is composed of several vSphere Pods. 

Image registries provide a centralized repository for an organization to store container images. 

Public image registries can be used to store container images. However, for security reasons, it is better to use an on-premises image registry. 

You use an on-premises image registry for the following purposes: 

Store container images securely. 

Control access to container images. 

vSphere with Kubernetes is integrated with Harbor and is also compatible with other container image registries. 


Harbor key features includes:

  • Security and vulnerability analysis.
  • Content signing and validation.
  • Extensible API and web UI.
  • Image replication.
  • Role-based access control.
  • Multitenant.


After Harbor installation has succeeded, access Harbor web console on https://registry_domain.

Then Login with:

  • Username: admin
  • Password: Set-in-harbor.yml

Top 5 Open-Source Load Balancers 2021

This article reviews the Best Open Source Load Balancers. Load Balancing software assists virtual appliances in monitoring and distributing excess traffic. It helps network administrators and data centers maintain constant loading speed. The load balancing software is used to convey network traffic to specific servers with accurate configurations. 


Load Balancing Algorithms Types:

  • Round-robin Algorithm – The simplest method involves moving the requests within the same order to available servers. 
  • Least-time Algorithm – It selects servers supported by the smallest amount of active requests and the fastest processing speed. The algorithm integrates with powerful algorithms to give the server higher memory, capacity, and power. 
  • Least-connections Algorithm – This algorithm sends requests to servers with the smallest amount of workload. The algorithm sends requests to the smallest amount of busy servers. 
  • Hash-based Algorithm – It assists a hash key to the client and server IP addresses. The algorithm ensures that user's requests are sent to the identical servers containing data from the previous sessions. This ends up in an efficient network resource delivery.

Access Google Drive on Ubuntu 20.04 LTS - Step by Step Process ?

This article covers how you can access Google drive on Ubuntu. Now you can easily access your Google Drive files directly from the File Manager.

The process involves adding your online Google account to your Gnome Online Accounts list. This configuration then lets you mount your Google Drive account to your Nautilus File Manager. 

After mounting, you can access your online files, edit them, and add new files directly from your Ubuntu to your Google Drive.


To Install and Open Gnome Online Accounts:

1. Open the Ubuntu command line, the Terminal, either through the system Dash or the Ctrl+Alt+T shortcut. 

2. Once the Terminal application opens, enter the following command as sudo:

$ sudo apt install gnome-online-accounts

Please remember that only an authorized user can add, remove and configure software on Ubuntu.

3.  Once the utility is installed on your system, you can open it through one of the following methods:

By entering the following command in your Terminal:
$ gnome-control-center online-accounts

Install and Use Docker Compose on CentOS 8 - Step by Step Process ?

This article covers how to install and use the Docker-compose on CentOS 8 system. By using the Docker-compose, you can deploy multiple Docker applications with container services through a single command. 

Basically, Docker Compose is a tool that can be used to define and run multiple containers as a single service. With Docker Compose, you can link multiple containers and deploy an application from a single command. 

It is mainly used in the development, testing and staging environment. 

Docker Compose uses a YAML file to define a complex stack in a file and running it with a single command.


To install Docker on CentOS:

1. Before starting, make sure you have Docker installed on your server. If not installed, you will need to add Docker-CE repository to your system. 

You can add it with the following command:

$ dnf config-manager --add-repo=https://download.docker.com/linux/centos/docker-ce.repo

2. Once the repository is added, install the latest version of docker using the following command:

# dnf install docker-ce --nobest -y
# systemctl start docker
# systemctl enable docker

3. You can verify the docker version with the following command:

$ docker --version

Install AWX on Ubuntu 20.04 - Step by Step Process ?

This article covers how to install ansible AWX on Ubuntu 20.04. You have got a basic idea about managing hosts, inventories, and projects from AWX ansible.

Ansible AWX is an open source tool which provides a web-based user interface, REST API, and task engine for easy and collaborative management of Ansible Playbooks and Inventories.

AWX allows you to centrally manage Ansible playbooks, inventories, Secrets, and scheduled jobs from a web interface. 


To Install Ansible on Ubuntu:

Run the following commands,

# sudo apt update
# sudo apt install ansible

Install KDE Plasma on Ubuntu 20.04 - Step by Step process to implement it ?

This article covers how to install KDE Plasma on Ubuntu 20.04 system. The Plasma desktop option will be available along with the standard Ubuntu system. Also, you will learn the steps to uninstall KDE Plasma from the Ubuntu system. 

KDE is the short form for K Desktop Environment.

Plasma is the desktop environment and KDE is the umbrella project responsible for the development of Plasma desktop and a bunch of other applications.


To Install KDE desktop environment on Ubuntu:

For KDE Full - This is the complete KDE pack. It comes with the complete package and core KDE plasma desktop. 

$ sudo apt install kde-full


For KDE Standard - It includes Plasma desktop with standard set of KDE apps such as Kate (default text editor), Konqueror (default web browser), Kget (Download Manager), KMail (email client), Dolphin (File Manager) and so on. 

$ sudo apt install kde-standard


To install KDE Plasma Desktop:

Run the command - $ sudo apt install kde-plasma-desktop

Who We Are ?

Most prominent, efficient, and well-performing IT companies and software solutions.

Related Posts