Nutanix Glossary

What Does Cloud Native Mean?

What Does it Mean to be “Cloud Native”?

Cloud native is a new degree of software abstraction, coupled with a set of dedicated technologies, tools, and processes that transform how applications can be built, deployed, and managed.

What are the characteristics of a cloud native application?

  1. Code is packaged as containers
  2. Architected as collections of microservices
  3. Developers and IT Operations work together in a tightly-integrated, agile capacity
  4. Deployed and managed across elastic cloud infrastructure
  5. Automated and policy-driven infrastructure resource allocation

Advantages of Cloud Native

  • Organisations embarking on cloud native journeys are adopting containers and Kubernetes, along with a variety of open-source and commercial technologies from an expansive cloud native ecosystem. Containers and cloud native technologies enable organisations to accelerate application development, upgrade applications non-disruptively, scale them efficiently, and easily port them across different environments. These capabilities ultimately translate to better business agility and competitive advantage.
  • Containers solve some critical issues for enterprises. Run anywhere, robust software packaging for developers, service upgrades, scalability, availability, and resource efficiency in a vendor neutral package. 
  • Kubernetes provides an infrastructure layer that IT can operate at scale in a programmatic, repeatable, fashion as opposed to more hands-on management that is typical in the enterprise. Because Kubernetes can be deployed on-prem and the public cloud it provides a common operating model across environments. This makes Kubernetes an ideal platform for companies that operate in both environments.

Cloud Native Architecture

Cloud native architecture uses containers and Kubernetes to build applications as sets of smaller, composable pieces that scale efficiently, upgrade non-disruptively, and migrate seamlessly to other environments. This technology is cost-efficient with better security and flexible automation.

The On-prem Kubernetes Stack

We’ll explore the different layers of the Kubernetes stack on-prem from the infrastructure that Kubernetes runs on to the services running on top of Kubernetes that developers can leverage to accelerate application delivery.  Note that some components can be included in multiple parts of the stack.  For instance, a Kubernetes distribution may include a Container Registry such as Harbor or Quay.  Alternatively, a Container Registry may be offered as a hosted or managed service for use with any distribution.

Compute Platform

Kubernetes is very flexible in where it can be deployed and innovation is expanding the options for how and where Kubernetes can be operated.  Typically Kubernetes requires a minimum 3 nodes with full network connectivity between them and attached storage.  Today Kubernetes is most commonly run on top of virtual infrastructure like VMware.  Bare Metal and HCI are two other common choices.

Additional capabilities can make provisioning and operating a Kubernetes cluster much easier. Infrastructure controlled via an API (e.g. Cluster API, Terraform, etc.) make it easy to provision and connect new nodes.  Custom hypervisors may offer workload isolation to plug security holes in containers.  A Load Balancer outside the Kubernetes cluster is useful for making workloads available and dynamically relocating and scaling them. Some workloads may benefit from custom hardware such as GPUs or TPUs. And if many clusters are to be operated at multiple locations (e.g. in retail stores) a distributed management capability for the underlying infrastructure would be useful.

Kubernetes Distributions

With the explosive growth in Kubernetes since its release in 2015 there are over 100 Kubernetes distributions and hosted platforms from tech giants and startups alike.  These companies take the core Kubernetes code and combine it with other projects such as a networking stack, cluster management tools, logging, monitoring, etc. to create a distribution.

Container Services

A useful Kubernetes environment requires additional infrastructure services beyond Kubernetes itself. These are sometimes bundled with distributions or are added on as separate services and include some basics such as a container registry, Ingress, load balancer, Secret Stores, certificate manager, to name a few.

As more companies build and deploy applications on Kubernetes there are many core services they depend on that typically also run in Kubernetes.  Examples include databases (e.g. Redis, Postgres), Pub Sub (Kafka), Indexing (ElasticSearch), CI/CD services (Gitlab, Jenkins) and storage (Portworx, MayaData, Minio, Red Hat).

Many companies have developed offerings that will manage or support these applications running on a customer’s Kubernetes distribution of choice. Most notable is probably Microsoft’s Azure Arc Data Services which provides Postgres and MS SQL as managed services in customer environments but there are many others including Confluent for Kafka, Crunchy Data for Postgres, Redis Enterprise for Redis, Percona, Elasticsearch, Minio, Mayadata, etc.

Containers

Containers are the atomic unit of cloud native software. A container comprises a piece of code packaged up with all of its dependencies (binaries, libraries, etc.) and is run as an isolated process. This translates to a new level of abstraction. Much like the virtual machine (VM) abstracts compute resources from the underlying hardware, a container abstracts an application from the underlying operating system (OS).

Containers are different from VMs in several important ways. They are significantly smaller in resource footprint, faster to spin up, and far easier to port across cloud environments. Unlike VMs, they are ephemeral in nature.

Containers pave the way for microservices software architecture. While legacy applications are monolithic in nature, microservices-based applications are built as collections of smaller, composable pieces (services) that can scale independently and allow for non-disruptive upgrades. A service comprises one or more containers that perform a common function.

Containerised/microservices-based applications enable streamlined, accelerated software development practices that require tighter integration and coordination between developers and IT operations, referred to as DevOps.

Kubernetes

Because containerised applications lend themselves to distributed environments and use compute, storage, and network resources differently than do legacy applications, they require a layer of infrastructure dedicated to orchestrating containerised workloads.

Kubernetes has emerged as the dominant container orchestrator and is often characterised as the operating system of the cloud. It is a portable, extensible, open-source platform for managing containerised workloads and services, that facilitates both declarative configuration and automation.

Specifically, Kubernetes:

  • assigns containers to machines (scheduling)

  • boots the specified containers through the container runtime

  • deals with upgrades, rollbacks, and the constantly changing nature of the system

  • responds to failures (container crashes, etc.)

  • creates cluster resources like service discovery, inter-VM networking, cluster ingress/egress, etc.

Kubernetes is designed for scalability, availability, security, and portability. It optimises the cost of infrastructure, distributing workloads across available resources. Each component of a Kubernetes cluster can also be configured for high availability.

Kubernetes functionality evolves rapidly as a result of continuous contributions from its active global community, and there now exists a variety of Kubernetes distributions available to users. While users are best served by CNCF-certified distributions (conformance enables interoperability), intelligent automation around Kubernetes lifecycle management features, coupled with easy integration of storage, networking, security, and monitoring capabilities are critical for production-grade cloud native environments.

Cloud Native Challenges

Cloud native technologies are the new currency of business agility and innovation, but configuring, deploying, and managing them are challenging for any type of enterprise. First, Kubernetes and its ecosystem of cloud native technologies is deep and evolves fast. Furthermore, legacy infrastructure isn’t architected for the way Kubernetes and containers use IT resources. This is a significant impediment to software developers, who often require resources on demand, along with easy-to-use services for their cloud native applications. Finally, every organisation ends up employing a mix of on-prem and public cloud-based Kubernetes environments, but benefitting from multicloud flexibility depends on being able to effectively manage and monitor all deployments.

Cloud Native Best Practices

Building an enterprise Kubernetes stack in the datacenter is a major undertaking for IT operations teams. It is essential to run Kubernetes and containerised applications on infrastructure that is resilient and can scale responsively in support of a dynamic, distributed system.

Nutanix hyperconverged infrastructure (HCI) is the ideal infrastructure foundation for cloud native workloads running on Kubernetes at scale. Nutanix offers better resilience for both Kubernetes platform components and application data, and superior scalability to bare metal infrastructure. Nutanix also simplifies infrastructure lifecycle management and persistent storage for stateful containers, eliminating some of the most difficult challenges organisations face in deploying and managing cloud native infrastructure.

Explore Our Top Resources

Test Drive Nutanix

Test Drive Nutanix Kubernetes Engine

Build a Best-in-class Hybrid Multicloud Stack with Red Hat and Nutanix

7 Ways to Simplify Kubernetes Lifecycle Management

Cloud Native Products & Solutions

Nutanix Kubernetes Engine

Fast-track your way to production-ready Kubernetes and simplify lifecycle management.

HCI for Kubernetes

Nutanix HCI is the ideal infrastructure foundation for Kubernetes and cloud native applications.

Hybrid Cloud Kubernetes

Through partnerships with Red Hat, Google Cloud, and Microsoft Azure, Nutanix offers a fast, reliable path to hybrid cloud Kubernetes.

Persistent Storage For Cloud Native Applications

Nutanix data services and CSI extends simplicity to configure and manage persistent storage in Kubernetes.

Get Started with Hyperconverged Infrastructure (HCI)

Get started with Nutanix today