The world of application packaging and deployment can be viewed as a spectrum where at one end we have hand-crafted installation/run scripts on physical infrastructure and on the other are fully containerized applications making the most of cloud-native services. Since most organisations don’t start out fully cloud-native or containerized, a commonly heard phrase for increasing cloud adoption is “lift and shift”. This is where applications are not re-architected, but taken straight to the cloud from an in-house deployment scenario. The main advantage of “lift and shift” is speed of cloud adoption and reduction in physical infrastructure costs. However, it is worth mentioning that this pattern often results in application deployments with a higher operational cost than those purpose-built to be there. In the longer-term, this can be offset by making more use of cloud-native services.
In this post, the advantages of Habitat for application packaging will be explored by deploying the same application across several different contexts, including:
There is extra effort required in packaging applications with Habitat. However, what you get in return is the ability to deploy almost anywhere with relative ease at a later stage. From installing your application on a new physical box to choosing an entirely new cloud provider, the process becomes simpler with Habitat-packaged applications.
For this demonstration, we will be using Table Setting, a single-file Python web application that exposes a REST API for managing forks. A Swagger user interface is exposed at the root URL to provide some interactivity. For those interested, the application makes use of: Flask; Flask-RESTPlus; SQLAlchemy; Flask Marshmallow and more. These frameworks are used regularly by many real-world Python applications and websites.
Table Setting can run with different backend databases. Here, we will use SQLite for an in-memory or file-based database, and MySQL for a more realistic production example. Whilst Table Setting is clearly not a production-ready application, it provides a realistic application-plus-database deployment scenario that could apply in many real-world cases. Since we are looking at a “lift and shift” example, we don’t want to re-architect our application. Therefore, following best practice, Table Setting uses environment variables to convey configuration changes.
Let’s explore how to run Table Setting locally before getting started with Habitat. As a prerequisite, Python 3.7 should be installed. This quickly brings us to the downside of shipping anything other than a single, executable artifact as disparity
can easily creep in! For example, there are two major versions of Python: 2.x
now seen as legacy, and 3.x
that we will use
here. Choosing Python 3.7 affects how we run Table Setting because we require the correct python
executable in our PATH
. Assuming that is the case, Table Setting can be run via:
$ git clone https://github.com/skpaterson/table-setting $ cd table-setting $ python -m venv tsenv $ . tsenv/bin/activate $ pip install -r requirements.txt $ python run_app.py
Navigate to http://localhost:5000 to see the Table Setting Swagger UI:
Feel free to try creating some forks! Standard CRUD operations (create, read, update, delete) are all available. By default, the application runs with a clean SQLite in-memory database that is wiped on restart.
A user running even this simple application locally is perhaps unnecessarily exposed to several things:
git
pip
A similar list could be formulated for applications written in other languages. In any case, this complicates life for performing any changes at a later stage e.g.,
In order to see what we need to add to this application to package it using Habitat, let’s look in the habitat directory:
$ tree habitat/ habitat/ ├── README.md ├── config ├── default.toml ├── hooks │ ├── init │ └── run └── plan.sh
The plan.sh
file defines how to build, and includes all runtime and buildtime dependencies, implementing similar steps to those taken when running locally. hooks
are scripts that control how an application is initialized on startup
and run. default.toml
can be used to define base configuration. As previously mentioned, Table Setting explicitly depends on a known good Python version and environment variables are used to convey runtime configuration. For example,
an environment variable can be set to specify which backend database type to run Table Setting against.
Let’s concentrate for now on creating deployment artifacts with Habitat. The Habitat Studio is a clean-room for building applications. Below we use the Studio to create and export different package formats:
$ cd table-setting $ hab studio enter [STUDIO] build [STUDIO] source results/last_build.env [STUDIO] hab pkg export tar "results/${pkg_artifact}" [STUDIO] hab pkg export docker "results/${pkg_artifact}"
We also use the Habitat-maintained core/mysql
package for some deployments, so let’s install and export that locally also:
[STUDIO] hab pkg install core/mysql » Installing core/mysql ☁ Determining latest version of core/mysql in the 'stable' channel ☛ Verifying core/mysql/5.7.21/20180609181124 ... output curtailed ★ Install of core/mysql/5.7.21/20180609181124 complete with 25 new packages installed.
We now use the package version from the installation command to export a Docker image locally:
[STUDIO] hab pkg export docker core/mysql/5.7.21/20180609181124
Note, however, that you can just hab pkg export docker core/mysql
without previously installing, and you’ll get the latest stable package automatically.
For reference, all of the plans built and maintained by the Habitat Core Team are available here.
The resulting artifacts are all we need for the deployment scenarios below.
Let’s assume we want to deploy Table Setting to an Ubuntu machine and use a file-based SQLite database backend. In a real-world setup, the versioned tar
deployment file might be stored in a dedicated artifact repository but we will
assume it is locally available on the target machine here.
Relying on the Habitat package naming convention, we can run the package as follows:
$ cat hab_tar_deployment_ubuntu.sh #!/bin/bash # one-time setup - packages can specify a linux user to run under, by default this is the hab user sudo adduser --disabled-password --gecos "" hab # assumes local package file sudo tar zxvf *-table-setting*.tar.gz -C / # run app in SQLite file mode export DB_TYPE=sqlite_file # install and run with the habitat supervisor sudo /hab/bin/hab sup run & sudo /hab/bin/hab svc load habskp/table-setting
Running this on a locally running Ubuntu VM looks as follows:
This is arguably simpler than the steps in the previous section. We’ve already successfully decoupled a lot of the package internals from the deployment process. Although some knowledge is required to run the tar
Habitat package locally,
this would be a similar set of steps for any Habitat package. Now let’s see what else we can do.
For this case, let’s take a look at the Docker Compose manifest in the root of the Table Setting repository:
$ cat docker-compose.yml version: '3' services: mysql: image: core/mysql environment: HAB_MYSQL: | bind = '0.0.0.0' app_username = 'tablesetting' app_password = 'tablesetting' root_password = 'notverysecure' tablesetting: image: habskp/table-setting ports: - 8000:8000 environment: - APP_PORT=8000 - DB_TYPE=mysql - DB_USER=tablesetting - DB_PASSWD=tablesetting command: --peer mysql --bind database:mysql.default restart: on-failure depends_on: - mysql
The core/mysql
image previously exported is used “as is” with some additional environment setup. For this we use HAB_MYSQL
and add the corresponding settings to the variables expected by the Table Setting application.
Run this locally from the root of the repository via the following command:
$ docker-compose up
We can now see the application running on port 8000
against a mysql
database backend:
At first this might seem like another variant of running locally but having a working setup with docker-compose
makes it easy to deploy to e.g. AWS ECS, see this post for more details.
Transitioning to a Kubernetes cluster at this point might seem like a big leap but to our advantage we have already exported the required Habitat package artifacts.
In the next two sections we will be deploying to Google Kubernetes Engine (GKE) with Habitat Operator already installed. Similarly to the steps described here we will upload our artifacts to Google Container Registry for convenience in our manifest. As an example, the following commands would upload our exported Table Setting image to GCR:
# assumes docker and gcloud are installed locally $ gcloud auth configure-docker $ docker tag habskp/table-setting:latest eu.gcr.io/spaterson-project/table-setting:latest $ docker push eu.gcr.io/spaterson-project/table-setting:latest
Also worthwhile to comment that although GCP was chosen for this post, the same artifacts could be deployed in the same way to other cloud provider container registries and Kubernetes clusters simply by supplying the appropriate deployment manifest files.
The below manifest deploys the Habitat-managed MySQL and Table Setting images uploaded to Google Container Registry. Similar environment variables as before are set to configure the database:
$ cat gke-gcr-hab-core-mysql-sample.yml --- apiVersion: habitat.sh/v1beta1 kind: Habitat metadata: name: database labels: app: table-setting-hab-db customVersion: v1beta2 spec: v1beta2: image: eu.gcr.io/spaterson-project/mysql:latest count: 1 env: - name: HAB_MYSQL value: "bind = '0.0.0.0'\napp_username = 'tablesetting'\napp_password = 'tablesetting'\nroot_password = 'notverysecure'" service: name: mysql topology: standalone --- apiVersion: habitat.sh/v1beta1 kind: Habitat metadata: name: table-setting-app labels: app: table-setting-hab customVersion: v1beta2 spec: v1beta2: image: eu.gcr.io/spaterson-project/table-setting:latest count: 1 env: - name: DB_TYPE value: mysql - name: DB_USER value: tablesetting - name: DB_PASSWD value: tablesetting service: name: table-setting topology: standalone bind: - name: database service: mysql group: default --- apiVersion: v1 kind: Service metadata: name: table-setting-lb spec: type: LoadBalancer selector: habitat-name: table-setting-app ports: - port: 80 targetPort: 5000
As a future improvement to the above, configuration could be conveyed via Habitat’s user.toml
mechanism, more details on that here.
Run the following to deploy this manifest on our Habitat enabled Kubernetes cluster:
$ kubectl create -f gke-gcr-hab-core-mysql-sample.yml habitat "database" created habitat "table-setting-app" created service "table-setting-lb" created
Next, wait for an external IP address to be assigned:
$ kubectl get service table-setting-lb NAME TYPE CLUSTER-IP EXTERNAL-IP PORT(S) AGE table-setting-lb LoadBalancer 10.51.245.180 35.189.83.69 80:30871/TCP 2m
Navigating to this external IP shows the expected UI:
From here we could explore setting up a cluster of MySQL databases as described here or alternatively look to becoming more cloud native.
Here we will demonstrate using Kubernetes Service Catalog
and a Google cloud Platform Service Broker,
an implementation of the Open Service Broker standard, to provision a cloud SQL (MySQL)
instance instead of relying on our own image from the last section.
In addition to the previous Kubernetes prerequisites, more steps are required to use cloud SQL. Thankfully these are described in detail here. For those interested, stopping at the end of Step 3.2 and replacing the service account and web container image link details in the below manifest should allow you to reproduce. The manifest itself now looks like this
and honours the cloud-mysql
namespace used in the Kubernetes Engine sample:
$ cat gke-gcr-cloud-sql-sample.yml apiVersion: apps/v1beta2 kind: Deployment metadata: name: table-setting namespace: cloud-mysql labels: app: table-setting spec: selector: matchLabels: app: table-setting template: metadata: labels: app: table-setting spec: containers: - name: web image: eu.gcr.io/spaterson-project/table-setting:latest ports: - containerPort: 5000 env: - name: DB_TYPE value: mysql - name: cloudsql-proxy image: gcr.io/cloudsql-docker/gce-proxy:1.11 env: - name: CONNECTION_NAME valueFrom: secretKeyRef: name: cloudsql-credentials key: connectionName command: ["/cloud_sql_proxy", "-instances=$(CONNECTION_NAME)=tcp:3306", "-credential_file=/secrets/cloudsql/privateKeyData"] volumeMounts: - name: mysql-service-account-6041 mountPath: /secrets/cloudsql readOnly: true volumes: - name: mysql-service-account-6041 secret: secretName: mysql-service-account-6041 - name: cloudsql emptyDir: --- apiVersion: v1 kind: Service metadata: name: cloudsql-user-service namespace: cloud-mysql labels: app: table-setting spec: selector: app: table-setting ports: - port: 80 targetPort: 5000 type: LoadBalancer
Deploy the application via:
$ kubectl create -f gke-gcr-cloud-sql-sample.yml deployment "table-setting" created service "cloudsql-user-service" created
As in the previous example, we wait until an external IP address is assigned:
$ kubectl get service cloudsql-user-service --namespace cloud-mysql NAME TYPE CLUSTER-IP EXTERNAL-IP PORT(S) AGE cloudsql-user-service LoadBalancer 10.51.247.227 35.189.84.130 80:30712/TCP 2m
Navigating to our very familiar UI, let’s create a fork:
Finally, let’s use the cloud shell in GCP UI to explore the database and confirm things are working as expected:
So we’re now successfully making use of a cloud SQL database with our Habitat-managed application. Congratulations if you made it this far!
At this point it would be remiss not to mention the many complimentary Habitat features that were completely skipped over in the above e.g. we didn’t touch upon Supervisor, Builder, Depot, Package updates, binds and exports … Thankfully there are plenty of blog posts and other documentation to remedy that!
In this post we explored several different deployment scenarios for a Habitat packaged application, from bare metal to running on a Kubernetes cluster with cloud based database. Whether your organisation is staying on-premise, in the middle of lifting and shifting, or even expanding to new cloud providers, hopefully this has shown how Habitat can make packaging and deployment easier. Thanks for reading!