SYS-CON MEDIA Authors: Pat Romanski, Liz McMillan, Yeshim Deniz, Elizabeth White, Courtney Abud

Blog Feed Post

Using DBaaS to Align Database Management with Your Application Delivery Processes


Today’s application development teams are moving fast, but is your team being held back by slow database management practices and tools? For many organizations, this is a significant problem that deserves as much attention as getting application code changes flowing smoothly into production. In fact, they are tightly interdependent challenges that need to be solved together.

Your Database Environment: Perception vs. Reality

Your Database Environment: Perception vs. Reality

Think about the application development process for a moment. How long does it take you to stand up a new copy of your database? Are you sure you have configured everything correctly and consistently? Do your developers and testers have access to production data for realistic testing? For most organizations, it’s unlikely that the answer to these questions are “minutes”, “very certain”, and “yes”.

It’s also critical to embrace the distributed and multi-instance nature of your organization’s data landscape. Gone are the days of a single, large and controlled database that sits underneath all of your applications. Even in a single application team, it’s a fairly universal truth that you will have multiple environments. From individual developer sandboxes, to integration and test infrastructure, and through to production clusters (and even DR copies), we aren’t really working with a single monolithic database environment. Each one of these environments means a new deployment of your database configuration.

Mapping Database Management to Application Delivery Processes

Data is different from code. Data represents the persistent state of your application, yet dynamically changes and grows as the application is used. Code, on the other hand, is much more of a static entity that can be versioned cleanly and deployed through configuration management tools. Many advanced approaches and tools have been developed to manage configuration drift, and ensuring that the code that is expected to be deployed is, in fact, the code deployed is now easier than ever. It’s the dynamic nature of data that makes managing it different from code deployments.

So what can you do about this difference? First, understand that there are aspects of your database environment that can be managed in a similar way as code and configuration. They fall into two buckets: the definition (configuration) of the database system itself and the definition of the schema / database code. Next, think about how the data being generated by your production application can be re-purposed efficiently for testing both new versions of your applications and the processes you will use to deploy your applications.

One great example of these concepts rolled into a consistent line of thinking is the work of Martin Fowler and Pramod Sadalage on Evolutionary Database Design. Evolutionary Database Design is a methodology for ensuring that your database environment is able to adapt and evolve along with your application. It advocates several specific approaches to ensuring that the database layer is an integral part of the entire application delivery pipeline.

While the Evolutionary Database Design pattern is certainly only one of many approaches to take, it’s well thought through and worth reviewing to find some key themes:

  1. Database instances need to be (re)generated on-demand – Development, integration, test, staging, and production environments should be consistent.
  2. Database schemas should be defined as “refactorings” – Your schema should be created by series of evolutionary migrations from one version to the next.
  3. Actual data matters – Use test data for earlier on in the application’s lifetime, but be prepared to use copies of production data later.

How does DBaaS Fit?

A Database-as-a-Service (DBaaS) platform can play a large role in how you advance your application delivery processes. Specifically, DBaaS can address two of the three themes from above: instance creation and data portability.

Instance creation is perhaps the easiest to comprehend, because that’s what most DBaaS solutions are solving at their core. Think about the definition of your database environment as configuration metadata (or a blueprint), which when used to deploy a new database instance from a DBaaS platform should result in a consistent database engine deployment for each instantiation. It shouldn’t matter if the environment’s specification is as simple as a single node or as complex as a multi tier cluster configured for sharding, it should be just as easy to get a new environment online.

The second role that a DBaaS platform can play is in helping to move the data itself around in the application lifecycle process. It should support operations like replica creation and database cloning, again via a self service UI and with APIs. These two functions are often though about as “operational” considerations for a database environment, but if you think about the whole application development system, they can become surprisingly useful to the entire team.

First, let’s define them:

Moving Data and DB Config through the Application Lifecycle

Moving Data and DB Config through the Application Lifecycle

  • Cloning is when a snapshot of the database system is created, and then transferred into a new “instance”. It’s a point in time copy created on-demand.
  • Replication is the synchronous or asynchronous replay of database changes against a “copy” of the database, which can then be detached from the “master” database to operate as a standalone system.

Both of these features offer a similar end result: you get a copy of the database that you can use for other purposes. Cloning takes a bit longer, but can easily happen off hours to minimize impact on the production environment. Replication can ensure a much faster “turn up” of the new environment with up-to-date data, but needs to be created ahead of the request for an instance. Regardless of the differences, they both provide a foundation with which each developer, tester and release manager can easily get a version of the production database when it’s needed.

Think about that in the context of bug-fixing: All too frequently, developers are handed user reports of an issue that is hard to reproduce without having a copy of the production data in place to replicate the situation. If it was easy for a developer to instantiate their own copy of the production environment (on demand) from a clone created each night, all of a sudden that developer will be able to resolve the bug much more effectively. Not only that, but the testing process will be able to ensure that the proposed patch does, in fact, solve the problem with real-world data as input.


You can move the code easily, but data and databases are a much harder a challenge without the right tools. A well designed Database-as-a-Service (DBaaS) environment can play a key role in helping you be more nimble in your treatment of data services within your application development pipeline. A DBaaS should offers solutions to both the “course grained” configuration management / automation challenges of the database engines themselves and the need to have self service access to copies of the dynamic data that your applications generate.

The right DBaaS solution should give you:

  • Self service access new DB instances for your development and testing teams
  • Move database engine configurations through the application development pipeline
  • Easily pull data back into staging, test and dev environments to give your teams access to real-world data
  • Support production rollout of application releases through replication and / or database cloning
  • APIs to make these operations an integral part of your continual integration and deployment processes

Read the original blog entry...

More Stories By Bob Gourley

Bob Gourley writes on enterprise IT. He is a founder of Crucial Point and publisher of

Latest Stories
The use of containers by developers -- and now increasingly IT operators -- has grown from infatuation to deep and abiding love. But as with any long-term affair, the honeymoon soon leads to needing to live well together ... and maybe even getting some relationship help along the way. And so it goes with container orchestration and automation solutions, which are rapidly emerging as the means to maintain the bliss between rapid container adoption and broad container use among multiple cloud host...
Today most companies are adopting or evaluating container technology - Docker in particular - to speed up application deployment, drive down cost, ease management and make application delivery more flexible overall. As with most new architectures, this dream takes significant work to become a reality. Even when you do get your application componentized enough and packaged properly, there are still challenges for DevOps teams to making the shift to continuous delivery and achieving that reducti...
Here to help unpack insights into the new era of using containers to gain ease with multi-cloud deployments are our panelists: Matt Baldwin, Founder and CEO at StackPointCloud, based in Seattle; Nic Jackson, Developer Advocate at HashiCorp, based in San Francisco, and Reynold Harbin, Director of Product Marketing at DigitalOcean, based in New York. The discussion is moderated by Dana Gardner, principal analyst at Interarbor Solutions.
Skeuomorphism usually means retaining existing design cues in something new that doesn’t actually need them. However, the concept of skeuomorphism can be thought of as relating more broadly to applying existing patterns to new technologies that, in fact, cry out for new approaches. In his session at DevOps Summit, Gordon Haff, Senior Cloud Strategy Marketing and Evangelism Manager at Red Hat, discussed why containers should be paired with new architectural practices such as microservices rathe...
In 2014, Amazon announced a new form of compute called Lambda. We didn't know it at the time, but this represented a fundamental shift in what we expect from cloud computing. Now, all of the major cloud computing vendors want to take part in this disruptive technology. In his session at 20th Cloud Expo, John Jelinek IV, a web developer at Linux Academy, will discuss why major players like AWS, Microsoft Azure, IBM Bluemix, and Google Cloud Platform are all trying to sidestep VMs and containers...
Using serverless computing has a number of obvious benefits over traditional application infrastructure - you pay only for what you use, scale up or down immediately to match supply with demand, and avoid operating any server infrastructure at all. However, implementing maintainable and scalable applications using serverless computing services like AWS Lambda poses a number of challenges. The absence of long-lived, user-managed servers means that states cannot be maintained by the service. Lo...
With the new Kubernetes offering, ClearDATA solves one of the largest challenges in healthcare IT around time-to-deployment. Using ClearDATA's Automated Safeguards for Kubernetes, healthcare organizations have access to the container orchestration to dynamically deploy new containers on demand, monitor the health of each container for threats and seamlessly roll back faulty application updates to a previous version, avoid system-wide downtime and ensure secure continuous access to patient data.
Isomorphic Software is the global leader in high-end, web-based business applications. We develop, market, and support the SmartClient & Smart GWT HTML5/Ajax platform, combining the productivity and performance of traditional desktop software with the simplicity and reach of the open web. With staff in 10 timezones, Isomorphic provides a global network of services related to our technology, with offerings ranging from turnkey application development to SLA-backed enterprise support. Leadin...
With the rise of Docker, Kubernetes, and other container technologies, the growth of microservices has skyrocketed among dev teams looking to innovate on a faster release cycle. This has enabled teams to finally realize their DevOps goals to ship and iterate quickly in a continuous delivery model. Why containers are growing in popularity is no surprise — they’re extremely easy to spin up or down, but come with an unforeseen issue. However, without the right foresight, DevOps and IT teams may lo...
Platform9, the open-source-as-a-service company making cloud infrastructure easy, today announced the general availability of its Managed Kubernetes service, the industry's first infrastructure-agnostic, SaaS-managed offering. Unlike legacy software distribution models, Managed Kubernetes is deployed and managed entirely as a SaaS solution, across on-premises and public cloud infrastructure. The company also introduced Fission, a new, open source, serverless framework built on Kubernetes. These ...
Emil Sayegh is an early pioneer of cloud computing and is recognized as one of the industry's true veterans. A cloud visionary, he is credited with launching and leading the cloud computing and hosting businesses for HP, Rackspace, and Codero. Emil built the Rackspace cloud business while serving as the company's GM of the Cloud Computing Division. Earlier at Rackspace he served as VP of the Product Group and launched the company's private cloud and hosted exchange services. He later moved o...
As you know, enterprise IT conversation over the past year have often centered upon the open-source Kubernetes container orchestration system. In fact, Kubernetes has emerged as the key technology -- and even primary platform -- of cloud migrations for a wide variety of organizations. Kubernetes is critical to forward-looking enterprises that continue to push their IT infrastructures toward maximum functionality, scalability, and flexibility. As they do so, IT professionals are also embr...
Kubernetes is a new and revolutionary open-sourced system for managing containers across multiple hosts in a cluster. Ansible is a simple IT automation tool for just about any requirement for reproducible environments. In his session at @DevOpsSummit at 18th Cloud Expo, Patrick Galbraith, a principal engineer at HPE, will discuss how to build a fully functional Kubernetes cluster on a number of virtual machines or bare-metal hosts. Also included will be a brief demonstration of running a Galer...
DevOps is under attack because developers don’t want to mess with infrastructure. They will happily own their code into production, but want to use platforms instead of raw automation. That’s changing the landscape that we understand as DevOps with both architecture concepts (CloudNative) and process redefinition (SRE). Rob Hirschfeld’s recent work in Kubernetes operations has led to the conclusion that containers and related platforms have changed the way we should be thinking about DevOps and...
Cloud-Native thinking and Serverless Computing are now the norm in financial services, manufacturing, telco, healthcare, transportation, energy, media, entertainment, retail and other consumer industries, as well as the public sector. The widespread success of cloud computing is driving the DevOps revolution in enterprise IT. Now as never before, development teams must communicate and collaborate in a dynamic, 24/7/365 environment. There is no time to wait for long development cycles that pro...