|By Andreas Grabner||
|March 26, 2014 01:15 PM EDT||
DevOps is speeding towards the IT world like a freight train and the hype around it is deafening. There is no reason to be afraid of this change as it is the natural reaction to the agile movement that revolutionized development just a few years ago. By definition, DevOps is the natural alignment of IT performance to business profitability. The relevance of this has yet to be quantified but it has been suggested that the route to the CEO's chair will come from the IT leaders that successfully make the transition to a DevOps model. If this still seems foreign to you, I recommend reading up on DevOps Blog from IT Revolution, the OpsCode Blog, and check out The Phoenix Project.
Despite all the talk around simple monitoring tools, breaking through the walls between Dev and Ops still poses a real challenge. This is because of a misunderstanding around Operations real purpose - extracting real value from its resources. According to Kevin Behr the definition of Operations is the act of harvesting value from IT resources. Anything that prevents this from happening is a detriment to the business. This means that firefighting and war room sessions are a hindrance to the DevOps model. The following screenshots are good examples of a war room scenario.
Unexpected crashes of websites upon new rollouts still lead to "War Room" situations - despite all the good efforts of DevOps and Agile Delivery/Deployment
Successfully riding the DevOps train: Many of our production customers that made it through the firefighting mode applied the principals of DevOps with a special focus on Application Performance. In this article we describe the steps and milestones companies need to go through in order to level-up their Operations and Engineering Teams to provide more value out of the existing resources.
The Foundation of DevOps
CAMS (Culture, Automation, Measurement, Sharing) are four key areas that are core to the DevOps movement. Culture is the hardest to change but is also the most important because it means a change in the way in which the different teams work together and share the responsibility for the end users of their application. It promotes the usage of development practices in operations to automate deployment. It also allows developers to learn from "the real world" Ops experience and with that mutual exchange it breaks down the walls.
The Lack of Performance Focus
An interesting fact based on the feedback we get from operation teams worldwide: The root cause for about 80% of site crashes or performance problems is related to only about 20% of problem patterns. Want to learn more? Check out blogs such as Top Performance Landmines in Production and I am sure you'll find some issues you already ran into yourself.
Looking at these common problem patterns it is clear that despite all the DevOps efforts lots of performance and scalability-related problems still make it into a release deployment. Why is that? Because our organizations are still very much driven by business requirements that need numerous new features being pushed in ever shorter release cycles. Teams keep growing and are being spread around the world. In order to keep up with the pace, third-party components are included in the code in place of in-house innovation. This "natural" evolution however is also the root cause for firefights and limiting the benefits of DevOps because there is too much focus on pushing functionality through the Deployment Pipeline but not enough focus on Performance.
More developers across more locations including more untested 3rd party code with less time to focus on performance
Plugging Performance into DevOps
In order to focus developers on performance to avoid War Room scenarios you must plug performance into the four pillars of CAMS:
- Culture: Performance as Key Requirement in Dev, Test and Ops
- Automation: Automated Performance Tests already in Continuous Integration
- Measurement: Measure Key Performance Metrics in CI, Test and Ops
- Sharing: Share the same tools and same performance data across Dev, Test and Ops
There are several key milestones to consider:
Milestone 1: Level-Up Performance to Increase Feedback Between Ops and Development
The first step in any DevOps initiative is to get the Ops teams and the Dev teams talking in order to relieve constraints on the business. This might be easy for small teams to accomplish but the larger the organization, the more difficult it becomes as constraints are greater in larger organizations. Operations has to diminish these constraints on the business. This is where APM solutions can help. Beware as not all solutions are created equal. As mentioned in the previous DevOps blog the drive to diminish constraint needs to be applied across the delivery chain. Monitoring just does not cut it here. There needs to be something that not only starts the process but allows the teams to continue to mature and grow. Simple monitoring tools fall short because they only help extinguish fires in operations. How does this continue to drive down constraints? Remember, firefighting is not a part of operations, which means operations should not be looking at fire extinguishers for their DevOps strategies.
Milestone 2: Level-Up Performance Thinking of Engineering
Both Operations and Test Teams have a good understanding of Performance as they deal with it every day. These teams need to educate engineering on the importance of performance as it is a key requirement to software engineering and how it plays a role in large-scale environments under heavy load.
The Ops team shares data with engineering to highlight the performance behavior of their applications under real production load. This helps engineers to prevent these top performance problems from entering production and with that eliminating the need for firefights.
The test teams do their share by providing automated performance test frameworks and educating engineering on how to automate testing for these performance problem patterns.
Milestone 3: Level-Up Load and Capacity Testing
With development executing its own performance tests it's time to level up the test team as well. On one side there is more time to focus on large-scale load tests that need to be executed in a production- like environment. This helps to find any "data-driven", scalability, and "third-party impacted" performance problems. Close collaboration with Ops ensures that tests can be executed either in the prod environment or in a staged environment that mirrors production. Executing these tests in collaboration with Ops allows the teams to become more confident when releasing a new version and also helps with proper capacity planning steps.
Running tests against the production system gives better input for capacity planning and uncovers heavy load application issues
Milestone 4: Level-Up Performance Test Automation
The "traditional" testing teams are used to execute performance and scalability tests in their own environments at the end of a milestone. The goal is to provide these test frameworks and environments to engineering so that these basic performance tests can be executed automatically in the CI environment. In order for this to work you need to make sure that:
- These test frameworks are easy to use and accepted by developers
- Deliver performance metrics to detect the common problem patterns
- These are fully integrated into continuous integration
Automatic Integration Tests run in C/I to detect performance regressions on metrics such as # of SQL Calls, Page Load Time, # of JS files or Images ...
What's Next? Build a Performance Center of Excellence
Many of our customers who jumped on the DevOps train a while back are now promoting a performance culture in their organizations. In the next few blogs we will cover their best practices and tips on either building a separate "Performance Center of Excellence" Team or up-leveling the existing DevOps teams to deliver software with high confidence and fewer War Room weekends.
As enterprises engage with Big Data technologies to develop applications needed to meet operational demands, new computation fabrics are continually being introduced. To leverage these new innovations, organizations are sacrificing market opportunities to gain expertise in learning new systems. In his session at Big Data Expo, Supreet Oberoi, Vice President of Field Engineering at Concurrent, Inc., discussed how to leverage existing infrastructure and investments and future-proof them against e...
May. 29, 2015 03:00 AM EDT Reads: 3,163
The consumption economy is here and so are cloud applications and solutions that offer more than subscription and flat fee models and at the same time are available on a pure consumption model, which not only reduces IT spend but also lowers infrastructure costs, and offers ease of use and availability. In their session at 15th Cloud Expo, Ermanno Bonifazi, CEO & Founder of Solgenia, and Ian Khan, Global Strategic Positioning & Brand Manager at Solgenia, discussed this shifting dynamic with an ...
May. 29, 2015 02:00 AM EDT Reads: 3,244
Due of the rise of Hadoop, many enterprises are now deploying their first small clusters of 10 to 20 servers. At this small scale, the complexity of operating the cluster looks and feels like general data center servers. It is not until the clusters scale, as they inevitably do, when the pain caused by the exponential complexity becomes apparent. We've seen this problem occur time and time again. In his session at Big Data Expo, Greg Bruno, Vice President of Engineering and co-founder of StackI...
May. 29, 2015 01:00 AM EDT Reads: 4,553
Once the decision has been made to move part or all of a workload to the cloud, a methodology for selecting that workload needs to be established. How do you move to the cloud? What does the discovery, assessment and planning look like? What workloads make sense? Which cloud model makes sense for each workload? What are the considerations for how to select the right cloud model? And how does that fit in with the overall IT transformation?
May. 29, 2015 12:00 AM EDT Reads: 4,291
You use an agile process; your goal is to make your organization more agile. But what about your data infrastructure? The truth is, today's databases are anything but agile - they are effectively static repositories that are cumbersome to work with, difficult to change, and cannot keep pace with application demands. Performance suffers as a result, and it takes far longer than it should to deliver new features and capabilities needed to make your organization competitive. As your application an...
May. 29, 2015 12:00 AM EDT Reads: 3,493
17th Cloud Expo, taking place Nov 3-5, 2015, at the Santa Clara Convention Center in Santa Clara, CA, will feature technical sessions from a rock star conference faculty and the leading industry players in the world. Cloud computing is now being embraced by a majority of enterprises of all sizes. Yesterday's debate about public vs. private has transformed into the reality of hybrid cloud: a recent survey shows that 74% of enterprises have a hybrid cloud strategy. Meanwhile, 94% of enterprises a...
May. 28, 2015 11:15 PM EDT Reads: 2,235
The recent trends like cloud computing, social, mobile and Internet of Things are forcing enterprises to modernize in order to compete in the competitive globalized markets. However, enterprises are approaching newer technologies with a more silo-ed way, gaining only sub optimal benefits. The Modern Enterprise model is presented as a newer way to think of enterprise IT, which takes a more holistic approach to embracing modern technologies.
May. 28, 2015 11:00 PM EDT Reads: 6,038
SYS-CON Events announced today that SUSE, a pioneer in open source software, will exhibit at SYS-CON's DevOps Summit 2015 New York, which will take place June 9-11, 2015, at the Javits Center in New York City, NY. SUSE provides reliable, interoperable Linux, cloud infrastructure and storage solutions that give enterprises greater control and flexibility. More than 20 years of engineering excellence, exceptional service and an unrivaled partner ecosystem power the products and support that help ...
May. 28, 2015 10:30 PM EDT Reads: 590
Move from reactive to proactive cloud management in a heterogeneous cloud infrastructure. In his session at 16th Cloud Expo, Manoj Khabe, Innovative Solution-Focused Transformation Leader at Vicom Computer Services, Inc., will show how to replace a help desk-centric approach with an ITIL-based service model and service-centric CMDB that’s tightly integrated with an event and incident management platform. Learn how to expand the scope of operations management to service management. He will al...
May. 28, 2015 10:00 PM EDT Reads: 1,445
The true value of the Internet of Things (IoT) lies not just in the data, but through the services that protect the data, perform the analysis and present findings in a usable way. With many IoT elements rooted in traditional IT components, Big Data and IoT isn’t just a play for enterprise. In fact, the IoT presents SMBs with the prospect of launching entirely new activities and exploring innovative areas. CompTIA research identifies several areas where IoT is expected to have the greatest impac...
May. 28, 2015 09:00 PM EDT Reads: 4,952
There's no doubt that the Internet of Things is driving the next wave of innovation. Google has spent billions over the past few months vacuuming up companies that specialize in smart appliances and machine learning. Already, Philips light bulbs, Audi automobiles, and Samsung washers and dryers can communicate with and be controlled from mobile devices. To take advantage of the opportunities the Internet of Things brings to your business, you'll want to start preparing now.
May. 28, 2015 07:00 PM EDT Reads: 5,814
In a world of ever-accelerating business cycles and fast-changing client expectations, the cloud increasingly serves as a growth engine and a path to new business models. Dynamic clouds enable businesses to continuously reinvent themselves, adapting their business processes, their service and software delivery and their operations to achieve speed-to-market and quick response to customer feedback. As the cloud evolves, the industry has multiple competing cloud technologies, offering on-premises ...
May. 28, 2015 06:00 PM EDT Reads: 4,010
As the world moves from DevOps to NoOps, application deployment to the cloud ought to become a lot simpler. However, applications have been architected with a much tighter coupling than it needs to be which makes deployment in different environments and migration between them harder. The microservices architecture, which is the basis of many new age distributed systems such as OpenStack, Netflix and so on is at the heart of CloudFoundry – a complete developer-oriented Platform as a Service (PaaS...
May. 28, 2015 05:00 PM EDT Reads: 1,658
T-Mobile has been transforming the wireless industry with its “Uncarrier” initiatives. Today as T-Mobile’s IT organization works to transform itself in a like manner, technical foundations built over the last couple of years are now key to their drive for more Agile delivery practices. In his session at DevOps Summit, Martin Krienke, Sr Development Manager at T-Mobile, will discuss where they started their Continuous Delivery journey, where they are today, and where they are going in an effort ...
May. 28, 2015 04:45 PM EDT Reads: 1,422
SAP is delivering break-through innovation combined with fantastic user experience powered by the market-leading in-memory technology, SAP HANA. In his General Session at 15th Cloud Expo, Thorsten Leiduck, VP ISVs & Digital Commerce, SAP, discussed how SAP and partners provide cloud and hybrid cloud solutions as well as real-time Big Data offerings that help companies of all sizes and industries run better. SAP launched an application challenge to award the most innovative SAP HANA and SAP HANA...
May. 28, 2015 04:00 PM EDT Reads: 3,191