|By Andreas Grabner||
|March 26, 2014 01:15 PM EDT||
DevOps is speeding towards the IT world like a freight train and the hype around it is deafening. There is no reason to be afraid of this change as it is the natural reaction to the agile movement that revolutionized development just a few years ago. By definition, DevOps is the natural alignment of IT performance to business profitability. The relevance of this has yet to be quantified but it has been suggested that the route to the CEO's chair will come from the IT leaders that successfully make the transition to a DevOps model. If this still seems foreign to you, I recommend reading up on DevOps Blog from IT Revolution, the OpsCode Blog, and check out The Phoenix Project.
Despite all the talk around simple monitoring tools, breaking through the walls between Dev and Ops still poses a real challenge. This is because of a misunderstanding around Operations real purpose - extracting real value from its resources. According to Kevin Behr the definition of Operations is the act of harvesting value from IT resources. Anything that prevents this from happening is a detriment to the business. This means that firefighting and war room sessions are a hindrance to the DevOps model. The following screenshots are good examples of a war room scenario.
Unexpected crashes of websites upon new rollouts still lead to "War Room" situations - despite all the good efforts of DevOps and Agile Delivery/Deployment
Successfully riding the DevOps train: Many of our production customers that made it through the firefighting mode applied the principals of DevOps with a special focus on Application Performance. In this article we describe the steps and milestones companies need to go through in order to level-up their Operations and Engineering Teams to provide more value out of the existing resources.
The Foundation of DevOps
CAMS (Culture, Automation, Measurement, Sharing) are four key areas that are core to the DevOps movement. Culture is the hardest to change but is also the most important because it means a change in the way in which the different teams work together and share the responsibility for the end users of their application. It promotes the usage of development practices in operations to automate deployment. It also allows developers to learn from "the real world" Ops experience and with that mutual exchange it breaks down the walls.
The Lack of Performance Focus
An interesting fact based on the feedback we get from operation teams worldwide: The root cause for about 80% of site crashes or performance problems is related to only about 20% of problem patterns. Want to learn more? Check out blogs such as Top Performance Landmines in Production and I am sure you'll find some issues you already ran into yourself.
Looking at these common problem patterns it is clear that despite all the DevOps efforts lots of performance and scalability-related problems still make it into a release deployment. Why is that? Because our organizations are still very much driven by business requirements that need numerous new features being pushed in ever shorter release cycles. Teams keep growing and are being spread around the world. In order to keep up with the pace, third-party components are included in the code in place of in-house innovation. This "natural" evolution however is also the root cause for firefights and limiting the benefits of DevOps because there is too much focus on pushing functionality through the Deployment Pipeline but not enough focus on Performance.
More developers across more locations including more untested 3rd party code with less time to focus on performance
Plugging Performance into DevOps
In order to focus developers on performance to avoid War Room scenarios you must plug performance into the four pillars of CAMS:
- Culture: Performance as Key Requirement in Dev, Test and Ops
- Automation: Automated Performance Tests already in Continuous Integration
- Measurement: Measure Key Performance Metrics in CI, Test and Ops
- Sharing: Share the same tools and same performance data across Dev, Test and Ops
There are several key milestones to consider:
Milestone 1: Level-Up Performance to Increase Feedback Between Ops and Development
The first step in any DevOps initiative is to get the Ops teams and the Dev teams talking in order to relieve constraints on the business. This might be easy for small teams to accomplish but the larger the organization, the more difficult it becomes as constraints are greater in larger organizations. Operations has to diminish these constraints on the business. This is where APM solutions can help. Beware as not all solutions are created equal. As mentioned in the previous DevOps blog the drive to diminish constraint needs to be applied across the delivery chain. Monitoring just does not cut it here. There needs to be something that not only starts the process but allows the teams to continue to mature and grow. Simple monitoring tools fall short because they only help extinguish fires in operations. How does this continue to drive down constraints? Remember, firefighting is not a part of operations, which means operations should not be looking at fire extinguishers for their DevOps strategies.
Milestone 2: Level-Up Performance Thinking of Engineering
Both Operations and Test Teams have a good understanding of Performance as they deal with it every day. These teams need to educate engineering on the importance of performance as it is a key requirement to software engineering and how it plays a role in large-scale environments under heavy load.
The Ops team shares data with engineering to highlight the performance behavior of their applications under real production load. This helps engineers to prevent these top performance problems from entering production and with that eliminating the need for firefights.
The test teams do their share by providing automated performance test frameworks and educating engineering on how to automate testing for these performance problem patterns.
Milestone 3: Level-Up Load and Capacity Testing
With development executing its own performance tests it's time to level up the test team as well. On one side there is more time to focus on large-scale load tests that need to be executed in a production- like environment. This helps to find any "data-driven", scalability, and "third-party impacted" performance problems. Close collaboration with Ops ensures that tests can be executed either in the prod environment or in a staged environment that mirrors production. Executing these tests in collaboration with Ops allows the teams to become more confident when releasing a new version and also helps with proper capacity planning steps.
Running tests against the production system gives better input for capacity planning and uncovers heavy load application issues
Milestone 4: Level-Up Performance Test Automation
The "traditional" testing teams are used to execute performance and scalability tests in their own environments at the end of a milestone. The goal is to provide these test frameworks and environments to engineering so that these basic performance tests can be executed automatically in the CI environment. In order for this to work you need to make sure that:
- These test frameworks are easy to use and accepted by developers
- Deliver performance metrics to detect the common problem patterns
- These are fully integrated into continuous integration
Automatic Integration Tests run in C/I to detect performance regressions on metrics such as # of SQL Calls, Page Load Time, # of JS files or Images ...
What's Next? Build a Performance Center of Excellence
Many of our customers who jumped on the DevOps train a while back are now promoting a performance culture in their organizations. In the next few blogs we will cover their best practices and tips on either building a separate "Performance Center of Excellence" Team or up-leveling the existing DevOps teams to deliver software with high confidence and fewer War Room weekends.
Explosive growth in connected devices. Enormous amounts of data for collection and analysis. Critical use of data for split-second decision making and actionable information. All three are factors in making the Internet of Things a reality. Yet, any one factor would have an IT organization pondering its infrastructure strategy. How should your organization enhance its IT framework to enable an Internet of Things implementation? In his session at @ThingsExpo, James Kirkland, Red Hat's Chief Arch...
Jul. 30, 2015 07:30 PM EDT Reads: 1,385
For IoT to grow as quickly as analyst firms’ project, a lot is going to fall on developers to quickly bring applications to market. But the lack of a standard development platform threatens to slow growth and make application development more time consuming and costly, much like we’ve seen in the mobile space. In his session at @ThingsExpo, Mike Weiner, Product Manager of the Omega DevCloud with KORE Telematics Inc., discussed the evolving requirements for developers as IoT matures and conducte...
Jul. 30, 2015 07:15 PM EDT Reads: 291
Providing the needed data for application development and testing is a huge headache for most organizations. The problems are often the same across companies - speed, quality, cost, and control. Provisioning data can take days or weeks, every time a refresh is required. Using dummy data leads to quality problems. Creating physical copies of large data sets and sending them to distributed teams of developers eats up expensive storage and bandwidth resources. And, all of these copies proliferating...
Jul. 30, 2015 06:30 PM EDT Reads: 877
Malicious agents are moving faster than the speed of business. Even more worrisome, most companies are relying on legacy approaches to security that are no longer capable of meeting current threats. In the modern cloud, threat diversity is rapidly expanding, necessitating more sophisticated security protocols than those used in the past or in desktop environments. Yet companies are falling for cloud security myths that were truths at one time but have evolved out of existence.
Jul. 30, 2015 06:00 PM EDT Reads: 1,795
Digital Transformation is the ultimate goal of cloud computing and related initiatives. The phrase is certainly not a precise one, and as subject to hand-waving and distortion as any high-falutin' terminology in the world of information technology. Yet it is an excellent choice of words to describe what enterprise IT—and by extension, organizations in general—should be working to achieve. Digital Transformation means: handling all the data types being found and created in the organizat...
Jul. 30, 2015 05:00 PM EDT Reads: 1,081
Public Cloud IaaS started its life in the developer and startup communities and has grown rapidly to a $20B+ industry, but it still pales in comparison to how much is spent worldwide on IT: $3.6 trillion. In fact, there are 8.6 million data centers worldwide, the reality is many small and medium sized business have server closets and colocation footprints filled with servers and storage gear. While on-premise environment virtualization may have peaked at 75%, the Public Cloud has lagged in adop...
Jul. 30, 2015 04:00 PM EDT Reads: 2,196
SYS-CON Events announced today that HPM Networks will exhibit at the 17th International Cloud Expo®, which will take place on November 3–5, 2015, at the Santa Clara Convention Center in Santa Clara, CA. For 20 years, HPM Networks has been integrating technology solutions that solve complex business challenges. HPM Networks has designed solutions for both SMB and enterprise customers throughout the San Francisco Bay Area.
Jul. 30, 2015 03:45 PM EDT Reads: 440
The time is ripe for high speed resilient software defined storage solutions with unlimited scalability. ISS has been working with the leading open source projects and developed a commercial high performance solution that is able to grow forever without performance limitations. In his session at Cloud Expo, Alex Gorbachev, President of Intelligent Systems Services Inc., shared foundation principles of Ceph architecture, as well as the design to deliver this storage to traditional SAN storage co...
Jul. 30, 2015 03:00 PM EDT Reads: 1,734
The Software Defined Data Center (SDDC), which enables organizations to seamlessly run in a hybrid cloud model (public + private cloud), is here to stay. IDC estimates that the software-defined networking market will be valued at $3.7 billion by 2016. Security is a key component and benefit of the SDDC, and offers an opportunity to build security 'from the ground up' and weave it into the environment from day one. In his session at 16th Cloud Expo, Reuven Harrison, CTO and Co-Founder of Tufin,...
Jul. 30, 2015 03:00 PM EDT Reads: 476
MuleSoft has announced the findings of its 2015 Connectivity Benchmark Report on the adoption and business impact of APIs. The findings suggest traditional businesses are quickly evolving into "composable enterprises" built out of hundreds of connected software services, applications and devices. Most are embracing the Internet of Things (IoT) and microservices technologies like Docker. A majority are integrating wearables, like smart watches, and more than half plan to generate revenue with ...
Jul. 30, 2015 02:30 PM EDT
The Internet of Everything (IoE) brings together people, process, data and things to make networked connections more relevant and valuable than ever before – transforming information into knowledge and knowledge into wisdom. IoE creates new capabilities, richer experiences, and unprecedented opportunities to improve business and government operations, decision making and mission support capabilities.
Jul. 30, 2015 01:45 PM EDT Reads: 259
The Cloud industry has moved from being more than just being able to provide infrastructure and management services on the Cloud. Enter a new era of Cloud computing where monetization’s services through the Cloud are an essential piece of strategy to feed your organizations bottom-line, your revenue and Profitability. In their session at 16th Cloud Expo, Ermanno Bonifazi, CEO & Founder of Solgenia, and Ian Khan, Global Strategic Positioning & Brand Manager at Solgenia, discussed how to easily o...
Jul. 30, 2015 01:45 PM EDT Reads: 373
In their session at 17th Cloud Expo, Hal Schwartz, CEO of Secure Infrastructure & Services (SIAS), and Chuck Paolillo, CTO of Secure Infrastructure & Services (SIAS), provide a study of cloud adoption trends and the power and flexibility of IBM Power and Pureflex cloud solutions. In his role as CEO of Secure Infrastructure & Services (SIAS), Hal Schwartz provides leadership and direction for the company.
Jul. 30, 2015 01:30 PM EDT
Rapid innovation, changing business landscapes, and new IT demands force businesses to make changes quickly. The DevOps approach is a way to increase business agility through collaboration, communication, and integration across different teams in the IT organization. In his session at DevOps Summit, Chris Van Tuin, Chief Technologist for the Western US at Red Hat, will discuss: The acceleration of application delivery for the business with DevOps
Jul. 30, 2015 12:45 PM EDT Reads: 1,111
The speed of software changes in growing and large scale rapid-paced DevOps environments presents a challenge for continuous testing. Many organizations struggle to get this right. Practices that work for small scale continuous testing may not be sufficient as the requirements grow. In his session at DevOps Summit, Marc Hornbeek, Sr. Solutions Architect of DevOps continuous test solutions at Spirent Communications, explained the best practices of continuous testing at high scale, which is rele...
Jul. 30, 2015 12:00 PM EDT Reads: 1,377