Welcome!

Blog Feed Post

AppDynamics interviews container expert Liz Rice

I was delighted to spend time with container guru Liz Rice recently, ahead of the presentation Liz will deliver at AppD Summit Europe.

There are over 460K Dockerized applications and more than 5 billion containers have been pulled so far. Why do you think containers as a concept have caught on so quickly?

Containerization-related technologies existed before Docker arrived on the scene, but Docker made it easy to use at the command line. This opened the advantages of containers up to the mainstream developer and allowed them to explore how containers make developers’ lives easier.

For example, a team can use container images to recreate exactly the same environment on their laptops as in production, simplifying the whole process of dependency management and avoiding the syndrome of “but it works fine on my machine.”

The developer community soon started to explore how well containers fit with a microservices architecture, and with CI/CD (continuous integration / continuous deployment) pipelines that make building and shipping code really easy. This really appeals to organizations who want to be able to ship features quickly. Businesses always want their software teams delivering faster!

What tips would you share with someone embarking on their container journey?

Containers can touch everything in the software life cycle, from the development process through continuous delivery approaches, orchestration, security, and site reliability — so the process of adopting containers can seem overwhelming at first. But you don’t necessarily have to do everything all at once!

It’s a good idea to remember why you are exploring container usage in the first place, and for many organizations, at a strategic level it’s all about increasing the speed of deployment. Containers are a helpful tool when your teams are working with an agile methodology and shipping code frequently through a CI/CD pipeline.

The journey to containerization can look very different in different companies, or even different teams within the same business. We increasingly see enterprises running large-scale containerized workloads in production, but there are also organizations where they are only using containers to simplify their development processes. In many cases developer teams have brought containers to their workflow simply to make their own lives easier, and usage within the organization has grown from there. In other enterprises the initial impetus comes from a platform project that builds an orchestrated cluster and offers it to the rest of the business. Individual software teams across the business are invited to start moving their workloads into containers so that they can run within that cluster. Fortunately, many people are willing to share stories about the approaches they took, talking about them in case study presentations at conferences and meetups.

At a practical level, there are a number of ways you can find out more about how and where to start. I’d recommend a useful site called Katacoda, an interactive learning platform which has courses and labs that introduce you to Docker and many other tools. Another option is to attend my session on May 4th, of course!

What are the typical challenges to container adoption? How can these be overcome?

One of the greatest challenges in advancing container usage is bringing the CISO on board. Security leads can have questions around how containers will affect security of deployment across the enterprise.

For example, if you’re shipping code more often, does this increase your risk profile? The good news there is that if you are embracing continuous delivery, you are shipping small incremental changes, so there isn’t a hugely different risk profile from one update to the next, and you can introduce automated checks like container image scanning to detect any known vulnerabilities in your dependencies.

Let’s also not forget microservices, which go hand in hand with containers. It’s much easier to reason about what an individual microservice should be doing in terms of accessing particular resources or user IDs than it is across a huge monolithic codebase. At Aqua we have tools that make it simple to apply these policies at the level of individual containers.  

Equally, CTOs need to be confident that they can ship well-tested code using containerized processes. You need to sell the technical and cultural benefits of being able to ship code more quickly, and follow a more agile path. Using containers can turn well-established deployment processes on their head and they are not necessarily right for all organizations. If your world is dominated by GANTT charts, then containers and CI/CD may not be for you!

Focusing in on container security, do you think there has been progress in this area?

Definitely. There have been concerns around container security in the past, which led to improvements in the Linux kernel to address those concerns, like user namespacing for example.

In some ways, containerizing workloads can help with security issues. For example, the ShellShock vulnerability affected a shell that’s present on most Linux machines. Most microservices don’t need a shell, so you can run them in containers that don’t include the shell. In the event of another ShellShock being discovered, you would still have to worry about patching the host, but you wouldn’t need to change those microservices code at all. That means less testing and less risk when the patch is applied.

Many sizable organizations are using containers in secure environments and have reached a level of robustness now. The Enterprise Edition of Docker with its focus on testing and certification is an example of this maturity, and of course tools like Aqua are now helping enterprises achieve really robust security for containerized deployments.  

How will skill sets need to evolve to take advantage of the potential containers can offer?

Containers are making the traditional developer role easier, but there are a lot of skills to be learned on the operations side. Orchestrated deployment is a learning curve area, as you won’t know which machine will exactly run which bit of code. As enterprises have bigger and bigger orchestrated deployments, there needs to be a mind shift from individual machines to a more holistic view. Monitoring, tracing, alerting, and diagnostics are all areas where the tooling is changing dramatically to take account of this. Traditional orchestration involves manual decisions, but in the container world, it’s more automated, less proprietary, and more collaborative.

Flash forward three years — how do you see containers developing?

Over the last year, I have seen a lot of growth in people moving from “playing” with containers to using them for real in production. Container fundamentals are already fairly stable, but there is still room for improvement in tooling for CI/CD, monitoring, diagnostics, and programmable infrastructure for example. Much like DevOps, containers are really moving into the mainstream.

Learn more

If you would like to learn more from Liz Rice and attend her Containers: Give me the hype not the facts session, please register for your place at AppD Summit Europe. The main conference day is free to attend.

The post AppDynamics interviews container expert Liz Rice appeared first on Application Performance Monitoring Blog | AppDynamics.

Read the original blog entry...

More Stories By Jyoti Bansal

In high-production environments where release cycles are measured in hours or minutes — not days or weeks — there's little room for mistakes and no room for confusion. Everyone has to understand what's happening, in real time, and have the means to do whatever is necessary to keep applications up and running optimally.

DevOps is a high-stakes world, but done well, it delivers the agility and performance to significantly impact business competitiveness.

Latest Stories
"We are still a relatively small software house and we are focusing on certain industries like FinTech, med tech, energy and utilities. We help our customers with their digital transformation," noted Piotr Stawinski, Founder and CEO of EARP Integration, in this SYS-CON.tv interview at 20th Cloud Expo, held June 6-8, 2017, at the Javits Center in New York City, NY.
"I think DevOps is now a rambunctious teenager – it’s starting to get a mind of its own, wanting to get its own things but it still needs some adult supervision," explained Thomas Hooker, VP of marketing at CollabNet, in this SYS-CON.tv interview at DevOps Summit at 20th Cloud Expo, held June 6-8, 2017, at the Javits Center in New York City, NY.
"DX encompasses the continuing technology revolution, and is addressing society's most important issues throughout the entire $78 trillion 21st-century global economy," said Roger Strukhoff, Conference Chair. "DX World Expo has organized these issues along 10 tracks with more than 150 of the world's top speakers coming to Istanbul to help change the world."
"We've been engaging with a lot of customers including Panasonic, we've been involved with Cisco and now we're working with the U.S. government - the Department of Homeland Security," explained Peter Jung, Chief Product Officer at Pulzze Systems, in this SYS-CON.tv interview at @ThingsExpo, held June 6-8, 2017, at the Javits Center in New York City, NY.
"We're here to tell the world about our cloud-scale infrastructure that we have at Juniper combined with the world-class security that we put into the cloud," explained Lisa Guess, VP of Systems Engineering at Juniper Networks, in this SYS-CON.tv interview at 20th Cloud Expo, held June 6-8, 2017, at the Javits Center in New York City, NY.
"I will be talking about ChatOps and ChatOps as a way to solve some problems in the DevOps space," explained Himanshu Chhetri, CTO of Addteq, in this SYS-CON.tv interview at @DevOpsSummit at 20th Cloud Expo, held June 6-8, 2017, at the Javits Center in New York City, NY.
"We are focused on SAP running in the clouds, to make this super easy because we believe in the tremendous value of those powerful worlds - SAP and the cloud," explained Frank Stienhans, CTO of Ocean9, Inc., in this SYS-CON.tv interview at 20th Cloud Expo, held June 6-8, 2017, at the Javits Center in New York City, NY.
Your homes and cars can be automated and self-serviced. Why can't your storage? From simply asking questions to analyze and troubleshoot your infrastructure, to provisioning storage with snapshots, recovery and replication, your wildest sci-fi dream has come true. In his session at @DevOpsSummit at 20th Cloud Expo, Dan Florea, Director of Product Management at Tintri, provided a ChatOps demo where you can talk to your storage and manage it from anywhere, through Slack and similar services with...
The financial services market is one of the most data-driven industries in the world, yet it’s bogged down by legacy CPU technologies that simply can’t keep up with the task of querying and visualizing billions of records. In his session at 20th Cloud Expo, Karthik Lalithraj, a Principal Solutions Architect at Kinetica, discussed how the advent of advanced in-database analytics on the GPU makes it possible to run sophisticated data science workloads on the same database that is housing the rich...
"We are an IT services solution provider and we sell software to support those solutions. Our focus and key areas are around security, enterprise monitoring, and continuous delivery optimization," noted John Balsavage, President of A&I Solutions, in this SYS-CON.tv interview at 20th Cloud Expo, held June 6-8, 2017, at the Javits Center in New York City, NY.
SYS-CON Events announced today that Massive Networks will exhibit at SYS-CON's 21st International Cloud Expo®, which will take place on Oct 31 – Nov 2, 2017, at the Santa Clara Convention Center in Santa Clara, CA. Massive Networks mission is simple. To help your business operate seamlessly with fast, reliable, and secure internet and network solutions. Improve your customer's experience with outstanding connections to your cloud.
Everything run by electricity will eventually be connected to the Internet. Get ahead of the Internet of Things revolution and join Akvelon expert and IoT industry leader, Sergey Grebnov, in his session at @ThingsExpo, for an educational dive into the world of managing your home, workplace and all the devices they contain with the power of machine-based AI and intelligent Bot services for a completely streamlined experience.
"We want to show that our solution is far less expensive with a much better total cost of ownership so we announced several key features. One is called geo-distributed erasure coding, another is support for KVM and we introduced a new capability called Multi-Part," explained Tim Desai, Senior Product Marketing Manager at Hitachi Data Systems, in this SYS-CON.tv interview at 20th Cloud Expo, held June 6-8, 2017, at the Javits Center in New York City, NY.
There is a huge demand for responsive, real-time mobile and web experiences, but current architectural patterns do not easily accommodate applications that respond to events in real time. Common solutions using message queues or HTTP long-polling quickly lead to resiliency, scalability and development velocity challenges. In his session at 21st Cloud Expo, Ryland Degnan, a Senior Software Engineer on the Netflix Edge Platform team, will discuss how by leveraging a reactive stream-based protocol,...
DevOps at Cloud Expo, taking place October 31 - November 2, 2017, at the Santa Clara Convention Center in Santa Clara, CA, is co-located with 21st Cloud Expo and will feature technical sessions from a rock star conference faculty and the leading industry players in the world. The widespread success of cloud computing is driving the DevOps revolution in enterprise IT. Now as never before, development teams must communicate and collaborate in a dynamic, 24/7/365 environment. There is no time to w...