Welcome!

Blog Feed Post

Learning by Design

I joined PagerDuty in 2014 when the company was, by many measures, a successful startup. The company was growing and had become the default choice for businesses that needed an effective and highly reliable IT alerting tool. I believe that early success was due, in part, to its founders choosing to solve a very specific problem that they had experienced first-hand. Their experience created a critical ingredient for successful product design: empathy. Having empathy means appreciating the target audience’s context and goals, their problems and priorities, their existing tools, constraints, and assumptions. PagerDuty’s founders developed knowledge over time through their work as on-call engineers, then leveraged this knowledge to build a cloud service for people in similar positions. That foundation in empathy is one of the qualities that convinced me to join the company as a user experience designer.

What convinced me to grow my career at PagerDuty was the company’s commitment to learning. Success in business is transient, and as I learned in my first few months on the product team, PagerDuty had no intention of resting on its laurels. The company aspired to go beyond on-call notifications and support the entire incident response process. To achieve that goal, we started by asking some big questions:

  • What are common models and tools for incident response? What is working well, and what is working poorly?
  • Who participates in incident response? What are their goals?
  • How do responders, managers, and impacted end-users differ in how they think about incidents?
  • How well does PagerDuty model and surface what people in different roles across an organization actually care about?

Desire to Make a Difference

To answer these questions, we could not rely on institutional knowledge and the occasional feature request to inform product design. To develop empathy for people who didn’t use PagerDuty and who faced a set of challenges unfamiliar to us, we had to devise a new way to learn. One of my trusted guidebooks in the discovery phase of user research is Interviewing Users: How to Uncover Compelling Insights by Steve Portigal. I recommend this book to every member of my team. For one, it’s a short read, and secondly, everyone who’s involved in the product-making process should understand how to talk to users in a way that yields valuable insights. Leveraging Portigal’s framework for conducting user research, we interviewed customers across a wide range of industries and organization sizes about their incident response practices. Not everyone attended every interview, but along the way, representatives from product management, engineering, user experience, support, marketing, and sales listened in and shared their key takeaways. We then met as a cross-functional team to synthesize the data and create a prioritized list of needs and opportunities. My colleagues’ engagement and desire to make a positive difference left a deep impression on me. I feel truly fortunate to work with people who genuinely care about our shared mission.

Swapping Engines in Flight

It was a good thing that my team was committed to the cause because we needed every ounce of enthusiasm to digest our findings. In talking to our users, we realized that building upon PagerDuty as it existed in 2014 would make solving their incident response problems extremely difficult. To drive an effective incident response, people need to know what is broken and the scale of impact. PagerDuty couldn’t convey what was broken because it tracked monitoring tool integrations, not the services/applications being monitored. PagerDuty couldn’t convey the scale of impact because it represented every critical alert from a monitoring tool as an incident, even though many customers told us that the way they respond to alerts is different than the way they respond to incidents. We realized that to achieve our goal of supporting the entire IT incident response process, we were going to have to evolve the fundamental constructs within our application without disrupting customers who were happily using the product. In essence, we had to perform the equivalent of swapping out our jet engines in flight.

It took multiple product development teams working in careful collaboration and months of iterative development, but we did it. Today, PagerDuty services can model our customers’ services and applications, no matter how many tools are monitoring them. PagerDuty incidents can model our customers’ incidents, whether there are zero alerts involved or thousands. I am astounded at how seamless the transition has been from our customers’ perspective. Customers were not forced to change their workflow or go through a complex migration process. If a customer wants to continue using PagerDuty for highly reliable IT alerting, they can. And when they’re ready to transform their digital operations, PagerDuty is ready to help them make the leap.

When people ask me what I like most about working at PagerDuty, I have to say it’s the feeling of making an impact. But the best part is that I’m not making that impact alone. I have close partners throughout the organization who have put in the effort to cultivate empathy for those we aspire to serve. They’ve demonstrated their willingness to act upon what we learn, even if it means ripping out and replacing the guts of the product with surgical precision. No one claimed that evolving PagerDuty from an IT alerting tool into a digital operations platform would be easy or straightforward. But here we are, two years later, and I know the possibilities for where we go next are limited only by our own vision.

If you’re interested in shaping that vision, come join us!

The post Learning by Design appeared first on PagerDuty.

Read the original blog entry...

More Stories By PagerDuty Blog

PagerDuty’s operations performance platform helps companies increase reliability. By connecting people, systems and data in a single view, PagerDuty delivers visibility and actionable intelligence across global operations for effective incident resolution management. PagerDuty has over 100 platform partners, and is trusted by Fortune 500 companies and startups alike, including Microsoft, National Instruments, Electronic Arts, Adobe, Rackspace, Etsy, Square and Github.

Latest Stories
Information technology (IT) advances are transforming the way we innovate in business, thereby disrupting the old guard and their predictable status-quo. It’s creating global market turbulence. Industries are converging, and new opportunities and threats are emerging, like never before. So, how are savvy chief information officers (CIOs) leading this transition? Back in 2015, the IBM Institute for Business Value conducted a market study that included the findings from over 1,800 CIO interviews ...
Virtualization over the past years has become a key strategy for IT to acquire multi-tenancy, increase utilization, develop elasticity and improve security. And virtual machines (VMs) are quickly becoming a main vehicle for developing and deploying applications. The introduction of containers seems to be bringing another and perhaps overlapped solution for achieving the same above-mentioned benefits. Are a container and a virtual machine fundamentally the same or different? And how? Is one techn...
DevOps is often described as a combination of technology and culture. Without both, DevOps isn't complete. However, applying the culture to outdated technology is a recipe for disaster; as response times grow and connections between teams are delayed by technology, the culture will die. A Nutanix Enterprise Cloud has many benefits that provide the needed base for a true DevOps paradigm.
What sort of WebRTC based applications can we expect to see over the next year and beyond? One way to predict development trends is to see what sorts of applications startups are building. In his session at @ThingsExpo, Arin Sime, founder of WebRTC.ventures, will discuss the current and likely future trends in WebRTC application development based on real requests for custom applications from real customers, as well as other public sources of information,
As businesses adopt functionalities in cloud computing, it’s imperative that IT operations consistently ensure cloud systems work correctly – all of the time, and to their best capabilities. In his session at @BigDataExpo, Bernd Harzog, CEO and founder of OpsDataStore, will present an industry answer to the common question, “Are you running IT operations as efficiently and as cost effectively as you need to?” He will expound on the industry issues he frequently came up against as an analyst, and...
Keeping pace with advancements in software delivery processes and tooling is taxing even for the most proficient organizations. Point tools, platforms, open source and the increasing adoption of private and public cloud services requires strong engineering rigor - all in the face of developer demands to use the tools of choice. As Agile has settled in as a mainstream practice, now DevOps has emerged as the next wave to improve software delivery speed and output. To make DevOps work, organization...
ChatOps is an emerging topic that has led to the wide availability of integrations between group chat and various other tools/platforms. Currently, HipChat is an extremely powerful collaboration platform due to the various ChatOps integrations that are available. However, DevOps automation can involve orchestration and complex workflows. In his session at @DevOpsSummit at 20th Cloud Expo, Himanshu Chhetri, CTO at Addteq, will cover practical examples and use cases such as self-provisioning infra...
The financial services market is one of the most data-driven industries in the world, yet it’s bogged down by legacy CPU technologies that simply can’t keep up with the task of querying and visualizing billions of records. In his session at 20th Cloud Expo, Jared Parker, Director of Financial Services at Kinetica, will discuss how the advent of advanced in-database analytics on the GPU makes it possible to run sophisticated data science workloads on the same database that is housing the rich inf...
For organizations that have amassed large sums of software complexity, taking a microservices approach is the first step toward DevOps and continuous improvement / development. Integrating system-level analysis with microservices makes it easier to change and add functionality to applications at any time without the increase of risk. Before you start big transformation projects or a cloud migration, make sure these changes won’t take down your entire organization.
Apache Hadoop is emerging as a distributed platform for handling large and fast incoming streams of data. Predictive maintenance, supply chain optimization, and Internet-of-Things analysis are examples where Hadoop provides the scalable storage, processing, and analytics platform to gain meaningful insights from granular data that is typically only valuable from a large-scale, aggregate view. One architecture useful for capturing and analyzing streaming data is the Lambda Architecture, represent...
My team embarked on building a data lake for our sales and marketing data to better understand customer journeys. This required building a hybrid data pipeline to connect our cloud CRM with the new Hadoop Data Lake. One challenge is that IT was not in a position to provide support until we proved value and marketing did not have the experience, so we embarked on the journey ourselves within the product marketing team for our line of business within Progress. In his session at @BigDataExpo, Sum...
Things are changing so quickly in IoT that it would take a wizard to predict which ecosystem will gain the most traction. In order for IoT to reach its potential, smart devices must be able to work together. Today, there are a slew of interoperability standards being promoted by big names to make this happen: HomeKit, Brillo and Alljoyn. In his session at @ThingsExpo, Adam Justice, vice president and general manager of Grid Connect, will review what happens when smart devices don’t work togethe...
SYS-CON Events announced today that Ocean9will exhibit at SYS-CON's 20th International Cloud Expo®, which will take place on June 6-8, 2017, at the Javits Center in New York City, NY. Ocean9 provides cloud services for Backup, Disaster Recovery (DRaaS) and instant Innovation, and redefines enterprise infrastructure with its cloud native subscription offerings for mission critical SAP workloads.
In his session at @ThingsExpo, Eric Lachapelle, CEO of the Professional Evaluation and Certification Board (PECB), will provide an overview of various initiatives to certifiy the security of connected devices and future trends in ensuring public trust of IoT. Eric Lachapelle is the Chief Executive Officer of the Professional Evaluation and Certification Board (PECB), an international certification body. His role is to help companies and individuals to achieve professional, accredited and worldw...
Adding public cloud resources to an existing application can be a daunting process. The tools that you currently use to manage the software and hardware outside the cloud aren’t always the best tools to efficiently grow into the cloud. All of the major configuration management tools have cloud orchestration plugins that can be leveraged, but there are also cloud-native tools that can dramatically improve the efficiency of managing your application lifecycle.