Welcome!

Related Topics: @CloudExpo, Java IoT, Linux Containers, @BigDataExpo, SDN Journal, @DevOpsSummit

@CloudExpo: Blog Post

Monitoring Magic and the Future of APM

The Cloud Precipitating New Agents for Monitoring

At the time when we were looking for a monitoring solution (2006-2007) APM as we know it today had yet to be defined.

There was no Gartner MQ, real-user-monitoring (RUM) was too high level, “agent monitoring” brought concerns of overhead and complexity, instrumenting the application meant to ARM it (i.e., Application Response Measurement), and transaction tagging was a pipe dream.

This created a fierce debate on the risks and rewards of agent vs. agentless monitoring, read The Monitoring Duality of APM. So, when we were developing our monitoring approach, our first priority was to do no harm, then collect performance metrics.

We first implemented agentless RUM technology (i.e. wire data analytics) to gain insight into the application behavior and build a baseline that captured a normal workload. Secondly we focused on synthetic transactions to provide visibility during off peak hours when transaction volumes were low.

Months later, after we had a basic end-user-experience (EUE) solidified and earned the trust of the stakeholders, we began working on deploying agent-based monitoring in production to get a runtime view of the J2EE stack tying it back to the business transactions.

Today’s APM has come a long way since then and has advanced to levels that some may refer to as “magic.” But with these new found powers, APM now comes with a paparazzi of peddlers trying to sell you something, and hucksters who are preaching how to reach the promised land of performance through marketing glossies.

Deciding on an APM solution that's right for your environment can be a daunting task with so many vendors using APM in their messaging today. It would be wise to start small, curate a POC, and get comfortable with the solution nuances, scrutinizing the integration touch points that will coalesce with your current operational processes.

Since there are so many facets to consider when selecting a modern-day APM solution, be wary of the “one size fits all” approach no matter how compelling the story. Remember, you are striving for a robust, real-time application monitoring solution which will most likely utilize a combination of active and passive monitoring techniques.

Consider that the Age of the Cloud has precipitated a new agent concept which is lightweight, deploys quickly, and goes in virtually undetected with zero configuration. These agents are built with a survival mode in mind including a self-healing option for hands-free maintenance.

Looking into the future, I see an evolution towards an intelligent transparent agent, that once given a “target application” to monitor, will run autonomously, and spawn automatically as virtual resources shift. The inherent behavioral learning capabilities will allow it to self-regulate and adjust system resources accordingly, utilizing adaptive controls to support predefined SLA's.

Continue reading, click here for the full article.

Image: Snugg LePup/Flickr

More Stories By Larry Dragich

Larry Dragich is actively involved with industry leaders, sharing knowledge of Application Performance Management (APM) technologies, from best practices and technical workflows, to resource allocation and approaches for implementation. He has been working in the APM space since 2006 where he built the Enterprise Systems Management team which is now the focal point for IT performance monitoring and capacity planning activities.

Latest Stories
In his session at 20th Cloud Expo, Scott Davis, CTO of Embotics, discussed how automation can provide the dynamic management required to cost-effectively deliver microservices and container solutions at scale. He also discussed how flexible automation is the key to effectively bridging and seamlessly coordinating both IT and developer needs for component orchestration across disparate clouds – an increasingly important requirement at today’s multi-cloud enterprise.
In his session at @ThingsExpo, Dr. Robert Cohen, an economist and senior fellow at the Economic Strategy Institute, presented the findings of a series of six detailed case studies of how large corporations are implementing IoT. The session explored how IoT has improved their economic performance, had major impacts on business models and resulted in impressive ROIs. The companies covered span manufacturing and services firms. He also explored servicification, how manufacturing firms shift from se...
As more and more companies are making the shift from on-premises to public cloud, the standard approach to DevOps is evolving. From encryption, compliance and regulations like GDPR, security in the cloud has become a hot topic. Many DevOps-focused companies have hired dedicated staff to fulfill these requirements, often creating further siloes, complexity and cost. This session aims to highlight existing DevOps cultural approaches, tooling and how security can be wrapped in every facet of the bu...
WebRTC is the future of browser-to-browser communications, and continues to make inroads into the traditional, difficult, plug-in web communications world. The 6th WebRTC Summit continues our tradition of delivering the latest and greatest presentations within the world of WebRTC. Topics include voice calling, video chat, P2P file sharing, and use cases that have already leveraged the power and convenience of WebRTC.
Translating agile methodology into real-world best practices within the modern software factory has driven widespread DevOps adoption, yet much work remains to expand workflows and tooling across the enterprise. As models evolve from pockets of experimentation into wholescale organizational reinvention, practitioners find themselves challenged to incorporate the culture and architecture necessary to support DevOps at scale.
SYS-CON Events announced today that CA Technologies has been named “Platinum Sponsor” of SYS-CON's 21st International Cloud Expo®, which will take place October 31-November 2, 2017, at the Santa Clara Convention Center in Santa Clara, CA. CA Technologies helps customers succeed in a future where every business – from apparel to energy – is being rewritten by software. From planning to development to management to security, CA creates software that fuels transformation for companies in the applic...
IT organizations are moving to the cloud in hopes to approve efficiency, increase agility and save money. Migrating workloads might seem like a simple task, but what many businesses don’t realize is that application migration criteria differs across organizations, making it difficult for architects to arrive at an accurate TCO number. In his session at 21st Cloud Expo, Joe Kinsella, CTO of CloudHealth Technologies, will offer a systematic approach to understanding the TCO of a cloud application...
"With Digital Experience Monitoring what used to be a simple visit to a web page has exploded into app on phones, data from social media feeds, competitive benchmarking - these are all components that are only available because of some type of digital asset," explained Leo Vasiliou, Director of Web Performance Engineering at Catchpoint Systems, in this SYS-CON.tv interview at DevOps Summit at 20th Cloud Expo, held June 6-8, 2017, at the Javits Center in New York City, NY.
SYS-CON Events announced today that Secure Channels, a cybersecurity firm, will exhibit at SYS-CON's 21st International Cloud Expo®, which will take place on Oct 31 – Nov 2, 2017, at the Santa Clara Convention Center in Santa Clara, CA. Secure Channels, Inc. offers several products and solutions to its many clients, helping them protect critical data from being compromised and access to computer networks from the unauthorized. The company develops comprehensive data encryption security strategie...
An increasing number of companies are creating products that combine data with analytical capabilities. Running interactive queries on Big Data requires complex architectures to store and query data effectively, typically involving data streams, an choosing efficient file format/database and multiple independent systems that are tied together through custom-engineered pipelines. In his session at @BigDataExpo at @ThingsExpo, Tomer Levi, a senior software engineer at Intel’s Advanced Analytics ...
The goal of Continuous Testing is to shift testing left to find defects earlier and release software faster. This can be achieved by integrating a set of open source functional and performance testing tools in the early stages of your software delivery lifecycle. There is one process that binds all application delivery stages together into one well-orchestrated machine: Continuous Testing. Continuous Testing is the conveyer belt between the Software Factory and production stages. Artifacts are m...
Cloud resources, although available in abundance, are inherently volatile. For transactional computing, like ERP and most enterprise software, this is a challenge as transactional integrity and data fidelity is paramount – making it a challenge to create cloud native applications while relying on RDBMS. In his session at 21st Cloud Expo, Claus Jepsen, Chief Architect and Head of Innovation Labs at Unit4, will explore that in order to create distributed and scalable solutions ensuring high availa...
Cloud adoption is often driven by a desire to increase efficiency, boost agility and save money. All too often, however, the reality involves unpredictable cost spikes and lack of oversight due to resource limitations. In his session at 20th Cloud Expo, Joe Kinsella, CTO and Founder of CloudHealth Technologies, tackled the question: “How do you build a fully optimized cloud?” He will examine: Why TCO is critical to achieving cloud success – and why attendees should be thinking holistically abo...
In IT, we sometimes coin terms for things before we know exactly what they are and how they’ll be used. The resulting terms may capture a common set of aspirations and goals – as “cloud” did broadly for on-demand, self-service, and flexible computing. But such a term can also lump together diverse and even competing practices, technologies, and priorities to the point where important distinctions are glossed over and lost.
Internet-of-Things discussions can end up either going down the consumer gadget rabbit hole or focused on the sort of data logging that industrial manufacturers have been doing forever. However, in fact, companies today are already using IoT data both to optimize their operational technology and to improve the experience of customer interactions in novel ways. In his session at @ThingsExpo, Gordon Haff, Red Hat Technology Evangelist, shared examples from a wide range of industries – including en...