Welcome!

Blog Feed Post

What’s the big deal with AppDynamics Azure Monitoring?

Let me guess: you’re probably expecting AppDynamics to be “another monitoring solution” for Windows Azure. You’re expecting it to show you basic server metrics like CPU, memory and disk I/O, along with a few CLR counters thrown in for good measure. Well, I’m sorry to disappoint you–but these metrics in isolation are about as useful as a chocolate teapot for monitoring your applications and business in the Cloud.

When end users interact with your Windows Azure application, they don’t experience servers, Azure roles, CPU, or CLR counters. Rather, they experience business transactions. Monitoring infrastructure metrics in the cloud is probably the worst KPI for managing Quality-of-Service (Qos) and application performance. Another problem in the cloud is application architectures, which have become virtual, dynamic and distributed. Applications are no longer just an app server and database. SOA design principles and Cloud Services mean that business transaction logic now executes across many distributed application tiers and services.

For example, the Windows Azure PaaS platform has several services that allow organizations to run their mission-critical applications in the Cloud. Azure Compute provides web and windows services, Azure AppFabric provides messaging, security and caching, and SQL Azure and Azure Storage provides data management capabilities.

A major challenge for organizations is gaining end-to-end visibility of how their business (transactions) perform in the Azure Cloud. AppDynamics was founded in 2008 to address this problem, and it’s a key reason why Microsoft announced AppDynamics as a partner to help their customers manage application performance within Windows Azure. For AppDynamics, it’s an enormous privilege and responsibility to be selected by Microsoft, given Windows Azure is a huge strategic play as Microsoft looks to dominate the Cloud Computing market.

So what’s the big deal with AppDynamics in Azure? Let’s take a look at what unique capabilities AppDynamics offers for Windows Azure customers:

Application Mapping for Windows Azure Roles and Services
AppDynamics is able to automatically discover, map, and visualize the living topology and performance of a production application running inside Windows Azure. This allows organizations to understand the application tiers, Azure roles, and service dependencies inside the Azure Cloud so they can rapidly isolate which components are responsible for slow performance or poor QoS. For example, take a look at the following screenshot of an application hosted inside Windows Azure:

Business Transaction-centric Monitoring
What’s unique about AppDynamics production monitoring is that our key unit of measurement is business transactions–rather than, say, web pages, servers, or CLR metrics. AppDynamics can auto-discover the business transactions that end users are experiencing in the application. This gives IT Operations and development teams the context and visibility they need to monitor and manage the end user experience, along with understanding how transactions flow across and inside each application tier, Azure Role, and Services.

For example, take a look at this screenshot, which shows what business transaction monitoring looks like in AppDynamics:

Deep Diagnostics for Rapid Root-Cause Analysis
For situations where end users experience poor QoS, slow downs, or outages, it’s possible for organizations to pinpoint the root cause of issues in minutes using AppDynamics. AppDynamics uses intelligent analytics to self-learn and baseline the normal performance of every business transaction in an application. When performance deviates, AppDynamics captures complete diagnostic snapshots of business transactions that breach their performance baseline. Snapshots represent the complete distributed flow and code execution of slow business transactions as they execute across and inside each application tier, Azure role and service. This means IT Operations and developers see right down to the line of code responsible for issues in production that impact their end users and QoS. Root cause is therefore only minutes away.

Here is a screenshot of how a single user business transaction executes within Windows Azure:

Here is the code execution of that same user business transaction within the 2nd Application Tier “LowValueOrders” worker role:

Auto–Scaling for provisioning Windows Azure resource on-the-fly
A key benefit of Cloud is resource elasticity – the ability for an application to leverage additional computing resource on-the-fly, so it’s able to scale up and down all by itself. For auto-scaling to happen in Windows Azure, the application needs to know when to scale up and when to scale down. AppDynamics provides a comprehensive policy wizard that can auto-scale an application in Windows Azure based on business transaction KPI like end user response time or throughput. For example, if order throughput in your application were to drop due to increased load on your application, you might want to spin up more Azure compute (web/worker roles) to compensate for the additional demand until throughput stabilizes or load reduces.

For example, here’s a screenshot of the AppDynamics Azure auto-scaling policy wizard:

If you want to try AppDynamics for Windows Azure you can sign-up at the marketplace right here.

App Man

Read the original blog entry...

More Stories By AppDynamics Blog

In high-production environments where release cycles are measured in hours or minutes — not days or weeks — there's little room for mistakes and no room for confusion. Everyone has to understand what's happening, in real time, and have the means to do whatever is necessary to keep applications up and running optimally.

DevOps is a high-stakes world, but done well, it delivers the agility and performance to significantly impact business competitiveness.

Latest Stories
Amazon has gradually rolled out parts of its IoT offerings in the last year, but these are just the tip of the iceberg. In addition to optimizing their back-end AWS offerings, Amazon is laying the ground work to be a major force in IoT – especially in the connected home and office. Amazon is extending its reach by building on its dominant Cloud IoT platform, its Dash Button strategy, recently announced Replenishment Services, the Echo/Alexa voice recognition control platform, the 6-7 strategic...
The best-practices for building IoT applications with Go Code that attendees can use to build their own IoT applications. In his session at @ThingsExpo, Indraneel Mitra, Senior Solutions Architect & Technology Evangelist at Cognizant, provided valuable information and resources for both novice and experienced developers on how to get started with IoT and Golang in a day. He also provided information on how to use Intel Arduino Kit, Go Robotics API and AWS IoT stack to build an application tha...
IoT generates lots of temporal data. But how do you unlock its value? You need to discover patterns that are repeatable in vast quantities of data, understand their meaning, and implement scalable monitoring across multiple data streams in order to monetize the discoveries and insights. Motif discovery and deep learning platforms are emerging to visualize sensor data, to search for patterns and to build application that can monitor real time streams efficiently. In his session at @ThingsExpo, ...
Adding public cloud resources to an existing application can be a daunting process. The tools that you currently use to manage the software and hardware outside the cloud aren’t always the best tools to efficiently grow into the cloud. All of the major configuration management tools have cloud orchestration plugins that can be leveraged, but there are also cloud-native tools that can dramatically improve the efficiency of managing your application lifecycle. In his session at 18th Cloud Expo, ...
Ovum, a leading technology analyst firm, has published an in-depth report, Ovum Decision Matrix: Selecting a DevOps Release Management Solution, 2016–17. The report focuses on the automation aspects of DevOps, Release Management and compares solutions from the leading vendors.
Continuous testing helps bridge the gap between developing quickly and maintaining high quality products. But to implement continuous testing, CTOs must take a strategic approach to building a testing infrastructure and toolset that empowers their team to move fast. Download our guide to laying the groundwork for a scalable continuous testing strategy.
SYS-CON Events announced today that LeaseWeb USA, a cloud Infrastructure-as-a-Service (IaaS) provider, will exhibit at the 19th International Cloud Expo, which will take place on November 1–3, 2016, at the Santa Clara Convention Center in Santa Clara, CA. LeaseWeb is one of the world's largest hosting brands. The company helps customers define, develop and deploy IT infrastructure tailored to their exact business needs, by combining various kinds cloud solutions.
Choosing the right cloud for your workloads is a balancing act that can cost your organization time, money and aggravation - unless you get it right the first time. Economics, speed, performance, accessibility, administrative needs and security all play a vital role in dictating your approach to the cloud. Without knowing the right questions to ask, you could wind up paying for capacity you'll never need or underestimating the resources required to run your applications.
Up until last year, enterprises that were looking into cloud services usually undertook a long-term pilot with one of the large cloud providers, running test and dev workloads in the cloud. With cloud’s transition to mainstream adoption in 2015, and with enterprises migrating more and more workloads into the cloud and in between public and private environments, the single-provider approach must be revisited. In his session at 18th Cloud Expo, Yoav Mor, multi-cloud solution evangelist at Cloudy...
SYS-CON Events announced today that 910Telecom will exhibit at the 19th International Cloud Expo, which will take place on November 1–3, 2016, at the Santa Clara Convention Center in Santa Clara, CA. Housed in the classic Denver Gas & Electric Building, 910 15th St., 910Telecom is a carrier-neutral telecom hotel located in the heart of Denver. Adjacent to CenturyLink, AT&T, and Denver Main, 910Telecom offers connectivity to all major carriers, Internet service providers, Internet backbones and ...
StackIQ has announced the release of Stacki 3.2. Stacki is an easy-to-use Linux server provisioning tool. Stacki 3.2 delivers new capabilities that simplify the automation and integration of site-specific requirements. StackIQ is the commercial entity behind this open source bare metal provisioning tool. Since the release of Stacki in June of 2015, the Stacki core team has been focused on making the Community Edition meet the needs of members of the community, adding features and value, while ...
Big Data, cloud, analytics, contextual information, wearable tech, sensors, mobility, and WebRTC: together, these advances have created a perfect storm of technologies that are disrupting and transforming classic communications models and ecosystems. In his session at @ThingsExpo, Erik Perotti, Senior Manager of New Ventures on Plantronics’ Innovation team, provided an overview of this technological shift, including associated business and consumer communications impacts, and opportunities it ...
SYS-CON Events announced today that Venafi, the Immune System for the Internet™ and the leading provider of Next Generation Trust Protection, will exhibit at @DevOpsSummit at 19th International Cloud Expo, which will take place on November 1–3, 2016, at the Santa Clara Convention Center in Santa Clara, CA. Venafi is the Immune System for the Internet™ that protects the foundation of all cybersecurity – cryptographic keys and digital certificates – so they can’t be misused by bad guys in attacks...
Security, data privacy, reliability, and regulatory compliance are critical factors when evaluating whether to move business applications from in-house, client-hosted environments to a cloud platform. Quality assurance plays a vital role in ensuring that the appropriate level of risk assessment, verification, and validation takes place to ensure business continuity during the migration to a new cloud platform.
It’s 2016: buildings are smart, connected and the IoT is fundamentally altering how control and operating systems work and speak to each other. Platforms across the enterprise are networked via inexpensive sensors to collect massive amounts of data for analytics, information management, and insights that can be used to continuously improve operations. In his session at @ThingsExpo, Brian Chemel, Co-Founder and CTO of Digital Lumens, will explore: The benefits sensor-networked systems bring to ...