Welcome!

Related Topics: SDN Journal, Containers Expo Blog

SDN Journal: Blog Feed Post

Data Center Architecture: Together and Apart

The datacenter represents a diverse set of orchestrated resources bound together by the applications they serve

The challenge in architecting, building, and managing data centers is one of balance. There are forces competing to both push together and pull apart datacenter resources. Finding an equilibrium point that is technological sustainable, operationally viable, and business friendly is challenging. The result is frequently a set of compromises that outweigh the advantages.

Logically together

The datacenter represents a diverse set of orchestrated resources bound together by the applications they serve. At its most simplest, these resources are physically co-located. At its extreme, these resources are geographically distributed across many sites. Whatever the physical layout, these resources are under pressure to be treated as a single logical group.

Resource collaboration - The datacenter is a collection of compute and storage resources that must work in concert in support of application workloads. The simple requirement of coordination creates an inward force pulling resources closer together, even if only logically. How can multiple elements work together towards a common goal if they are completely separate?

The answer is that they cannot. And as IT moves increasingly towards distributed applications, the interdependence between resources only grows.

Interestingly, the performance advantages of distributed architectures are only meaningful when communication between servers is uninhibited. If the network that makes communication possible slows down, the efficacy of the distributed architecture decreases. This means that datacenter architects must solve simultaneously for compute and storage demand, and the interconnect capacity required between them.

Resource availability - Building out a datacenter is an exercise in matching resource capacity to demand. But not just in aggregate.

Individual applications, tenants, and geographies all place localized demands on datacenter resources. If the aggregate demand is sufficient but the resources exist in separate resource pools, you end up in a perpetual state of mismatch. There is always too much or too little workload capacity. The former means you have overbuilt. The latter leaves you wanting for more, which oddly enough means you end up having to overbuild.

Combatting these resource islands requires pulling resources closer together. In the most simple case, this is a physical act. But even if resources cannot be physically co-located, there are entire classes of technologies whose primary function is to allow physically separate resources to behave as if they are in close proximity.

Of course this does not come without a cost. The complexity of managing the disparate technologies required to logically pool physically separate resources can be prohibitively difficult. Even the most skilled specialists have to invest time in creating a properly engineered fabric between sites that accounts for queuing, prioritization, load balancing, and so on. The number of protocols and technologies required is high, and the volume of devices over which they must be applied can be huge. The result is a level of complexity that makes the network more expensive to manage and more difficult to change.

Organizational process - Friction is greatest at boundaries. Whenever a task requires involvement across different organizations or teams, the act of human coordination imposes a tax on both effort and time. In larger organizations, the handoff between teams might be automated to reduce communication mistakes (as with a ticketing system), but the shift in context is still expensive.

This creates organizational pressure to pull together things that might otherwise be separate. If distributed resources can be logically centralized and managed within a common organization, it reduces the dependence on outside teams. The removal of boundaries from common workflows lowers organizational friction and makes easier the overall task of managing the infrastructure.

Physically separate

At the same time that forces are pulling things together, there are equally strong oppositional forces exerting outward pressure on datacenter resources.

Business continuity - For many companies, the datacenter represents a mission critical element of their infrastructure. For companies whose existence depends on the presence of the resources within the datacenter (be they data, servers, or applications), it is untenably risky to rely on a single physical site. This exerts an outward force on resources as companies must create multiple physical sites, typically separated by enough distance that a disaster would not meaningfully impact all sites.

Despite the operational desire to keep things together, the risk to the business dictates that resources be physically separate.

Natural expansion - As resources are added to a datacenter, they are typically installed in racks in relative close proximity to each other. When racks are empty, there is no reason to unnecessarily create physical separation between resources working in concert. Over time, adjacent rack space is filled through the natural expansion of compute, storage, and networking capacity.

As equipment expands, available rack space is depleted, and new racks and rows are populated. Eventually, the device sprawl can occupy entire data centers.

Imagine now that a cluster of servers occupies a rack in one corner of the datacenter. If that cluster is to be expanded, where does the next server go? If the nearby racks are already built out, that resource must be installed some physical distance away from the resources with which it must coordinate.

It is near impossible to plan for all future growth at the time of datacenter inception. Leaving enough space in adjacent racks to account for a decade of growth is impractically expensive. A sparsely populated datacenter suffers from poor space utilization, challenging power distribution, and difficult cabling. Thus, the mere act of expansion actually exerts an outward force leading to physically distributed resources.

Real estate - Sometimes, even when architects want to keep resources together, physical limitations create problems. There is no more immovable object than real estate (which serves as a proxy for all of space, power, and HVAC). In some cases, it is impossible to build out either laterally or even up. In other cases, there is no additional power to be had from the grid. Either of these scenarios forces an expansion to another site, which requires the physical separation of resources that might be expected to function in concert.

Additionally, as land rates change and technologies evolve, the best spots for data centers are not always known. It is difficult at best to predict with enough certainty how a physical site will evolve over an arbitrarily long time horizon. For example, not long ago, the thought of building cooling-hungry data centers in the hot desert was foreign. Today, Las Vegas is home to some of the most cutting edge facilities in the world. This means that geographical dispersion is likely a certainty for large companies. The forces pulling resources physically apart are unlikely to be neutralized.

Finding a balance

Given the strong forces working to keep resources logically together and the equally strong forces keeping them physically separate, how does anyone find a balance?

The price for balance is cost and complexity. You pay for reach directly, and control requires complexity. Both translate into higher carrying costs for the infrastructure. The push-pull dynamic in datacenters is not going away anytime soon. In fact, a move towards more distributed applications will only make harder the balancing act that already exists.

Newer technology offerings like SDN and datacenter fabrics offer some hope, but only insofar as they offer alternatives to the existing problems. Whatever the solution, architects will need to evaluate approaches based not just on the features but on the long-term costs of those features.

[Today’s fun fact: “Way” is the most frequently used noun in the English language. No way!]

The post Datacenter architecture: Together and apart appeared first on Plexxi.

More Stories By Michael Bushong

The best marketing efforts leverage deep technology understanding with a highly-approachable means of communicating. Plexxi's Vice President of Marketing Michael Bushong has acquired these skills having spent 12 years at Juniper Networks where he led product management, product strategy and product marketing organizations for Juniper's flagship operating system, Junos. Michael spent the last several years at Juniper leading their SDN efforts across both service provider and enterprise markets. Prior to Juniper, Michael spent time at database supplier Sybase, and ASIC design tool companies Synopsis and Magma Design Automation. Michael's undergraduate work at the University of California Berkeley in advanced fluid mechanics and heat transfer lend new meaning to the marketing phrase "This isn't rocket science."

Latest Stories
With Cloud Foundry you can easily deploy and use apps utilizing websocket technology, but not everybody realizes that scaling them out is not that trivial. In his session at 21st Cloud Expo, Roman Swoszowski, CTO and VP, Cloud Foundry Services, at Grape Up, will show you an example of how to deal with this issue. He will demonstrate a cloud-native Spring Boot app running in Cloud Foundry and communicating with clients over websocket protocol that can be easily scaled horizontally and coordinate...
In his session at 20th Cloud Expo, Scott Davis, CTO of Embotics, discussed how automation can provide the dynamic management required to cost-effectively deliver microservices and container solutions at scale. He also discussed how flexible automation is the key to effectively bridging and seamlessly coordinating both IT and developer needs for component orchestration across disparate clouds – an increasingly important requirement at today’s multi-cloud enterprise.
Any startup has to have a clear go –to-market strategy from the beginning. Similarly, any data science project has to have a go to production strategy from its first days, so it could go beyond proof-of-concept. Machine learning and artificial intelligence in production would result in hundreds of training pipelines and machine learning models that are continuously revised by teams of data scientists and seamlessly connected with web applications for tenants and users.
Recently, WebRTC has a lot of eyes from market. The use cases of WebRTC are expanding - video chat, online education, online health care etc. Not only for human-to-human communication, but also IoT use cases such as machine to human use cases can be seen recently. One of the typical use-case is remote camera monitoring. With WebRTC, people can have interoperability and flexibility for deploying monitoring service. However, the benefit of WebRTC for IoT is not only its convenience and interopera...
Vulnerability management is vital for large companies that need to secure containers across thousands of hosts, but many struggle to understand how exposed they are when they discover a new high security vulnerability. In his session at 21st Cloud Expo, John Morello, CTO of Twistlock, will address this pressing concern by introducing the concept of the “Vulnerability Risk Tree API,” which brings all the data together in a simple REST endpoint, allowing companies to easily grasp the severity of t...
DevOps at Cloud Expo, taking place October 31 - November 2, 2017, at the Santa Clara Convention Center in Santa Clara, CA, is co-located with 21st Cloud Expo and will feature technical sessions from a rock star conference faculty and the leading industry players in the world. The widespread success of cloud computing is driving the DevOps revolution in enterprise IT. Now as never before, development teams must communicate and collaborate in a dynamic, 24/7/365 environment. There is no time to w...
IT organizations are moving to the cloud in hopes to approve efficiency, increase agility and save money. Migrating workloads might seem like a simple task, but what many businesses don’t realize is that application migration criteria differs across organizations, making it difficult for architects to arrive at an accurate TCO number. In his session at 21st Cloud Expo, Joe Kinsella, CTO of CloudHealth Technologies, will offer a systematic approach to understanding the TCO of a cloud application...
For financial firms, the cloud is going to increasingly become a crucial part of dealing with customers over the next five years and beyond, particularly with the growing use and acceptance of virtual currencies. There are new data storage paradigms on the horizon that will deliver secure solutions for storing and moving sensitive financial data around the world without touching terrestrial networks. In his session at 20th Cloud Expo, Cliff Beek, President of Cloud Constellation Corporation, d...
Most companies are adopting or evaluating container technology - Docker in particular - to speed up application deployment, drive down cost, ease management and make application delivery more flexible overall. As with most new architectures, this dream takes a lot of work to become a reality. Even when you do get your application componentized enough and packaged properly, there are still challenges for DevOps teams to making the shift to continuous delivery and achieving that reduction in cost ...
"With Digital Experience Monitoring what used to be a simple visit to a web page has exploded into app on phones, data from social media feeds, competitive benchmarking - these are all components that are only available because of some type of digital asset," explained Leo Vasiliou, Director of Web Performance Engineering at Catchpoint Systems, in this SYS-CON.tv interview at DevOps Summit at 20th Cloud Expo, held June 6-8, 2017, at the Javits Center in New York City, NY.
SYS-CON Events announced today that Secure Channels, a cybersecurity firm, will exhibit at SYS-CON's 21st International Cloud Expo®, which will take place on Oct 31 – Nov 2, 2017, at the Santa Clara Convention Center in Santa Clara, CA. Secure Channels, Inc. offers several products and solutions to its many clients, helping them protect critical data from being compromised and access to computer networks from the unauthorized. The company develops comprehensive data encryption security strategie...
Connecting to major cloud service providers is becoming central to doing business. But your cloud provider’s performance is only as good as your connectivity solution. Massive Networks will place you in the driver's seat by exposing how you can extend your LAN from any location to include any cloud platform through an advanced high-performance connection that is secure and dedicated to your business-critical data. In his session at 21st Cloud Expo, Paul Mako, CEO & CIO of Massive Networks, wil...
From 2013, NTT Communications has been providing cPaaS service, SkyWay. Its customer’s expectations for leveraging WebRTC technology are not only typical real-time communication use cases such as Web conference, remote education, but also IoT use cases such as remote camera monitoring, smart-glass, and robotic. Because of this, NTT Communications has numerous IoT business use-cases that its customers are developing on top of PaaS. WebRTC will lead IoT businesses to be more innovative and address...
When shopping for a new data processing platform for IoT solutions, many development teams want to be able to test-drive options before making a choice. Yet when evaluating an IoT solution, it’s simply not feasible to do so at scale with physical devices. Building a sensor simulator is the next best choice; however, generating a realistic simulation at very high TPS with ease of configurability is a formidable challenge. When dealing with multiple application or transport protocols, you would be...
Deep learning has been very successful in social sciences and specially areas where there is a lot of data. Trading is another field that can be viewed as social science with a lot of data. With the advent of Deep Learning and Big Data technologies for efficient computation, we are finally able to use the same methods in investment management as we would in face recognition or in making chat-bots. In his session at 20th Cloud Expo, Gaurav Chakravorty, co-founder and Head of Strategy Development ...