|By Michael Bushong||
|April 24, 2014 11:30 AM EDT||
From a cost perspective, the networking dialogue is dominated by CapEx. Acquisition costs for new networking gear have historically been tied to hardware, and despite the relatively recent rise of bare metal switching, networking hardware remains a high-stakes business. But SDN is changing this dynamic in potentially significant ways.
The first point to clarify when talking about CapEx is that CapEx does not necessarily mean hardware (at least not the way that most people mean). While there is a strict financial definition for CapEx, in the networking industry it has become shorthand for Procurement Costs. Because networking solutions have been predominantly monetized through hardware, we associate procurement costs with hardware, but this is changing.
The fact that the ’S’ in SDN stands for software is reason enough for people to look beyond the chassis. But the reality is that while vendors have monetized the hardware, the value has been increasingly moving to the software side for more than a decade. So long as everyone was selling hardware, it didn’t really matter that much whether the cost was tied to the hardware or the software, so we have been a little bit lazy collectively in determining a deliberate pricing mix.
More recently, however, there have been additional solutions that are offered entirely through software. With virtual networking devices, for example, there is no physical hardware (unless you count the servers and the network that connects the servers). A common sales tactic for these types of solutions is to point out how expensive physical solutions are. Why pay for all that sheet metal when you can get the same functionality in a virtual form factor? Of course, you are not really paying for the sheet metal; your check also pays for the software and all the features that go into that sheet metal. But the argument is pretty compelling.
The point here is that the only thing that really matters is how much you pay for the whole solution. Whether the price is affixed to hardware or software is an accounting detail – important for some people, but not really the most important thing for the majority of buyers. Rather than calling it CapEx, we ought to be referring more broadly to procurement or acquisition costs. All in, Solution A costs X dollars to bring in house, and Solution B costs Y dollars.
This would certainly simplify the conversation some. But even then, it isn’t all about procurement costs anymore either.
Depending on the solution, the procurement costs account for roughly one-third of the total cost of ownership. The remaining two-thirds of the cost is ongoing operating expense (power, cooling, space, management, support, and so on). The models here for most solutions start to get pretty squishy. While we can fairly formulaically determine things like power, space, and support, when it comes to estimating the cost of managing a device, the models are so dependent on uncontrollable things that they border on useless. And even when the models are sound, most companies have not sufficiently instrumented their network operations to really know what they are spending.
But just because it is difficult to model OpEx does not mean that network teams should ignore it.
If there is one thing that the gaming industry has taught us, it is that there are all kinds of creative ways to separate someone from their money. In the early days of video games, 100% of the cost was procurement cost. After you bought the install media, you had paid everything you were ever going to pay. Before long, some of the more popular games figured out that they could lower initial costs (make the barrier to entry lower) and then charge for ongoing use through subscriptions.
As the networking world adjusts the pricing mix – associating more of the cost with the software – we should expect that charge models will mirror what we have seen on the consumer side. It is not a big stretch (and in fact already happening) to see massive up-front hardware costs replaced with more palatable hardware pricing combined with either higher software or potentially support costs. This has the dual benefit of making it easier for customers to select a vendor, and creating annuities for said vendor.
But the evolution of game pricing models did not end with subscriptions.
For anyone who has gotten sucked into the hell that is Candy Crush, you are already well aware of in-app purchases. The initial game is free, but if you want to get a special advantage or unlock a level, you can make an in-app purchase. They have cleverly priced the in-app purchases to feel like you are hardly spending anything. It’s less than a dollar. I should just go ahead and get that spotted donut thingy! Of course, by the time you add up all those just a dollar moments, you end up paying far more than you ever would have up front.
The magic of this type of pricing is that most of this is not really known up front. When you first get Candy Crush, you don’t really think you are going to buy the special extras. And Candy Crush doesn’t tell you that the levels get progressively harder to the point that they are nigh impossible without a little extra help.
Before you write this off as not applicable to networking, consider a few points.
First, despite the huge open source push, there are still a lot of companies pursuing commercial grade versions of the otherwise free software. Sure, you might buy into the open source controller, but if you need the networking version of the spotted donut thing, what do you do? This is essentially the networking equivalent of the in-app purchase. Call it the in-arch purchase. Once you buy into a particular architecture, the switching costs are prohibitively high. If you have to pay more for the commercial software, can you really say no?
Second, some of the tiered pricing models that are taking root make it more difficult to accurately model ongoing license costs. If you are not thinking about how the costs will scale with the number of ports, users, VMs, or whatever, you might find out down the road that your solution is contributing more ongoing costs than anticipated. For example, buying one VM from Amazon might seem easy enough, but what if you need thousands? It doesn’t stay cheap forever.
Maybe the in-arch costs are just extra features or capabilities. Or ongoing support and services. Whatever the source, these types of costs contribute to the ongoing operating expenses. And because the primary purchasing criterion is CapEx (procurement costs), burying some of these costs a little later in the product lifecycle and making them a bit smaller in magnitude (but larger in volume) will be attractive.
The punch line here is that we are on the cusp of a change in monetization strategies. You might think that pricing and costs will be transparent, but has the networking community given us a real reason to believe that to date? If you think so, consider this: why do buyers celebrate 50% discounts? It’s because pricing is ridiculously obfuscated in this industry. Until we all start expecting more, I just don’t know why this would change.
Along those lines, my colleague Bill Koss posted some facts about Plexxi costs. In the interest of transparency, it’s worth taking a look here.
[Today’s fun fact: The wettest spot in the world is located on the island of Kauai. Mt. Waialeale consistently records rainfall at the rate of nearly 500 inches per year. That’s enough so drown 7 6-foot-tall men standing on each other’s heads.]
Amazon has gradually rolled out parts of its IoT offerings in the last year, but these are just the tip of the iceberg. In addition to optimizing their back-end AWS offerings, Amazon is laying the ground work to be a major force in IoT – especially in the connected home and office. Amazon is extending its reach by building on its dominant Cloud IoT platform, its Dash Button strategy, recently announced Replenishment Services, the Echo/Alexa voice recognition control platform, the 6-7 strategic...
Dec. 8, 2016 07:15 PM EST Reads: 313
"Qosmos has launched L7Viewer, a network traffic analysis tool, so it analyzes all the traffic between the virtual machine and the data center and the virtual machine and the external world," stated Sebastien Synold, Product Line Manager at Qosmos, in this SYS-CON.tv interview at 19th Cloud Expo, held November 1-3, 2016, at the Santa Clara Convention Center in Santa Clara, CA.
Dec. 8, 2016 06:45 PM EST Reads: 919
Without a clear strategy for cost control and an architecture designed with cloud services in mind, costs and operational performance can quickly get out of control. To avoid multiple architectural redesigns requires extensive thought and planning. Boundary (now part of BMC) launched a new public-facing multi-tenant high resolution monitoring service on Amazon AWS two years ago, facing challenges and learning best practices in the early days of the new service. In his session at 19th Cloud Exp...
Dec. 8, 2016 05:45 PM EST Reads: 928
Everyone knows that truly innovative companies learn as they go along, pushing boundaries in response to market changes and demands. What's more of a mystery is how to balance innovation on a fresh platform built from scratch with the legacy tech stack, product suite and customers that continue to serve as the business' foundation. In his General Session at 19th Cloud Expo, Michael Chambliss, Head of Engineering at ReadyTalk, discussed why and how ReadyTalk diverted from healthy revenue and mor...
Dec. 8, 2016 05:45 PM EST Reads: 1,714
As data explodes in quantity, importance and from new sources, the need for managing and protecting data residing across physical, virtual, and cloud environments grow with it. Managing data includes protecting it, indexing and classifying it for true, long-term management, compliance and E-Discovery. Commvault can ensure this with a single pane of glass solution – whether in a private cloud, a Service Provider delivered public cloud or a hybrid cloud environment – across the heterogeneous enter...
Dec. 8, 2016 05:00 PM EST Reads: 1,810
You have great SaaS business app ideas. You want to turn your idea quickly into a functional and engaging proof of concept. You need to be able to modify it to meet customers' needs, and you need to deliver a complete and secure SaaS application. How could you achieve all the above and yet avoid unforeseen IT requirements that add unnecessary cost and complexity? You also want your app to be responsive in any device at any time. In his session at 19th Cloud Expo, Mark Allen, General Manager of...
Dec. 8, 2016 04:45 PM EST Reads: 1,839
Financial Technology has become a topic of intense interest throughout the cloud developer and enterprise IT communities. Accordingly, attendees at the upcoming 20th Cloud Expo at the Javits Center in New York, June 6-8, 2017, will find fresh new content in a new track called FinTech.
Dec. 8, 2016 04:45 PM EST Reads: 2,243
"We are a leader in the market space called network visibility solutions - it enables monitoring tools and Big Data analysis to access the data and be able to see the performance," explained Shay Morag, VP of Sales and Marketing at Niagara Networks, in this SYS-CON.tv interview at 19th Cloud Expo, held November 1-3, 2016, at the Santa Clara Convention Center in Santa Clara, CA.
Dec. 8, 2016 04:30 PM EST Reads: 356
@DevOpsSummit taking place June 6-8, 2017 at Javits Center, New York City, is co-located with the 20th International Cloud Expo and will feature technical sessions from a rock star conference faculty and the leading industry players in the world. @DevOpsSummit at Cloud Expo New York Call for Papers is now open.
Dec. 8, 2016 04:30 PM EST Reads: 1,963
Keeping pace with advancements in software delivery processes and tooling is taxing even for the most proficient organizations. Point tools, platforms, open source and the increasing adoption of private and public cloud services requires strong engineering rigor - all in the face of developer demands to use the tools of choice. As Agile has settled in as a mainstream practice, now DevOps has emerged as the next wave to improve software delivery speed and output. To make DevOps work, organization...
Dec. 8, 2016 04:15 PM EST Reads: 279
The 20th International Cloud Expo has announced that its Call for Papers is open. Cloud Expo, to be held June 6-8, 2017, at the Javits Center in New York City, brings together Cloud Computing, Big Data, Internet of Things, DevOps, Containers, Microservices and WebRTC to one location. With cloud computing driving a higher percentage of enterprise IT budgets every year, it becomes increasingly important to plant your flag in this fast-expanding business opportunity. Submit your speaking proposal ...
Dec. 8, 2016 04:15 PM EST Reads: 2,309
"We are a modern development application platform and we have a suite of products that allow you to application release automation, we do version control, and we do application life cycle management," explained Flint Brenton, CEO of CollabNet, in this SYS-CON.tv interview at DevOps at 19th Cloud Expo, held November 1-3, 2016, at the Santa Clara Convention Center in Santa Clara, CA.
Dec. 8, 2016 04:00 PM EST Reads: 1,074
Regulatory requirements exist to promote the controlled sharing of information, while protecting the privacy and/or security of the information. Regulations for each type of information have their own set of rules, policies, and guidelines. Cloud Service Providers (CSP) are faced with increasing demand for services at decreasing prices. Demonstrating and maintaining compliance with regulations is a nontrivial task and doing so against numerous sets of regulatory requirements can be daunting task...
Dec. 8, 2016 03:45 PM EST Reads: 1,081
Bert Loomis was a visionary. This general session will highlight how Bert Loomis and people like him inspire us to build great things with small inventions. In their general session at 19th Cloud Expo, Harold Hannon, Architect at IBM Bluemix, and Michael O'Neill, Strategic Business Development at Nvidia, discussed the accelerating pace of AI development and how IBM Cloud and NVIDIA are partnering to bring AI capabilities to "every day," on-demand. They also reviewed two "free infrastructure" pr...
Dec. 8, 2016 03:30 PM EST Reads: 1,225
"This is specifically designed to accommodate some of the needs for high availability and failover in a network managed system for the major Korean corporations," stated Thomas Masters, Managing Director at InfranicsUSA, in this SYS-CON.tv interview at 19th Cloud Expo, held November 1-3, 2016, at the Santa Clara Convention Center in Santa Clara, CA.
Dec. 8, 2016 03:15 PM EST Reads: 313