|By Marten Terpstra||
|May 16, 2014 12:45 PM EDT||
IP Multicast is one of those technologies that most everyone loves to hate. It’s almost the perfect example of how complicated we have made networking. Getting IP Multicast to run depends on several protocols that are all somewhat intertwined or dependent on each, their relationship sometimes explicit, sometimes implicit.
Even trying to describe the basic operation is complicated.
When an application or service provides information using IP multicast, it simply starts sending it onto a specific multicast group. The multicast router for the subnet of the sender sees the incoming multicast packet and will initially have no forwarding information for that stream in its forwarding hardware. The packet is passed onto the CPU of that router, which will encapsulate this packet and send it towards a special multicast router designated the Rendez-vous Point (RP). When the RP has installed the multicast routes for this group, it will tell the multicast router on the sender’s segment to stop sending. When it does, this router installs its own multicast routes for the source tree (the tree specific to this sender) and the shared tree (the one towards the RP) without any outgoing interfaces, and the traffic is dropped at this first router. But, the network (well at least the part between the sender and the RP) is now aware of this multicast stream. And who is sending.
Now when we want to join this IP Multicast group, the first action is send an IGMP join out on the subnet you are attached to. The IP Multicast router that serves this subnet sees the join and determines where RP can be found. It takes the client join, and sends it towards the RP, using the unicast routing table as its guide. Every multicast router along the way registers that there is a listener on the interface this join came in on and passes it along towards the IP. All along this path, the unicast routing entry for the RP is used to create the tree towards the listener.
Once received by the RP, the shared tree and the source tree towards the sender have been joined. We have an end to end path between sender and receiver, with the RP in the middle of it all. All that is left is to send a join from the RP towards the router on the sender’s subnet to essentially tell it to start passing the actual multicast along the path towards the RP (the source tree), where the RP will then push it out onto the shared tree towards the destination. Voila, it’s as simple as that.
But wait, we are not done. Once the packets start to flow from source to destination, the multicast router closest to the destination will send another join message for this group, but this time towards the sender. It is only now that it can do this because those first few data packets actually indicate who the sender is. That join is passed router to router to router towards the router on the sender’s subnet, and once arrived, that router will now also start sending the multicast data along that path towards the receiver. The receiving subnet router sees that stream appearing and will now send a prune message onto the shared tree towards the RP, indicating it no longer needs the multicast stream through the RP.
If you are not familiar with IP Multicast and after reading the above are not confused, congratulations, your brain is very well wired for complex networking.
If you step away from how IGMP and PIM implement this today as above, the most fundamental of IP multicast topologies is that you need to build a forwarding tree that is rooted in the source, with the destinations as its leaves. At each intermediate node in the tree, the packets are replicated to its branches, therefore creating the least amount of duplication. And by using a tree, it is loop free, packets won’t swirl around the network bringing it to its knees.
The challenging part though is that the tree is based on the unicast forwarding topology. From a leaf on this tree towards the sender, each step is identical to how a unicast IP packet would be forwarded. The forwarding topologies are connected and dependent on each other. IP Multicast is built on top of a unicast routed infrastructure, and unicast routing changes can have dramatic impacts to the multicast forwarding topologies.
I mentioned here before that I once spent a wonderful 2 weeks in Delhi working on a network where surveillance cameras created an aggregate 8Gbit/sec worth of multicast data, with a requirement that any unicast change would have limited impact to these streams. Believe me, it is extremely hard to engineer and tune, and we had the luxury of hijacking a really large network night after night to simulate failures.
SDN based architectures have the opportunity to change all this. Multicast forwarding was designed the way it was designed to work on arbitrary network topologies, with random senders and receivers coming and going. It builds trees on the fly and on demand. For many networks, topologies are not arbitrary, and those applications that consume/produce lots of multicast do not have randomly placed senders and receivers that come and go as they please. Many of them are well known or placed in fairly static and fixed topologies.
A controller with a global view of the network can create multicast topologies ahead of time. It knows all possible replication points and can create distribution trees among them. It can create different distribution trees for different multicast groups. It can create them independent of the unicast forwarding. It can calculate backup topologies in case portions of the tree fail. And it can do all of that guaranteeing there are no loops and optimal replication. When applications indicate their participation in specific multicast streams as senders or listeners to this controller, it can optimize very specifically based on those participants. The possibilities are endless.
We had a customer visit us yesterday that has very significant multicast needs and we walked him through some of these possibilities. He left with a huge smile on his face. And that smile on his face was not because he really liked what we built (even though he did), but it was because we showed him that if you remove legacy network thinking and constraints, networking can yet again be extremely exciting and creates solutions that he did not think were possible, in a fairly simple and straightforward way. And that, in turn, is truly exciting to us.
Eighty percent of a data scientist’s time is spent gathering and cleaning up data, and 80% of all data is unstructured and almost never analyzed. Cognitive computing, in combination with Big Data, is changing the equation by creating data reservoirs and using natural language processing to enable analysis of unstructured data sources. This is impacting every aspect of the analytics profession from how data is mined (and by whom) to how it is delivered. This is not some futuristic vision: it's ha...
Feb. 13, 2016 04:45 AM EST Reads: 464
The cloud promises new levels of agility and cost-savings for Big Data, data warehousing and analytics. But it’s challenging to understand all the options – from IaaS and PaaS to newer services like HaaS (Hadoop as a Service) and BDaaS (Big Data as a Service). In her session at @BigDataExpo at @ThingsExpo, Hannah Smalltree, a director at Cazena, will provide an educational overview of emerging “as-a-service” options for Big Data in the cloud. This is critical background for IT and data profes...
Feb. 13, 2016 03:45 AM EST Reads: 241
Father business cycles and digital consumers are forcing enterprises to respond faster to customer needs and competitive demands. Successful integration of DevOps and Agile development will be key for business success in today’s digital economy. In his session at DevOps Summit, Pradeep Prabhu, Co-Founder & CEO of Cloudmunch, covered the critical practices that enterprises should consider to seamlessly integrate Agile and DevOps processes, barriers to implementing this in the enterprise, and pr...
Feb. 13, 2016 03:00 AM EST Reads: 463
The principles behind DevOps are not new - for decades people have been automating system administration and decreasing the time to deploy apps and perform other management tasks. However, only recently did we see the tools and the will necessary to share the benefits and power of automation with a wider circle of people. In his session at DevOps Summit, Bernard Sanders, Chief Technology Officer at CloudBolt Software, explored the latest tools including Puppet, Chef, Docker, and CMPs needed to...
Feb. 13, 2016 02:30 AM EST Reads: 368
Sensors and effectors of IoT are solving problems in new ways, but small businesses have been slow to join the quantified world. They’ll need information from IoT using applications as varied as the businesses themselves. In his session at @ThingsExpo, Roger Meike, Distinguished Engineer, Director of Technology Innovation at Intuit, showed how IoT manufacturers can use open standards, public APIs and custom apps to enable the Quantified Small Business. He used a Raspberry Pi to connect sensors...
Feb. 13, 2016 02:30 AM EST Reads: 368
Let’s face it, embracing new storage technologies, capabilities and upgrading to new hardware often adds complexity and increases costs. In his session at 18th Cloud Expo, Seth Oxenhorn, Vice President of Business Development & Alliances at FalconStor, will discuss how a truly heterogeneous software-defined storage approach can add value to legacy platforms and heterogeneous environments. The result reduces complexity, significantly lowers cost, and provides IT organizations with improved effi...
Feb. 13, 2016 12:45 AM EST Reads: 273
It's easy to assume that your app will run on a fast and reliable network. The reality for your app's users, though, is often a slow, unreliable network with spotty coverage. What happens when the network doesn't work, or when the device is in airplane mode? You get unhappy, frustrated users. An offline-first app is an app that works, without error, when there is no network connection.
Feb. 12, 2016 10:00 PM EST Reads: 245
Data-as-a-Service is the complete package for the transformation of raw data into meaningful data assets and the delivery of those data assets. In her session at 18th Cloud Expo, Lakshmi Randall, an industry expert, analyst and strategist, will address: What is DaaS (Data-as-a-Service)? Challenges addressed by DaaS Vendors that are enabling DaaS Architecture options for DaaS
Feb. 12, 2016 09:45 PM EST Reads: 386
One of the bewildering things about DevOps is integrating the massive toolchain including the dozens of new tools that seem to crop up every year. Part of DevOps is Continuous Delivery and having a complex toolchain can add additional integration and setup to your developer environment. In his session at @DevOpsSummit at 18th Cloud Expo, Miko Matsumura, Chief Marketing Officer of Gradle Inc., will discuss which tools to use in a developer stack, how to provision the toolchain to minimize onboa...
Feb. 12, 2016 09:00 PM EST Reads: 135
SYS-CON Events announced today that Catchpoint Systems, Inc., a provider of innovative web and infrastructure monitoring solutions, has been named “Silver Sponsor” of SYS-CON's DevOps Summit at 18th Cloud Expo New York, which will take place June 7-9, 2016, at the Javits Center in New York City, NY. Catchpoint is a leading Digital Performance Analytics company that provides unparalleled insight into customer-critical services to help consistently deliver an amazing customer experience. Designed...
Feb. 12, 2016 06:00 PM EST Reads: 404
Companies can harness IoT and predictive analytics to sustain business continuity; predict and manage site performance during emergencies; minimize expensive reactive maintenance; and forecast equipment and maintenance budgets and expenditures. Providing cost-effective, uninterrupted service is challenging, particularly for organizations with geographically dispersed operations.
Feb. 12, 2016 06:00 PM EST
When building large, cloud-based applications that operate at a high scale, it’s important to maintain a high availability and resilience to failures. In order to do that, you must be tolerant of failures, even in light of failures in other areas of your application. “Fly two mistakes high” is an old adage in the radio control airplane hobby. It means, fly high enough so that if you make a mistake, you can continue flying with room to still make mistakes. In his session at 18th Cloud Expo, Lee...
Feb. 12, 2016 04:45 PM EST
DevOps is not just last year’s buzzword. Companies with DevOps practices are 2.5x more likely to exceed profitability, market share, and productivity goals. But how do you enable high performance? What can you do right now to start? Find out from DevOps experts including Gene Kim, co-author of "The Phoenix Project," and the Dynatrace Center of Excellence.
Feb. 12, 2016 04:30 PM EST
With the proliferation of both SQL and NoSQL databases, organizations can now target specific fit-for-purpose database tools for their different application needs regarding scalability, ease of use, ACID support, etc. Platform as a Service offerings make this even easier now, enabling developers to roll out their own database infrastructure in minutes with minimal management overhead. However, this same amount of flexibility also comes with the challenges of picking the right tool, on the right ...
Feb. 12, 2016 04:30 PM EST Reads: 192
SYS-CON Events announced today that Column Technologies will exhibit at SYS-CON's @DevOpsSummit at Cloud Expo, which will take place on June 7-9, 2016, at the Javits Center in New York City, NY. Established in 1998, Column Technologies is a global technology solutions provider with over 400 employees, headquartered in the United States with offices in Canada, India, and the United Kingdom. Column Technologies provides “Best of Breed” technology solutions that automate the key DevOps principal...
Feb. 12, 2016 04:15 PM EST