|By Nikita Ivanov||
|December 30, 2014 05:00 PM EST||
The Facts and Fiction of In-Memory Computing
In the last year, conversations about In-Memory Computing (IMC) have become more and more prevalent in enterprise IT circles, especially with organizations feeling the pressure to process massive quantities of data at the speed that is now being demanded by the Internet. The hype around IMC is justified: tasks that once took hours to execute are streamlined down to seconds by moving the computation and data from disk, directly to RAM. Through this simple adjustment, analytics are happening in real-time, and applications (as well as the development of applications) are working at-pace with this new standard of technology and speed.
Despite becoming both more cost-effective and accepted within enterprise computing, there are still a small handful of falsehoods that confuse even the most technical of individuals in enterprise IT.
Myth: In-memory computing is about databases, so this isn't really relevant to my business.
The best way to clear the air around IMC is to start with a simple explanation of what, in fact, in-memory computing is. While many assume that because we are talking about RAM, we are having a conversation about databases and storage, but this is not the case.
IMC, at its most basic level, is using a middleware software that allows one to store data in RAM - across a broad cluster of computers - and do any and all processing where it resides (in the memory). With traditional methods, data processing is often confined to spinning disks.
By comparison, in-memory computing speeds up this process by roughly 5,000 times. Now you can see that we're not talking about storage only - instead active, fluid data and computing.
Which brings me to another, more tangible point about computing efficiency. By incorporating in-memory, a handful of processes are streamlined in order to save time, resources, and money.
To start, in-memory requires much less hardware; the result - significantly decreased capital, operational and infrastructure overhead.
Moreover, IT departments can also significantly extend the life of existing hardware and software through the increased performance that is inherent with IMC - thus amplifying the ROI on the machines that have already been purchased.
Surprisingly, in-memory computing is not a new phenomenon. Since the inception of RAM, IMC has been viewed as reliable accelerant for high-performance computing, bringing us to the next crucial misconception about this technology.
Myth: In-memory computing is expensive, therefore not practical for my operation.
There is a reason that this is one of the most common misunderstandings about IMC, because there was a point in time where the cost of memory was once quite high. That being said, the cost of RAM has been dropping consistently, at a rate of about 30% - for the last 5 years.
Today, the price of a 1 Terabyte RAM cluster can go for anywhere between $20 and $40 thousand - including all of the CPUs, networking, etc. A few years from now that same setup will likely be available for half that price.
Regardless of the future price of RAM, which based upon current projections will likely continue to fall, the current economics have already placed this technology well within the reaches of the enterprise computing budgets that require this level of scale.
Myth: My needs are already being met by Flash.
There are three different reasons why this mentality is held by IT folks, each of which are highly misinformed. I'll start with the most common, which is the idea that your business doesn't need the Lambourgini-esque super-computing power of IMC.
The hard yet obvious reality is that if your business is in any way data-driven, you likely cannot survive without speed and agility in this department. As time goes on, the amount of data that businesses accumulate compounds with new streams and variances. This is a sink-or-swim reality.
Another myth commonly used to dispel IMC is that if businesses are able to just effectively mount RAM disk, they will get in-memory processing. Unfortunately, it's not that easy. As mentioned earlier, IMC works through middleware to effectively unlock its power.
Finally, there's the notion that one can just replace their HDDs with SSDs in order to get this super-charged performance. For SSDs - in certain situations - the performance gain that you can pull from flash storage in lieu of spinning disk is enough.
However, speed matters - and is rapidly becoming more of a requirement every day. At this point, it's like comparing apples to oranges with speed improvements of 10 to 100x over SSDs.
Myth: Memory is not durable enough to be truly sustainable.
This is another notion that for whatever reason has been both widely perpetuated - and is entirely false.
The fact is - almost all in-memory computing middleware (apart from very simplistic ones) offer one or multiple strategies for in-memory backups, durable storage backups, disk-based swap space overflow, etc.
More sophisticated vendors provide a comprehensive tiered storage approach where users can decide what portion of the overall data set is stored in RAM, local disk swap space or RDBMS/HDFS - where each tier can store progressively more data but with progressively longer latencies.
Yet another source of confusion is the difference between operational datasets and historical datasets. In-memory computing is not aimed at replacing enterprise data warehouse (EDW), backup or offline storage services - like Hadoop, for example. The goal of IMC is to improve the operational datasets that require mixed OLTP and OLAP processing and in most cases are less than 10TB in size. That is to say, in-memory computing is not "all or nothing" - and does not require that every aspect of data be housed in memory.
The in-memory computing revolution is by no means intended to obliterate disks from the enterprise. For now, the disk still serves a well-defined role for offline/backup use cases - tasks that are not the focus of IMC.
Myth: In-memory is inaccessible to my business because so few developers actually know how to use it.
Yes indeed, In-memory computing is a highly complex technology, that for now, only a few vendors have even been able to successfully develop offerings for. However, like much of high-technology, in-memory computing has entered the world of open source - bringing its capabilities and power to the fingertips of developers around the world.
Currently, with GridGain, developers have the ability to get their hands on IMC with a simple download at http://gridgain.org/.
In-memory computing is already being tapped across a broad range of functions and industries including (but not limited to) financial trading systems, online game, bioinformatics, hyper-local advertising, cognitive computing, and geospatial analysis.
By raising awareness, and bringing the capabilities of IMC to more developers and organizations - industries around the globe are poised to experience entirely new standards of speed, computing, and performance.
For basic one-to-one voice or video calling solutions, WebRTC has proven to be a very powerful technology. Although WebRTC’s core functionality is to provide secure, real-time p2p media streaming, leveraging native platform features and server-side components brings up new communication capabilities for web and native mobile applications, allowing for advanced multi-user use cases such as video broadcasting, conferencing, and media recording.
Oct. 24, 2016 07:30 PM EDT Reads: 3,190
SYS-CON Events announced today that Cloudbric, a leading website security provider, will exhibit at the 19th International Cloud Expo, which will take place on November 1–3, 2016, at the Santa Clara Convention Center in Santa Clara, CA. Cloudbric is an elite full service website protection solution specifically designed for IT novices, entrepreneurs, and small and medium businesses. First launched in 2015, Cloudbric is based on the enterprise level Web Application Firewall by Penta Security Sys...
Oct. 24, 2016 07:15 PM EDT Reads: 1,153
Established in 1998, Calsoft is a leading software product engineering Services Company specializing in Storage, Networking, Virtualization and Cloud business verticals. Calsoft provides End-to-End Product Development, Quality Assurance Sustenance, Solution Engineering and Professional Services expertise to assist customers in achieving their product development and business goals. The company's deep domain knowledge of Storage, Virtualization, Networking and Cloud verticals helps in delivering ...
Oct. 24, 2016 07:15 PM EDT Reads: 1,043
The best way to leverage your Cloud Expo presence as a sponsor and exhibitor is to plan your news announcements around our events. The press covering Cloud Expo and @ThingsExpo will have access to these releases and will amplify your news announcements. More than two dozen Cloud companies either set deals at our shows or have announced their mergers and acquisitions at Cloud Expo. Product announcements during our show provide your company with the most reach through our targeted audiences.
Oct. 24, 2016 06:15 PM EDT Reads: 4,745
In the next five to ten years, millions, if not billions of things will become smarter. This smartness goes beyond connected things in our homes like the fridge, thermostat and fancy lighting, and into heavily regulated industries including aerospace, pharmaceutical/medical devices and energy. “Smartness” will embed itself within individual products that are part of our daily lives. We will engage with smart products - learning from them, informing them, and communicating with them. Smart produc...
Oct. 24, 2016 05:45 PM EDT Reads: 1,496
Extreme Computing is the ability to leverage highly performant infrastructure and software to accelerate Big Data, machine learning, HPC, and Enterprise applications. High IOPS Storage, low-latency networks, in-memory databases, GPUs and other parallel accelerators are being used to achieve faster results and help businesses make better decisions. In his session at 18th Cloud Expo, Michael O'Neill, Strategic Business Development at NVIDIA, focused on some of the unique ways extreme computing is...
Oct. 24, 2016 05:00 PM EDT Reads: 3,901
SYS-CON Events announced today that 910Telecom will exhibit at the 19th International Cloud Expo, which will take place on November 1–3, 2016, at the Santa Clara Convention Center in Santa Clara, CA. Housed in the classic Denver Gas & Electric Building, 910 15th St., 910Telecom is a carrier-neutral telecom hotel located in the heart of Denver. Adjacent to CenturyLink, AT&T, and Denver Main, 910Telecom offers connectivity to all major carriers, Internet service providers, Internet backbones and ...
Oct. 24, 2016 05:00 PM EDT Reads: 3,629
SYS-CON Events announced today that Coalfire will exhibit at the 19th International Cloud Expo, which will take place on November 1–3, 2016, at the Santa Clara Convention Center in Santa Clara, CA. Coalfire is the trusted leader in cybersecurity risk management and compliance services. Coalfire integrates advisory and technical assessments and recommendations to the corporate directors, executives, boards, and IT organizations for global brands and organizations in the technology, cloud, health...
Oct. 24, 2016 04:45 PM EDT Reads: 1,567
SYS-CON Events announced today that Transparent Cloud Computing (T-Cloud) Consortium will exhibit at the 19th International Cloud Expo®, which will take place on November 1–3, 2016, at the Santa Clara Convention Center in Santa Clara, CA. The Transparent Cloud Computing Consortium (T-Cloud Consortium) will conduct research activities into changes in the computing model as a result of collaboration between "device" and "cloud" and the creation of new value and markets through organic data proces...
Oct. 24, 2016 04:30 PM EDT Reads: 1,366
In his session at 19th Cloud Expo, Claude Remillard, Principal Program Manager in Developer Division at Microsoft, will contrast how his team used config as code and immutable patterns for continuous delivery of microservices and apps to the cloud. He will show the immutable patterns helps developers do away with most of the complexity of config as code-enabling scenarios such as rollback, zero downtime upgrades with far greater simplicity. He will also have live demos of building immutable pipe...
Oct. 24, 2016 04:30 PM EDT Reads: 1,588
WebRTC defines no default signaling protocol, causing fragmentation between WebRTC silos. SIP and XMPP provide possibilities, but come with considerable complexity and are not designed for use in a web environment. In his session at @ThingsExpo, Matthew Hodgson, technical co-founder of the Matrix.org, discussed how Matrix is a new non-profit Open Source Project that defines both a new HTTP-based standard for VoIP & IM signaling and provides reference implementations.
Oct. 24, 2016 04:15 PM EDT Reads: 2,770
The Internet of Things (IoT), in all its myriad manifestations, has great potential. Much of that potential comes from the evolving data management and analytic (DMA) technologies and processes that allow us to gain insight from all of the IoT data that can be generated and gathered. This potential may never be met as those data sets are tied to specific industry verticals and single markets, with no clear way to use IoT data and sensor analytics to fulfill the hype being given the IoT today.
Oct. 24, 2016 04:15 PM EDT Reads: 2,604
In his general session at 18th Cloud Expo, Lee Atchison, Principal Cloud Architect and Advocate at New Relic, discussed cloud as a ‘better data center’ and how it adds new capacity (faster) and improves application availability (redundancy). The cloud is a ‘Dynamic Tool for Dynamic Apps’ and resource allocation is an integral part of your application architecture, so use only the resources you need and allocate /de-allocate resources on the fly.
Oct. 24, 2016 04:00 PM EDT Reads: 3,724
As data explodes in quantity, importance and from new sources, the need for managing and protecting data residing across physical, virtual, and cloud environments grow with it. Managing data includes protecting it, indexing and classifying it for true, long-term management, compliance and E-Discovery. Commvault can ensure this with a single pane of glass solution – whether in a private cloud, a Service Provider delivered public cloud or a hybrid cloud environment – across the heterogeneous enter...
Oct. 24, 2016 03:45 PM EDT Reads: 1,403
Traditional on-premises data centers have long been the domain of modern data platforms like Apache Hadoop, meaning companies who build their business on public cloud were challenged to run Big Data processing and analytics at scale. But recent advancements in Hadoop performance, security, and most importantly cloud-native integrations, are giving organizations the ability to truly gain value from all their data. In his session at 19th Cloud Expo, David Tishgart, Director of Product Marketing ...
Oct. 24, 2016 03:45 PM EDT Reads: 2,558