Welcome!

Related Topics: @BigDataExpo, Java IoT, Microservices Expo, Containers Expo Blog, @CloudExpo, SDN Journal

@BigDataExpo: Article

Consolidating Big Data

How to make your data center more cost-effective while improving performance

Cloud computing has opened the doors to a vast array of online services. With the emergence of new cloud technologies, both public and private companies are seeing increases in performance gains, elasticity and convenience. However, maintaining a competitive advantage has become increasingly difficult. Service providers are taking a closer look at their data storage infrastructure for ways to improve performance and cut costs.

If the status quo remains, maintaining low-cost cloud services will become increasingly difficult. Service providers will incur higher costs, while consumers become burdened with storage capacity restrictions. Such obstacles are influencing service providers to find new ways to scale cost-effectively and increase performance in the data center.

Cost-Benefit Analysis
In response to the increase of online account activity, service providers are consolidating their data centers to a centralized environment. By doing so, they are able to cut costs while increasing efficiency, allowing data to be accessible from any location. Centralizing equipment enables providers the ability to deliver enhanced Internet connections, performance and reliability.

However, with these added benefits also come disadvantages. For instance, scalability becomes more expensive and difficult to achieve. Improving efficiency within a centralized data center requires the purchase of additional high-performance, specialized equipment, which increases costs and energy consumption, challenging endeavors to control at scale. In an economy where cost-cutting is becoming a necessity for large and small enterprises alike, these added expenses are unacceptable.

Characteristics of the Cloud
Solving performance problems, like data bottlenecks, is a growing concern for cloud providers who must oversee significantly more users and accompanying performance demands, than do enterprises. Although the average user of an enterprise system requires elevated performance, these systems generally manage fewer users who are able to access their data directly through the network. Moreover, enterprise system users are accessing, saving and sending comparatively relatively small files that require less storage capacity and performance.

Outside the internal enterprise network, however, it's a different story. Cloud systems are simultaneously being accessed by a multitude of users across the Internet, which itself becomes a performance bottleneck. The average cloud user stores relatively larger files than the average enterprise user placing greater strains on data center resources. The cloud provider's storage system not only has to scale to each user, but must also sustain performance across all users as well.

Best Practices
In response to growing storage demands, cloud providers are faced with profound business implications. Service providers need to scale quickly in order to meet the booming demand for more data storage. The following best practices can help optimize data center ROI in a period of significant IT cutbacks:

  • Opt for commodity components when possible: Low-energy hardware makes good business sense. Commodity hardware is not only cost-effective, but also energy-efficient, which significantly reduces both setup and operating costs in one move.
  • Seek out a distributed storage system: Distributed storage presents the best way to build at scale even though the data center trend has been moving toward centralization. Increased performance at the software level counterbalances the performance advantage of a centralized data storage approach.
  • Avoid bottlenecks: A single point of entry can easily lead to a performance bottleneck. Adding caches to relieve the bottleneck, as most data center infrastructures currently do, quickly adds cost and complexity to a system. On the other hand, a horizontally scalable system that distributes data among all nodes delivers a high level of redundancy.

Moving Forward
Currently, Big Data storage consists mainly of high performance, vertically scaled storage systems. Since these infrastructures can only scale to a single petabyte and are costly, they are not a sustainable solution. Moving to a horizontally scaled data storage model that distributes data evenly onto energy-efficient hardware can reduce costs and increase performance in the cloud. With these insights, cloud service providers can take the necessary steps to improve the efficiency, scalability and performance of their data storage centers.

More Stories By Stefan Bernbo

Stefan Bernbo is the founder and CEO of Compuverde. For 20 years, he has designed and built numerous enterprise scale data storage solutions designed to be cost effective for storing huge data sets. From 2004 to 2010 Stefan worked within this field for Storegate, the wide-reaching Internet based storage solution for consumer and business markets, with the highest possible availability and scalability requirements. Previously, Stefan has worked with system and software architecture on several projects with Swedish giant Ericsson, the world-leading provider of telecommunications equipment and services to mobile and fixed network operators.

Comments (0)

Share your thoughts on this story.

Add your comment
You must be signed in to add a comment. Sign-in | Register

In accordance with our Comment Policy, we encourage comments that are on topic, relevant and to-the-point. We will remove comments that include profanity, personal attacks, racial slurs, threats of violence, or other inappropriate material that violates our Terms and Conditions, and will block users who make repeated violations. We ask all readers to expect diversity of opinion and to treat one another with dignity and respect.


Latest Stories
With 15% of enterprises adopting a hybrid IT strategy, you need to set a plan to integrate hybrid cloud throughout your infrastructure. In his session at 18th Cloud Expo, Steven Dreher, Director of Solutions Architecture at Green House Data, discussed how to plan for shifting resource requirements, overcome challenges, and implement hybrid IT alongside your existing data center assets. Highlights included anticipating workload, cost and resource calculations, integrating services on both sides...
In his session at @DevOpsSummit at 19th Cloud Expo, Yoseph Reuveni, Director of Software Engineering at Jet.com, will discuss Jet.com's journey into containerizing Microsoft-based technologies like C# and F# into Docker. He will talk about lessons learned and challenges faced, the Mono framework tryout and how they deployed everything into Azure cloud. Yoseph Reuveni is a technology leader with unique experience developing and running high throughput (over 1M tps) distributed systems with extre...
Manufacturers are embracing the Industrial Internet the same way consumers are leveraging Fitbits – to improve overall health and wellness. Both can provide consistent measurement, visibility, and suggest performance improvements customized to help reach goals. Fitbit users can view real-time data and make adjustments to increase their activity. In his session at @ThingsExpo, Mark Bernardo Professional Services Leader, Americas, at GE Digital, discussed how leveraging the Industrial Internet a...
Big Data engines are powering a lot of service businesses right now. Data is collected from users from wearable technologies, web behaviors, purchase behavior as well as several arbitrary data points we’d never think of. The demand for faster and bigger engines to crunch and serve up the data to services is growing exponentially. You see a LOT of correlation between “Cloud” and “Big Data” but on Big Data and “Hybrid,” where hybrid hosting is the sanest approach to the Big Data Infrastructure pro...
In his session at 18th Cloud Expo, Sagi Brody, Chief Technology Officer at Webair Internet Development Inc., and Logan Best, Infrastructure & Network Engineer at Webair, focused on real world deployments of DDoS mitigation strategies in every layer of the network. He gave an overview of methods to prevent these attacks and best practices on how to provide protection in complex cloud platforms. He also outlined what we have found in our experience managing and running thousands of Linux and Unix ...
Cloud analytics is dramatically altering business intelligence. Some businesses will capitalize on these promising new technologies and gain key insights that’ll help them gain competitive advantage. And others won’t. Whether you’re a business leader, an IT manager, or an analyst, we want to help you and the people you need to influence with a free copy of “Cloud Analytics for Dummies,” the essential guide to this explosive new space for business intelligence.
"My role is working with customers, helping them go through this digital transformation. I spend a lot of time talking to banks, big industries, manufacturers working through how they are integrating and transforming their IT platforms and moving them forward," explained William Morrish, General Manager Product Sales at Interoute, in this SYS-CON.tv interview at 18th Cloud Expo, held June 7-9, 2016, at the Javits Center in New York City, NY.
A critical component of any IoT project is what to do with all the data being generated. This data needs to be captured, processed, structured, and stored in a way to facilitate different kinds of queries. Traditional data warehouse and analytical systems are mature technologies that can be used to handle certain kinds of queries, but they are not always well suited to many problems, particularly when there is a need for real-time insights.
Choosing the right cloud for your workloads is a balancing act that can cost your organization time, money and aggravation - unless you get it right the first time. Economics, speed, performance, accessibility, administrative needs and security all play a vital role in dictating your approach to the cloud. Without knowing the right questions to ask, you could wind up paying for capacity you'll never need or underestimating the resources required to run your applications.
Enterprise networks are complex. Moreover, they were designed and deployed to meet a specific set of business requirements at a specific point in time. But, the adoption of cloud services, new business applications and intensifying security policies, among other factors, require IT organizations to continuously deploy configuration changes. Therefore, enterprises are looking for better ways to automate the management of their networks while still leveraging existing capabilities, optimizing perf...
"Software-defined storage is a big problem in this industry because so many people have different definitions as they see fit to use it," stated Peter McCallum, VP of Datacenter Solutions at FalconStor Software, in this SYS-CON.tv interview at 18th Cloud Expo, held June 7-9, 2016, at the Javits Center in New York City, NY.
IoT generates lots of temporal data. But how do you unlock its value? You need to discover patterns that are repeatable in vast quantities of data, understand their meaning, and implement scalable monitoring across multiple data streams in order to monetize the discoveries and insights. Motif discovery and deep learning platforms are emerging to visualize sensor data, to search for patterns and to build application that can monitor real time streams efficiently. In his session at @ThingsExpo, ...
The best-practices for building IoT applications with Go Code that attendees can use to build their own IoT applications. In his session at @ThingsExpo, Indraneel Mitra, Senior Solutions Architect & Technology Evangelist at Cognizant, provided valuable information and resources for both novice and experienced developers on how to get started with IoT and Golang in a day. He also provided information on how to use Intel Arduino Kit, Go Robotics API and AWS IoT stack to build an application tha...
You think you know what’s in your data. But do you? Most organizations are now aware of the business intelligence represented by their data. Data science stands to take this to a level you never thought of – literally. The techniques of data science, when used with the capabilities of Big Data technologies, can make connections you had not yet imagined, helping you discover new insights and ask new questions of your data. In his session at @ThingsExpo, Sarbjit Sarkaria, data science team lead ...
Extracting business value from Internet of Things (IoT) data doesn’t happen overnight. There are several requirements that must be satisfied, including IoT device enablement, data analysis, real-time detection of complex events and automated orchestration of actions. Unfortunately, too many companies fall short in achieving their business goals by implementing incomplete solutions or not focusing on tangible use cases. In his general session at @ThingsExpo, Dave McCarthy, Director of Products...