Welcome!

News Feed Item

Pivotal Advances Hadoop Offering, Integrates In-Memory Processing and Delivers Business Data Lake Architecture for the Enterprise

Pivotal HD 2.0 and Pivotal GemFire XD deliver the most advanced big data analytics platform on the market; Fortifies next-generation capabilities of data driven applications and analytics

SAN FRANCISCO, March 17, 2014 /PRNewswire/ --

Summary

  • Building on its architecture for Business Data Lakes, Pivotal today announced the availability of Pivotal HD 2.0, a commercially supported enterprise distribution of the Apache Hadoop stack, now rebased and hardened on Apache Hadoop 2.2.
  • Pivotal also announced the general availability of Pivotal GemFire XD, an in-memory database seamlessly integrated with Pivotal HD 2.0 that brings high concurrency, real-time transactions and in-memory analytical processing to power next generation big data applications.
  • Pivotal HD 2.0 expands analytic use cases with integration and support of GraphLab, MADlib, and popular languages and formats such as R, Python, Java, and Parquet to create a powerful and easy to use analytical platform for data scientists and analysts in Hadoop.
  • Pivotal is hosting a webinar on March 27, 2014 that expands on today's news.  Registration to Pivotal's Next Generation Business Data Lake webinar can be found at http://bit.ly/1i8on7g.

 

www.gopivotal.com.

Pivotal, the software company at the intersection of big data, PaaS, and agile development, today announced the release of Pivotal™ HD 2.0 and the general availability of Pivotal GemFire™ XD. The combination of Pivotal HD 2.0, the HAWQ™ query engine, and Gemfire XD constitute the foundation for the Business Data Lake architecture, the big data application framework for enterprises, data scientists, analysts and developers that provides a more flexible, faster way to develop data savvy software than what they can do with Hadoop alone.

Breaking New Ground In Real-Time for Apps, Data and Analytics

Behind every leading enterprise there is real-time analytics that drives real-time advantages and intelligence. Available today, Pivotal GemFire XD bridges GemFire's proven in-memory intelligence and integrates it with Pivotal HD 2.0 and HAWQ. Pivotal GemFire technology enables businesses to make prescriptive decisions in real-time, such as stock trading, fraud detection, intelligence for energy companies, routing for the telecom industries, or for scaling reservations for the world's largest annual movement of humans on the planet.

Also new within Pivotal HD is the world's first enterprise integration of GraphLab, an advanced set of algorithms for graph analytics that enables data scientists and analysts to leverage popular algorithms for insight, i.e. page rank, collaborative filtering and computer vision.

Pivotal released Pivotal HD with HAWQ last year to radically increase the speed of analysis for Hadoop queries; it was designed from the ground up as a massively parallel SQL processing engine optimized specifically for analytics.

With Pivotal HD 2.0, new improvements to HAWQ include:  

  • MADlib Machine Learning Library – Unlock deeper predictive insights faster, better with over 50 in-database analytic algorithms;
  • Language Translation – Leverage the full power of R, Python and Java to enable business logic and procedures otherwise cumbersome with SQL;
  • Parquet Support – Beta support for read and write Parquet files, opening the power of HAWQ's SQL query engine on this popular open file type

Supporting Quotes

Nik Rouda, Senior Analyst, ESG

"The combined release of Pivotal HD 2.0 and GemFireXD, introducing real-time SQL queries, makes it easier for developers to get live insights on streaming data sources without having to learn new tools. Not least, the range of new programming language support gives an even richer portfolio of advanced analytics capabilities. Pivotal is well-positioned to change the nature of how big data gets done in the enterprise."

Josh Klahr, Vice President, Product Management, Pivotal

"When it comes to Hadoop, other approaches in the market have left customers with a mishmash of un-integrated products and processes. Building on our industry-leading SQL-on-Hadoop offer, HAWQ, Pivotal HD 2.0 is the first platform to fully integrate proven enterprise in-memory technology, Pivotal GemFire XD, with advanced services on Hadoop 2.2 that provide native support for a comprehensive data science toolset. Data driven businesses now have the capabilities they need to gain a massive head start toward developing analytics and applications for more intelligent and innovative products and services."

Resources

About Pivotal
Pivotal, committed to open source and open standards, recently introduced Pivotal One, the world's first comprehensive multi-cloud Enterprise PaaS. The company is also a leading provider of application and data infrastructure software, agile development services, and data science consulting. Follow Pivotal on Twitter @gopivotal, LinkedIn, and G+.

©2014 Pivotal Software, Inc.  All rights reserved.  Pivotal, GemFire and HAWQ are trademarks and/or registered trademarks of Pivotal Software, Inc. in the United States and/or other Countries. This release contains "forward-looking statements" as defined under the Federal Securities Laws. Actual results could differ materially from those projected in the forward-looking statements as a result of certain risk factors, including but not limited to: (i) adverse changes in general economic or market conditions; (ii) delays or reductions in information technology spending; (iii) the relative and varying rates of product price and component cost declines and the volume and mixture of product and services revenues; (iv) competitive factors, including but not limited to pricing pressures and new product introductions; (v) component and product quality and availability; (vi) fluctuations in VMware, Inc.'s operating results and risks associated with trading of VMware stock; (vii) the transition to new products, the uncertainty of customer acceptance of new product offerings and rapid technological and market change; (viii) risks associated with managing the growth of our business, including risks associated with acquisitions and investments and the challenges and costs of integration, restructuring and achieving anticipated synergies; (ix) the ability to attract and retain highly qualified employees; (x) insufficient, excess or obsolete inventory; (xi) fluctuating currency exchange rates; (xii) threats and other disruptions to our secure data centers or networks; (xiii) our ability to protect our proprietary technology; (xiv) war or acts of terrorism; and (xv) other one-time events and other important factors disclosed previously and from time to time in the filings EMC Corporation, the parent company of Pivotal, with the U.S. Securities and Exchange Commission.  EMC and Pivotal disclaim any obligation to update any such forward-looking statements after the date of this release.

Logo - http://photos.prnewswire.com/prnh/20130910/SF76762LOGO

SOURCE Pivotal

More Stories By PR Newswire

Copyright © 2007 PR Newswire. All rights reserved. Republication or redistribution of PRNewswire content is expressly prohibited without the prior written consent of PRNewswire. PRNewswire shall not be liable for any errors or delays in the content, or for any actions taken in reliance thereon.

Latest Stories
The revocation of Safe Harbor has radically affected data sovereignty strategy in the cloud. In his session at 17th Cloud Expo, Jeff Miller, Product Management at Cavirin Systems, discussed how to assess these changes across your own cloud strategy, and how you can mitigate risks previously covered under the agreement.
CloudEXPO New York 2018, colocated with DXWorldEXPO New York 2018 will be held November 11-13, 2018, in New York City and will bring together Cloud Computing, FinTech and Blockchain, Digital Transformation, Big Data, Internet of Things, DevOps, AI, Machine Learning and WebRTC to one location.
The Internet of Things will challenge the status quo of how IT and development organizations operate. Or will it? Certainly the fog layer of IoT requires special insights about data ontology, security and transactional integrity. But the developmental challenges are the same: People, Process and Platform and how we integrate our thinking to solve complicated problems. In his session at 19th Cloud Expo, Craig Sproule, CEO of Metavine, demonstrated how to move beyond today's coding paradigm and sh...
The best way to leverage your Cloud Expo presence as a sponsor and exhibitor is to plan your news announcements around our events. The press covering Cloud Expo and @ThingsExpo will have access to these releases and will amplify your news announcements. More than two dozen Cloud companies either set deals at our shows or have announced their mergers and acquisitions at Cloud Expo. Product announcements during our show provide your company with the most reach through our targeted audiences.
More and more companies are looking to microservices as an architectural pattern for breaking apart applications into more manageable pieces so that agile teams can deliver new features quicker and more effectively. What this pattern has done more than anything to date is spark organizational transformations, setting the foundation for future application development. In practice, however, there are a number of considerations to make that go beyond simply “build, ship, and run,” which changes how...
Traditional on-premises data centers have long been the domain of modern data platforms like Apache Hadoop, meaning companies who build their business on public cloud were challenged to run Big Data processing and analytics at scale. But recent advancements in Hadoop performance, security, and most importantly cloud-native integrations, are giving organizations the ability to truly gain value from all their data. In his session at 19th Cloud Expo, David Tishgart, Director of Product Marketing ...
DevOpsSummit New York 2018, colocated with CloudEXPO | DXWorldEXPO New York 2018 will be held November 11-13, 2018, in New York City. Digital Transformation (DX) is a major focus with the introduction of DXWorldEXPO within the program. Successful transformation requires a laser focus on being data-driven and on using all the tools available that enable transformation if they plan to survive over the long term.
As Enterprise business moves from Monoliths to Microservices, adoption and successful implementations of Microservices become more evident. The goal of Microservices is to improve software delivery speed and increase system safety as scale increases. Documenting hurdles and problems for the use of Microservices will help consultants, architects and specialists to avoid repeating the same mistakes and learn how and when to use (or not use) Microservices at the enterprise level. The circumstance w...
Enterprises are striving to become digital businesses for differentiated innovation and customer-centricity. Traditionally, they focused on digitizing processes and paper workflow. To be a disruptor and compete against new players, they need to gain insight into business data and innovate at scale. Cloud and cognitive technologies can help them leverage hidden data in SAP/ERP systems to fuel their businesses to accelerate digital transformation success.
@DevOpsSummit at Cloud Expo, taking place November 12-13 in New York City, NY, is co-located with 22nd international CloudEXPO | first international DXWorldEXPO and will feature technical sessions from a rock star conference faculty and the leading industry players in the world. The widespread success of cloud computing is driving the DevOps revolution in enterprise IT. Now as never before, development teams must communicate and collaborate in a dynamic, 24/7/365 environment. There is no time t...
What are the new priorities for the connected business? First: businesses need to think differently about the types of connections they will need to make – these span well beyond the traditional app to app into more modern forms of integration including SaaS integrations, mobile integrations, APIs, device integration and Big Data integration. It’s important these are unified together vs. doing them all piecemeal. Second, these types of connections need to be simple to design, adapt and configure...
Cell networks have the advantage of long-range communications, reaching an estimated 90% of the world. But cell networks such as 2G, 3G and LTE consume lots of power and were designed for connecting people. They are not optimized for low- or battery-powered devices or for IoT applications with infrequently transmitted data. Cell IoT modules that support narrow-band IoT and 4G cell networks will enable cell connectivity, device management, and app enablement for low-power wide-area network IoT. B...
Transformation Abstract Encryption and privacy in the cloud is a daunting yet essential task for both security practitioners and application developers, especially as applications continue moving to the cloud at an exponential rate. What are some best practices and processes for enterprises to follow that balance both security and ease of use requirements? What technologies are available to empower enterprises with code, data and key protection from cloud providers, system administrators, inside...
With the proliferation of both SQL and NoSQL databases, organizations can now target specific fit-for-purpose database tools for their different application needs regarding scalability, ease of use, ACID support, etc. Platform as a Service offerings make this even easier now, enabling developers to roll out their own database infrastructure in minutes with minimal management overhead. However, this same amount of flexibility also comes with the challenges of picking the right tool, on the right ...
Organizations planning enterprise data center consolidation and modernization projects are faced with a challenging, costly reality. Requirements to deploy modern, cloud-native applications simultaneously with traditional client/server applications are almost impossible to achieve with hardware-centric enterprise infrastructure. Compute and network infrastructure are fast moving down a software-defined path, but storage has been a laggard. Until now.