Click here to close now.




















Welcome!

News Feed Item

Bringing New Intelligence to Cloud Orchestration with Cloudify 3.0

GigaSpaces Technologies today announced it has completely re-architected its Cloudify offering to provide Intelligent Orchestration of applications on the cloud. With this product rewrite, the new Cloudify orchestration platform simplifies the application deployment, management and scaling experience on OpenStack, VMware vSphere and other clouds and environments.

In current orchestration models, most tools focus primarily on application installation, while much of application management occurs after deployment. As a result, vast custom tool chains are often used to manually manage post-deployment processes such as monitoring and logging, leading to significant overhead, complexity and inconsistency across systems. Cloudify’s redesign provides a simple solution for managing the full application lifecycle. The new intelligent orchestration model introduces a feedback loop that automates fixes and updates without manual intervention, all with a single platform that integrates with any tool chain. Cloudify 3.0 reduces the complexity of cloud application management and ensures that managed applications meet their desired SLA.

“To deliver this next generation, intelligent orchestration, we needed to rethink Cloudify’s design,” said Yaron Parasol, VP of Product at GigaSpaces. “With a new language of code, adoption of industry standards and development of scalable and custom workflows, we created something that few are doing today - orchestration of the entire app lifecycle that encompasses both pre-deployment and post-deployment management with a single platform.”

Watch the Cloudify product team share what’s new with Cloudify 3.

Cloudify 3.0 Highlights:

  • New intelligence in orchestration: Cloudify 3.0 eliminates the boundaries between orchestration and monitoring, providing a mechanism that automatically reacts to monitored events with appropriate corrective measures. Version 3.0 includes these building blocks of custom workflows, a workflow engine and a modeling language that enables the automation of any process and any stack. The subsequent release (due in Q4 2014) will introduce monitoring and custom policies for automated triggering of such corrective measures to provide auto-healing and auto-scaling capabilities.
  • Integrating the entire automation and monitoring tool chain: Cloudify 3.0 brings together a variety of tools that are used throughout the various stages of the application lifecycle. In doing so, Cloudify promotes common industry best practices, making it easy to integrate and use best-of-breed tools to manage the environment. Cloudify has a new plug-in architecture that enables easy integration of a wide range of tools for monitoring, configuration management and cloud infrastructure. Examples of such integration include Chef, Puppet, Fabric and Docker for configuration management, OpenStack Heat for infrastructure orchestration, logstash and Elasticsearch for logging and monitoring, and Reimann.IO for real-time analytics.
  • Native integration with OpenStack: As OpenStack is fast becoming the de facto standard for private clouds, Cloudify 3.0 offers even tighter integration with OpenStack technology and core services, including KeyStone, Neutron, Nova and Heat. The underlying design of Cloudify was re-architected to match the design principles of OpenStack services, including the rewriting of the core services in Python and leveraging common infrastructure building blocks such as RabbitMQ.
  • Support for VMware, CloudStack, SoftLayer and other clouds: Cloudify 3.0 contains built-in plug-ins for VMware vSphere and Apache CloudStack, and will soon have plug-ins for vCloud and SoftLayer. It comes with open plug-in architecture to support other clouds, including Amazon AWS, GCE and Linux containers such as Docker (plug-ins for all of which will be released in the coming few weeks). With Cloudify 3.0, users can span the same application across multiple cloud environments without creating a new Cloudify setup per environment. This is useful for users who are transitioning from their existing environment into a cloud environment and also for allowing bursting and hybrid deployments between OpenStack, VMware, Amazon and other clouds.
  • New topology-driven monitoring: Cloudify 3.0 introduces a new concept of topology-driven monitoring in which the entire application management and tracking system is centralized around the application topology, rather than the infrastructure. This makes it possible at any given moment to track not only the state of the application, but also the status of deployment, updates and scaling processes through a single view. As the monitoring system is integrated with the orchestration engine, the two systems are always in sync and up to date, eliminating the need to rely on external discover services.
  • Support for the TOSCA specification: TOSCA (Topology and Orchestration Specification for Cloud Applications) allows users to describe any set of automation processes on cloud applications with an extendable set of hooks and component types. Cloudify 3.0 uses a YAML-based orchestration template driven by the TOSCA specification. The next release of Cloudify will include full syntax compatibility with the TOSCA specification as soon as it becomes officially supported.
  • Multiple applications that can span thousands of nodes: Cloudify 3.0 allows for management and monitoring of large scale applications using a message broker to manage the communication with its managed instances and a logging and analysis engine for big data scale.

About GigaSpaces

GigaSpaces Technologies provides software middleware for deployment, management and scaling of mission-critical applications on cloud environments through two main product lines, XAP In-Memory Computing and Cloudify. Hundreds of Tier-1 organizations worldwide are leveraging GigaSpaces’ technology to enhance IT efficiency and performance, from top financial firms, e-commerce companies, online gaming providers, healthcare organizations and telecom carriers.

GigaSpaces was founded in 2000 and has offices in the US, Europe and Asia. For more information, please visit www.gigaspaces.com or our blog at blog.gigaspaces.com.

More Stories By Business Wire

Copyright © 2009 Business Wire. All rights reserved. Republication or redistribution of Business Wire content is expressly prohibited without the prior written consent of Business Wire. Business Wire shall not be liable for any errors or delays in the content, or for any actions taken in reliance thereon.

Latest Stories
In a recent research, analyst firm IDC found that the average cost of a critical application failure is $500,000 to $1 million per hour and the average total cost of unplanned application downtime is $1.25 billion to $2.5 billion per year for Fortune 1000 companies. In addition to the findings on the cost of the downtime, the research also highlighted best practices for development, testing, application support, infrastructure, and operations teams.
"We specialize in testing. DevOps is all about continuous delivery and accelerating the delivery pipeline and there is no continuous delivery without testing," noted Marc Hornbeek, Sr. Solutions Architect at Spirent Communications, in this SYS-CON.tv interview at @DevOpsSummit, held June 9-11, 2015, at the Javits Center in New York City.
How do you securely enable access to your applications in AWS without exposing any attack surfaces? The answer is usually very complicated because application environments morph over time in response to growing requirements from your employee base, your partners and your customers. In his session at @DevOpsSummit, Haseeb Budhani, CEO and Co-founder of Soha, shared five common approaches that DevOps teams follow to secure access to applications deployed in AWS, Azure, etc., and the friction an...
"Alert Logic is a managed security service provider that basically deploys technologies, but we support those technologies with the people and process behind it," stated Stephen Coty, Chief Security Evangelist at Alert Logic, in this SYS-CON.tv interview at 16th Cloud Expo, held June 9-11, 2015, at the Javits Center in New York City.
Digital Transformation is the ultimate goal of cloud computing and related initiatives. The phrase is certainly not a precise one, and as subject to hand-waving and distortion as any high-falutin' terminology in the world of information technology. Yet it is an excellent choice of words to describe what enterprise IT—and by extension, organizations in general—should be working to achieve. Digital Transformation means: handling all the data types being found and created in the organizat...
The essence of cloud computing is that all consumable IT resources are delivered as services. In his session at 15th Cloud Expo, Yung Chou, Technology Evangelist at Microsoft, demonstrated the concepts and implementations of two important cloud computing deliveries: Infrastructure as a Service (IaaS) and Platform as a Service (PaaS). He discussed from business and technical viewpoints what exactly they are, why we care, how they are different and in what ways, and the strategies for IT to tran...
The Internet of Everything (IoE) brings together people, process, data and things to make networked connections more relevant and valuable than ever before – transforming information into knowledge and knowledge into wisdom. IoE creates new capabilities, richer experiences, and unprecedented opportunities to improve business and government operations, decision making and mission support capabilities.
The Software Defined Data Center (SDDC), which enables organizations to seamlessly run in a hybrid cloud model (public + private cloud), is here to stay. IDC estimates that the software-defined networking market will be valued at $3.7 billion by 2016. Security is a key component and benefit of the SDDC, and offers an opportunity to build security 'from the ground up' and weave it into the environment from day one. In his session at 16th Cloud Expo, Reuven Harrison, CTO and Co-Founder of Tufin,...
The Internet of Things is not only adding billions of sensors and billions of terabytes to the Internet. It is also forcing a fundamental change in the way we envision Information Technology. For the first time, more data is being created by devices at the edge of the Internet rather than from centralized systems. What does this mean for today's IT professional? In this Power Panel at @ThingsExpo, moderated by Conference Chair Roger Strukhoff, panelists addressed this very serious issue of pro...
Container technology is sending shock waves through the world of cloud computing. Heralded as the 'next big thing,' containers provide software owners a consistent way to package their software and dependencies while infrastructure operators benefit from a standard way to deploy and run them. Containers present new challenges for tracking usage due to their dynamic nature. They can also be deployed to bare metal, virtual machines and various cloud platforms. How do software owners track the usag...
With SaaS use rampant across organizations, how can IT departments track company data and maintain security? More and more departments are commissioning their own solutions and bypassing IT. A cloud environment is amorphous and powerful, allowing you to set up solutions for all of your user needs: document sharing and collaboration, mobile access, e-mail, even industry-specific applications. In his session at 16th Cloud Expo, Shawn Mills, President and a founder of Green House Data, discussed h...
Discussions about cloud computing are evolving into discussions about enterprise IT in general. As enterprises increasingly migrate toward their own unique clouds, new issues such as the use of containers and microservices emerge to keep things interesting. In this Power Panel at 16th Cloud Expo, moderated by Conference Chair Roger Strukhoff, panelists addressed the state of cloud computing today, and what enterprise IT professionals need to know about how the latest topics and trends affect t...
"Our biggest growth area has been the security services, the managed services - the things that differentiate us in the market that there is no client that's too small and there's no client that's too big," explained Paul Mazzucco, Chief Security Officer at TierPoint, in this SYS-CON.tv interview at 16th Cloud Expo, held June 9-11, 2015, at the Javits Center in New York City.
SYS-CON Events announced today that HPM Networks will exhibit at the 17th International Cloud Expo®, which will take place on November 3–5, 2015, at the Santa Clara Convention Center in Santa Clara, CA. For 20 years, HPM Networks has been integrating technology solutions that solve complex business challenges. HPM Networks has designed solutions for both SMB and enterprise customers throughout the San Francisco Bay Area.
Containers are changing the security landscape for software development and deployment. As with any security solutions, security approaches that work for developers, operations personnel and security professionals is a requirement. In his session at DevOps Summit, Kevin Gilpin, CTO and Co-Founder of Conjur, will discuss various security considerations for container-based infrastructure and related DevOps workflows.