Welcome!

News Feed Item

Java Concurrency and Scalability Platform Akka Celebrates Fifth Anniversary

Akka Raised the Standard for Handling Scale and Failure on the JVM; Today Its Developer Community Continues to Grow Across Data Streaming and Internet of Things Use Cases

SAN FRANCISCO, CA -- (Marketwired) -- 07/10/14 -- Typesafe, provider of the world's leading Reactive platform, today announced that July 12 will mark the five year anniversary of Akka, the popular run-time and toolkit for concurrency and scalability on the JVM ("Java Virtual Machine"), supported through the years by developers at high-growth and Blue-Chip companies like Amazon, BBC, Cisco, Credit Suisse, eBay, Groupon, Huffington Post and many more.

The Akka Creation Story (click here for a full interactive timeline on the history of Akka):

Akka was originally created by Swedish programmer Jonas Bonér -- who had built compilers, runtimes and open source frameworks for distributed applications at vendors like BEA and Terracotta. He'd experienced the scale and resilience limitations of CORBA, RPC, XA, EJBs, SOA, and the various Web Services standards and abstraction techniques that Java developers used to approach the overall problem set over the last 20 years. He'd lost faith in those ways of doing things.

This time he looked outside of the Java and enterprise space for answers. He spent some time with the Oz and Erlang programming languages. He saw a lot that he liked about how Erlang managed failure for services that simply could not go down (things like telecom switches for emergency calls), and how principles from Erlang and Oz could be applied towards the concurrency and distributed computing frontiers for mainstream enterprises. In particular he saw the Actor Model -- which emphasizes loose coupling and embracing failure in software systems and dataflow concurrency -- as the bridge to the future.

After months of intense thinking and hacking, Bonér shared his vision for the Akka Actor Kernel (now simply "Akka") on the Scala mailing list, and about a month later (on July 12, 2009) shared the first public release of Akka 0.5 on GitHub. Today Akka is the open source platform that major financial institutions use to handle billions of transactions, and that massively trafficked sites like Walmart and Gilt use to scale their services for peak usage. A full interactive timeline of the history of Akka (including a list of contributors) may be viewed here.

Recent Akka Highlights

As the Akka community has grown, the platform has been leveraged to power highly trafficked web sites, data and analytics, shuffling large amounts of data around, batch processing, real-time processing, and other distributed computing use cases where success means achieving low latency and high throughput. In recent years, several key growth areas have emerged for Akka:

Akka Cluster
In July 2013, version 2.2 of Akka shipped under the code name "Coltrane" and included full support of clustering. Akka Cluster provides a fault-tolerant decentralized peer-to-peer based cluster membership service with no single point of failure or single point of bottleneck. It does this using gossip protocols and an automatic failure detector. It also ships with a suite of high-level modules on top providing things like clustered Pub/Sub, clustered singleton, cluster sharding and more.

Akka Persistence
Predictably handling failure across distributed systems is Akka's calling card. But what happens to the Actor's state when things start failing? In October 2013, Akka Persistence was introduced to allow stateful actors to recover from JVM crashes in a way that Actors themselves are persisted in memory. The key concept in Akka Persistence is called Event Sourcing and allows you to -- instead of storing an actor's state directly -- persist the state-changing events that are sent to the Actor. These changes are immutable facts that are appended to a journal (backed by a pluggable durable storage), which allows for very high transaction rates, efficient replication, migration, replay, auditing, and another powerful layer of failure management.

Akka Streams
Historically, stream-based processing on the JVM ("Java Virtual Machine") has been perilous for both developers and operations, because when data is streamed at higher rates than recipients can handle, it builds up in the system until no space is left, leading to system failures in production. In April 2014, Typesafe announced the release of Akka Streams -- designed to help developers more easily achieve truly asynchronous, non-blocking data streaming on the JVM.

Akka HTTP
In June 2013, Typesafe acquired Spray.io, one of the best performing REST / HTTP libraries in the Java ecosystem. Then in June 2014, Typesafe announced the first preview of the core module of Akka HTTP -- a suite of lightweight Scala libraries providing client/server RESTful support on top of Akka. It fully embraces the Actor-, Future-, and Stream-based programming models used by the underlying platform. This lets developers build high-performant and scalable on RESTful applications with idiomatic Java and Scala code without worrying about wrapping around other Java libraries.

Recent Akka Presentations at Scala Days 2014:

Additional Resources:

About Typesafe
Typesafe (Twitter: @Typesafe) is dedicated to helping developers build Reactive applications on the JVM. With the Typesafe Reactive Platform, you can create modern, event-driven applications that scale on multicore and cloud computing architectures. Typesafe Activator, a browser-based tool with reusable templates, makes it easy to get started with Play Framework, Akka and Scala. Backed by Greylock Partners, Shasta Ventures, Bain Capital Ventures and Juniper Networks, Typesafe is headquartered in San Francisco with offices in Switzerland and Sweden. To start building Reactive applications today, download Typesafe Activator!

Image Available: http://www2.marketwire.com/mw/frame_mw?attachid=2636189

More Stories By Marketwired .

Copyright © 2009 Marketwired. All rights reserved. All the news releases provided by Marketwired are copyrighted. Any forms of copying other than an individual user's personal reference without express written permission is prohibited. Further distribution of these materials is strictly forbidden, including but not limited to, posting, emailing, faxing, archiving in a public database, redistributing via a computer network or in a printed form.

Latest Stories
"There is a huge interest in Kubernetes. People are now starting to use Kubernetes and implement it," stated Sebastian Scheele, co-founder of Loodse, in this SYS-CON.tv interview at DevOps at 19th Cloud Expo, held November 1-3, 2016, at the Santa Clara Convention Center in Santa Clara, CA.
In 2014, Amazon announced a new form of compute called Lambda. We didn't know it at the time, but this represented a fundamental shift in what we expect from cloud computing. Now, all of the major cloud computing vendors want to take part in this disruptive technology. In his session at 20th Cloud Expo, John Jelinek IV, a web developer at Linux Academy, will discuss why major players like AWS, Microsoft Azure, IBM Bluemix, and Google Cloud Platform are all trying to sidestep VMs and containers...
SYS-CON Media announced today that @WebRTCSummit Blog, the largest WebRTC resource in the world, has been launched. @WebRTCSummit Blog offers top articles, news stories, and blog posts from the world's well-known experts and guarantees better exposure for its authors than any other publication. @WebRTCSummit Blog can be bookmarked ▸ Here @WebRTCSummit conference site can be bookmarked ▸ Here
A critical component of any IoT project is what to do with all the data being generated. This data needs to be captured, processed, structured, and stored in a way to facilitate different kinds of queries. Traditional data warehouse and analytical systems are mature technologies that can be used to handle certain kinds of queries, but they are not always well suited to many problems, particularly when there is a need for real-time insights.
Providing secure, mobile access to sensitive data sets is a critical element in realizing the full potential of cloud computing. However, large data caches remain inaccessible to edge devices for reasons of security, size, format or limited viewing capabilities. Medical imaging, computer aided design and seismic interpretation are just a few examples of industries facing this challenge. Rather than fighting for incremental gains by pulling these datasets to edge devices, we need to embrace the i...
All organizations that did not originate this moment have a pre-existing culture as well as legacy technology and processes that can be more or less amenable to DevOps implementation. That organizational culture is influenced by the personalities and management styles of Executive Management, the wider culture in which the organization is situated, and the personalities of key team members at all levels of the organization. This culture and entrenched interests usually throw a wrench in the work...
Web Real-Time Communication APIs have quickly revolutionized what browsers are capable of. In addition to video and audio streams, we can now bi-directionally send arbitrary data over WebRTC's PeerConnection Data Channels. With the advent of Progressive Web Apps and new hardware APIs such as WebBluetooh and WebUSB, we can finally enable users to stitch together the Internet of Things directly from their browsers while communicating privately and securely in a decentralized way.
IoT is at the core or many Digital Transformation initiatives with the goal of re-inventing a company's business model. We all agree that collecting relevant IoT data will result in massive amounts of data needing to be stored. However, with the rapid development of IoT devices and ongoing business model transformation, we are not able to predict the volume and growth of IoT data. And with the lack of IoT history, traditional methods of IT and infrastructure planning based on the past do not app...
In his session at DevOps Summit, Tapabrata Pal, Director of Enterprise Architecture at Capital One, will tell a story about how Capital One has embraced Agile and DevOps Security practices across the Enterprise – driven by Enterprise Architecture; bringing in Development, Operations and Information Security organizations together. Capital Ones DevOpsSec practice is based upon three "pillars" – Shift-Left, Automate Everything, Dashboard Everything. Within about three years, from 100% waterfall, C...
Fifty billion connected devices and still no winning protocols standards. HTTP, WebSockets, MQTT, and CoAP seem to be leading in the IoT protocol race at the moment but many more protocols are getting introduced on a regular basis. Each protocol has its pros and cons depending on the nature of the communications. Does there really need to be only one protocol to rule them all? Of course not. In his session at @ThingsExpo, Chris Matthieu, co-founder and CTO of Octoblu, walked through how Octob...
The Internet of Things can drive efficiency for airlines and airports. In their session at @ThingsExpo, Shyam Varan Nath, Principal Architect with GE, and Sudip Majumder, senior director of development at Oracle, discussed the technical details of the connected airline baggage and related social media solutions. These IoT applications will enhance travelers' journey experience and drive efficiency for the airlines and the airports.
"We're bringing out a new application monitoring system to the DevOps space. It manages large enterprise applications that are distributed throughout a node in many enterprises and we manage them as one collective," explained Kevin Barnes, President of eCube Systems, in this SYS-CON.tv interview at DevOps at 18th Cloud Expo, held June 7-9, 2016, at the Javits Center in New York City, NY.
Containers have changed the mind of IT in DevOps. They enable developers to work with dev, test, stage and production environments identically. Containers provide the right abstraction for microservices and many cloud platforms have integrated them into deployment pipelines. DevOps and containers together help companies achieve their business goals faster and more effectively. In his session at DevOps Summit, Ruslan Synytsky, CEO and Co-founder of Jelastic, reviewed the current landscape of Dev...
SYS-CON Events announced today that Catchpoint, a leading digital experience intelligence company, has been named “Silver Sponsor” of SYS-CON's 20th International Cloud Expo®, which will take place on June 6-8, 2017, at the Javits Center in New York City, NY. Catchpoint Systems is a leading Digital Performance Analytics company that provides unparalleled insight into your customer-critical services to help you consistently deliver an amazing customer experience. Designed for digital business, C...
"We formed Formation several years ago to really address the need for bring complete modernization and software-defined storage to the more classic private cloud marketplace," stated Mark Lewis, Chairman and CEO of Formation Data Systems, in this SYS-CON.tv interview at 18th Cloud Expo, held June 7-9, 2016, at the Javits Center in New York City, NY.