Akka Streams Integration, codename Alpakka

We believe that Akka Streams can be the tool for building a modern alternative to Apache Camel. That will not happen by itself overnight and this is a call for arms for the community to join us on this mission. The biggest asset of Camel is its rich set of endpoint components. We would like to see that similar endpoints are developed for Akka Streams. Our goal is to build a strong and healthy community around such integrations. We’ve already seen quite some up-take in the community, including connectors to S3, Kafka and more. Akka Streams are built around the core concept of being simple to extend using the powerful yet simple to use APIs. Added components can be used together with all other great things in Akka Streams, such as easy transformation and manipulation of the data stream.

Don’t hesitate to get involved!

In upcoming blog posts we will describe how to use the GraphStage API for building Sinks and Sources to connect to external data sources over various integration protocols. We will show how to handle challenges such as blocking and asynchronous communication. Transformations are also important for integration scenarios and we will illustrate how to implement a streaming XML parser as an example of such encoder/decoder stages.

Akka Streams already has a lot that are useful for integrations. Defining processing pipelines is what the Akka Streams DSL is all about and that is exactly what you need for operating on streaming data that cannot fit in memory as a whole. It handles backpressure in an efficient non-blocking way that prevents out-of-memory errors, which is a typical problem when using unbounded buffering with producers that are faster than consumers.

The following are examples of things that are readily available for building your integrations with Akka Streams today (all available with Java and Scala APIs).

Using Akka Streams and the currently available connectors, an ETL example that deals with multiple data sources and destinations is as straightforward as this:

// Read huge file with Wikipedia content Source < WikipediaEntry , CompletionStage < IOResult >> wikipediaEntries = FileIO . fromPath ( Paths . get ( "/tmp" , "wiki" )) . via ( parseWikiEntries ()); // Enrich the data by fetching matching image from a // web service with HTTP Source < RichWikipediaEntry , CompletionStage < IOResult >> enrichedData = wikipediaEntries . via ( enrichWithImageData ); // Store content in Kafka and corresponding image in AWS S3 enrichedData . alsoTo ( s3ImageStorage ()) . to ( kafkaTopic ) . run ( materializer );

In the above example we use Akka Http to enrich the data:

// parallel fetching of additional data using Akka HTTP, the response is an image final int parallelism = 8 ; final Http http = Http . get ( system ); Flow < WikipediaEntry , RichWikipediaEntry , NotUsed > enrichWithImageData = Flow . of ( WikipediaEntry . class ) . mapAsyncUnordered ( parallelism , w -> { final HttpRequest request = HttpRequest . create ( "http://images.example.com/?query=" + w . title ()); return http . singleRequest ( request , mat ) . thenCompose ( response -> { final CompletionStage < HttpEntity . Strict > entity = response . entity (). toStrict ( 1000 , materializer ); return entity . thenApply ( e -> new RichWikipediaEntry ( w , e . getData ())); } ); });

We use Akka Stream Kafka to publish the content to a Kafka topic:

Sink<RichWikipediaEntry, NotUsed> kafkaTopic = Flow.of(RichWikipediaEntry.class) .map(entry -> entry.wikipediaEntry().content()) .map(elem -> new ProducerRecord("contents", elem)) .to(Producer.plainSink(producerSettings));

The Github repository to use for contributing your favorite integration component is Alpakka. Please create issues and pull requests for discussion and proposals. Take a look at the list of Camel components for inspiration. Implementations in Java or Scala are welcome.

This will be great fun, we are looking forward to your contributions!