Akka Java
Akka Java
Release 2.3.1
Typesafe Inc
CONTENTS
Introduction
1.1 What is Akka? . . . . . . . . . . . .
1.2 Why Akka? . . . . . . . . . . . . . .
1.3 Getting Started . . . . . . . . . . . .
1.4 The Obligatory Hello World . . . . .
1.5 Use-case and Deployment Scenarios .
1.6 Examples of use-cases for Akka . . .
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
1
1
2
3
6
7
7
General
2.1 Terminology, Concepts . . . . . . . . .
2.2 Actor Systems . . . . . . . . . . . . .
2.3 What is an Actor? . . . . . . . . . . .
2.4 Supervision and Monitoring . . . . . .
2.5 Actor References, Paths and Addresses
2.6 Location Transparency . . . . . . . . .
2.7 Akka and the Java Memory Model . . .
2.8 Message Delivery Reliability . . . . .
2.9 Configuration . . . . . . . . . . . . . .
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
9
9
11
13
15
18
24
25
27
32
Actors
3.1 Actors . . . . . . . . . . . . . . . .
3.2 Typed Actors . . . . . . . . . . . . .
3.3 Fault Tolerance . . . . . . . . . . . .
3.4 Dispatchers . . . . . . . . . . . . . .
3.5 Mailboxes . . . . . . . . . . . . . .
3.6 Routing . . . . . . . . . . . . . . . .
3.7 Building Finite State Machine Actors
3.8 Persistence . . . . . . . . . . . . . .
3.9 Testing Actor Systems . . . . . . . .
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
69
69
88
93
107
110
115
132
135
153
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
167
167
186
200
209
.
.
.
.
.
.
.
.
.
Networking
240
6.1 Cluster Specification . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 240
6.2 Cluster Usage . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 248
6.3 Remoting . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 269
i
6.4
6.5
6.6
6.7
6.8
6.9
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
278
283
285
296
299
304
Utilities
7.1 Event Bus . . . .
7.2 Logging . . . . .
7.3 Scheduler . . . .
7.4 Duration . . . .
7.5 Circuit Breaker .
7.6 Akka Extensions
7.7 Microkernel . .
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
317
317
322
328
331
332
336
338
340
340
341
344
Experimental Modules
9.1 Persistence . . . . . . . . . . . . .
9.2 Multi Node Testing . . . . . . . . .
9.3 Actors (Java with Lambda Support)
9.4 FSM (Java with Lambda Support) .
9.5 External Contributions . . . . . . .
Serialization
I/O . . . . .
Using TCP .
Using UDP .
ZeroMQ . .
Camel . . . .
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
345
345
362
367
386
395
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
422
422
424
427
429
430
433
11 Project Information
11.1 Migration Guides
11.2 Issue Tracking .
11.3 Licenses . . . .
11.4 Sponsors . . . .
11.5 Project . . . . .
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
434
434
442
442
443
443
12 Additional Information
12.1 Frequently Asked Questions . . . . . .
12.2 Books . . . . . . . . . . . . . . . . . .
12.3 Other Language Bindings . . . . . . .
12.4 Akka in OSGi . . . . . . . . . . . . .
12.5 Incomplete List of HTTP Frameworks .
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
445
445
448
448
448
449
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
ii
CHAPTER
ONE
INTRODUCTION
1.1 What is Akka?
Scalable real-time transaction processing
We believe that writing correct concurrent, fault-tolerant and scalable applications is too hard. Most of the time its
because we are using the wrong tools and the wrong level of abstraction. Akka is here to change that. Using the
Actor Model we raise the abstraction level and provide a better platform to build scalable, resilient and responsive
applicationssee the Reactive Manifesto for more details. For fault-tolerance we adopt the let it crash model
which the telecom industry has used with great success to build applications that self-heal and systems that never
stop. Actors also provide the abstraction for transparent distribution and the basis for truly scalable and faulttolerant applications.
Akka is Open Source and available under the Apache 2 License.
Download from http://akka.io/downloads.
Please note that all code samples compile, so if you want direct access to the sources, have a look over at the Akka
Docs subproject on github: for Java and Scala.
Location Transparency
Everything in Akka is designed to work in a distributed environment: all interactions of actors use pure message
passing and everything is asynchronous.
For an overview of the cluster support see the Java and Scala documentation chapters.
Persistence
Messages received by an actor can optionally be persisted and replayed when the actor is started or restarted. This
allows actors to recover their state, even after JVM crashes or when being migrated to another node.
You can find more details in the respective chapter for Java or Scala.
how
they
are
using
Akka:
1.3.3 Download
There are several ways to download Akka. You can download it as part of the Typesafe Platform (as described
above). You can download the full distribution with microkernel, which includes all modules. Or you can use a
build tool like Maven or SBT to download dependencies from the Akka Maven repository.
1.3.4 Modules
Akka is very modular and consists of several JARs containing different features.
akka-actor Classic Actors, Typed Actors, IO Actor etc.
akka-agent Agents, integrated with Scala STM
akka-camel Apache Camel integration
akka-cluster Cluster membership management, elastic routers.
akka-kernel Akka microkernel for running a bare-bones mini application server
akka-osgi base bundle for using Akka in OSGi containers, containing the akka-actor classes
akka-osgi-aries Aries blueprint for provisioning actor systems
akka-remote Remote Actors
akka-slf4j SLF4J Logger (event bus listener)
akka-testkit Toolkit for testing Actor systems
akka-zeromq ZeroMQ integration
In addition to these stable modules there are several which are on their way into the stable core but are still marked
experimental at this point. This does not mean that they do not function as intended, it primarily means that
their API has not yet solidified enough in order to be considered frozen. You can help accelerating this process by
giving feedback on these modules on our mailing list.
akka-contrib an assortment of contributions which may or may not be moved into core modules, see
External Contributions for more details.
The filename of the actual JAR is for example akka-actor_2.10-2.3.1.jar (and analog for the other
modules).
How to see the JARs dependencies of each Akka module is described in the Dependencies section.
1.3.7 Microkernel
The Akka distribution includes the microkernel. To run the microkernel put your application jar in the deploy
directory and use the scripts in the bin directory.
More information is available in the documentation of the Microkernel (Scala) / Microkernel (Java).
Note: for snapshot versions both SNAPSHOT and timestamped versions are published.
Note: the libraryDependencies setting above is specific to SBT v0.12.x and higher. If you are using an older
version of SBT, the libraryDependencies should look like this:
libraryDependencies +=
"com.typesafe.akka" % "akka-actor_2.10" % "2.3.1"
1.6.1 Here are some of the areas where Akka is being deployed into production
Transaction processing (Online Gaming, Finance/Banking, Trading, Statistics, Betting, Social
Media, Telecom)
Scale up, scale out, fault-tolerance / HA
Service backend (any industry, any app)
Service REST, SOAP, Cometd, WebSockets etc Act as message hub / integration layer Scale up, scale
out, fault-tolerance / HA
Concurrency/parallelism (any app)
Correct Simple to work with and understand Just add the jars to your existing JVM project (use Scala,
Java, Groovy or JRuby)
Simulation
Master/Worker, Compute Grid, MapReduce etc.
CHAPTER
TWO
GENERAL
2.1 Terminology, Concepts
In this chapter we attempt to establish a common terminology to define a solid ground for communicating about
concurrent, distributed systems which Akka targets. Please note that, for many of these terms, there is no single agreed definition. We simply seek to give working definitions that will be used in the scope of the Akka
documentation.
affected subsystems stall. Deadlock is closely related to blocking, as it is necessary that a participant thread be
able to delay the progression of other threads indefinitely.
In the case of deadlock, no participants can make progress, while in contrast Starvation happens, when there are
participants that can make progress, but there might be one or more that cannot. Typical scenario is the case
of a naive scheduling algorithm that always selects high-priority tasks over low-priority ones. If the number of
incoming high-priority tasks is constantly high enough, no low-priority ones will be ever finished.
Livelock is similar to deadlock as none of the participants make progress. The difference though is that instead
of being frozen in a state of waiting for others to progress, the participants continuously change their state. An
example scenario when two participants have two identical resources available. They each try to get the resource,
but they also check if the other needs the resource, too. If the resource is requested by the other participant, they
try to get the other instance of the resource. In the unfortunate case it might happen that the two participants
bounce between the two resources, never acquiring it, but always yielding to the other.
10
it finishes in a bounded number of steps. All lock-free objects are obstruction-free, but the opposite is generally
not true.
Optimistic concurrency control (OCC) methods are usually obstruction-free. The OCC approach is that every
participant tries to execute its operation on the shared object, but if a participant detects conflicts from others, it
rolls back the modifications, and tries again according to some schedule. If there is a point in time, where one of
the participants is the only one trying, the operation will succeed.
11
which simplifies state management for collecting the replies. This is known as the Error Kernel Pattern
from Erlang.
If one actor depends on another actor for carrying out its duty, it should watch that other actors liveness
and act upon receiving a termination notice. This is different from supervision, as the watching party has
no influence on the supervisor strategy, and it should be noted that a functional dependency alone is not a
criterion for deciding where to place a certain child actor in the hierarchy.
There are of course always exceptions to these rules, but no matter whether you follow the rules or break them,
you should always have a reason.
12
Dedicate a single thread to manage a set of blocking resources (e.g. a NIO selector driving multiple channels) and dispatch events as they occur as actor messages.
The first possibility is especially well-suited for resources which are single-threaded in nature, like database handles which traditionally can only execute one outstanding query at a time and use internal synchronization to
ensure this. A common pattern is to create a router for N actors, each of which wraps a single DB connection and
handles queries as sent to the router. The number N must then be tuned for maximum throughput, which will vary
depending on which DBMS is deployed on what hardware.
Note: Configuring thread pools is a task best delegated to Akka, simply configure in the application.conf
and instantiate through an ActorSystem [Java, Scala]
2.3.2 State
Actor objects will typically contain some variables which reflect possible states the actor may be in. This can be an
explicit state machine (e.g. using the fsm-scala module), or it could be a counter, set of listeners, pending requests,
etc. These data are what make an actor valuable, and they must be protected from corruption by other actors. The
good news is that Akka actors conceptually each have their own light-weight thread, which is completely shielded
from the rest of the system. This means that instead of having to synchronize access using locks you can just write
your actor code without worrying about concurrency at all.
Behind the scenes Akka will run sets of actors on sets of real threads, where typically many actors share one
thread, and subsequent invocations of one actor may end up being processed on different threads. Akka ensures
that this implementation detail does not affect the single-threadedness of handling the actors state.
13
Because the internal state is vital to an actors operations, having inconsistent state is fatal. Thus, when the actor
fails and is restarted by its supervisor, the state will be created from scratch, like upon first creating the actor. This
is to enable the ability of self-healing of the system.
Optionally, an actors state can be automatically recovered to the state before a restart by persisting received
messages and replaying them after restart (see Persistence).
2.3.3 Behavior
Every time a message is processed, it is matched against the current behavior of the actor. Behavior means a
function which defines the actions to be taken in reaction to the message at that point in time, say forward a
request if the client is authorized, deny it otherwise. This behavior may change over time, e.g. because different
clients obtain authorization over time, or because the actor may go into an out-of-service mode and later come
back. These changes are achieved by either encoding them in state variables which are read from the behavior
logic, or the function itself may be swapped out at runtime, see the become and unbecome operations. However,
the initial behavior defined during construction of the actor object is special in the sense that a restart of the actor
will reset its behavior to this initial one.
2.3.4 Mailbox
An actors purpose is the processing of messages, and these messages were sent to the actor from other actors (or
from outside the actor system). The piece which connects sender and receiver is the actors mailbox: each actor
has exactly one mailbox to which all senders enqueue their messages. Enqueuing happens in the time-order of
send operations, which means that messages sent from different actors may not have a defined order at runtime
due to the apparent randomness of distributing actors across threads. Sending multiple messages to the same target
from the same actor, on the other hand, will enqueue them in the same order.
There are different mailbox implementations to choose from, the default being a FIFO: the order of the messages
processed by the actor matches the order in which they were enqueued. This is usually a good default, but
applications may need to prioritize some messages over others. In this case, a priority mailbox will enqueue not
always at the end but at a position as given by the message priority, which might even be at the front. While using
such a queue, the order of messages processed will naturally be defined by the queues algorithm and in general
not be FIFO.
An important feature in which Akka differs from some other actor model implementations is that the current
behavior must always handle the next dequeued message, there is no scanning the mailbox for the next matching
one. Failure to handle a message will typically be treated as a failure, unless this behavior is overridden.
2.3.5 Children
Each actor is potentially a supervisor: if it creates children for delegating sub-tasks, it will automatically supervise
them. The list of children is maintained within the actors context and the actor has access to it. Modifications to
the list are done by creating (context.actorOf(...)) or stopping (context.stop(child)) children
and these actions are reflected immediately. The actual creation and termination actions happen behind the scenes
in an asynchronous way, so they do not block their supervisor.
14
15
Warning: Supervision related parent-child communication happens by special system messages that have
their own mailboxes separate from user messages. This implies that supervision related events are not deterministically ordered relative to ordinary messages. In general, the user cannot influence the order of normal
messages and failure notifications. For details and example see the Discussion: Message Ordering section.
An actor system will during its creation start at least three actors, shown in the image above. For more information
about the consequences for actor paths see Top-Level Scopes for Actor Paths.
/user: The Guardian Actor
The actor which is probably most interacted with is the parent of all user-created actors, the guardian named
"/user". Actors created using system.actorOf() are children of this actor. This means that when this
guardian terminates, all normal actors in the system will be shutdown, too. It also means that this guardians
supervisor strategy determines how the top-level normal actors are supervised. Since Akka 2.1 it is possible to
configure this using the setting akka.actor.guardian-supervisor-strategy, which takes the fullyqualified class-name of a SupervisorStrategyConfigurator. When the guardian escalates a failure, the
root guardians response will be to terminate the guardian, which in effect will shut down the whole actor system.
/system: The System Guardian
This special guardian has been introduced in order to achieve an orderly shut-down sequence where logging remains active while all normal actors terminate, even though logging itself is implemented using actors. This
is realized by having the system guardian watch the user guardian and initiate its own shut-down upon reception of the Terminated message. The top-level system actors are supervised using a strategy which
will restart indefinitely upon all types of Exception except for ActorInitializationException and
ActorKilledException, which will terminate the child in question. All other throwables are escalated,
which will shut down the whole actor system.
/: The Root Guardian
The root guardian is the grand-parent of all so-called top-level actors and supervises all the special actors
mentioned in Top-Level Scopes for Actor Paths using the SupervisorStrategy.stoppingStrategy,
whose purpose is to terminate the child upon any type of Exception. All other throwables will be escalated
. . . but to whom? Since every real actor has a supervisor, the supervisor of the root guardian cannot be a real
16
actor. And because this means that it is outside of the bubble, it is called the bubble-walker. This is a
synthetic ActorRef which in effect stops its child upon the first sign of trouble and sets the actor systems
isTerminated status to true as soon as the root guardian is fully terminated (all children recursively stopped).
17
Monitoring is particularly useful if a supervisor cannot simply restart its children and has to terminate them, e.g.
in case of errors during actor initialization. In that case it should monitor those children and re-create them or
schedule itself to retry this at a later time.
Another common use case is that an actor needs to fail in the absence of an external resource, which may also be
one of its own children. If a third party terminates a child by way of the system.stop(child) method or
sending a PoisonPill, the supervisor might well be affected.
18
The above image displays the relationship between the most important entities within an actor system, please read
on for the details.
19
Here, akka.tcp is the default remote transport for the 2.2 release; other transports are pluggable. A remote host using UDP would be accessible by using akka.udp. The interpretation of the host and port part
(i.e.serv.example.com:5678 in the example) depends on the transport mechanism used, but it must abide by the
URI structural rules.
Logical Actor Paths
The unique path obtained by following the parental supervision links towards the root guardian is called the logical
actor path. This path matches exactly the creation ancestry of an actor, so it is completely deterministic as soon as
the actor systems remoting configuration (and with it the address component of the path) is set.
Physical Actor Paths
While the logical actor path describes the functional location within one actor system, configuration-based remote
deployment means that an actor may be created on a different network host than its parent, i.e. within a different
actor system. In this case, following the actor path from the root guardian up entails traversing the network, which
is a costly operation. Therefore, each actor also has a physical path, starting at the root guardian of the actor
system where the actual actor object resides. Using this path as sender reference when querying other actors will
let them reply directly to this actor, minimizing delays incurred by routing.
One important aspect is that a physical actor path never spans multiple actor systems or JVMs. This means that
the logical path (supervision hierarchy) and the physical path (actor deployment) of an actor may diverge if one
of its ancestors is remotely supervised.
20
Creating Actors
An actor system is typically started by creating actors beneath the guardian actor using the
ActorSystem.actorOf method and then using ActorContext.actorOf from within the created
actors to spawn the actor tree. These methods return a reference to the newly created actor. Each actor has direct
access (through its ActorContext) to references for its parent, itself and its children. These references may be
sent within messages to other actors, enabling those to reply directly.
Looking up Actors by Concrete Path
In addition, actor references may be looked up using the ActorSystem.actorSelection method. The
selection can be used for communicating with said actor and the actor corresponding to the selection is looked up
when delivering each message.
To acquire an ActorRef that is bound to the life-cycle of a specific actor you need to send a message, such as
the built-in Identify message, to the actor and use the sender() reference of a reply from the actor.
Note: actorFor is deprecated in favor of actorSelection because actor references acquired with
actorFor behave differently for local and remote actors. In the case of a local actor reference, the named
actor needs to exist before the lookup, or else the acquired reference will be an EmptyLocalActorRef. This
will be true even if an actor with that exact path is created after acquiring the actor reference. For remote actor
references acquired with actorFor the behaviour is different and sending messages to such a reference will under
the hood look up the actor by path on the remote system for every message send.
Absolute paths may of course also be looked up on context in the usual way, i.e.
context.actorSelection("/user/serviceA") ! msg
will send msg to all siblings including the current actor. As for references obtained using actorFor, a traversal of
the supervision hierarchy is done in order to perform the message send. As the exact set of actors which match
a selection may change even while a message is making its way to the recipients, it is not possible to watch a
selection for liveliness changes. In order to do that, resolve the uncertainty by sending a request and gathering all
answers, extracting the sender references, and then watch all discovered concrete actors. This scheme of resolving
a selection may be improved upon in a future release.
2.5. Actor References, Paths and Addresses
21
Note: What the above sections described in some detail can be summarized and memorized easily as follows:
actorOf only ever creates a new actor, and it creates it as a direct child of the context on which this method
is invoked (which may be any actor or actor system).
actorSelection only ever looks up existing actors when messages are delivered, i.e. does not create
actors, or verify existence of actors when the selection is created.
actorFor (deprecated in favor of actorSelection) only ever looks up an existing actor, i.e. does not create
one.
22
"/system" is the guardian actor for all system-created top-level actors, e.g. logging listeners or actors
automatically deployed by configuration at the start of the actor system.
"/deadLetters" is the dead letter actor, which is where all messages sent to stopped or non-existing
actors are re-routed (on a best-effort basis: messages may be lost even within the local JVM).
"/temp" is the guardian for all short-lived system-created actors, e.g. those which are used in the implementation of ActorRef.ask.
"/remote" is an artificial path below which all actors reside whose supervisors are remote actor references
The need to structure the name space for actors like this arises from a central and very simple design goal:
everything in the hierarchy is an actor, and all actors function in the same way. Hence you can not only look
up the actors you created, you can also look up the system guardian and send it a message (which it will dutifully
2.5. Actor References, Paths and Addresses
23
discard in this case). This powerful principle means that there are no quirks to remember, it makes the whole
system more uniform and consistent.
If you want to read more about the top-level structure of an actor system, have a look at The Top-Level Supervisors.
24
25
The actor send rule: the send of the message to an actor happens before the receive of that message by the
same actor.
The actor subsequent processing rule: processing of one message happens before processing of the next
message by the same actor.
Note: In laymans terms this means that changes to internal fields of the actor are visible when the next message
is processed by that actor. So fields in your actor need not be volatile or equivalent.
Both rules only apply for the same actor instance and are not valid if different actors are used.
26
Messages should be immutable, this is to avoid the shared mutable state trap.
27
at-least-once delivery means that for each message handed to the mechanism potentially multiple attempts
are made at delivering it, such that at least one succeeds; again, in more casual terms this means that
messages may be duplicated but not lost.
exactly-once delivery means that for each message handed to the mechanism exactly one delivery is made
to the recipient; the message can neither be lost nor duplicated.
The first one is the cheapesthighest performance, least implementation overheadbecause it can be done in
a fire-and-forget fashion without keeping state at the sending end or in the transport mechanism. The second
one requires retries to counter transport losses, which means keeping state at the sending end and having an
acknowledgement mechanism at the receiving end. The third is most expensiveand has consequently worst
performancebecause in addition to the second it requires state to be kept at the receiving end in order to filter
out duplicate deliveries.
Discussion: Why No Guaranteed Delivery?
At the core of the problem lies the question what exactly this guarantee shall mean:
1. The message is sent out on the network?
2. The message is received by the other host?
3. The message is put into the target actors mailbox?
4. The message is starting to be processed by the target actor?
5. The message is processed successfully by the target actor?
Each one of these have different challenges and costs, and it is obvious that there are conditions under which
any message passing library would be unable to comply; think for example about configurable mailbox types
and how a bounded mailbox would interact with the third point, or even what it would mean to decide upon the
successfully part of point five.
Along those same lines goes the reasoning in Nobody Needs Reliable Messaging. The only meaningful way for a
sender to know whether an interaction was successful is by receiving a business-level acknowledgement message,
which is not something Akka could make up on its own (neither are we writing a do what I mean framework
nor would you want us to).
Akka embraces distributed computing and makes the fallibility of communication explicit through message passing, therefore it does not try to lie and emulate a leaky abstraction. This is a model that has been used with great
success in Erlang and requires the users to design their applications around it. You can read more about this
approach in the Erlang documentation (section 10.9 and 10.10), Akka follows it closely.
Another angle on this issue is that by providing only basic guarantees those use cases which do not need stronger
reliability do not pay the cost of their implementation; it is always possible to add stronger reliability on top of
basic ones, but it is not possible to retro-actively remove reliability in order to gain more performance.
Discussion: Message Ordering
The rule more specifically is that for a given pair of actors, messages sent from the first to the second will not be
received out-of-order. This is illustrated in the following:
Actor A1 sends messages M1, M2, M3 to A2
Actor A3 sends messages M4, M5, M6 to A2
This means that:
1. If M1 is delivered it must be delivered before M2 and M3
2. If M2 is delivered it must be delivered before M3
3. If M4 is delivered it must be delivered before M5 and M6
4. If M5 is delivered it must be delivered before M6
28
Communication of failure
Please note, that the ordering guarantees discussed above only hold for user messages between actors. Failure
of a child of an actor is communicated by special system messages that are not ordered relative to ordinary user
messages. In particular:
Child actor C sends message M to its parent P
Child actor fails with failure F
Parent actor P might receive the two events either in order M, F or F, M
The reason for this is that internal system messages has their own mailboxes therefore the ordering of enqueue
calls of a user and system message cannot guarantee the ordering of their dequeue times.
29
StackOverflowError
OutOfMemoryError
other VirtualMachineError
In addition, local sends can fail in Akka-specific ways:
if the mailbox does not accept the message (e.g. full BoundedMailbox)
if the receiving actor fails while processing the message or is already terminated
While the first is clearly a matter of configuration the second deserves some thought: the sender of a message does
not get feedback if there was an exception while processing, that notification goes to the supervisor instead. This
is in general not distinguishable from a lost message for an outside observer.
Ordering of Local Message Sends
Assuming strict FIFO mailboxes the abovementioned caveat of non-transitivity of the message ordering guarantee
is eliminated under certain conditions. As you will note, these are quite subtle as it stands, and it is even possible
that future performance optimizations will invalidate this whole paragraph. The possibly non-exhaustive list of
counter-indications is:
Before receiving the first reply from a top-level actor, there is a lock which protects an internal interim
queue, and this lock is not fair; the implication is that enqueue requests from different senders which arrive
during the actors construction (figuratively, the details are more involved) may be reordered depending on
low-level thread scheduling. Since completely fair locks do not exist on the JVM this is unfixable.
The same mechanism is used during the construction of a Router, more precisely the routed ActorRef, hence
the same problem exists for actors deployed with Routers.
As mentioned above, the problem occurs anywhere a lock is involved during enqueueing, which may also
apply to custom mailboxes.
This list has been compiled carefully, but other problematic scenarios may have escaped our analysis.
How does Local Ordering relate to Network Ordering
As explained in the previous paragraph local message sends obey transitive causal ordering under certain conditions. If the remote message transport would respect this ordering as well, that would translate to transitive causal
ordering across one network link, i.e. if exactly two network hosts are involved. Involving multiple links, e.g. the
three actors on three different nodes mentioned above, then no guarantees can be made.
The current remote transport does not support this (again this is caused by non-FIFO wake-up order of a lock, this
time serializing connection establishment).
As a speculative view into the future it might be possible to support this ordering guarantee by re-implementing
the remote transport layer based completely on actors; at the same time we are looking into providing other lowlevel transport protocols like UDP or SCTP which would enable higher throughput or lower latency by removing
this guarantee again, which would mean that choosing between different implementations would allow trading
guarantees versus performance.
30
31
2.9 Configuration
You can start using Akka without defining any configuration, since sensible default values are provided. Later on
you might need to amend the settings to change the default behavior or adapt for specific runtime environments.
Typical examples of settings that you might amend:
log level and logger backend
enable remoting
message serializers
definition of routers
tuning of dispatchers
Akka uses the Typesafe Config Library, which might also be a good choice for the configuration of your own application or library built with or without Akka. This library is implemented in Java with no external dependencies;
you should have a look at its documentation (in particular about ConfigFactory), which is only summarized in the
following.
Warning: If you use Akka from the Scala REPL from the 2.9.x series, and you do not provide your own
ClassLoader to the ActorSystem, start the REPL with -Yrepl-sync to work around a deficiency in the REPLs
provided Context ClassLoader.
2.9. Configuration
32
The philosophy is that code never contains default values, but instead relies upon their presence in the
reference.conf supplied with the library in question.
Highest precedence is given to overrides given as system properties, see the HOCON specification (near the
bottom). Also noteworthy is that the application configurationwhich defaults to applicationmay be
overridden using the config.resource property (there are more, please refer to the Config docs).
Note: If you are writing an Akka application, keep you configuration in application.conf at the root of
the class path. If you are writing an Akka-based library, keep its configuration in reference.conf at the root
of the JAR file.
2.9. Configuration
33
More advanced include and substitution mechanisms are explained in the HOCON specification.
2.9. Configuration
34
used.
If in doubt, you can also easily and nicely inspect configuration objects before or after using them to construct an
actor system:
Welcome to Scala version 2.10.3 (Java HotSpot(TM) 64-Bit Server VM, Java 1.6.0_27).
Type in expressions to have them evaluated.
Type :help for more information.
scala> import com.typesafe.config._
import com.typesafe.config._
scala> ConfigFactory.parseString("a.b=12")
res0: com.typesafe.config.Config = Config(SimpleConfigObject({"a" : {"b" : 12}}))
scala> res0.root.render
res1: java.lang.String =
{
# String: 1
"a" : {
# String: 1
"b" : 12
}
}
The comments preceding every item give detailed information about the origin of the setting (file & line number)
plus possible comments which were present, e.g. in the reference configuration. The settings as merged with the
reference and parsed by the actor system can be displayed like this:
final ActorSystem system = ActorSystem.create();
System.out.println(system.settings());
// this is a shortcut for system.settings().config().root().render()
2.9. Configuration
35
myapp1 {
akka.loglevel = "WARNING"
my.own.setting = 43
}
myapp2 {
akka.loglevel = "ERROR"
app2.setting = "appname"
}
my.own.setting = 42
my.other.setting = "hello"
val config = ConfigFactory.load()
val app1 = ActorSystem("MyApp1", config.getConfig("myapp1").withFallback(config))
val app2 = ActorSystem("MyApp2",
config.getConfig("myapp2").withOnlyPath("akka").withFallback(config))
These two samples demonstrate different variations of the lift-a-subtree trick: in the first case, the configuration
accessible from within the actor system is this
akka.loglevel = "WARNING"
my.own.setting = 43
my.other.setting = "hello"
// plus myapp1 and myapp2 subtrees
while in the second one, only the akka subtree is lifted, with the following result
akka.loglevel = "ERROR"
my.own.setting = 42
my.other.setting = "hello"
// plus myapp1 and myapp2 subtrees
Note: The configuration library is really powerful, explaining all features exceeds the scope affordable here.
In particular not covered are how to include other configuration files within other files (see a small example at
Including files) and copying parts of the configuration tree by way of path substitutions.
You may also specify and parse the configuration programmatically in other ways when instantiating the
ActorSystem.
import akka.actor.ActorSystem
import com.typesafe.config.ConfigFactory
val customConf = ConfigFactory.parseString("""
akka.actor.deployment {
/my-service {
router = round-robin-pool
nr-of-instances = 3
}
}
""")
// ConfigFactory.load sandwiches customConfig between default reference
// config and default overrides, and then resolves it.
val system = ActorSystem("MySystem", ConfigFactory.load(customConf))
2.9. Configuration
36
From inside your replacement file specified with -Dconfig.resource and friends, you can include
"application" if you still want to use application.{conf,json,properties} as well. Settings
specified before include "application" would be overridden by the included file, while those after would
override the included file.
In code, there are many customization options.
There are several overloads of ConfigFactory.load(); these allow you to specify something to be sandwiched between system properties (which override) and the defaults (from reference.conf), replacing the
usual application.{conf,json,properties} and replacing -Dconfig.file and friends.
The simplest variant of ConfigFactory.load() takes a resource basename (instead of application);
myname.conf, myname.json, and myname.properties would then be used instead of
application.{conf,json,properties}.
The most flexible variant takes a Config object, which you can load using any method in ConfigFactory.
For example you could put a config string in code using ConfigFactory.parseString() or you could
make a map and ConfigFactory.parseMap(), or you could load a file.
You can also combine your custom config with the usual config, that might look like:
// make a Config with just your special setting
Config myConfig =
ConfigFactory.parseString("something=somethingElse");
// load the normal config stack (system props,
// then application.conf, then reference.conf)
Config regularConfig =
ConfigFactory.load();
// override regular stack with myConfig
Config combined =
myConfig.withFallback(regularConfig);
// put the result in between the overrides
// (system props) and defaults again
Config complete =
ConfigFactory.load(combined);
// create ActorSystem
ActorSystem system =
ActorSystem.create("myname", complete);
When working with Config objects, keep in mind that there are three layers in the cake:
ConfigFactory.defaultOverrides() (system properties)
the apps settings
ConfigFactory.defaultReference() (reference.conf)
The normal goal is to customize the middle layer while leaving the other two alone.
ConfigFactory.load() loads the whole stack
the overloads of ConfigFactory.load() let you specify a different middle layer
the ConfigFactory.parse() variations load single files or resources
To stack two layers, use override.withFallback(fallback); try to keep system props
(defaultOverrides()) on top and reference.conf (defaultReference()) on the bottom.
Do keep in mind, you can often just add another include statement in application.conf rather than writing code. Includes at the top of application.conf will be overridden by the rest of application.conf,
while those at the bottom will override the earlier stuff.
2.9. Configuration
37
remote deployment. Configuration of these features are described in the chapters detailing corresponding topics.
An example may look like this:
akka.actor.deployment {
# /user/actorA/actorB is a remote deployed actor
/actorA/actorB {
remote = "akka.tcp://[email protected]:2553"
}
# all direct children of /user/actorC have a dedicated dispatcher
"/actorC/*" {
dispatcher = my-dispatcher
}
# /user/actorD/actorE has a special priority mailbox
/actorD/actorE {
mailbox = prio-mailbox
}
# /user/actorF/actorG/actorH is a random pool
/actorF/actorG/actorH {
router = random-pool
nr-of-instances = 5
}
}
my-dispatcher {
fork-join-executor.parallelism-min = 10
fork-join-executor.parallelism-max = 10
}
prio-mailbox {
mailbox-type = "a.b.MyPrioMailbox"
}
The deployment section for a specific actor is identified by the path of the actor relative to /user.
You can use asterisks as wildcard matches for the actor path sections, so you could specify: /*/sampleActor
and that would match all sampleActor on that level in the hierarchy. You can also use wildcard in the last
position to match all actors at a certain level: /someParent/*. Non-wildcard matches always have higher
priority to match than wildcards, so: /foo/bar is considered more specific than /foo/* and only the highest
priority match is used. Please note that it cannot be used to partially match section, like this: /foo*/bar,
/f*o/bar etc.
2.9. Configuration
38
2.9. Configuration
39
# Frequency with which stopping actors are prodded in case they had to be
# removed from their parents
reaper-interval = 5s
# Serializes and deserializes (non-primitive) messages to ensure immutability,
# this is only intended for testing.
serialize-messages = off
# Serializes and deserializes creators (in Props) to ensure that they can be
# sent over the network, this is only intended for testing. Purely local deployments
# as marked with deploy.scope == LocalScope are exempt from verification.
serialize-creators = off
# Timeout for send operations to top-level actors which are in the process
# of being started. This is only relevant if using a bounded mailbox or the
# CallingThreadDispatcher for a top-level actor.
unstarted-push-timeout = 10s
typed {
# Default timeout for typed actor methods with non-void return type
timeout = 5s
}
# Mapping between deployment.router short names to fully qualified class names
router.type-mapping {
from-code = "akka.routing.NoRouter"
round-robin-pool = "akka.routing.RoundRobinPool"
round-robin-group = "akka.routing.RoundRobinGroup"
random-pool = "akka.routing.RandomPool"
random-group = "akka.routing.RandomGroup"
balancing-pool = "akka.routing.BalancingPool"
smallest-mailbox-pool = "akka.routing.SmallestMailboxPool"
broadcast-pool = "akka.routing.BroadcastPool"
broadcast-group = "akka.routing.BroadcastGroup"
scatter-gather-pool = "akka.routing.ScatterGatherFirstCompletedPool"
scatter-gather-group = "akka.routing.ScatterGatherFirstCompletedGroup"
consistent-hashing-pool = "akka.routing.ConsistentHashingPool"
consistent-hashing-group = "akka.routing.ConsistentHashingGroup"
}
deployment {
# deployment id pattern - on the format: /parent/child etc.
default {
# The id of the dispatcher to use for this actor.
# If undefined or empty the dispatcher specified in code
# (Props.withDispatcher) is used, or default-dispatcher if not
# specified at all.
dispatcher = ""
# The id of the mailbox to use for this actor.
# If undefined or empty the default mailbox of the configured dispatcher
# is used or if there is no mailbox configuration the mailbox specified
# in code (Props.withMailbox) is used.
# If there is a mailbox defined in the configured dispatcher then that
# overrides this setting.
mailbox = ""
# routing (load-balance) scheme to use
# - available: "from-code", "round-robin", "random", "smallest-mailbox",
#
"scatter-gather", "broadcast"
2.9. Configuration
40
# - or:
Fully qualified class name of the router class.
#
The class must extend akka.routing.CustomRouterConfig and
#
have a public constructor with com.typesafe.config.Config
#
and optional akka.actor.DynamicAccess parameter.
# - default is "from-code";
# Whether or not an actor is transformed to a Router is decided in code
# only (Props.withRouter). The type of router can be overridden in the
# configuration; specifying "from-code" means that the values specified
# in the code shall be used.
# In case of routing, the actors to be routed to can be specified
# in several ways:
# - nr-of-instances: will create that many children
# - routees.paths: will route messages to these paths using ActorSelection,
#
i.e. will not create children
# - resizer: dynamically resizable number of routees as specified in
#
resizer below
router = "from-code"
# number of children to create in case of a router;
# this setting is ignored if routees.paths is given
nr-of-instances = 1
# within is the timeout used for routers containing future calls
within = 5 seconds
# number of virtual nodes per node for consistent-hashing router
virtual-nodes-factor = 10
routees {
# Alternatively to giving nr-of-instances you can specify the full
# paths of those actors which should be routed to. This setting takes
# precedence over nr-of-instances
paths = []
}
#
#
#
#
#
#
#
#
To use a dedicated dispatcher for the routees of the pool you can
define the dispatcher configuration inline with the property name
pool-dispatcher in the deployment section of the router.
For example:
pool-dispatcher {
fork-join-executor.parallelism-min = 5
fork-join-executor.parallelism-max = 5
}
2.9. Configuration
41
#
messages in their mailbox. Note that estimating mailbox size of
#
default UnboundedMailbox is O(N) operation.
pressure-threshold = 1
# Percentage to increase capacity whenever all routees are busy.
# For example, 0.2 would increase 20% (rounded up), i.e. if current
# capacity is 6 it will request an increase of 2 more routees.
rampup-rate = 0.2
# Minimum fraction of busy routees before backing off.
# For example, if this is 0.3, then well remove some routees only when
# less than 30% of routees are busy, i.e. if current capacity is 10 and
# 3 are busy then the capacity is unchanged, but if 2 or less are busy
# the capacity is decreased.
# Use 0.0 or negative to avoid removal of routees.
backoff-threshold = 0.3
# Fraction of routees to be removed when the resizer reaches the
# backoffThreshold.
# For example, 0.1 would decrease 10% (rounded up), i.e. if current
# capacity is 9 it will request an decrease of 1 routee.
backoff-rate = 0.1
# Number of messages between resize operation.
# Use 1 to resize before each message.
messages-per-resize = 10
}
}
}
default-dispatcher {
# Must be one of the following
# Dispatcher, PinnedDispatcher, or a FQCN to a class inheriting
# MessageDispatcherConfigurator with a public constructor with
# both com.typesafe.config.Config parameter and
# akka.dispatch.DispatcherPrerequisites parameters.
# PinnedDispatcher must be used together with executor=thread-pool-executor.
type = "Dispatcher"
# Which kind of ExecutorService to use for this dispatcher
# Valid options:
# - "default-executor" requires a "default-executor" section
# - "fork-join-executor" requires a "fork-join-executor" section
# - "thread-pool-executor" requires a "thread-pool-executor" section
# - A FQCN of a class extending ExecutorServiceConfigurator
executor = "default-executor"
# This will be used if you have set "executor = "default-executor"".
# If an ActorSystem is created with a given ExecutionContext, this
# ExecutionContext will be used as the default executor for all
# dispatchers in the ActorSystem configured with
# executor = "default-executor". Note that "default-executor"
# is the default value for executor, and therefore used if not
# specified otherwise. If no ExecutionContext is given,
# the executor configured in "fallback" will be used.
default-executor {
fallback = "fork-join-executor"
}
# This will be used if you have set "executor = "fork-join-executor""
fork-join-executor {
# Min number of threads to cap factor-based parallelism number to
parallelism-min = 8
2.9. Configuration
42
# The parallelism factor is used to determine thread pool size using the
# following formula: ceil(available processors * factor). Resulting size
# is then bounded by the parallelism-min and parallelism-max values.
parallelism-factor = 3.0
# Max number of threads to cap factor-based parallelism number to
parallelism-max = 64
}
# This will be used if you have set "executor = "thread-pool-executor""
thread-pool-executor {
# Keep alive time for threads
keep-alive-time = 60s
# Min number of threads to cap factor-based core number to
core-pool-size-min = 8
# The core pool size factor is used to determine thread pool core size
# using the following formula: ceil(available processors * factor).
# Resulting size is then bounded by the core-pool-size-min and
# core-pool-size-max values.
core-pool-size-factor = 3.0
# Max number of threads to cap factor-based number to
core-pool-size-max = 64
# Minimum number of threads to cap factor-based max number to
# (if using a bounded task queue)
max-pool-size-min = 8
# Max no of threads (if using a bounded task queue) is determined by
# calculating: ceil(available processors * factor)
max-pool-size-factor = 3.0
# Max number of threads to cap factor-based max number to
# (if using a bounded task queue)
max-pool-size-max = 64
# Specifies the bounded capacity of the task queue (< 1 == unbounded)
task-queue-size = -1
# Specifies which type of task queue will be used, can be "array" or
# "linked" (default)
task-queue-type = "linked"
# Allow core threads to time out
allow-core-timeout = on
}
# How long time the dispatcher will wait for new actors until it shuts down
shutdown-timeout = 1s
# Throughput defines the number of messages that are processed in a batch
# before the thread is returned to the pool. Set to 1 for as fair as possible.
throughput = 5
# Throughput deadline for Dispatcher, set to 0 or negative for no deadline
throughput-deadline-time = 0ms
# For BalancingDispatcher: If the balancing dispatcher should attempt to
# schedule idle actors using the same dispatcher when a message comes in,
# and the dispatchers ExecutorService is not fully busy already.
2.9. Configuration
43
attempt-teamwork = on
# If this dispatcher requires a specific type of mailbox, specify the
# fully-qualified class name here; the actually created mailbox will
# be a subtype of this type. The empty string signifies no requirement.
mailbox-requirement = ""
}
default-mailbox {
# FQCN of the MailboxType. The Class of the FQCN must have a public
# constructor with
# (akka.actor.ActorSystem.Settings, com.typesafe.config.Config) parameters.
mailbox-type = "akka.dispatch.UnboundedMailbox"
# If the mailbox is bounded then it uses this setting to determine its
# capacity. The provided value must be positive.
# NOTICE:
# Up to version 2.1 the mailbox type was determined based on this setting;
# this is no longer the case, the type must explicitly be a bounded mailbox.
mailbox-capacity = 1000
# If the mailbox is bounded then this is the timeout for enqueueing
# in case the mailbox is full. Negative values signify infinite
# timeout, which should be avoided as it bears the risk of dead-lock.
mailbox-push-timeout-time = 10s
# For Actor with Stash: The default capacity of the stash.
# If negative (or zero) then an unbounded stash is used (default)
# If positive then a bounded stash is used and the capacity is set using
# the property
stash-capacity = -1
}
mailbox {
# Mapping between message queue semantics and mailbox configurations.
# Used by akka.dispatch.RequiresMessageQueue[T] to enforce different
# mailbox types on actors.
# If your Actor implements RequiresMessageQueue[T], then when you create
# an instance of that actor its mailbox type will be decided by looking
# up a mailbox configuration via T in this mapping
requirements {
"akka.dispatch.UnboundedMessageQueueSemantics" =
akka.actor.mailbox.unbounded-queue-based
"akka.dispatch.BoundedMessageQueueSemantics" =
akka.actor.mailbox.bounded-queue-based
"akka.dispatch.DequeBasedMessageQueueSemantics" =
akka.actor.mailbox.unbounded-deque-based
"akka.dispatch.UnboundedDequeBasedMessageQueueSemantics" =
akka.actor.mailbox.unbounded-deque-based
"akka.dispatch.BoundedDequeBasedMessageQueueSemantics" =
akka.actor.mailbox.bounded-deque-based
"akka.dispatch.MultipleConsumerSemantics" =
akka.actor.mailbox.unbounded-queue-based
}
unbounded-queue-based {
# FQCN of the MailboxType, The Class of the FQCN must have a public
# constructor with (akka.actor.ActorSystem.Settings,
# com.typesafe.config.Config) parameters.
mailbox-type = "akka.dispatch.UnboundedMailbox"
}
bounded-queue-based {
2.9. Configuration
44
# FQCN of the MailboxType, The Class of the FQCN must have a public
# constructor with (akka.actor.ActorSystem.Settings,
# com.typesafe.config.Config) parameters.
mailbox-type = "akka.dispatch.BoundedMailbox"
}
unbounded-deque-based {
# FQCN of the MailboxType, The Class of the FQCN must have a public
# constructor with (akka.actor.ActorSystem.Settings,
# com.typesafe.config.Config) parameters.
mailbox-type = "akka.dispatch.UnboundedDequeBasedMailbox"
}
bounded-deque-based {
# FQCN of the MailboxType, The Class of the FQCN must have a public
# constructor with (akka.actor.ActorSystem.Settings,
# com.typesafe.config.Config) parameters.
mailbox-type = "akka.dispatch.BoundedDequeBasedMailbox"
}
}
debug {
# enable function of Actor.loggable(), which is to log any received message
# at DEBUG level, see the Testing Actor Systems section of the Akka
# Documentation at http://akka.io/docs
receive = off
# enable DEBUG logging of all AutoReceiveMessages (Kill, PoisonPill et.c.)
autoreceive = off
# enable DEBUG logging of actor lifecycle changes
lifecycle = off
# enable DEBUG logging of all LoggingFSMs for events, transitions and timers
fsm = off
# enable DEBUG logging of subscription changes on the eventStream
event-stream = off
# enable DEBUG logging of unhandled messages
unhandled = off
# enable WARN logging of misconfigured routers
router-misconfiguration = off
}
# Entries for pluggable serializers and their bindings.
serializers {
java = "akka.serialization.JavaSerializer"
bytes = "akka.serialization.ByteArraySerializer"
}
# Class to Serializer binding. You only need to specify the name of an
# interface or abstract base class of the messages. In case of ambiguity it
# is using the most specific configured class, or giving a warning and
# choosing the first one.
#
# To disable one of the default serializers, assign its class to "none", like
# "java.io.Serializable" = none
serialization-bindings {
"[B" = bytes
"java.io.Serializable" = java
}
2.9. Configuration
45
2.9. Configuration
46
2.9. Configuration
47
# may block for a long time when network IO is faster than file IO.
# Decreasing the value may improve fairness while increasing may improve
# throughput.
file-io-transferTo-limit = 512 KiB
# The number of times to retry the finishConnect call after being notified about
# OP_CONNECT. Retries are needed if the OP_CONNECT notification doesnt imply that
# finishConnect will succeed, which is the case on Android.
finish-connect-retries = 5
}
udp {
# The number of selectors to stripe the served channels over; each of
# these will use one select loop on the selector-dispatcher.
nr-of-selectors = 1
# Maximum number of open channels supported by this UDP module Generally
# UDP does not require a large number of channels, therefore it is
# recommended to keep this setting low.
max-channels = 4096
# The select loop can be used in two modes:
# - setting "infinite" will select without a timeout, hogging a thread
# - setting a positive timeout will do a bounded select call,
#
enabling sharing of a single thread between multiple selectors
#
(in this case you will have to use a different configuration for the
#
selector-dispatcher, e.g. using "type=Dispatcher" with size 1)
# - setting it to zero means polling, i.e. calling selectNow()
select-timeout = infinite
# When trying to assign a new connection to a selector and the chosen
# selector is at full capacity, retry selector choosing and assignment
# this many times before giving up
selector-association-retries = 10
# The maximum number of datagrams that are read in one go,
# higher numbers decrease latency, lower numbers increase fairness on
# the worker-dispatcher
receive-throughput = 3
# The number of bytes per direct buffer in the pool used to read or write
# network data from the kernel.
direct-buffer-size = 128 KiB
# The maximal number of direct buffers kept in the direct buffer pool for
# reuse.
direct-buffer-pool-limit = 1000
# The maximum number of bytes delivered by a Received message. Before
# more data is read from the network the connection actor will try to
# do other work.
received-message-size-limit = unlimited
# Enable fine grained logging of what goes on inside the implementation.
# Be aware that this may log more than once per message sent to the actors
# of the tcp implementation.
trace-logging = off
# Fully qualified config path which holds the dispatcher configuration
# to be used for running the select() calls in the selectors
selector-dispatcher = "akka.io.pinned-dispatcher"
2.9. Configuration
48
2.9. Configuration
49
worker-dispatcher = "akka.actor.default-dispatcher"
# Fully qualified config path which holds the dispatcher configuration
# for the selector management actors
management-dispatcher = "akka.actor.default-dispatcher"
}
}
akka-agent
####################################
# Akka Agent Reference Config File #
####################################
# This is the reference config file that contains all the default settings.
# Make your edits/overrides in your application.conf.
akka {
agent {
# The dispatcher used for agent-send-off actor
send-off-dispatcher {
executor = thread-pool-executor
type = PinnedDispatcher
}
# The dispatcher used for agent-alter-off actor
alter-off-dispatcher {
executor = thread-pool-executor
type = PinnedDispatcher
}
}
}
akka-camel
####################################
# Akka Camel Reference Config File #
####################################
# This is the reference config file that contains all the default settings.
# Make your edits/overrides in your application.conf.
akka {
camel {
# FQCN of the ContextProvider to be used to create or locate a CamelContext
# it must implement akka.camel.ContextProvider and have a no-arg constructor
# the built-in default create a fresh DefaultCamelContext
context-provider = akka.camel.DefaultContextProvider
# Whether JMX should be enabled or disabled for the Camel Context
jmx = off
# enable/disable streaming cache on the Camel Context
streamingCache = on
consumer {
# Configured setting which determines whether one-way communications
2.9. Configuration
50
akka-cluster
######################################
# Akka Cluster Reference Config File #
######################################
# This is the reference config file that contains all the default settings.
# Make your edits/overrides in your application.conf.
akka {
cluster {
# Initial contact points of the cluster.
# The nodes to join automatically at startup.
# Comma separated full URIs defined by a string on the form of
# "akka.tcp://system@hostname:port"
# Leave as empty if the node is supposed to be joined manually.
seed-nodes = []
# how long to wait for one of the seed nodes to reply to initial join request
seed-node-timeout = 5s
# If a join request fails it will be retried after this period.
# Disable join retry by specifying "off".
retry-unsuccessful-join-after = 10s
# Should the leader in the cluster be allowed to automatically mark
# unreachable nodes as DOWN after a configured time of unreachability?
# Using auto-down implies that two separate clusters will automatically be
# formed in case of network partition.
# Disable with "off" or specify a duration to enable auto-down.
auto-down-unreachable-after = off
# deprecated in 2.3, use auto-down-unreachable-after instead
auto-down = off
# The roles of this member. List of strings, e.g. roles = ["A", "B"].
# The roles are part of the membership information and can be used by
# routers or other services to distribute work to certain member types,
2.9. Configuration
51
2.9. Configuration
52
2.9. Configuration
53
2.9. Configuration
54
serializers {
akka-cluster = "akka.cluster.protobuf.ClusterMessageSerializer"
}
serialization-bindings {
"akka.cluster.ClusterMessage" = akka-cluster
}
router.type-mapping {
adaptive-pool = "akka.cluster.routing.AdaptiveLoadBalancingPool"
adaptive-group = "akka.cluster.routing.AdaptiveLoadBalancingGroup"
}
}
}
akka-multi-node-testkit
#############################################
# Akka Remote Testing Reference Config File #
#############################################
# This is the reference config file that contains all the default settings.
# Make your edits/overrides in your application.conf.
akka {
testconductor {
# Timeout for joining a barrier: this is the maximum time any participants
# waits for everybody else to join a named barrier.
barrier-timeout = 30s
# Timeout for interrogation of TestConductors Controller actor
query-timeout = 5s
# Threshold for packet size in time unit above which the failure injector will
# split the packet and deliver in smaller portions; do not give value smaller
# than HashedWheelTimer resolution (would not make sense)
packet-split-threshold = 100ms
# amount of time for the ClientFSM to wait for the connection to the conductor
# to be successful
connect-timeout = 20s
# Number of connect attempts to be made to the conductor controller
client-reconnects = 10
# minimum time interval which is to be inserted between reconnect attempts
reconnect-backoff = 1s
netty {
# (I&O) Used to configure the number of I/O worker threads on server sockets
server-socket-worker-pool {
# Min number of threads to cap factor-based number to
pool-size-min = 1
# The pool size factor is used to determine thread pool size
# using the following formula: ceil(available processors * factor).
# Resulting size is then bounded by the pool-size-min and
# pool-size-max values.
pool-size-factor = 1.0
2.9. Configuration
55
akka-persistence
##########################################
# Akka Persistence Reference Config File #
##########################################
akka {
# Protobuf serialization for persistent messages
actor {
serializers {
akka-persistence-snapshot = "akka.persistence.serialization.SnapshotSerializer"
akka-persistence-message = "akka.persistence.serialization.MessageSerializer"
}
serialization-bindings {
"akka.persistence.serialization.Snapshot" = akka-persistence-snapshot
"akka.persistence.serialization.Message" = akka-persistence-message
}
}
persistence {
journal {
# Maximum size of a persistent message batch written to the journal.
# Only applies to internally created batches by processors that receive
# persistent messages individually. Application-defined batches, even if
# larger than this setting, are always written as a single isolated batch.
max-message-batch-size = 200
# Maximum size of a confirmation batch written to the journal.
max-confirmation-batch-size = 10000
2.9. Configuration
56
2.9. Configuration
57
2.9. Configuration
58
executor = "fork-join-executor"
fork-join-executor {
parallelism-min = 2
parallelism-max = 8
}
}
}
}
}
akka-remote
#####################################
# Akka Remote Reference Config File #
#####################################
# This is the reference config file that contains all the default settings.
# Make your edits/overrides in your application.conf.
# comments about akka.actor settings left out where they are already in akka# actor.jar, because otherwise they would be repeated in config rendering.
akka {
actor {
serializers {
akka-containers = "akka.remote.serialization.MessageContainerSerializer"
proto = "akka.remote.serialization.ProtobufSerializer"
daemon-create = "akka.remote.serialization.DaemonMsgCreateSerializer"
}
serialization-bindings {
# Since com.google.protobuf.Message does not extend Serializable but
# GeneratedMessage does, need to use the more specific one here in order
# to avoid ambiguity
"akka.actor.ActorSelectionMessage" = akka-containers
"com.google.protobuf.GeneratedMessage" = proto
"akka.remote.DaemonMsgCreate" = daemon-create
}
deployment {
default {
# if this is set to a valid remote address, the named actor will be
# deployed at that node e.g. "akka.tcp://sys@host:port"
remote = ""
target {
#
#
#
#
#
#
#
#
#
2.9. Configuration
59
2.9. Configuration
60
2.9. Configuration
61
2.9. Configuration
62
2.9. Configuration
63
trttl = "akka.remote.transport.ThrottlerProvider"
}
### Default configuration for the Netty based transport drivers
netty.tcp {
# The class given here must implement the akka.remote.transport.Transport
# interface and offer a public constructor which takes two arguments:
# 1) akka.actor.ExtendedActorSystem
# 2) com.typesafe.config.Config
transport-class = "akka.remote.transport.netty.NettyTransport"
# Transport drivers can be augmented with adapters by adding their
# name to the applied-adapters list. The last adapter in the
# list is the adapter immediately above the driver, while
# the first one is the top of the stack below the standard
# Akka protocol
applied-adapters = []
transport-protocol = tcp
# The default remote server port clients should connect to.
# Default is 2552 (AKKA), use 0 if you want a random available port
# This port needs to be unique for each actor system on the same machine.
port = 2552
# The hostname or ip to bind the remoting to,
# InetAddress.getLocalHost.getHostAddress is used if empty
hostname = ""
# Enables SSL support on this transport
enable-ssl = false
# Sets the connectTimeoutMillis of all outbound connections,
# i.e. how long a connect may take until it is timed out
connection-timeout = 15 s
# If set to "<id.of.dispatcher>" then the specified dispatcher
# will be used to accept inbound connections, and perform IO. If "" then
# dedicated threads will be used.
# Please note that the Netty driver only uses this configuration and does
# not read the "akka.remote.use-dispatcher" entry. Instead it has to be
# configured manually to point to the same dispatcher if needed.
use-dispatcher-for-io = ""
# Sets the high water mark for the in and outbound sockets,
# set to 0b for platform default
write-buffer-high-water-mark = 0b
# Sets the low water mark for the in and outbound sockets,
# set to 0b for platform default
write-buffer-low-water-mark = 0b
# Sets the send buffer size of the Sockets,
# set to 0b for platform default
send-buffer-size = 256000b
# Sets the receive buffer size of the Sockets,
# set to 0b for platform default
receive-buffer-size = 256000b
# Maximum message size the transport will accept, but at least
# 32000 bytes.
2.9. Configuration
64
# Please note that UDP does not support arbitrary large datagrams,
# so this setting has to be chosen carefully when using UDP.
# Both send-buffer-size and receive-buffer-size settings has to
# be adjusted to be able to buffer messages of maximum size.
maximum-frame-size = 128000b
# Sets the size of the connection backlog
backlog = 4096
# Enables the TCP_NODELAY flag, i.e. disables Nagles algorithm
tcp-nodelay = on
# Enables TCP Keepalive, subject to the O/S kernels configuration
tcp-keepalive = on
}
netty.udp = ${akka.remote.netty.tcp}
netty.udp {
transport-protocol = udp
}
netty.ssl = ${akka.remote.netty.tcp}
netty.ssl = {
# Enable SSL/TLS encryption.
2.9. Configuration
65
2.9. Configuration
66
}
}
akka-testkit
######################################
# Akka Testkit Reference Config File #
######################################
# This is the reference config file that contains all the default settings.
# Make your edits/overrides in your application.conf.
akka {
test {
# factor by which to scale timeouts during tests, e.g. to account for shared
# build system load
timefactor = 1.0
# duration of EventFilter.intercept waits after the block is finished until
# all required messages are received
filter-leeway = 3s
# duration to wait in expectMsg and friends outside of within() block
# by default
single-expect-default = 3s
# The timeout that is added as an implicit by DefaultTimeout trait
default-timeout = 5s
calling-thread-dispatcher {
type = akka.testkit.CallingThreadDispatcherConfigurator
}
}
}
akka-zeromq
#####################################
# Akka ZeroMQ Reference Config File #
#####################################
# This is the reference config file that contains all the default settings.
# Make your edits/overrides in your application.conf.
akka {
zeromq {
# The default timeout for a poll on the actual zeromq socket.
poll-timeout = 100ms
# Timeout for creating a new socket
2.9. Configuration
67
new-socket-timeout = 5s
socket-dispatcher {
# A zeromq socket needs to be pinned to the thread that created it.
# Changing this value results in weird errors and race conditions within
# zeromq
executor = thread-pool-executor
type = "PinnedDispatcher"
thread-pool-executor.allow-core-timeout = off
}
}
}
2.9. Configuration
68
CHAPTER
THREE
ACTORS
3.1 Actors
The Actor Model provides a higher level of abstraction for writing concurrent and distributed systems. It alleviates
the developer from having to deal with explicit locking and thread management, making it easier to write correct
concurrent and parallel systems. Actors were defined in the 1973 paper by Carl Hewitt but have been popularized
by the Erlang language, and used for example at Ericsson with great success to build highly concurrent and reliable
telecom systems.
The API of Akkas Actors is similar to Scala Actors which has borrowed some of its syntax from Erlang.
69
Props
Props is a configuration class to specify options for the creation of actors, think of it as an immutable and thus
freely shareable recipe for creating an actor including associated deployment information (e.g. which dispatcher
to use, see more below). Here are some examples of how to create a Props instance.
import akka.actor.Props;
import akka.japi.Creator;
static class MyActorC implements Creator<MyActor> {
@Override public MyActor create() {
return new MyActor("...");
}
}
Props props1 = Props.create(MyUntypedActor.class);
Props props2 = Props.create(MyActor.class, "...");
Props props3 = Props.create(new MyActorC());
The second line shows how to pass constructor arguments to the Actor being created. The presence of a matching
constructor is verified during construction of the Props object, resulting in an IllegalArgumentEception
if no or multiple matching constructors are found.
The third line demonstrates the use of a Creator. The creator class must be static, which is verified during
Props construction. The type parameters upper bound is used to determine the produced actor class, falling
back to Actor if fully erased. An example of a parametric factory could be:
static class ParametricCreator<T extends MyActor> implements Creator<T> {
@Override public T create() {
// ... fabricate actor here
}
}
Note: In order for mailbox requirementslike using a deque-based mailbox for actors using the stashto be
picked up, the actor type needs to be known before creating it, which is what the Creator type argument allows.
Therefore make sure to use the specific type for your actors wherever possible.
Recommended Practices
It is a good idea to provide static factory methods on the UntypedActor which help keeping the creation of
suitable Props as close to the actor definition as possible. This also allows usage of the Creator-based methods
which statically verify that the used constructor actually exists instead relying only on a runtime check.
public class DemoActor extends UntypedActor {
/**
* Create Props for an actor of this type.
* @param magicNumber The magic number to be passed to this actors constructor.
* @return a Props for creating this actor, which can then be further configured
(e.g. calling .withDispatcher() on it)
*
*/
public static Props props(final int magicNumber) {
return Props.create(new Creator<DemoActor>() {
private static final long serialVersionUID = 1L;
@Override
public DemoActor create() throws Exception {
return new DemoActor(magicNumber);
}
});
3.1. Actors
70
}
final int magicNumber;
public DemoActor(int magicNumber) {
this.magicNumber = magicNumber;
}
@Override
public void onReceive(Object msg) {
// some behavior here
}
}
system.actorOf(DemoActor.props(42), "demo");
Using the ActorSystem will create top-level actors, supervised by the actor systems provided guardian actor,
while using an actors context will create a child actor.
class A extends UntypedActor {
final ActorRef child =
getContext().actorOf(Props.create(MyUntypedActor.class), "myChild");
// plus some behavior ...
}
It is recommended to create a hierarchy of children, grand-children and so on such that it fits the logical failurehandling structure of the application, see Actor Systems.
The call to actorOf returns an instance of ActorRef. This is a handle to the actor instance and the only way to
interact with it. The ActorRef is immutable and has a one to one relationship with the Actor it represents. The
ActorRef is also serializable and network-aware. This means that you can serialize it, send it over the wire and
use it on a remote host and it will still be representing the same Actor on the original node, across the network.
The name parameter is optional, but you should preferably name your actors, since that is used in log messages
and for identifying actors. The name must not be empty or start with $, but it may contain URL encoded characters (eg. %20 for a blank space). If the given name is already in use by another child to the same parent an
InvalidActorNameException is thrown.
Actors are automatically started asynchronously when created.
Dependency Injection
If your UntypedActor has a constructor that takes parameters then those need to be part of the Props as well, as
described above. But there are cases when a factory method must be used, for example when the actual constructor
arguments are determined by a dependency injection framework.
3.1. Actors
71
import akka.actor.Actor;
import akka.actor.IndirectActorProducer;
class DependencyInjector implements IndirectActorProducer {
final Object applicationContext;
final String beanName;
public DependencyInjector(Object applicationContext, String beanName) {
this.applicationContext = applicationContext;
this.beanName = beanName;
}
@Override
public Class<? extends Actor> actorClass() {
return MyActor.class;
}
@Override
public MyActor produce() {
MyActor result;
// obtain fresh Actor instance from DI framework ...
return result;
}
}
final ActorRef myActor = getContext().actorOf(
Props.create(DependencyInjector.class, applicationContext, "MyActor"),
"myactor3");
Warning: You might be tempted at times to offer an IndirectActorProducer which always returns
the same instance, e.g. by using a static field. This is not supported, as it goes against the meaning of an actor
restart, which is described here: What Restarting Means.
When using a dependency injection framework, actor beans MUST NOT have singleton scope.
Techniques for dependency injection and integration with dependency injection frameworks are described in more
depth in the Using Akka with Dependency Injection guideline and the Akka Java Spring tutorial in Typesafe
Activator.
The Inbox
When writing code outside of actors which shall communicate with actors, the ask pattern can be a solution (see
below), but there are two thing it cannot do: receiving multiple replies (e.g. by subscribing an ActorRef to a
notification service) and watching other actors lifecycle. For these purposes there is the Inbox class:
final Inbox inbox = Inbox.create(system);
inbox.send(target, "hello");
assert inbox.receive(Duration.create(1, TimeUnit.SECONDS)).equals("world");
The send method wraps a normal tell and supplies the internal actors reference as the sender. This allows the
reply to be received on the last line. Watching an actor is quite simple as well:
final Inbox inbox = Inbox.create(system);
inbox.watch(target);
target.tell(PoisonPill.getInstance(), ActorRef.noSender());
assert inbox.receive(Duration.create(1, TimeUnit.SECONDS)) instanceof Terminated;
3.1. Actors
72
The implementations shown above are the defaults provided by the UntypedActor class.
3.1. Actors
73
Actor Lifecycle
A path in an actor system represents a place which might be occupied by a living actor. Initially (apart from
system initialized actors) a path is empty. When actorOf() is called it assigns an incarnation of the actor
described by the passed Props to the given path. An actor incarnation is identified by the path and a UID. A
restart only swaps the Actor instance defined by the Props but the incarnation and hence the UID remains the
same.
The lifecycle of an incarnation ends when the actor is stopped. At that point the appropriate lifecycle events are
called and watching actors are notified of the termination. After the incarnation is stopped, the path can be reused
again by creating an actor with actorOf(). In this case the name of the new incarnation will be the same as the
previous one but the UIDs will differ.
An ActorRef always represents an incarnation (path and UID) not just a given path. Therefore if an actor is
stopped and a new one with the same name is created an ActorRef of the old incarnation will not point to the
new one.
ActorSelection on the other hand points to the path (or multiple paths if wildcards are used) and is completely
oblivious to which incarnation is currently occupying it. ActorSelection cannot be watched for this reason.
It is possible to resolve the current incarnations ActorRef living under the path by sending an Identify
message to the ActorSelection which will be replied to with an ActorIdentity containing the correct
reference (see Identifying Actors via Actor Selection). This can also be done with the resolveOne method of
the ActorSelection, which returns a Future of the matching ActorRef.
Lifecycle Monitoring aka DeathWatch
In order to be notified when another actor terminates (i.e. stops permanently, not temporary failure and restart), an
actor may register itself for reception of the Terminated message dispatched by the other actor upon termination
(see Stopping Actors). This service is provided by the DeathWatch component of the actor system.
3.1. Actors
74
Registering a monitor is easy (see fourth line, the rest is for demonstrating the whole functionality):
import akka.actor.Terminated;
public class WatchActor extends UntypedActor {
final ActorRef child = this.getContext().actorOf(Props.empty(), "child");
{
this.getContext().watch(child); // <-- the only call needed for registration
}
ActorRef lastSender = getContext().system().deadLetters();
@Override
public void onReceive(Object message) {
if (message.equals("kill")) {
getContext().stop(child);
lastSender = getSender();
} else if (message instanceof Terminated) {
final Terminated t = (Terminated) message;
if (t.getActor() == child) {
lastSender.tell("finished", getSelf());
}
} else {
unhandled(message);
}
}
}
It should be noted that the Terminated message is generated independent of the order in which registration and
termination occur. In particular, the watching actor will receive a Terminated message even if the watched
actor has already been terminated at the time of registration.
Registering multiple times does not necessarily lead to multiple messages being generated, but there is no guarantee that only exactly one such message is received: if termination of the watched actor has generated and queued
the message, and another registration is done before this message has been processed, then a second message will
be queued, because registering for monitoring of an already terminated actor leads to the immediate generation of
the Terminated message.
It
is
also
possible
to
deregister
from
watching
another
actors
liveliness
using
getContext().unwatch(target).
This works even if the Terminated message has already
been enqueued in the mailbox; after calling unwatch no Terminated message for that actor will be processed
anymore.
Start Hook
Right after starting the actor, its preStart method is invoked.
@Override
public void preStart() {
child = getContext().actorOf(Props.empty());
}
This method is called when the actor is first created. During restarts it is called by the default implementation of
postRestart, which means that by overriding that method you can choose whether the initialization code in
this method is called only exactly once for this actor or for every restart. Initialization code which is part of the
actors constructor will always be called when an instance of the actor class is created, which happens at every
restart.
Restart Hooks
All actors are supervised, i.e. linked to another actor with a fault handling strategy. Actors may be restarted in
case an exception is thrown while processing a message (see Supervision and Monitoring). This restart involves
3.1. Actors
75
Stop Hook
After stopping an actor, its postStop hook is called, which may be used e.g. for deregistering this actor from
other services. This hook is guaranteed to run after message queuing has been disabled for this actor, i.e. messages
sent to a stopped actor will be redirected to the deadLetters of the ActorSystem.
The supplied path is parsed as a java.net.URI, which basically means that it is split on / into path elements.
If the path starts with /, it is absolute and the look-up starts at the root guardian (which is the parent of "/user");
otherwise it starts at the current actor. If a path element equals .., the look-up will take a step up towards the
supervisor of the currently traversed actor, otherwise it will step down to the named child. It should be noted
that the .. in actor paths here always means the logical structure, i.e. the supervisor.
The path elements of an actor selection may contain wildcard patterns allowing for broadcasting of messages to
that section:
// will look all children to serviceB with names starting with worker
getContext().actorSelection("/user/serviceB/worker*");
// will look up all siblings beneath same supervisor
getContext().actorSelection("../*");
Messages can be sent via the ActorSelection and the path of the ActorSelection is looked up when
delivering each message. If the selection does not match any actors the message will be dropped.
3.1. Actors
76
To acquire an ActorRef for an ActorSelection you need to send a message to the selection and use the
getSender reference of the reply from the actor. There is a built-in Identify message that all Actors will
understand and automatically reply to with a ActorIdentity message containing the ActorRef. This message is handled specially by the actors which are traversed in the sense that if a concrete name lookup fails (i.e.
a non-wildcard path element does not correspond to a live actor) then a negative result is generated. Please note
that this does not mean that delivery of that reply is guaranteed, it still is a normal message.
import akka.actor.ActorIdentity;
import akka.actor.ActorSelection;
import akka.actor.Identify;
public class Follower extends UntypedActor {
final String identifyId = "1";
{
ActorSelection selection =
getContext().actorSelection("/user/another");
selection.tell(new Identify(identifyId), getSelf());
}
ActorRef another;
final ActorRef probe;
public Follower(ActorRef probe) {
this.probe = probe;
}
@Override
public void onReceive(Object message) {
if (message instanceof ActorIdentity) {
ActorIdentity identity = (ActorIdentity) message;
if (identity.correlationId().equals(identifyId)) {
ActorRef ref = identity.getRef();
if (ref == null)
getContext().stop(getSelf());
else {
another = ref;
getContext().watch(another);
probe.tell(ref, getSelf());
}
}
} else if (message instanceof Terminated) {
final Terminated t = (Terminated) message;
if (t.getActor().equals(another)) {
getContext().stop(getSelf());
}
} else {
unhandled(message);
}
}
}
You can also acquire an ActorRef for an ActorSelection with the resolveOne method of the
ActorSelection. It returns a Future of the matching ActorRef if such an actor exists. It is completed
with failure [[akka.actor.ActorNotFound]] if no such actor exists or the identification didnt complete within the
supplied timeout.
Remote actor addresses may also be looked up, if remoting is enabled:
getContext().actorSelection("akka.tcp://app@otherhost:1234/user/serviceB");
77
actor needs to exist before the lookup, or else the acquired reference will be an EmptyLocalActorRef. This
will be true even if an actor with that exact path is created after acquiring the actor reference. For remote actor
references acquired with actorFor the behaviour is different and sending messages to such a reference will under
the hood look up the actor by path on the remote system for every message send.
The sender reference is passed along with the message and available within the receiving actor via its getSender
method while processing this message. Inside of an actor it is usually getSelf who shall be the sender, but there
3.1. Actors
78
can be cases where replies shall be routed to some other actore.g. the parentin which the second argument to
tell would be a different one. Outside of an actor and if no reply is needed the second argument can be null;
if a reply is needed outside of an actor you can use the ask-pattern described next..
Ask: Send-And-Receive-Future
The ask pattern involves actors as well as futures, hence it is offered as a use pattern rather than a method on
ActorRef:
import
import
import
import
import
import
import
static akka.pattern.Patterns.ask;
static akka.pattern.Patterns.pipe;
scala.concurrent.Future;
scala.concurrent.duration.Duration;
akka.dispatch.Futures;
akka.dispatch.Mapper;
akka.util.Timeout;
This example demonstrates ask together with the pipe pattern on futures, because this is likely to be a common
combination. Please note that all of the above is completely non-blocking and asynchronous: ask produces a
Future, two of which are composed into a new future using the Futures.sequence and map methods and
then pipe installs an onComplete-handler on the future to effect the submission of the aggregated Result to
another actor.
Using ask will send a message to the receiving Actor as with tell, and the receiving actor must reply with
getSender().tell(reply, getSelf()) in order to complete the returned Future with a value. The
ask operation involves creating an internal actor for handling this reply, which needs to have a timeout after
which it is destroyed in order not to leak resources; see more below.
Warning: To complete the future with an exception you need send a Failure message to the sender. This is
not done automatically when an actor throws an exception while processing a message.
try {
String result = operation();
getSender().tell(result, getSelf());
} catch (Exception e) {
getSender().tell(new akka.actor.Status.Failure(e), getSelf());
throw e;
}
3.1. Actors
79
If the actor does not complete the future, it will expire after the timeout period, specified as parameter to the ask
method; this will complete the Future with an AskTimeoutException.
See Futures for more information on how to await or query a future.
The onComplete, onSuccess, or onFailure methods of the Future can be used to register a callback to
get a notification when the Future completes. Gives you a way to avoid blocking.
Warning: When using future callbacks, inside actors you need to carefully avoid closing over the containing
actors reference, i.e. do not call methods or access mutable state on the enclosing actor from within the callback. This would break the actor encapsulation and may introduce synchronization bugs and race conditions
because the callback will be scheduled concurrently to the enclosing actor. Unfortunately there is not yet a
way to detect these illegal accesses at compile time. See also: Actors and shared mutable state
Forward message
You can forward a message from one actor to another. This means that the original sender address/reference is
maintained even though the message is going through a mediator. This can be useful when writing actors that
work as routers, load-balancers, replicators etc. You need to pass along your context variable as well.
target.forward(result, getContext());
An alternative to using if-instanceof checks is to use Apache Commons MethodUtils to invoke a named method
whose parameter type matches the message type.
3.1. Actors
80
akka.actor.ActorRef;
akka.actor.ReceiveTimeout;
akka.actor.UntypedActor;
scala.concurrent.duration.Duration;
81
Upon ActorSystem.shutdown, the system guardian actors will be stopped, and the aforementioned process
will ensure proper termination of the whole system.
The postStop hook is invoked after an actor is fully stopped. This enables cleaning up of resources:
@Override
public void postStop() {
// clean up resources here ...
}
Note: Since stopping an actor is asynchronous, you cannot immediately reuse the name of the child you just
stopped; this will result in an InvalidActorNameException. Instead, watch the terminating actor and
create its replacement in response to the Terminated message which will eventually arrive.
PoisonPill
You can also send an actor the akka.actor.PoisonPill message, which will stop the actor when the message is processed. PoisonPill is enqueued as ordinary messages and will be handled after messages that were
already queued in the mailbox.
Use it like this:
myActor.tell(akka.actor.PoisonPill.getInstance(), sender);
Graceful Stop
gracefulStop is useful if you need to wait for termination or compose ordered termination of several actors:
import
import
import
import
import
static akka.pattern.Patterns.gracefulStop;
scala.concurrent.Await;
scala.concurrent.Future;
scala.concurrent.duration.Duration;
akka.pattern.AskTimeoutException;
try {
Future<Boolean> stopped =
gracefulStop(actorRef, Duration.create(5, TimeUnit.SECONDS), Manager.SHUTDOWN);
Await.result(stopped, Duration.create(6, TimeUnit.SECONDS));
// the actor has been stopped
} catch (AskTimeoutException e) {
// the actor wasnt stopped within 5 seconds
}
public class Manager extends UntypedActor {
public static final String SHUTDOWN = "shutdown";
ActorRef worker = getContext().watch(getContext().actorOf(
Props.create(Cruncher.class), "worker"));
public void onReceive(Object message) {
if (message.equals("job")) {
worker.tell("crunch", getSelf());
} else if (message.equals(SHUTDOWN)) {
worker.tell(PoisonPill.getInstance(), getSelf());
getContext().become(shuttingDown);
}
}
Procedure<Object> shuttingDown = new Procedure<Object>() {
3.1. Actors
82
@Override
public void apply(Object message) {
if (message.equals("job")) {
getSender().tell("service unavailable, shutting down", getSelf());
} else if (message instanceof Terminated) {
getContext().stop(getSelf());
}
}
};
}
When gracefulStop() returns successfully, the actors postStop() hook will have been executed: there
exists a happens-before edge between the end of postStop() and the return of gracefulStop().
In the above example a custom Manager.SHUTDOWN message is sent to the target actor to initiate the process
of stopping the actor. You can use PoisonPill for this, but then you have limited possibilities to perform
interactions with other actors before stopping the target actor. Simple cleanup tasks can be handled in postStop.
Warning: Keep in mind that an actor stopping and its name being deregistered are separate events which
happen asynchronously from each other. Therefore it may be that you will find the name still in use after
gracefulStop() returned. In order to guarantee proper deregistration, only reuse names from within a
supervisor you control and only in response to a Terminated message, i.e. not for top-level actors.
3.1.10 HotSwap
Upgrade
Akka supports hotswapping the Actors message loop (e.g. its implementation) at runtime. Use the
getContext().become method from within the Actor. The hotswapped code is kept in a Stack which can be
pushed (replacing or adding at the top) and popped.
Warning: Please note that the actor will revert to its original behavior when restarted by its Supervisor.
To hotswap the Actor using getContext().become:
import akka.japi.Procedure;
public class HotSwapActor extends UntypedActor {
Procedure<Object> angry = new Procedure<Object>() {
@Override
public void apply(Object message) {
if (message.equals("bar")) {
getSender().tell("I am already angry?", getSelf());
} else if (message.equals("foo")) {
getContext().become(happy);
}
}
};
Procedure<Object> happy = new Procedure<Object>() {
@Override
public void apply(Object message) {
if (message.equals("bar")) {
getSender().tell("I am already happy :-)", getSelf());
} else if (message.equals("foo")) {
getContext().become(angry);
}
}
3.1. Actors
83
};
public void onReceive(Object message) {
if (message.equals("bar")) {
getContext().become(angry);
} else if (message.equals("foo")) {
getContext().become(happy);
} else {
unhandled(message);
}
}
}
This variant of the become method is useful for many different things, such as to implement a Finite State
Machine (FSM). It will replace the current behavior (i.e. the top of the behavior stack), which means that you do
not use unbecome, instead always the next behavior is explicitly installed.
The other way of using become does not replace but add to the top of the behavior stack. In this case care must
be taken to ensure that the number of pop operations (i.e. unbecome) matches the number of push ones in
the long run, otherwise this amounts to a memory leak (which is why this behavior is not the default).
public class UntypedActorSwapper {
public static class Swap {
public static Swap SWAP = new Swap();
private Swap() {
}
}
public static class Swapper extends UntypedActor {
LoggingAdapter log = Logging.getLogger(getContext().system(), this);
public void onReceive(Object message) {
if (message == SWAP) {
log.info("Hi");
getContext().become(new Procedure<Object>() {
@Override
public void apply(Object message) {
log.info("Ho");
getContext().unbecome(); // resets the latest become
}
}, false); // this signals stacking of the new behavior
} else {
unhandled(message);
}
}
}
public static void main(String... args) {
ActorSystem system = ActorSystem.create("MySystem");
ActorRef swap = system.actorOf(Props.create(Swapper.class));
swap.tell(SWAP, ActorRef.noSender()); // logs Hi
swap.tell(SWAP, ActorRef.noSender()); // logs Ho
swap.tell(SWAP, ActorRef.noSender()); // logs Hi
swap.tell(SWAP, ActorRef.noSender()); // logs Ho
swap.tell(SWAP, ActorRef.noSender()); // logs Hi
swap.tell(SWAP, ActorRef.noSender()); // logs Ho
}
}
3.1. Actors
84
3.1.11 Stash
The UntypedActorWithStash class enables an actor to temporarily stash away messages that can not or
should not be handled using the actors current behavior. Upon changing the actors message handler, i.e., right
before invoking getContext().become() or getContext().unbecome(), all stashed messages can
be unstashed, thereby prepending them to the actors mailbox. This way, the stashed messages can be processed
in the same order as they have been received originally. An actor that extends UntypedActorWithStash will
automatically get a deque-based mailbox.
Note:
The abstract class UntypedActorWithStash implements the marker interface
RequiresMessageQueue<DequeBasedMessageQueueSemantics> which requests the system
to automatically choose a deque based mailbox implementation for the actor. If you want more control over the
mailbox, see the documentation on mailboxes: Mailboxes.
Here is an example of the UntypedActorWithStash class in action:
import akka.actor.UntypedActorWithStash;
public class ActorWithProtocol extends UntypedActorWithStash {
public void onReceive(Object msg) {
if (msg.equals("open")) {
unstashAll();
getContext().become(new Procedure<Object>() {
public void apply(Object msg) throws Exception {
if (msg.equals("write")) {
// do writing...
} else if (msg.equals("close")) {
unstashAll();
getContext().unbecome();
} else {
stash();
}
}
}, false); // add behavior on top instead of replacing
} else {
stash();
}
}
}
Invoking stash() adds the current message (the message that the actor received last) to the actors stash.
It is typically invoked when handling the default case in the actors message handler to stash messages that
arent handled by the other cases. It is illegal to stash the same message twice; to do so results in an
IllegalStateException being thrown. The stash may also be bounded in which case invoking stash()
may lead to a capacity violation, which results in a StashOverflowException. The capacity of the stash
can be configured using the stash-capacity setting (an Int) of the mailboxs configuration.
Invoking unstashAll() enqueues messages from the stash to the actors mailbox until the capacity of the mailbox (if any) has been reached (note that messages from the stash are prepended to the mailbox). In case a bounded
mailbox overflows, a MessageQueueAppendFailedException is thrown. The stash is guaranteed to be
empty after calling unstashAll().
The stash is backed by a scala.collection.immutable.Vector. As a result, even a very large number
of messages may be stashed without a major impact on performance.
Note that the stash is part of the ephemeral actor state, unlike the mailbox. Therefore, it should be managed like
other parts of the actors state which have the same property. The UntypedActorWithStash implementation
of preRestart will call unstashAll(), which is usually the desired behavior.
Note: If you want to enforce that your actor can only work with an unbounded stash, then you should use the
UntypedActorWithUnboundedStash class instead.
3.1. Actors
85
3.1. Actors
86
when one would like to avoid reinitializing internals on restart. For example, it is often useful to preserve child
actors across restarts. The following section provides a pattern for this case.
Initialization via preStart
The method preStart() of an actor is only called once directly during the initialization of the first instance,
that is, at creation of its ActorRef. In the case of restarts, preStart() is called from postRestart(),
therefore if not overridden, preStart() is called on every incarnation. However, overriding postRestart()
one can disable this behavior, and ensure that there is only one call to preStart().
One useful usage of this pattern is to disable creation of new ActorRefs for children during restarts. This can
be achieved by overriding preRestart():
@Override
public void preStart() {
// Initialize children here
}
// Overriding postRestart to disable the call to preStart()
// after restarts
@Override
public void postRestart(Throwable reason) {
}
// The default implementation of preRestart() stops all the children
// of the actor. To opt-out from stopping the children, we
// have to override preRestart()
@Override
public void preRestart(Throwable reason, Option<Object> message)
throws Exception {
// Keep the call to postStop(), but no stopping of children
postStop();
}
Please note, that the child actors are still restarted, but no new ActorRef is created. One can recursively apply
the same principles for the children, ensuring that their preStart() method is called only at the creation of
their refs.
For more information see What Restarting Means.
Initialization via message passing
There are cases when it is impossible to pass all the information needed for actor initialization in the constructor,
for example in the presence of circular dependencies. In this case the actor should listen for an initialization
message, and use become() or a finite state-machine state transition to encode the initialized and uninitialized
states of the actor.
private String initializeMe = null;
@Override
public void onReceive(Object message) throws Exception {
if (message.equals("init")) {
initializeMe = "Up and running";
getContext().become(new Procedure<Object>() {
@Override
public void apply(Object message) throws Exception {
if (message.equals("U OK?"))
getSender().tell(initializeMe, getSelf());
}
});
3.1. Actors
87
}
}
If the actor may receive messages before it has been initialized, a useful tool can be the Stash to save messages
until the initialization finishes, and replaying them after the actor became initialized.
Warning: This pattern should be used with care, and applied only when none of the patterns above are
applicable. One of the potential issues is that messages might be lost when sent to remote actors. Also,
publishing an ActorRef in an uninitialized state might lead to the condition that it receives a user message
before the initialization has been done.
88
Warning: Same as not exposing this of an Akka Actor, its important not to expose this of a Typed
Actor, instead you should pass the external proxy reference, which is obtained from within your Typed Actor
as TypedActor.self(), this is your external identity, as the ActorRef is the external identity of an
Akka Actor.
akka.actor.TypedActor;
akka.actor.*;
akka.japi.*;
akka.dispatch.Futures;
scala.concurrent.Await;
scala.concurrent.Future;
scala.concurrent.duration.Duration;
java.util.concurrent.TimeUnit;
The most trivial way of creating a Typed Actor instance of our Squarer:
Squarer mySquarer =
TypedActor.get(system).typedActorOf(
new TypedProps<SquarerImpl>(Squarer.class, SquarerImpl.class));
First type is the type of the proxy, the second type is the type of the implementation. If you need to call a specific
constructor you do it like this:
89
Squarer otherSquarer =
TypedActor.get(system).typedActorOf(
new TypedProps<SquarerImpl>(Squarer.class,
new Creator<SquarerImpl>() {
public SquarerImpl create() { return new SquarerImpl("foo"); }
}),
"name");
Since you supply a Props, you can specify which dispatcher to use, what the default timeout should be used and
more. Now, our Squarer doesnt have any methods, so wed better add those.
public interface Squarer {
void squareDontCare(int i); //fire-forget
Future<Integer> square(int i); //non-blocking send-request-reply
Option<Integer> squareNowPlease(int i);//blocking send-request-reply
int squareNow(int i); //blocking send-request-reply
}
Alright, now weve got some methods we can call, but we need to implement those in SquarerImpl.
class SquarerImpl implements Squarer {
private String name;
public SquarerImpl() {
this.name = "default";
}
public SquarerImpl(String name) {
this.name = name;
}
Excellent, now we have an interface and an implementation of that interface, and we know how to create a Typed
Actor from that, so lets look at calling these methods.
90
akka.japi.Option<?> will use send-request-reply semantics, but will block to wait for an
answer, and return akka.japi.Option.None if no answer was produced within the timeout, or
akka.japi.Option.Some<?> containing the result otherwise. Any exception that was thrown during this call will be rethrown.
Any other type of value will use send-request-reply semantics, but will block to wait for an answer, throwing java.util.concurrent.TimeoutException if there was a timeout or rethrow
any exception that was thrown during this call.
This will block for as long as the timeout that was set in the Props of the Typed Actor, if needed. It will return
None if a timeout occurs.
int iSquare = mySquarer.squareNow(10); //Int
This will block for as long as the timeout that was set in the Props of the Typed Actor, if needed. It will throw a
java.util.concurrent.TimeoutException if a timeout occurs.
Request-reply-with-future message send
Future<Integer> fSquare = mySquarer.square(10); //A Future[Int]
This call is asynchronous, and the Future returned can be used for asynchronous composition.
This asynchronously stops the Typed Actor associated with the specified proxy ASAP.
TypedActor.get(system).poisonPill(otherSquarer);
This asynchronously stops the Typed Actor associated with the specified proxy after its done with all calls that
were made prior to this call.
91
Squarer childSquarer =
TypedActor.get(TypedActor.context()).
typedActorOf(
new TypedProps<SquarerImpl>(Squarer.class, SquarerImpl.class)
);
//Use "childSquarer" as a Squarer
You can also create a child Typed Actor in regular Akka Actors by giving the UntypedActorContext as an
input parameter to TypedActor.get(. . . ).
3.2.11 Proxying
You can use the typedActorOf that takes a TypedProps and an ActorRef to proxy the given ActorRef as a
TypedActor. This is usable if you want to communicate remotely with TypedActors on other machines, just pass
the ActorRef to typedActorOf.
92
93
Description
The progress Listener starts the work.
The Worker schedules work by sending Do messages periodically to itself
When receiving Do the Worker tells the CounterService to increment the counter, three times.
The Increment message is forwarded to the Counter, which updates its counter variable and sends
current value to the Storage.
The Worker asks the CounterService of current value of the counter and pipes the result back to
the Listener.
94
Description
The Storage throws StorageException.
The CounterService is supervisor of the Storage and restarts the Storage when
StorageException is thrown.
The Storage continues to fail and is restarted.
After 3 failures and restarts within 5 seconds the Storage is stopped by its supervisor, i.e. the
CounterService.
The CounterService is also watching the Storage for termination and receives the
Terminated message when the Storage has been stopped ...
and tells the Counter that there is no Storage.
The CounterService schedules a Reconnect message to itself.
When it receives the Reconnect message it creates a new Storage ...
and tells the Counter to use the new Storage
java.util.ArrayList;
java.util.HashMap;
java.util.List;
java.util.Map;
import
import
import
import
import
import
import
import
import
akka.actor.*;
akka.dispatch.Mapper;
akka.japi.Function;
scala.concurrent.duration.Duration;
akka.util.Timeout;
akka.event.Logging;
akka.event.LoggingAdapter;
com.typesafe.config.Config;
com.typesafe.config.ConfigFactory;
static akka.actor.SupervisorStrategy.resume;
static akka.actor.SupervisorStrategy.restart;
static akka.actor.SupervisorStrategy.stop;
static akka.actor.SupervisorStrategy.escalate;
akka.actor.SupervisorStrategy.Directive;
static akka.pattern.Patterns.ask;
static akka.pattern.Patterns.pipe;
import
import
import
import
static
static
static
static
docs.actor.japi.FaultHandlingDocSample.WorkerApi.*;
docs.actor.japi.FaultHandlingDocSample.CounterServiceApi.*;
docs.actor.japi.FaultHandlingDocSample.CounterApi.*;
docs.actor.japi.FaultHandlingDocSample.StorageApi.*;
95
96
/**
* Worker performs some work when it receives the Start message. It will
* continuously notify the sender of the Start message of current Progress.
* The Worker supervise the CounterService.
*/
public static class Worker extends UntypedActor {
final LoggingAdapter log = Logging.getLogger(getContext().system(), this);
final Timeout askTimeout = new Timeout(Duration.create(5, "seconds"));
// The sender of the initial Start message will continuously be notified
// about progress
ActorRef progressListener;
final ActorRef counterService = getContext().actorOf(
Props.create(CounterService.class), "counter");
final int totalCount = 51;
// Stop the CounterService child if it throws ServiceUnavailable
private static SupervisorStrategy strategy = new OneForOneStrategy(-1,
Duration.Inf(), new Function<Throwable, Directive>() {
@Override
public Directive apply(Throwable t) {
if (t instanceof ServiceUnavailable) {
return stop();
} else {
return escalate();
}
}
});
@Override
public SupervisorStrategy supervisorStrategy() {
return strategy;
}
public void onReceive(Object msg) {
log.debug("received message {}", msg);
if (msg.equals(Start) && progressListener == null) {
progressListener = getSender();
getContext().system().scheduler().schedule(
Duration.Zero(), Duration.create(1, "second"), getSelf(), Do,
getContext().dispatcher(), null
);
} else if (msg.equals(Do)) {
counterService.tell(new Increment(1), getSelf());
counterService.tell(new Increment(1), getSelf());
counterService.tell(new Increment(1), getSelf());
// Send current progress to the initial sender
pipe(ask(counterService, GetCurrentCount, askTimeout)
.mapTo(classTag(CurrentCount.class))
.map(new Mapper<CurrentCount, Progress>() {
public Progress apply(CurrentCount c) {
return new Progress(100.0 * c.count / totalCount);
}
}, getContext().dispatcher()), getContext().dispatcher())
.to(progressListener);
} else {
unhandled(msg);
}
}
}
public interface CounterServiceApi {
97
/**
* Adds the value received in Increment message to a persistent counter.
* Replies with CurrentCount when it is asked for CurrentCount. CounterService
* supervise Storage and Counter.
*/
public static class CounterService extends UntypedActor {
// Reconnect message
static final Object Reconnect = "Reconnect";
private static class SenderMsgPair {
final ActorRef sender;
final Object msg;
SenderMsgPair(ActorRef sender, Object msg) {
this.msg = msg;
this.sender = sender;
}
}
final LoggingAdapter log = Logging.getLogger(getContext().system(), this);
final String key = getSelf().path().name();
ActorRef storage;
ActorRef counter;
98
99
forwardOrPlaceInBacklog(Object msg) {
We need the initial value from storage before we can start delegate to
the counter. Before that we place the messages in a backlog, to be sent
to the counter when it is initialized.
(counter == null) {
if (backlog.size() >= MAX_BACKLOG)
throw new ServiceUnavailable("CounterService not available," +
" lack of initial value");
backlog.add(new SenderMsgPair(getSender(), msg));
} else {
counter.forward(msg, getContext());
}
}
}
public interface CounterApi {
public static class UseStorage {
public final ActorRef storage;
public UseStorage(ActorRef storage) {
this.storage = storage;
}
public String toString() {
return String.format("%s(%s)", getClass().getSimpleName(), storage);
}
}
}
/**
* The in memory count variable that will send current value to the Storage,
* if there is any storage available at the moment.
*/
public static class Counter extends UntypedActor {
final LoggingAdapter log = Logging.getLogger(getContext().system(), this);
final String key;
long count;
ActorRef storage;
public Counter(String key, long initialValue) {
this.key = key;
this.count = initialValue;
}
100
@Override
public void onReceive(Object msg) {
log.debug("received message {}", msg);
if (msg instanceof UseStorage) {
storage = ((UseStorage) msg).storage;
storeCount();
} else if (msg instanceof Increment) {
count += ((Increment) msg).n;
storeCount();
} else if (msg.equals(GetCurrentCount)) {
getSender().tell(new CurrentCount(key, count), getSelf());
} else {
unhandled(msg);
}
}
void
//
//
if
storeCount() {
Delegate dangerous work, to protect our valuable state.
We can continue without storage.
(storage != null) {
storage.tell(new Store(new Entry(key, count)), getSelf());
}
}
}
public interface StorageApi {
public static class Store {
public final Entry entry;
public Store(Entry entry) {
this.entry = entry;
}
public String toString() {
return String.format("%s(%s)", getClass().getSimpleName(), entry);
}
}
public static class Entry {
public final String key;
public final long value;
public Entry(String key, long value) {
this.key = key;
this.value = value;
}
public String toString() {
return String.format("%s(%s, %s)", getClass().getSimpleName(), key, value);
}
}
public static class Get {
public final String key;
public Get(String key) {
this.key = key;
}
public String toString() {
return String.format("%s(%s)", getClass().getSimpleName(), key);
}
101
}
public static class StorageException extends RuntimeException {
private static final long serialVersionUID = 1L;
public StorageException(String msg) {
super(msg);
}
}
}
/**
* Saves key/value pairs to persistent storage when receiving Store message.
* Replies with current value when receiving Get message. Will throw
* StorageException if the underlying data store is out of order.
*/
public static class Storage extends UntypedActor {
final LoggingAdapter log = Logging.getLogger(getContext().system(), this);
final DummyDB db = DummyDB.instance;
@Override
public void onReceive(Object msg) {
log.debug("received message {}", msg);
if (msg instanceof Store) {
Store store = (Store) msg;
db.save(store.entry.key, store.entry.value);
} else if (msg instanceof Get) {
Get get = (Get) msg;
Long value = db.load(get.key);
getSender().tell(new Entry(get.key, value == null ?
Long.valueOf(0L) : value), getSelf());
} else {
unhandled(msg);
}
}
}
public static class DummyDB {
public static final DummyDB instance = new DummyDB();
private final Map<String, Long> db = new HashMap<String, Long>();
private DummyDB() {
}
public synchronized void save(String key, Long value) throws StorageException {
if (11 <= value && value <= 14)
throw new StorageException("Simulated store failure " + value);
db.put(key, value);
}
public synchronized Long load(String key) throws StorageException {
return db.get(key);
}
}
}
102
I have chosen a few well-known exception types in order to demonstrate the application of the fault handling
directives described in Supervision and Monitoring. First off, it is a one-for-one strategy, meaning that each child
is treated separately (an all-for-one strategy works very similarly, the only difference is that any decision is applied
to all children of the supervisor, not only the failing one). There are limits set on the restart frequency, namely
maximum 10 restarts per minute. -1 and Duration.Inf() means that the respective limit does not apply,
leaving the possibility to specify an absolute upper limit on the restarts or to make the restarts work infinitely. The
child actor is stopped if the limit is exceeded.
Note: If the strategy is declared inside the supervising actor (as opposed to a separate class) its decider has access
to all internal state of the actor in a thread-safe fashion, including obtaining a reference to the currently failed child
(available as the getSender of the failure message).
103
This supervisor will be used to create a child, with which we can experiment:
public class Child extends UntypedActor {
int state = 0;
104
The test is easier by using the utilities described in akka-testkit, where TestProbe provides an actor ref useful
for receiving and inspecting replies.
import
import
import
import
import
import
import
import
import
import
import
import
import
import
import
import
import
import
akka.actor.ActorRef;
akka.actor.ActorSystem;
akka.actor.SupervisorStrategy;
static akka.actor.SupervisorStrategy.resume;
static akka.actor.SupervisorStrategy.restart;
static akka.actor.SupervisorStrategy.stop;
static akka.actor.SupervisorStrategy.escalate;
akka.actor.SupervisorStrategy.Directive;
akka.actor.OneForOneStrategy;
akka.actor.Props;
akka.actor.Terminated;
akka.actor.UntypedActor;
scala.collection.immutable.Seq;
scala.concurrent.Await;
static akka.pattern.Patterns.ask;
scala.concurrent.duration.Duration;
akka.testkit.AkkaSpec;
akka.testkit.TestProbe;
105
The first test shall demonstrate the Resume directive, so we try it out by setting some non-initial state in the actor
and have it fail:
child.tell(42, ActorRef.noSender());
assert Await.result(ask(child, "get", 5000), timeout).equals(42);
child.tell(new ArithmeticException(), ActorRef.noSender());
assert Await.result(ask(child, "get", 5000), timeout).equals(42);
As you can see the value 42 survives the fault handling directive. Now, if we change the failure to a more serious
NullPointerException, that will no longer be the case:
child.tell(new NullPointerException(), ActorRef.noSender());
assert Await.result(ask(child, "get", 5000), timeout).equals(0);
And finally in case of the fatal IllegalArgumentException the child will be terminated by the supervisor:
final TestProbe probe = new TestProbe(system);
probe.watch(child);
child.tell(new IllegalArgumentException(), ActorRef.noSender());
probe.expectMsgClass(Terminated.class);
Up to now the supervisor was completely unaffected by the childs failure, because the directives set did handle it.
In case of an Exception, this is not true anymore and the supervisor escalates the failure.
child = (ActorRef) Await.result(ask(supervisor,
Props.create(Child.class), 5000), timeout);
probe.watch(child);
assert Await.result(ask(child, "get", 5000), timeout).equals(0);
child.tell(new Exception(), ActorRef.noSender());
probe.expectMsgClass(Terminated.class);
The supervisor itself is supervised by the top-level actor provided by the ActorSystem, which
has the default policy to restart in case of all Exception cases (with the notable exceptions of
ActorInitializationException and ActorKilledException). Since the default directive in case
of a restart is to kill all children, we expected our poor child not to survive this failure.
In case this is not desired (which depends on the use case), we need to use a different supervisor which overrides
this behavior.
public class Supervisor2 extends UntypedActor {
private static SupervisorStrategy strategy = new OneForOneStrategy(10,
Duration.create("1 minute"),
new Function<Throwable, Directive>() {
@Override
public Directive apply(Throwable t) {
if (t instanceof ArithmeticException) {
return resume();
} else if (t instanceof NullPointerException) {
return restart();
} else if (t instanceof IllegalArgumentException) {
return stop();
} else {
return escalate();
}
}
});
@Override
public SupervisorStrategy supervisorStrategy() {
return strategy;
}
106
With this parent, the child survives the escalated restart, as demonstrated in the last test:
superprops = Props.create(Supervisor2.class);
supervisor = system.actorOf(superprops);
child = (ActorRef) Await.result(ask(supervisor,
Props.create(Child.class), 5000), timeout);
child.tell(23, ActorRef.noSender());
assert Await.result(ask(child, "get", 5000), timeout).equals(23);
child.tell(new Exception(), ActorRef.noSender());
assert Await.result(ask(child, "get", 5000), timeout).equals(0);
3.4 Dispatchers
An Akka MessageDispatcher is what makes Akka Actors tick, it is the engine of the machine so to speak.
All MessageDispatcher implementations are also an ExecutionContext, which means that they can be
used to execute arbitrary code, for instance Futures.
3.4. Dispatchers
107
my-dispatcher {
# Dispatcher is the name of the event-based dispatcher
type = Dispatcher
# What kind of ExecutionService to use
executor = "fork-join-executor"
# Configuration for the fork join pool
fork-join-executor {
# Min number of threads to cap factor-based parallelism
parallelism-min = 2
# Parallelism (threads) ... ceil(available processors *
parallelism-factor = 2.0
# Max number of threads to cap factor-based parallelism
parallelism-max = 10
}
# Throughput defines the maximum number of messages to be
# processed per actor before the thread jumps to the next
# Set to 1 for as fair as possible.
throughput = 100
}
number to
factor)
number to
actor.
An alternative to the deployment configuration is to define the dispatcher in code. If you define the dispatcher
in the deployment configuration then this value will be used instead of programmatically provided parameter.
ActorRef myActor =
system.actorOf(Props.create(MyUntypedActor.class).withDispatcher("my-dispatcher"),
"myactor3");
3.4. Dispatchers
108
Note: The dispatcher you specify in withDispatcher and the dispatcher property in the deployment configuration is in fact a path into your configuration. So in this example its a top-level section, but
you could for instance put it as a sub-section, where youd use periods to denote sub-sections, like this:
"foo.bar.my-dispatcher"
the
using
FQCN
of
exan
PinnedDispatcher
This dispatcher dedicates a unique thread for each actor using it; i.e. each actor will have its own
thread pool with only one thread in the pool.
Sharability: None
Mailboxes: Any, creates one per Actor
Use cases: Bulkheading
Driven by: Any akka.dispatch.ThreadPoolExecutorConfigurator by
thread-pool-executor
default
CallingThreadDispatcher
This dispatcher runs invocations on the current thread only. This dispatcher does not create any new
threads, but it can be used from different threads concurrently for the same actor. See CallingThreadDispatcher for details and restrictions.
Sharability: Unlimited
Mailboxes: Any, creates one per Actor per Thread (on demand)
Use cases: Testing
Driven by: The calling thread (duh)
More dispatcher configuration examples
Configuring a PinnedDispatcher:
my-pinned-dispatcher {
executor = "thread-pool-executor"
type = PinnedDispatcher
}
3.4. Dispatchers
109
3.5 Mailboxes
An Akka Mailbox holds the messages that are destined for an Actor. Normally each Actor has its own
mailbox, but with for example a BalancingPool all routees will share a single mailbox instance.
The type parameter to the RequiresMessageQueue interface needs to be mapped to a mailbox in configuration like this:
bounded-mailbox {
mailbox-type = "akka.dispatch.BoundedMailbox"
mailbox-capacity = 1000
mailbox-push-timeout-time = 10s
}
akka.actor.mailbox.requirements {
"akka.dispatch.BoundedMessageQueueSemantics" = bounded-mailbox
}
Now every time you create an actor of type MyBoundedUntypedActor it will try to get a bounded mailbox.
If the actor has a different mailbox configured in deployment, either directly or via a dispatcher with a specified
mailbox type, then that will override this mapping.
Note: The type of the queue in the mailbox created for an actor will be checked against the required type in the
interface and if the queue doesnt implement the required type then actor creation will fail.
3.5. Mailboxes
110
The given requirement names a class or interface which will then be ensured to be a supertype of the message
queues implementation. In case of a conflicte.g. if the actor requires a mailbox type which does not satisfy this
requirementthen actor creation will fail.
How the Mailbox Type is Selected
When an actor is created, the ActorRefProvider first determines the dispatcher which will execute it. Then
the mailbox is determined as follows:
1. If the actors deployment configuration section contains a mailbox key then that names a configuration
section describing the mailbox type to be used.
2. If the actors Props contains a mailbox selectioni.e. withMailbox was called on itthen that names
a configuration section describing the mailbox type to be used.
3. If the dispatchers configuration section contains a mailbox-type key the same section will be used to
configure the mailbox type.
4. If the actor requires a mailbox type as described above then the mapping for that requirement will be used
to determine the mailbox type to be used; if that fails then the dispatchers requirementif anywill be
tried instead.
5. If the dispatcher requires a mailbox type as described above then the mapping for that requirement will be
used to determine the mailbox type to be used.
6. The default mailbox akka.actor.default-mailbox will be used.
Default Mailbox
When the mailbox is not specified as described above the default mailbox is used. By default it is an unbounded
mailbox, which is backed by a java.util.concurrent.ConcurrentLinkedQueue.
SingleConsumerOnlyUnboundedMailbox is an even more efficient mailbox, and it can be used as the
default mailbox, but it cannot be used with a BalancingDispatcher.
Configuration of SingleConsumerOnlyUnboundedMailbox as default mailbox:
akka.actor.default-mailbox {
mailbox-type = "akka.dispatch.SingleConsumerOnlyUnboundedMailbox"
}
3.5. Mailboxes
111
SingleConsumerOnlyUnboundedMailbox
Backed by a very efficient Multiple Producer Single Consumer queue, cannot be used with BalancingDispatcher
Blocking: No
Bounded: No
Configuration name: akka.dispatch.SingleConsumerOnlyUnboundedMailbox
BoundedMailbox
Backed by a java.util.concurrent.LinkedBlockingQueue
Blocking: Yes
Bounded: Yes
Configuration name: bounded or akka.dispatch.BoundedMailbox
UnboundedPriorityMailbox
Backed by a java.util.concurrent.PriorityBlockingQueue
Blocking: Yes
Bounded: No
Configuration name: akka.dispatch.UnboundedPriorityMailbox
BoundedPriorityMailbox
Backed
by
a
java.util.PriorityBlockingQueue
akka.util.BoundedBlockingQueue
wrapped
in
an
Blocking: Yes
Bounded: Yes
Configuration name: akka.dispatch.BoundedPriorityMailbox
3.5. Mailboxes
112
prio-dispatcher {
mailbox-type = "docs.dispatcher.DispatcherDocSpec$MyPrioMailbox"
//Other dispatcher configuration goes here
}
3.5. Mailboxes
113
akka.actor.ActorRef;
akka.actor.ActorSystem;
akka.dispatch.Envelope;
akka.dispatch.MailboxType;
akka.dispatch.MessageQueue;
akka.dispatch.ProducesMessageQueue;
com.typesafe.config.Config;
java.util.concurrent.ConcurrentLinkedQueue;
java.util.Queue;
scala.Option;
And then you just specify the FQCN of your MailboxType as the value of the mailbox-type in the dispatcher
configuration, or the mailbox configuration.
Note:
Make sure to include a constructor which takes akka.actor.ActorSystem.Settings and
com.typesafe.config.Config arguments, as this constructor is invoked reflectively to construct your
mailbox type. The config passed in as second argument is that section from the configuration which describes
the dispatcher or mailbox setting using this mailbox type; the mailbox type will be instantiated once for each
dispatcher or mailbox setting using it.
3.5. Mailboxes
114
You can also use the mailbox as a requirement on the dispatcher like this:
custom-dispatcher {
mailbox-requirement =
"docs.dispatcher.MyUnboundedJMessageQueueSemantics"
}
akka.actor.mailbox.requirements {
"docs.dispatcher.MyUnboundedJMessageQueueSemantics" =
custom-dispatcher-mailbox
}
custom-dispatcher-mailbox {
mailbox-type = "docs.dispatcher.MyUnboundedJMailbox"
}
will probably fail; you will have to allow for some time to pass and retry the check la TestKit.awaitCond.
3.6 Routing
Messages can be sent via a router to efficiently route them to destination actors, known as its routees. A Router
can be used inside or outside of an actor, and you can manage the routees yourselves or use a self contained router
actor with configuration capabilities.
Different routing strategies can be used, according to your applications needs. Akka comes with several useful
routing strategies right out of the box. But, as you will see in this chapter, it is also possible to create your own.
3.6. Routing
115
We create a Router and specify that it should use RoundRobinRoutingLogic when routing the messages
to the routees.
The routing logic shipped with Akka are:
akka.routing.RoundRobinRoutingLogic
akka.routing.RandomRoutingLogic
akka.routing.SmallestMailboxRoutingLogic
akka.routing.BroadcastRoutingLogic
akka.routing.ScatterGatherFirstCompletedRoutingLogic
akka.routing.ConsistentHashingRoutingLogic
We create the routees as ordinary child actors wrapped in ActorRefRoutee. We watch the routees to be able
to replace them if they are terminated.
Sending messages via the router is done with the route method, as is done for the Work messages in the example
above.
The Router is immutable and the RoutingLogic is thread safe; meaning that they can also be used outside
of actors.
Note: In general, any message sent to a router will be sent onwards to its routees, but there is one exception. The
special Broadcast Messages will send to all of a routers routees
3.6. Routing
116
Group - The routee actors are created externally to the router and the router sends messages to the specified
path using actor selection, without watching for termination.
The settings for a router actor can be defined in configuration or programmatically. Although router actors can
be defined in the configuration file, they must still be created programmatically, i.e. you cannot make a router
through external configuration alone. If you define the router actor in the configuration file then these settings will
be used instead of any programmatically provided parameters.
You send messages to the routees via the router actor in the same way as for ordinary actors, i.e. via its ActorRef.
The router actor forwards messages onto its routees without changing the original sender. When a routee replies
to a routed message, the reply will be sent to the original sender, not to the router actor.
Note: In general, any message sent to a router will be sent onwards to its routees, but there are a few exceptions.
These are documented in the Specially Handled Messages section below.
Pool
The following code and configuration snippets show how to create a round-robin router that forwards messages to
five Worker routees. The routees will be created as the routers children.
akka.actor.deployment {
/parent/router1 {
router = round-robin-pool
nr-of-instances = 5
}
}
ActorRef router1 =
getContext().actorOf(FromConfig.getInstance().props(Props.create(Worker.class)),
"router1");
Here is the same example, but with the router configuration provided programmatically instead of from configuration.
ActorRef router2 =
getContext().actorOf(new RoundRobinPool(5).props(Props.create(Worker.class)),
"router2");
In addition to being able to create local actors as routees, you can instruct the router to deploy its created children
on a set of remote hosts. Routees will be deployed in round-robin fashion. In order to deploy routees remotely,
wrap the router configuration in a RemoteRouterConfig, attaching the remote addresses of the nodes to
deploy to. Remote deployment requires the akka-remote module to be included in the classpath.
Address[] addresses = {
new Address("akka.tcp", "remotesys", "otherhost", 1234),
AddressFromURIString.parse("akka.tcp://othersys@anotherhost:1234")};
ActorRef routerRemote = system.actorOf(
new RemoteRouterConfig(new RoundRobinPool(5), addresses).props(
Props.create(Echo.class)));
Senders
3.6. Routing
117
However, it is often useful for routees to set the router as a sender. For example, you might want to set the router
as the sender if you want to hide the details of the routees behind the router. The following code snippet shows
how to set the parent router as sender.
getSender().tell("reply", getContext().parent());
Supervision
Routees that are created by a pool router will be created as the routers children. The router is therefore also the
childrens supervisor.
The supervision strategy of the router actor can be configured with the supervisorStrategy property of the
Pool. If no configuration is provided, routers default to a strategy of always escalate. This means that errors are
passed up to the routers supervisor for handling. The routers supervisor will decide what to do about any errors.
Note the routers supervisor will treat the error as an error with the router itself. Therefore a directive to stop or
restart will cause the router itself to stop or restart. The router, in turn, will cause its children to stop and restart.
It should be mentioned that the routers restart behavior has been overridden so that a restart, while still re-creating
the children, will still preserve the same number of actors in the pool.
This means that if you have not specified supervisorStrategy of the router or its parent a failure in a routee
will escalate to the parent of the router, which will by default restart the router, which will restart all routees (it
uses Escalate and does not stop routees during restart). The reason is to make the default behave such that adding
withRouter to a childs definition does not change the supervision strategy applied to the child. This might be
an inefficiency that you can avoid by specifying the strategy when defining the router.
Setting the strategy is easily done:
final SupervisorStrategy strategy =
new OneForOneStrategy(5, Duration.create(1, TimeUnit.MINUTES),
Collections.<Class<? extends Throwable>>singletonList(Exception.class));
final ActorRef router = system.actorOf(new RoundRobinPool(5).
withSupervisorStrategy(strategy).props(Props.create(Echo.class)));
Note: If the child of a pool router terminates, the pool router will not automatically spawn a new child. In the
event that all children of a pool router have terminated the router will terminate itself unless it is a dynamic router,
e.g. using a resizer.
Group
Sometimes, rather than having the router actor create its routees, it is desirable to create routees separately and provide them to the router for its use. You can do this by passing an paths of the routees to the routers configuration.
Messages will be sent with ActorSelection to these paths.
The example below shows how to create a router by providing it with the path strings of three routee actors.
akka.actor.deployment {
/parent/router3 {
router = round-robin-group
routees.paths = ["/user/workers/w1", "/user/workers/w2", "/user/workers/w3"]
}
}
ActorRef router3 =
getContext().actorOf(FromConfig.getInstance().props(), "router3");
Here is the same example, but with the router configuration provided programmatically instead of from configuration.
3.6. Routing
118
ActorRef router4 =
getContext().actorOf(new RoundRobinGroup(paths).props(), "router4");
The paths may contain protocol and address information for actors running on remote hosts. Remoting requires
the akka-remote module to be included in the classpath.
akka.actor.deployment {
/parent/remoteGroup {
router = round-robin-group
routees.paths = [
"akka.tcp://[email protected]:2552/user/workers/w1",
"akka.tcp://[email protected]:2552/user/workers/w1",
"akka.tcp://[email protected]:2552/user/workers/w1"]
}
}
3.6. Routing
119
akka.actor.deployment {
/parent/router3 {
router = round-robin-group
routees.paths = ["/user/workers/w1", "/user/workers/w2", "/user/workers/w3"]
}
}
ActorRef router3 =
getContext().actorOf(FromConfig.getInstance().props(), "router3");
BalancingPool
A Router that will try to redistribute work from busy routees to idle routees. All routees share the same mailbox.
3.6. Routing
120
Addition configuration for the balancing dispatcher, which is used by the pool, can be configured in the
pool-dispatcher section of the router deployment configuration.
akka.actor.deployment {
/parent/router9b {
router = balancing-pool
nr-of-instances = 5
pool-dispatcher {
attempt-teamwork = off
}
}
}
3.6. Routing
121
There is no Group variant of the SmallestMailboxPool because the size of the mailbox and the internal dispatching
state of the actor is not practically available from the paths of the routees.
BroadcastPool and BroadcastGroup
A broadcast router forwards the message it receives to all its routees.
BroadcastPool defined in configuration:
akka.actor.deployment {
/parent/router13 {
router = broadcast-pool
nr-of-instances = 5
}
}
ActorRef router13 =
getContext().actorOf(FromConfig.getInstance().props(
Props.create(Worker.class)), "router13");
Note: Broadcast routers always broadcast every message to their routees. If you do not want to broadcast every
message, then you can use a non-broadcasting router and use Broadcast Messages as needed.
3.6. Routing
122
nr-of-instances = 5
within = 10 seconds
}
}
ActorRef router17 =
getContext().actorOf(FromConfig.getInstance().props(
Props.create(Worker.class)), "router17");
3.6. Routing
123
3.6. Routing
124
cache.tell(new ConsistentHashableEnvelope(
new Entry("hello", "HELLO"), "hello"), getRef());
cache.tell(new ConsistentHashableEnvelope(
new Entry("hi", "HI"), "hi"), getRef());
cache.tell(new Get("hello"), getRef());
expectMsgEquals("HELLO");
cache.tell(new Get("hi"), getRef());
expectMsgEquals("HI");
cache.tell(new Evict("hi"), getRef());
cache.tell(new Get("hi"), getRef());
expectMsgEquals(NOT_FOUND);
In the above example you see that the Get message implements ConsistentHashable itself, while the
Entry message is wrapped in a ConsistentHashableEnvelope. The Evict message is handled by
the hashMapping partial function.
ConsistentHashingPool defined in configuration:
akka.actor.deployment {
/parent/router21 {
router = consistent-hashing-pool
nr-of-instances = 5
virtual-nodes-factor = 10
}
}
ActorRef router21 =
getContext().actorOf(FromConfig.getInstance().props(Props.create(Worker.class)),
"router21");
virtual-nodes-factor is the number of virtual nodes per routee that is used in the consistent hash node
ring to make the distribution more uniform.
3.6. Routing
125
For a router, which normally passes on messages to routees, it is important to realise that PoisonPill messages
are processed by the router only. PoisonPill messages sent to a router will not be sent on to routees.
However, a PoisonPill message sent to a router may still affect its routees, because it will stop the router and
when the router stops it also stops its children. Stopping children is normal actor behavior. The router will stop
routees that it has created as children. Each child will process its current message and then stop. This may lead to
some messages being unprocessed. See the documentation on Stopping actors for more information.
If you wish to stop a router and its routees, but you would like the routees to first process all the messages
currently in their mailboxes, then you should not send a PoisonPill message to the router. Instead you should
wrap a PoisonPill message inside a Broadcast message so that each routee will receive the PoisonPill
message. Note that this will stop all routees, even if the routees arent children of the router, i.e. even routees
programmatically provided to the router.
router.tell(new Broadcast(PoisonPill.getInstance()), getTestActor());
With the code shown above, each routee will receive a PoisonPill message. Each routee will continue to
process its messages as normal, eventually processing the PoisonPill. This will cause the routee to stop. After
all routees have stopped the router will itself be stopped automatically unless it is a dynamic router, e.g. using a
resizer.
Note: Brendan W McAdams excellent blog post Distributing Akka Workloads - And Shutting Down Afterwards
discusses in more detail how PoisonPill messages can be used to shut down routers and routees.
Kill Messages
Kill messages are another type of message that has special handling. See Killing an Actor for general information about how actors handle Kill messages.
3.6. Routing
126
When a Kill message is sent to a router the router processes the message internally, and does not send it on to its
routees. The router will throw an ActorKilledException and fail. It will then be either resumed, restarted
or terminated, depending how it is supervised.
Routees that are children of the router will also be suspended, and will be affected by the supervision directive
that is applied to the router. Routees that are not the routers children, i.e. those that were created externally to the
router, will not be affected.
router.tell(Kill.getInstance(), getTestActor());
As with the PoisonPill message, there is a distinction between killing a router, which indirectly kills its
children (who happen to be routees), and killing routees directly (some of whom may not be children.) To kill
routees directly the router should be sent a Kill message wrapped in a Broadcast message.
router.tell(new Broadcast(Kill.getInstance()), getTestActor());
Managagement Messages
Sending akka.routing.GetRoutees to a router actor will make it send back its currently used routees
in a akka.routing.Routees message.
Sending akka.routing.AddRoutee to a router actor will add that routee to its collection of routees.
Sending akka.routing.RemoveRoutee to a router actor will remove that routee to its collection of
routees.
Sending akka.routing.AdjustPoolSize to a pool router actor will add or remove that number of
routees to its collection of routees.
These management messages may be handled after other messages, so if you send AddRoutee immediately
followed an ordinary message you are not guaranteed that the routees have been changed when the ordinary
message is routed. If you need to know when the change has been applied you can send AddRoutee followed by
GetRoutees and when you receive the Routees reply you know that the preceeding change has been applied.
3.6. Routing
127
It is also worth pointing out that if you define the router in the configuration file then this value will be used
instead of any programmatically sent parameters.
Note: Resizing is triggered by sending messages to the actor pool, but it is not completed synchronously; instead
a message is sent to the head RouterActor to perform the size change. Thus you cannot rely on resizing
to instantaneously create new workers when all others are busy, because the message just sent will be queued to
the mailbox of a busy actor. To remedy this, configure the pool to use a balancing dispatcher, see Configuring
Dispatchers for more information.
3.6. Routing
128
select will be called for each message and in this example pick a few destinations by round-robin, by
reusing the existing RoundRobinRoutingLogic and wrap the result in a SeveralRoutees instance.
SeveralRoutees will send the message to all of the supplied routues.
The implementation of the routing logic must be thread safe, since it might be used outside of actors.
A unit test of the routing logic:
public final class TestRoutee implements Routee {
public final int n;
public TestRoutee(int n) {
this.n = n;
}
@Override
public void send(Object message, ActorRef sender) {
}
@Override
public int hashCode() {
return n;
}
@Override
public boolean equals(Object obj) {
return (obj instanceof TestRoutee) &&
n == ((TestRoutee) obj).n;
}
}
RedundancyRoutingLogic logic = new RedundancyRoutingLogic(3);
List<Routee> routeeList = new ArrayList<Routee>();
for (int n = 1; n <= 7; n++) {
routeeList.add(new TestRoutee(n));
}
IndexedSeq<Routee> routees = immutableIndexedSeq(routeeList);
SeveralRoutees r1 = (SeveralRoutees)
assertEquals(r1.getRoutees().get(0),
assertEquals(r1.getRoutees().get(1),
assertEquals(r1.getRoutees().get(2),
logic.select("msg", routees);
routeeList.get(0));
routeeList.get(1));
routeeList.get(2));
SeveralRoutees r2 = (SeveralRoutees)
assertEquals(r2.getRoutees().get(0),
assertEquals(r2.getRoutees().get(1),
assertEquals(r2.getRoutees().get(2),
logic.select("msg", routees);
routeeList.get(3));
routeeList.get(4));
routeeList.get(5));
SeveralRoutees r3 = (SeveralRoutees)
assertEquals(r3.getRoutees().get(0),
assertEquals(r3.getRoutees().get(1),
assertEquals(r3.getRoutees().get(2),
logic.select("msg", routees);
routeeList.get(6));
routeeList.get(0));
routeeList.get(1));
You could stop here and use the RedundancyRoutingLogic with a akka.routing.Router as described
in A Simple Router.
Let us continue and make this into a self contained, configurable, router actor.
Create a class that extends PoolBase, GroupBase or CustomRouterConfig. That class is a factory for
3.6. Routing
129
the routing logic and holds the configuration for the router. Here we make it a Group.
import java.util.List;
import
import
import
import
import
import
import
import
import
import
import
import
import
scala.Option;
scala.collection.immutable.Iterable;
akka.actor.ActorContext;
akka.actor.ActorPath;
akka.actor.ActorSystem;
akka.actor.Props;
akka.dispatch.Dispatchers;
akka.routing.Group;
akka.routing.Routee;
akka.routing.Router;
akka.routing.RouterActor;
akka.routing.RouterConfig;
akka.routing.RoutingLogic;
import com.typesafe.config.Config;
import akka.routing.GroupBase;
import static docs.jrouting.CustomRouterDocTest.RedundancyRoutingLogic;
public class RedundancyGroup extends GroupBase {
private final List<String> paths;
private final int nbrCopies;
public RedundancyGroup(List<String> paths, int nbrCopies) {
this.paths = paths;
this.nbrCopies = nbrCopies;
}
public RedundancyGroup(Config config) {
this(config.getStringList("routees.paths"),
config.getInt("nbr-copies"));
}
@Override
public java.lang.Iterable<String> getPaths() {
return paths;
}
@Override
public Router createRouter(ActorSystem system) {
return new Router(new RedundancyRoutingLogic(nbrCopies));
}
@Override
public String routerDispatcher() {
return Dispatchers.DefaultDispatcherId();
}
}
3.6. Routing
130
ActorRef redundancy1 =
system.actorOf(new RedundancyGroup(paths, 3).props(),
"redundancy1");
redundancy1.tell("important", getTestActor());
Note that we added a constructor in RedundancyGroup that takes a Config parameter. That makes it possible
to define it in configuration.
akka.actor.deployment {
/redundancy2 {
router = "docs.jrouting.RedundancyGroup"
routees.paths = ["/user/s1", "/user/s2", "/user/s3"]
nbr-copies = 5
}
}
That is the only thing you need to do enable a dedicated dispatcher for a pool.
Note: If you use a group of actors and route to their paths, then they will still use the same dispatcher that was
configured for them in their Props, it is not possible to change an actors dispatcher after it has been created.
The head router cannot always run on the same dispatcher, because it does not process the same type
of messages, hence this special actor does not use the dispatcher configured in Props, but takes the
routerDispatcher from the RouterConfig instead, which defaults to the actor systems default dispatcher. All standard routers allow setting this property in their constructor or factory method, custom routers
have to implement the method in a suitable way.
Props props =
// head router actor will run on "router-dispatcher" dispatcher
// Worker routees will run on "pool-dispatcher" dispatcher
new RandomPool(5).withDispatcher("router-dispatcher").props(
Props.create(Worker.class));
ActorRef router = system.actorOf(props, "poolWithDispatcher");
3.6. Routing
131
Note:
It is not allowed to configure the routerDispatcher to be a
akka.dispatch.BalancingDispatcherConfigurator since the messages meant for the special router actor cannot be processed by any other actor.
132
}
}
protected void enqueue(Object o) {
if (queue != null)
queue.add(o);
}
protected List<Object> drainQueue() {
final List<Object> q = queue;
if (q == null)
throw new IllegalStateException("drainQueue(): not yet initialized");
queue = new ArrayList<Object>();
return q;
}
/*
* Here are the interrogation methods:
*/
protected boolean isInitialized() {
return target != null;
}
protected State getState() {
return state;
}
protected ActorRef getTarget() {
if (target == null)
throw new IllegalStateException("getTarget(): not yet initialized");
return target;
}
/*
* And finally the callbacks (only one in this example: react to state change)
*/
abstract protected void transition(State old, State next);
}
The benefit of this approach is that state changes can be acted upon in one central place, which makes it impossible
to forget inserting code for reacting to state transitions when adding to the FSMs machinery.
133
}
public final Object flush = new Object();
public final class Batch {
final List<Object> objects;
public Batch(List<Object> objects) {
this.objects = objects;
}
}
This actor has only the two states IDLE and ACTIVE, making their handling quite straight-forward in the concrete
actor derived from the base class:
import akka.event.LoggingAdapter;
import akka.event.Logging;
import akka.actor.UntypedActor;
public class MyFSM extends MyFSMBase {
private final LoggingAdapter log =
Logging.getLogger(getContext().system(), this);
@Override
public void onReceive(Object o) {
if (getState() == State.IDLE) {
if (o instanceof SetTarget)
init(((SetTarget) o).ref);
else
whenUnhandled(o);
} else if (getState() == State.ACTIVE) {
if (o == flush)
setState(State.IDLE);
else
whenUnhandled(o);
}
}
@Override
public void transition(State old, State next) {
if (old == State.ACTIVE) {
getTarget().tell(new Batch(drainQueue()), getSelf());
}
}
private void whenUnhandled(Object o) {
if (o instanceof Queue && isInitialized()) {
enqueue(((Queue) o).o);
setState(State.ACTIVE);
} else {
log.warning("received unknown message {} in state {}", o, getState());
}
}
}
134
The trick here is to factor out common functionality like whenUnhandled and transition in order to obtain
a few well-defined points for reacting to change or insert logging.
3.8 Persistence
Java 8 lambda expressions are also supported now. (See section Persistence (Java with Lambda Support))
Akka persistence enables stateful actors to persist their internal state so that it can be recovered when an actor is
started, restarted after a JVM crash or by a supervisor, or migrated in a cluster. The key concept behind Akka
persistence is that only changes to an actors internal state are persisted but never its current state directly (except
for optional snapshots). These changes are only ever appended to storage, nothing is ever mutated, which allows
for very high transaction rates and efficient replication. Stateful actors are recovered by replaying stored changes
to these actors from which they can rebuild internal state. This can be either the full history of changes or starting
from a snapshot which can dramatically reduce recovery times. Akka persistence also provides point-to-point
communication channels with at-least-once message delivery semantics.
Warning: This module is marked as experimental as of its introduction in Akka 2.3.0. We will continue to
improve this API based on our users feedback, which implies that while we try to keep incompatible changes
to a minimum the binary compatibility guarantee for maintenance releases does not apply to the contents of
the akka.persistence package.
Akka persistence is inspired by and the official replacement of the eventsourced library. It follows the same
concepts and architecture of eventsourced but significantly differs on API and implementation level. See also
Migration Guide Eventsourced to Akka Persistence 2.3.x
3.8.1 Dependencies
Akka persistence is a separate jar file. Make sure that you have the following dependency in your project:
<dependency>
<groupId>com.typesafe.akka</groupId>
<artifactId>akka-persistence-experimental_2.10</artifactId>
<version>2.3.1</version>
</dependency>
3.8.2 Architecture
Processor: A processor is a persistent, stateful actor. Messages sent to a processor are written to a journal
before its onReceive method is called. When a processor is started or restarted, journaled messages are
replayed to that processor, so that it can recover internal state from these messages.
View: A view is a persistent, stateful actor that receives journaled messages that have been written by
another processor. A view itself does not journal new messages, instead, it updates internal state only from
a processors replicated message stream.
Channel: Channels are used by processors and views to communicate with other actors. They prevent that
replayed messages are redundantly delivered to these actors and provide at-least-once message delivery
semantics, also in case of sender and receiver JVM crashes.
3.8. Persistence
135
Journal: A journal stores the sequence of messages sent to a processor. An application can control which
messages are journaled and which are received by the processor without being journaled. The storage
backend of a journal is pluggable. The default journal storage plugin writes to the local filesystem, replicated
journals are available as Community plugins.
Snapshot store: A snapshot store persists snapshots of a processors or a views internal state. Snapshots
are used for optimizing recovery times. The storage backend of a snapshot store is pluggable. The default
snapshot storage plugin writes to the local filesystem.
Event sourcing. Based on the building blocks described above, Akka persistence provides abstractions for
the development of event sourced applications (see section Event sourcing)
3.8.3 Processors
A processor can be implemented by extending the abstract UntypedProcessor class and implementing the
onReceive method.
class MyProcessor extends UntypedProcessor {
public void onReceive(Object message) throws Exception {
if (message instanceof Persistent) {
// message successfully written to journal
Persistent persistent = (Persistent)message;
Object payload = persistent.payload();
Long sequenceNr = persistent.sequenceNr();
// ...
} else if (message instanceof PersistenceFailure) {
// message failed to be written to journal
PersistenceFailure failure = (PersistenceFailure)message;
Object payload = failure.payload();
Long sequenceNr = failure.sequenceNr();
Throwable cause = failure.cause();
// ...
} else {
// message not written to journal
}
}
}
Processors only write messages of type Persistent to the journal, others are received without being persisted.
When a processors onReceive method is called with a Persistent message it can safely assume that this
message has been successfully written to the journal. If a journal fails to write a Persistent message then
the processor is stopped, by default. If a processor should continue running on persistence failures it must handle
PersistenceFailure messages. In this case, a processor may want to inform the sender about the failure, so
that the sender can re-send the message, if needed.
An UntypedProcessor itself is an Actor and can therefore be instantiated with actorOf.
processor = getContext().actorOf(Props.create(MyProcessor.class), "myProcessor");
processor.tell(Persistent.create("foo"), null);
processor.tell("bar", null);
Recovery
By default, a processor is automatically recovered on start and on restart by replaying journaled messages. New
messages sent to a processor during recovery do not interfere with replayed messages. New messages will only
be received by a processor after recovery completes.
3.8. Persistence
136
Recovery customization
Automated recovery on start can be disabled by overriding preStart with an empty implementation.
@Override
public void preStart() {}
If not overridden, preStart sends a Recover message to getSelf(). Applications may also override
preStart to define further Recover parameters such as an upper sequence number bound, for example.
@Override
public void preStart() {
getSelf().tell(Recover.create(457L), null);
}
Upper sequence number bounds can be used to recover a processor to past state instead of current state. Automated
recovery on restart can be disabled by overriding preRestart with an empty implementation.
@Override
public void preRestart(Throwable reason, Option<Object> message) {}
Recovery status
A processor can query its own recovery status via the methods
public boolean recoveryRunning();
public boolean recoveryFinished();
Sometimes there is a need for performing additional initialization when the recovery has completed, before processing any other message sent to the processor. The processor can send itself a message from preStart. It will
be stashed and received after recovery. The mailbox may contain other messages that are queued in front of that
message and therefore you need to stash until you receive that message.
@Override
public void preStart() throws Exception {
super.preStart();
self().tell("FIRST", self());
}
public void onReceive(Object message) throws Exception {
if (message.equals("FIRST")) {
recoveryCompleted();
getContext().become(active);
unstashAll();
} else if (recoveryFinished()) {
stash();
} else {
active.apply(message);
}
}
private void recoveryCompleted() {
// perform init after recovery, before any other messages
// ...
}
Procedure<Object> active = new Procedure<Object>() {
@Override
3.8. Persistence
137
Failure handling
A persistent message that caused an exception will be received again by a processor after restart. To prevent a
replay of that message during recovery it can be deleted.
@Override
public void preRestart(Throwable reason, Option<Object> message) {
if (message.isDefined() && message.get() instanceof Persistent) {
deleteMessage(((Persistent) message.get()).sequenceNr());
}
super.preRestart(reason, message);
}
Message deletion
A processor can delete a single message by calling the deleteMessage method with the sequence number of
that message as argument. An optional permanent parameter specifies whether the message shall be permanently deleted from the journal or only marked as deleted. In both cases, the message wont be replayed. Later
extensions to Akka persistence will allow to replay messages that have been marked as deleted which can be useful
for debugging purposes, for example. To delete all messages (journaled by a single processor) up to a specified
sequence number, processors should call the deleteMessages method.
Identifiers
A processor must have an identifier that doesnt change across different actor incarnations. It defaults to the
String representation of processors path without the address part and can be obtained via the processorId
method.
public String processorId();
Applications can customize a processors id by specifying an actor name during processor creation as shown in
section Processors. This changes that processors name in its actor hierarchy and hence influences only part of the
processor id. To fully customize a processors id, the processorId method must be overridden.
@Override
public String processorId() {
return "my-stable-processor-id";
}
3.8.4 Views
Views can be implemented by extending the UntypedView trait and implementing the onReceive and the
processorId methods.
class MyView extends UntypedView {
@Override
public String processorId() {
return "some-processor-id";
3.8. Persistence
138
}
@Override
public void onReceive(Object message) throws Exception {
if (message instanceof Persistent) {
// ...
}
}
}
The processorId identifies the processor from which the view receives journaled messages. It is not necessary
the referenced processor is actually running. Views read messages from a processors journal directly. When a
processor is started later and begins to write new messages, the corresponding view is updated automatically, by
default.
Updates
The default update interval of all views of an actor system is configurable:
akka.persistence.view.auto-update-interval = 5s
View implementation classes may also override the autoUpdateInterval method to return a custom update
interval for a specific view class or view instance. Applications may also trigger additional updates at any time by
sending a view an Update message.
final ActorRef view = system.actorOf(Props.create(MyView.class));
view.tell(Update.create(true), null);
If the await parameter is set to true, messages that follow the Update request are processed when the incremental message replay, triggered by that update request, completed. If set to false (default), messages following
the update request may interleave with the replayed message stream. Automated updates always run with await
= false.
Automated updates of all views of an actor system can be turned off by configuration:
akka.persistence.view.auto-update = off
Implementation classes may override the configured default value by overriding the autoUpdate
method.
To limit the number of replayed messages per update request, applications can configure a custom akka.persistence.view.auto-update-replay-max value or override the
autoUpdateReplayMax method. The number of replayed messages for manual updates can be limited with
the replayMax parameter of the Update message.
Recovery
Initial recovery of views works in the very same way as for Processors (i.e. by sending a Recover
message to self). The maximum number of replayed messages during initial recovery is determined by
autoUpdateReplayMax. Further possibilities to customize initial recovery are explained in section Processors.
Identifiers
A view must have an identifier that doesnt change across different actor incarnations. It defaults to the String
representation of the actor path without the address part and can be obtained via the viewId method.
Applications can customize a views id by specifying an actor name during view creation. This changes that
views name in its actor hierarchy and hence influences only part of the view id. To fully customize a views
id, the viewId method must be overridden. Overriding viewId is the recommended way to generate stable
identifiers.
3.8. Persistence
139
The viewId must differ from the referenced processorId, unless Snapshots of a view and its processor shall
be shared (which is what applications usually do not want).
3.8.5 Channels
Channels are special actors that are used by processors or views to communicate with other actors (channel destinations). The following discusses channels in context of processors but this is also applicable to views.
Channels prevent redundant delivery of replayed messages to destinations during processor recovery. A replayed
message is retained by a channel if its delivery has been confirmed by a destination.
class MyProcessor extends UntypedProcessor {
private final ActorRef destination;
private final ActorRef channel;
public MyProcessor() {
this.destination = getContext().actorOf(Props.create(MyDestination.class));
this.channel = getContext().actorOf(Channel.props(), "myChannel");
}
public void onReceive(Object message) throws Exception {
if (message instanceof Persistent) {
Persistent p = (Persistent)message;
Persistent out = p.withPayload("done " + p.payload());
channel.tell(Deliver.create(out, destination.path()), getSelf());
}
}
}
class MyDestination extends UntypedActor {
public void onReceive(Object message) throws Exception {
if (message instanceof ConfirmablePersistent) {
ConfirmablePersistent p = (ConfirmablePersistent)message;
Object payload = p.payload();
Long sequenceNr = p.sequenceNr();
int redeliveries = p.redeliveries();
// ...
p.confirm();
}
}
}
A channel is ready to use once it has been created, no recovery or further activation is needed. A Deliver request
instructs a channel to send a Persistent message to a destination. A destination is provided as ActorPath
and messages are sent by the channel via that paths ActorSelection. Sender references are preserved by a
channel, therefore, a destination can reply to the sender of a Deliver request.
Note: Sending via a channel has at-least-once delivery semanticsby virtue of either the sending actor or
the channel being persistentwhich means that the semantics do not match those of a normal ActorRef send
operation:
it is not at-most-once delivery
message order for the same senderreceiver pair is not retained due to possible resends
after a crash and restart of the destination messages are still deliveredto the new actor incarnation
These semantics match precisely what an ActorPath represents (see Actor Lifecycle), therefore you need to
supply a path and not a reference when constructing Deliver messages.
If a processor wants to reply to a Persistent message sender it should use the getSender() path as channel
destination.
3.8. Persistence
140
A channel keeps messages in memory until their successful delivery has been confirmed or the maximum number
of re-deliveries is reached. To be notified about messages that have reached the maximum number of re-deliveries,
applications can register a listener at channel creation.
class MyListener extends UntypedActor {
@Override
public void onReceive(Object message) throws Exception {
if (message instanceof RedeliverFailure) {
Iterable<ConfirmablePersistent> messages =
((RedeliverFailure)message).getMessages();
// ...
}
}
}
final ActorRef myListener = getContext().actorOf(Props.create(MyListener.class));
getContext().actorOf(Channel.props(
ChannelSettings.create().withRedeliverFailureListener(null)));
A listener receives RedeliverFailure notifications containing all messages that could not be delivered. On
receiving a RedeliverFailure message, an application may decide to restart the sending processor to enforce
a re-send of these messages to the channel or confirm these messages to prevent further re-sends. The sending
processor can also be restarted any time later to re-send unconfirmed messages.
This combination of
message persistence by sending processors
message replays by sending processors
message re-deliveries by channels and
application-level confirmations (acknowledgements) by destinations
enables channels to provide at-least-once message delivery semantics. Possible duplicates can be detected by
destinations by tracking message sequence numbers. Message sequence numbers are generated per sending processor. Depending on how a processor routes outbound messages to destinations, they may either see a contiguous
message sequence or a sequence with gaps.
3.8. Persistence
141
Warning: If a processor emits more than one outbound message per inbound Persistent message it
must use a separate channel for each outbound message to ensure that confirmations are uniquely identifiable,
otherwise, at-least-once message delivery semantics do not apply. This rule has been introduced to avoid
writing additional outbound message identifiers to the journal which would decrease the overall throughput.
It is furthermore recommended to collapse multiple outbound messages to the same destination into a single
outbound message, otherwise, if sent via multiple channels, their ordering is not defined.
If an application wants to have more control how sequence numbers are assigned to messages it should use an
application-specific sequence number generator and include the generated sequence numbers into the payload
of Persistent messages.
Persistent channels
Channels created with Channel.props do not persist messages. These channels are usually used in combination with a sending processor that takes care of persistence, hence, channel-specific persistence is not necessary in
this case. They are referred to as transient channels in the following.
Persistent channels are like transient channels but additionally persist messages before delivering them. Messages that have been persisted by a persistent channel are deleted when destinations confirm their delivery. A persistent channel can be created with PersistentChannel.props and configured with a
PersistentChannelSettings object.
final ActorRef channel = getContext().actorOf(PersistentChannel.props(
PersistentChannelSettings.create()
.withRedeliverInterval(Duration.create(30, TimeUnit.SECONDS))
.withRedeliverMax(15)), "myPersistentChannel");
channel.tell(Deliver.create(Persistent.create("example"), destination.path()), getSelf());
A persistent channel is useful for delivery of messages to slow destinations or destinations that are unavailable for
a long time. It can constrain the number of pending confirmations based on the pendingConfirmationsMax
and pendingConfirmationsMin parameters of PersistentChannelSettings.
PersistentChannelSettings.create()
.withPendingConfirmationsMax(10000)
.withPendingConfirmationsMin(2000);
It suspends delivery when the number of pending confirmations reaches pendingConfirmationsMax and resumes delivery again when this number falls below pendingConfirmationsMin. This prevents both, flooding destinations with more messages than they can process and unlimited memory consumption by the channel.
A persistent channel continues to persist new messages even when message delivery is temporarily suspended.
Standalone usage
Applications may also use channels standalone. Transient channels can be used standalone if re-delivery attempts
to destinations are required but message loss in case of a sender JVM crash is not an issue. If message loss in
case of a sender JVM crash is an issue, persistent channels should be used. In this case, applications may want to
receive replies from the channel whether messages have been successfully persisted or not. This can be enabled
by creating the channel with the replyPersistent configuration parameter set to true:
PersistentChannelSettings.create().withReplyPersistent(true);
With this setting, either the successfully persisted message is replied to the sender or a PersistenceFailure
message. In case the latter case, the sender should re-send the message.
Identifiers
In the same way as Processors and Views, channels also have an identifier that defaults to a channels path.
A channel identifier can therefore be customized by using a custom actor name at channel creation. This
3.8. Persistence
142
changes that channels name in its actor hierarchy and hence influences only part of the channel identifier.
To fully customize a channel identifier, it should be provided as argument Channel.props(String) or
PersistentChannel.props(String) (recommended to generate stable identifiers).
this.channel = getContext().actorOf(Channel.props("my-stable-channel-id"));
This is necessary for delivery confirmations to work properly. Both ways are equivalent but we recommend using
p.withPayload(...) for clarity. It is not allowed to send a message via a channel that has been created with
Persistent.create(...). This would redeliver the message on every replay even though its delivery was
confirmed by a destination.
Sequence number
The sequence number of a Persistent message can be obtained via its sequenceNr method. Persistent
messages are assigned sequence numbers on a per-processor basis (or per channel basis if used standalone). A
sequence starts at 1L and doesnt contain gaps unless a processor deletes messages.
3.8.7 Snapshots
Snapshots can dramatically reduce recovery times of processors and views. The following discusses snapshots in
context of processors but this is also applicable to views.
Processors can save snapshots of internal state by calling the saveSnapshot method. If saving of a snapshot
succeeds, the processor receives a SaveSnapshotSuccess message, otherwise a SaveSnapshotFailure
message
class MyProcessor extends UntypedProcessor {
private Object state;
@Override
public void onReceive(Object message) throws Exception {
if (message.equals("snap")) {
saveSnapshot(state);
} else if (message instanceof SaveSnapshotSuccess) {
SnapshotMetadata metadata = ((SaveSnapshotSuccess)message).metadata();
// ...
} else if (message instanceof SaveSnapshotFailure) {
SnapshotMetadata metadata = ((SaveSnapshotFailure)message).metadata();
// ...
}
}
}
During recovery, the processor is offered a previously saved snapshot via a SnapshotOffer message from
which it can initialize internal state.
3.8. Persistence
143
The replayed messages that follow the SnapshotOffer message, if any, are younger than the offered snapshot.
They finally recover the processor to its current (i.e. latest) state.
In general, a processor is only offered a snapshot if that processor has previously saved one or more snapshots
and at least one of these snapshots matches the SnapshotSelectionCriteria that can be specified for
recovery.
processor.tell(Recover.create(SnapshotSelectionCriteria.create(457L, System.currentTimeMillis())),
3.8. Persistence
144
import akka.actor.*;
import akka.japi.Procedure;
import akka.persistence.*;
import static java.util.Arrays.asList;
class Cmd implements Serializable {
private final String data;
public Cmd(String data) {
this.data = data;
}
public String getData() {
return data;
}
}
class Evt implements Serializable {
private final String data;
public Evt(String data) {
this.data = data;
}
public String getData() {
return data;
}
}
class ExampleState implements Serializable {
private final ArrayList<String> events;
public ExampleState() {
this(new ArrayList<String>());
}
public ExampleState(ArrayList<String> events) {
this.events = events;
}
public ExampleState copy() {
return new ExampleState(new ArrayList<String>(events));
}
public void update(Evt evt) {
events.add(evt.getData());
}
public int size() {
return events.size();
}
@Override
public String toString() {
return events.toString();
}
}
class ExampleProcessor extends UntypedEventsourcedProcessor {
private ExampleState state = new ExampleState();
public int getNumEvents() {
3.8. Persistence
145
return state.size();
}
public void onReceiveRecover(Object msg) {
if (msg instanceof Evt) {
state.update((Evt) msg);
} else if (msg instanceof SnapshotOffer) {
state = (ExampleState)((SnapshotOffer)msg).snapshot();
}
}
public void onReceiveCommand(Object msg) {
if (msg instanceof Cmd) {
final String data = ((Cmd)msg).getData();
final Evt evt1 = new Evt(data + "-" + getNumEvents());
final Evt evt2 = new Evt(data + "-" + (getNumEvents() + 1));
persist(asList(evt1, evt2), new Procedure<Evt>() {
public void apply(Evt evt) throws Exception {
state.update(evt);
if (evt.equals(evt2)) {
getContext().system().eventStream().publish(evt);
}
}
});
} else if (msg.equals("snap")) {
// IMPORTANT: create a copy of snapshot
// because ExampleState is mutable !!!
saveSnapshot(state.copy());
} else if (msg.equals("print")) {
System.out.println(state);
}
}
}
The example defines two data types, Cmd and Evt to represent commands and events, respectively. The state
of the ExampleProcessor is a list of persisted event data contained in ExampleState.
The processors onReceiveRecover method defines how state is updated during recovery by handling Evt
and SnapshotOffer messages. The processors onReceiveCommand method is a command handler. In
this example, a command is handled by generating two events which are then persisted and handled. Events are
persisted by calling persist with an event (or a sequence of events) as first argument and an event handler as
second argument.
The persist method persists events asynchronously and the event handler is executed for successfully persisted
events. Successfully persisted events are internally sent back to the processor as individual messages that trigger
event handler executions. An event handler may close over processor state and mutate it. The sender of a persisted
event is the sender of the corresponding command. This allows event handlers to reply to the sender of a command
(not shown).
The main responsibility of an event handler is changing processor state using event data and notifying others about
successful state changes by publishing events.
When persisting events with persist it is guaranteed that the processor will not receive further commands
between the persist call and the execution(s) of the associated event handler. This also holds for multiple
persist calls in context of a single command.
The easiest way to run this example yourself is to download Typesafe Activator and open the tutorial named Akka
Persistence Samples with Java. It contains instructions on how to run the EventsourcedExample.
Note: Its also possible to switch between different command handlers during normal processing and recovery
with getContext().become() and getContext().unbecome(). To get the actor into the same state
after recovery you need to take special care to perform the same state transitions with become and unbecome
in the receiveRecover method as you would have done in the command handler.
3.8. Persistence
146
In larger integration scenarios, channel destinations may be actors that submit received events to an external
message broker, for example. After having successfully submitted an event, they should call confirm() on the
received ConfirmablePersistent message.
A new batch write is triggered by a processor as soon as a batch reaches the maximum size or if the journal
completed writing the previous batch. Batch writes are never timer-based which keeps latencies at a minimum.
Applications that want to have more explicit control over batch writes and batch sizes can send processors
PersistentBatch messages.
3.8. Persistence
147
Persistent messages contained in a PersistentBatch are always written atomically, even if the batch
size is greater than max-message-batch-size. Also, a PersistentBatch is written isolated from other
batches. Persistent messages contained in a PersistentBatch are received individually by a processor.
PersistentBatch messages, for example, are used internally by an UntypedEventsourcedProcessor
to ensure atomic writes of events. All events that are persisted in context of a single command are written as
a single batch to the journal (even if persist is called multiple times per command). The recovery of an
UntypedEventsourcedProcessor will therefore never be done partially (with only a subset of events
persisted by a single command).
Confirmation
and
deletion
operations
performed
by
Channels
batched.
The
maximum
confirmation
and
deletion
batch
sizes
are
with
akka.persistence.journal.max-confirmation-batch-size
akka.persistence.journal.max-deletion-batch-size, respectively.
are
also
configurable
and
akka.actor.UntypedActor;
scala.concurrent.Future;
akka.japi.Option;
akka.japi.Procedure;
akka.persistence.*;
akka.persistence.journal.japi.*;
akka.persistence.snapshot.japi.*;
3.8. Persistence
148
/**
* Java API, Plugin API: synchronously writes a batch of persistent messages to the
* journal. The batch write must be atomic i.e. either all persistent messages in the
* batch are written or none.
*/
void doWriteMessages(Iterable<PersistentRepr> messages);
/**
* Java API, Plugin API: synchronously writes a batch of delivery confirmations to
* the journal.
*/
void doWriteConfirmations(Iterable<PersistentConfirmation> confirmations);
/**
* Java API, Plugin API: synchronously deletes messages identified by messageIds
* from the journal. If permanent is set to false, the persistent messages are
* marked as deleted, otherwise they are permanently deleted.
*/
void doDeleteMessages(Iterable<PersistentId> messageIds, boolean permanent);
/**
* Java API, Plugin API: synchronously deletes all persistent messages up to
* toSequenceNr. If permanent is set to false, the persistent messages are
* marked as deleted, otherwise they are permanently deleted.
*
* @see AsyncRecoveryPlugin
*/
void doDeleteMessagesTo(String processorId, long toSequenceNr, boolean permanent);
AsyncWriteJournal is an actor that should be extended if the storage backend API supports asynchronous,
non-blocking writes. In this case, the methods to be implemented are:
/**
* Java API, Plugin API: synchronously writes a batch of persistent messages to the
* journal. The batch write must be atomic i.e. either all persistent messages in the
* batch are written or none.
*/
Future<Void> doAsyncWriteMessages(Iterable<PersistentRepr> messages);
/**
* Java API, Plugin API: synchronously writes a batch of delivery confirmations to
* the journal.
*/
Future<Void> doAsyncWriteConfirmations(Iterable<PersistentConfirmation> confirmations);
/**
* Java API, Plugin API: synchronously deletes messages identified by messageIds
* from the journal. If permanent is set to false, the persistent messages are
* marked as deleted, otherwise they are permanently deleted.
*/
Future<Void> doAsyncDeleteMessages(Iterable<PersistentId> messageIds, boolean permanent);
/**
* Java API, Plugin API: synchronously deletes all persistent messages up to
* toSequenceNr. If permanent is set to false, the persistent messages are
* marked as deleted, otherwise they are permanently deleted.
*
* @see AsyncRecoveryPlugin
*/
Future<Void> doAsyncDeleteMessagesTo(String processorId, long toSequenceNr, boolean permanent);
Message replays and sequence number recovery are always asynchronous, therefore, any journal plugin must
implement:
3.8. Persistence
149
/**
* Java API, Plugin API: asynchronously replays persistent messages.
* Implementations replay a message by calling replayCallback. The returned
* future must be completed when all messages (matching the sequence number
* bounds) have been replayed. The future must be completed with a failure if
* any of the persistent messages could not be replayed.
*
* The replayCallback must also be called with messages that have been marked
* as deleted. In this case a replayed messages deleted method must return
* true.
*
* The channel ids of delivery confirmations that are available for a replayed
* message must be contained in that messages confirms sequence.
*
* @param processorId processor id.
* @param fromSequenceNr sequence number where replay should start (inclusive).
* @param toSequenceNr sequence number where replay should end (inclusive).
* @param max maximum number of messages to be replayed.
* @param replayCallback called to replay a single message. Can be called from any
thread.
*
*/
Future<Void> doAsyncReplayMessages(String processorId, long fromSequenceNr, long toSequenceNr, lon
/**
* Java API, Plugin API: asynchronously reads the highest stored sequence number
* for the given processorId.
*
* @param processorId processor id.
* @param fromSequenceNr hint where to start searching for the highest sequence
number.
*
*/
Future<Long> doAsyncReadHighestSequenceNr(String processorId, long fromSequenceNr);
/**
* Java API, Plugin API: asynchronously loads a snapshot.
*
* @param processorId processor id.
* @param criteria selection criteria for loading.
*/
Future<Option<SelectedSnapshot>> doLoadAsync(String processorId, SnapshotSelectionCriteria criteri
3.8. Persistence
150
/**
* Java API, Plugin API: asynchronously saves a snapshot.
*
* @param metadata snapshot metadata.
* @param snapshot snapshot.
*/
Future<Void> doSaveAsync(SnapshotMetadata metadata, Object snapshot);
/**
* Java API, Plugin API: called after successful saving of a snapshot.
*
* @param metadata snapshot metadata.
*/
void onSaved(SnapshotMetadata metadata) throws Exception;
/**
* Java API, Plugin API: deletes the snapshot identified by metadata.
*
* @param metadata snapshot metadata.
*/
void doDelete(SnapshotMetadata metadata) throws Exception;
/**
* Java API, Plugin API: deletes all snapshots matching criteria.
*
* @param processorId processor id.
* @param criteria selection criteria for deleting.
*/
void doDelete(String processorId, SnapshotSelectionCriteria criteria) throws Exception;
A snapshot store plugin can be activated with the following minimal configuration:
# Path to the snapshot store plugin to be used
akka.persistence.snapshot-store.plugin = "my-snapshot-store"
# My custom snapshot store plugin
my-snapshot-store {
# Class name of the plugin.
class = "docs.persistence.MySnapshotStore"
# Dispatcher for the plugin actor.
plugin-dispatcher = "akka.persistence.dispatchers.default-plugin-dispatcher"
}
With this plugin, each actor system runs its own private LevelDB instance.
3.8. Persistence
151
By default, the shared instance writes journaled messages to a local directory named journal in the current
working directory. The storage location can be changed by configuration:
akka.persistence.journal.leveldb-shared.store.dir = "target/shared"
Actor systems that use a shared LevelDB store must activate the akka.persistence.journal.leveldb-shared
plugin.
akka.persistence.journal.plugin = "akka.persistence.journal.leveldb-shared"
This plugin must be initialized by injecting the (remote) SharedLeveldbStore actor reference. Injection is
done by calling the SharedLeveldbJournal.setStore method with the actor reference as argument.
class SharedStorageUsage extends UntypedActor {
@Override
public void preStart() throws Exception {
String path = "akka.tcp://[email protected]:2552/user/store";
ActorSelection selection = getContext().actorSelection(path);
selection.tell(new Identify(1), getSelf());
}
@Override
public void onReceive(Object message) throws Exception {
if (message instanceof ActorIdentity) {
ActorIdentity identity = (ActorIdentity) message;
if (identity.correlationId().equals(1)) {
ActorRef store = identity.getRef();
if (store != null) {
SharedLeveldbJournal.setStore(store, getContext().system());
}
}
}
}
}
Internal journal commands (sent by processors) are buffered until injection completes. Injection is idempotent i.e.
only the first injection is used.
Local snapshot store
The default snapshot store plugin is akka.persistence.snapshot-store.local. It writes snapshot
files to the local filesystem. The default storage location is a directory named snapshots in the current working
directory. This can be changed by configuration where the specified path can be relative or absolute:
akka.persistence.snapshot-store.local.dir = "target/snapshots"
3.8. Persistence
152
3.8.13 Testing
When running tests with LevelDB default settings in sbt, make sure to set fork := true in your sbt project
otherwise, youll see an UnsatisfiedLinkError. Alternatively, you can switch to a LevelDB Java port by
setting
akka.persistence.journal.leveldb.native = off
or
akka.persistence.journal.leveldb-shared.store.native = off
in your Akka configuration. The LevelDB Java port is for testing purposes only.
3.8.14 Configuration
There are several configuration properties for the persistence module, please refer to the reference configuration.
153
Testing (multiple) encapsulated actors including multi-threaded scheduling; this implies non-deterministic
order of events but shielding from concurrency concerns by the actor model and will be called Integration
Testing in the following.
There are of course variations on the granularity of tests in both categories, where unit testing reaches down to
white-box tests and integration testing can encompass functional tests of complete actor networks. The important
distinction lies in whether concurrency concerns are part of the test or not. The tools offered are described in detail
in the following sections.
Note: Be sure to add the module akka-testkit to your dependencies.
Since TestActorRef is generic in the actor type it returns the underlying actor with its proper static type. From
this point on you may bring any unit testing tool to bear on your actor as usual.
Testing the Actors Behavior
When the dispatcher invokes the processing behavior of an actor on a message, it actually calls apply on the
current behavior registered for the actor. This starts out with the return value of the declared receive method,
but it may also be changed using become and unbecome in response to external messages. All of this contributes to the overall actor behavior and it does not lend itself to easy testing on the Actor itself. Therefore the
TestActorRef offers a different mode of operation to complement the Actor testing: it supports all operations
3.9. Testing Actor Systems
154
also valid on normal ActorRef. Messages sent to the actor are processed synchronously on the current thread
and answers may be sent back as usual. This trick is made possible by the CallingThreadDispatcher
described below (see CallingThreadDispatcher); this dispatcher is set implicitly for any actor instantiated into a
TestActorRef.
final Props props = Props.create(MyActor.class);
final TestActorRef<MyActor> ref = TestActorRef.create(system, props, "testB");
final Future<Object> future = akka.pattern.Patterns.ask(ref, "say42", 3000);
assertTrue(future.isCompleted());
assertEquals(42, Await.result(future, Duration.Zero()));
As the TestActorRef is a subclass of LocalActorRef with a few special extras, also aspects like supervision and restarting work properly, but beware that execution is only strictly synchronous as long as all actors
involved use the CallingThreadDispatcher. As soon as you add elements which include more sophisticated scheduling you leave the realm of unit testing as you then need to think about asynchronicity again (in most
cases the problem will be to wait until the desired effect had a chance to happen).
One more special aspect which is overridden for single-threaded tests is the receiveTimeout, as including
that would entail asynchronous queuing of ReceiveTimeout messages, violating the synchronous contract.
Note:
To summarize:
TestActorRef overwrites two fields:
it sets the dispatcher to
CallingThreadDispatcher.global and it sets the receiveTimeout to None.
Use Cases
You may of course mix and match both modi operandi of TestActorRef as suits your test needs:
one common use case is setting up the actor into a specific internal state before sending the test message
another is to verify correct internal state transitions after having sent the test message
Feel free to experiment with the possibilities, and if you find useful patterns, dont hesitate to let the Akka forums
know about them! Who knows, common operations might even be worked into nice DSLs.
155
The JavaTestKit class contains a collection of tools which makes this common task easy.
import
import
import
import
org.junit.AfterClass;
org.junit.Assert;
org.junit.BeforeClass;
org.junit.Test;
import
import
import
import
import
import
akka.actor.ActorRef;
akka.actor.ActorSystem;
akka.actor.Props;
akka.actor.UntypedActor;
akka.testkit.JavaTestKit;
scala.concurrent.duration.Duration;
156
The JavaTestKit contains an actor named testActor which is the entry point for messages to be examined
with the various expectMsg... assertions detailed below. The test actors reference is obtained using the
getRef method as demonstrated above. The testActor may also be passed to other actors as usual, usually
subscribing it as notification listener. There is a whole set of examination methods, e.g. receiving all consecutive
messages matching certain criteria, receiving a whole sequence of fixed messages or classes, receiving nothing for
some time, etc.
The ActorSystem passed in to the constructor of JavaTestKit is accessible via the getSystem method.
Note: Remember to shut down the actor system after the test is finished (also in case of failure) so that all
actorsincluding the test actorare stopped.
Built-In Assertions
The above mentioned expectMsgEquals is not the only method for formulating assertions concerning received
messages, the full set is this:
final String hello
final Object
any
final Object[] all
final int i
final Number j
expectNoMsg();
final Object[] two
=
=
=
=
=
expectMsgEquals("hello");
expectMsgAnyOf("hello", "world");
expectMsgAllOf("hello", "world");
expectMsgClass(Integer.class);
expectMsgAnyClassOf(Integer.class, Long.class);
= receiveN(2);
In these examples, the maximum durations you will find mentioned below are left out, in which case they use
the default value from configuration item akka.test.single-expect-default which itself defaults to 3
seconds (or they obey the innermost enclosing Within as detailed below). The full signatures are:
public <T> T expectMsgEquals(Duration max, T msg)
The given message object must be received within the specified time; the object will be returned.
157
The match(Object in) method will be evaluated once a message has been received within the allotted
time (which may be given as constructor argument). If it throws noMatch() (where it is sufficient to call
that method; the throw keyword is only needed in cases where the compiler would otherwise complain
about wrong return typesJava is lacking Scalas notion of a type which signifies will not ever return
normally), then the expectation fails with an AssertionError, otherwise the matched and possibly
transformed object is stored for retrieval using the get method.
ReceiveWhile<T>
new JavaTestKit(system) {{
getRef().tell(42, ActorRef.noSender());
getRef().tell(43, ActorRef.noSender());
158
getRef().tell("hello", ActorRef.noSender());
final String[] out =
new ReceiveWhile<String>(String.class, duration("1 second")) {
// do not put code outside this method, will run afterwards
protected String match(Object in) {
if (in instanceof Integer) {
return in.toString();
} else {
throw noMatch();
}
}
}.get(); // this extracts the received messages
assertArrayEquals(new String[] {"42", "43"}, out);
expectMsgEquals("hello");
}};
This construct works like ExpectMsg, but it continually collects messages as long as they match the criteria,
and it does not fail when a non-matching one is encountered. Collecting messages also ends when the time
is up, when too much time passes between messages or when enough messages have been received.
new ReceiveWhile<String>(
String.class,
duration("100 millis"),
duration("50 millis"),
12
) {
// match elided ...
};
//
//
//
//
//
The need to specify the String result type twice results from the need to create a correctly typed array
and Javas inability to infer the classs type argument.
AwaitCond
new JavaTestKit(system) {{
getRef().tell(42, ActorRef.noSender());
new AwaitCond(
duration("1 second"), // maximum wait time
duration("100 millis") // interval at which to check the condition
) {
// do not put code outside this method, will run afterwards
protected boolean cond() {
// typically used to wait for something to start up
return msgAvailable();
}
};
}};
This general construct is not connected with the test kits message reception, the embedded condition can
compute the boolean result from anything in scope.
AwaitAssert
new JavaTestKit(system) {{
getRef().tell(42, ActorRef.noSender());
new AwaitAssert(
duration("1 second"), // maximum wait time
duration("100 millis") // interval at which to check the condition
) {
// do not put code outside this method, will run afterwards
protected void check() {
assertEquals(msgAvailable(), true);
}
};
159
}};
This general construct is not connected with the test kits message reception, the embedded assert can check
anything in scope.
There are also cases where not all messages sent to the test kit are actually relevant to the test, but removing them
would mean altering the actors under test. For this purpose it is possible to ignore certain messages:
IgnoreMsg
new JavaTestKit(system) {{
// ignore all Strings
new IgnoreMsg() {
protected boolean ignore(Object msg) {
return msg instanceof String;
}
};
getRef().tell("hello", ActorRef.noSender());
getRef().tell(42, ActorRef.noSender());
expectMsgEquals(42);
// remove message filter
ignoreNoMsg();
getRef().tell("hello", ActorRef.noSender());
expectMsgEquals("hello");
}};
If a number of occurrences is specificas demonstrated abovethen exec() will block until that number of
matching messages have been received or the timeout configured in akka.test.filter-leeway is used up
(time starts counting after the run() method returns). In case of a timeout the test fails.
Note: Be sure to exchange the default logger with the TestEventListener in your application.conf
to enable this function:
akka.loggers = [akka.testkit.TestEventListener]
Timing Assertions
Another important part of functional testing concerns timing: certain events must not happen immediately (like a
timer), others need to happen before a deadline. Therefore, all examination methods accept an upper time limit
3.9. Testing Actor Systems
160
within the positive or negative result must be obtained. Lower time limits need to be checked external to the
examination, which is facilitated by a new construct for managing time constraints:
new JavaTestKit(system) {{
getRef().tell(42, ActorRef.noSender());
new Within(Duration.Zero(), Duration.create(1, "second")) {
// do not put code outside this method, will run afterwards
public void run() {
assertEquals((Integer) 42, expectMsgClass(Integer.class));
}
};
}};
The block in Within.run must complete after a Duration which is between min and max, where the former
defaults to zero. The deadline calculated by adding the max parameter to the blocks start time is implicitly
available within the block to all examination methods, if you do not specify it, it is inherited from the innermost
enclosing within block.
It should be noted that if the last message-receiving assertion of the block is expectNoMsg or receiveWhile,
the final check of the within is skipped in order to avoid false positives due to wake-up latencies. This means
that while individual contained assertions still use the maximum time bound, the overall block may take arbitrarily
longer in this case.
Note: All times are measured using System.nanoTime, meaning that they describe wall time, not CPU time
or system time.
The tight timeouts you use during testing on your lightning-fast notebook will invariably lead to spurious test
failures on the heavily loaded Jenkins server (or similar). To account for this situation, all maximum durations are
internally scaled by a factor taken from the Configuration, akka.test.timefactor, which defaults to 1.
You can scale other durations with the same factor by using dilated method in JavaTestKit.
new JavaTestKit(system) {{
final Duration original = duration("1 second");
final Duration stretched = dilated(original);
assertTrue("dilated", stretched.gteq(original));
}};
161
This simple test verifies an equally simple Forwarder actor by injecting a probe as the forwarders target. Another
example would be two actors A and B which collaborate by A sending messages to B. In order to verify this
message flow, a TestProbe could be inserted as target of A, using the forwarding capabilities or auto-pilot
described below to include a real B in the test setup.
Probes may also be equipped with custom assertions to make your test code even more concise and clear:
new JavaTestKit(system) {{
class MyProbe extends JavaTestKit {
public MyProbe() {
super(system);
}
public void assertHello() {
expectMsgEquals("hello");
}
}
final MyProbe probe = new MyProbe();
probe.getRef().tell("hello", ActorRef.noSender());
probe.assertHello();
}};
You have complete flexibility here in mixing and matching the JavaTestKit facilities with your own checks
and choosing an intuitive name for it. In real life your code will probably be a bit more complicated than the
example given above; just use the power!
Warning: Any message send from a TestProbe to another actor which runs on the CallingThreadDispatcher runs the risk of dead-lock, if that other actor might also send to this probe. The implementation of
TestProbe.watch and TestProbe.unwatch will also send a message to the watchee, which means
that it is dangerous to try watching e.g. TestActorRef from a TestProbe.
162
The probe stores the sender of the last dequeued message (i.e. after its expectMsg* reception), which may be
retrieved using the getLastSender method. This information can also implicitly be used for having the probe
reply to the last received message:
new JavaTestKit(system) {{
final JavaTestKit probe = new JavaTestKit(system);
probe.getRef().tell("hello", getRef());
probe.expectMsgEquals("hello");
probe.reply("world");
expectMsgEquals("world");
assertEquals(probe.getRef(), getLastSender());
}};
The probe can also forward a received message (i.e. after its expectMsg* reception), retaining the original
sender:
new JavaTestKit(system) {{
final JavaTestKit probe = new JavaTestKit(system);
probe.getRef().tell("hello", getRef());
probe.expectMsgEquals("hello");
probe.forward(getRef());
expectMsgEquals("hello");
assertEquals(getRef(), getLastSender());
}};
Auto-Pilot
Receiving messages in a queue for later inspection is nice, but in order to keep a test running and verify traces later
you can also install an AutoPilot in the participating test probes (actually in any TestKit) which is invoked
before enqueueing to the inspection queue. This code can be used to forward messages, e.g. in a chain A -->
Probe --> B, as long as a certain protocol is obeyed.
new JavaTestKit(system) {{
final JavaTestKit probe = new JavaTestKit(system);
// install auto-pilot
probe.setAutoPilot(new TestActor.AutoPilot() {
public AutoPilot run(ActorRef sender, Object msg) {
sender.tell(msg, ActorRef.noSender());
return noAutoPilot();
}
});
// first one is replied to directly ...
probe.getRef().tell("hello", getRef());
expectMsgEquals("hello");
// ... but then the auto-pilot switched itself off
probe.getRef().tell("world", getRef());
expectNoMsg();
}};
The run method must return the auto-pilot for the next message, wrapped in an Option; setting it to None
terminates the auto-pilot.
163
The behavior of within blocks when using test probes might be perceived as counter-intuitive: you need to
remember that the nicely scoped deadline as described above is local to each probe. Hence, probes do not react to
each others deadlines or to the deadline set in an enclosing JavaTestKit instance:
new JavaTestKit(system) {{
final JavaTestKit probe = new JavaTestKit(system);
new Within(duration("1 second")) {
public void run() {
probe.expectMsgEquals("hello");
}
};
}};
3.9.3 CallingThreadDispatcher
The CallingThreadDispatcher serves good purposes in unit testing, as described above, but originally
it was conceived in order to allow contiguous stack traces to be generated in case of an error. As this special
dispatcher runs everything which would normally be queued directly on the current thread, the full history of a
messages processing chain is recorded on the call stack, so long as all intervening actors run on this dispatcher.
How to use it
Just set the dispatcher as you normally would:
system.actorOf(
Props.create(MyActor.class)
.withDispatcher(CallingThreadDispatcher.Id()));
How it works
When receiving an invocation, the CallingThreadDispatcher checks whether the receiving actor is already
active on the current thread. The simplest example for this situation is an actor which sends a message to itself.
In this case, processing cannot continue immediately as that would violate the actor model, so the invocation
is queued and will be processed when the active invocation on that actor finishes its processing; thus, it will
be processed on the calling thread, but simply after the actor finishes its previous work. In the other case, the
invocation is simply processed immediately on the current thread. Futures scheduled via this dispatcher are also
executed immediately.
This scheme makes the CallingThreadDispatcher work like a general purpose dispatcher for any actors
which never block on external events.
In the presence of multiple threads it may happen that two invocations of an actor running on this dispatcher
happen on two different threads at the same time. In this case, both will be processed directly on their respective
threads, where both compete for the actors lock and the loser has to wait. Thus, the actor model is left intact, but
the price is loss of concurrency due to limited scheduling. In a sense this is equivalent to traditional mutex style
concurrency.
The other remaining difficulty is correct handling of suspend and resume: when an actor is suspended, subsequent
invocations will be queued in thread-local queues (the same ones used for queuing in the normal case). The call
to resume, however, is done by one specific thread, and all other threads in the system will probably not be
executing this specific actor, which leads to the problem that the thread-local queues cannot be emptied by their
native threads. Hence, the thread calling resume will collect all currently queued invocations from all threads
into its own queue and process them.
164
Limitations
Warning: In case the CallingThreadDispatcher is used for top-level actors, but without going through TestActorRef, then there is a time window during which the actor is awaiting construction by the user guardian actor.
Sending messages to the actor during this time period will result in them being enqueued and then executed on
the guardians thread instead of the callers thread. To avoid this, use TestActorRef.
If an actors behavior blocks on a something which would normally be affected by the calling actor after having
sent the message, this will obviously dead-lock when using this dispatcher. This is a common scenario in actor
tests based on CountDownLatch for synchronization:
val latch = new CountDownLatch(1)
actor ! startWorkAfter(latch)
// actor will call latch.await() before proceeding
doSomeSetupStuff()
latch.countDown()
The example would hang indefinitely within the message processing initiated on the second line and never reach
the fourth line, which would unblock it on a normal dispatcher.
Thus, keep in mind that the CallingThreadDispatcher is not a general-purpose replacement for the normal
dispatchers. On the other hand it may be quite useful to run your actor network on it for testing, because if it runs
without dead-locking chances are very high that it will not dead-lock in production.
Warning: The above sentence is unfortunately not a strong guarantee, because your code might directly or
indirectly change its behavior when running on a different dispatcher. If you are looking for a tool to help
you debug dead-locks, the CallingThreadDispatcher may help with certain error scenarios, but keep
in mind that it has may give false negatives as well as false positives.
Thread Interruptions
If the CallingThreadDispatcher sees that the current thread has its isInterrupted() flag set when message
processing returns, it will throw an InterruptedException after finishing all its processing (i.e. all messages which need processing as described above are processed before this happens). As tell cannot throw
exceptions due to its contract, this exception will then be caught and logged, and the threads interrupted status
will be set again.
If during message processing an InterruptedException is thrown then it will be caught inside the CallingThreadDispatchers message handling loop, the threads interrupted flag will be set and processing continues
normally.
Note: The summary of these two paragraphs is that if the current thread is interrupted while doing work under
the CallingThreadDispatcher, then that will result in the isInterrupted flag to be true when the message
send returns and no InterruptedException will be thrown.
Benefits
To summarize, these are the features with the CallingThreadDispatcher has to offer:
Deterministic execution of single-threaded tests while retaining nearly full actor semantics
Full message processing history leading up to the point of failure in exception stack traces
Exclusion of certain classes of dead-lock scenarios
165
3.9.5 Configuration
There are several configuration properties for the TestKit module, please refer to the reference configuration.
166
CHAPTER
FOUR
akka.actor.AbstractActor;
akka.event.Logging;
akka.event.LoggingAdapter;
akka.japi.pf.ReceiveBuilder;
167
Please note that the Akka Actor receive message loop is exhaustive, which is different compared to Erlang and
the late Scala Actors. This means that you need to provide a pattern match for all messages that it can accept
and if you want to be able to handle unknown messages then you need to have a default case as in the example
above. Otherwise an akka.actor.UnhandledMessage(message, sender, recipient) will be
published to the ActorSystems EventStream.
Note further that the return type of the behavior defined above is Unit; if the actor shall reply to the received
message then this must be done explicitly as explained below.
The argument to the receive method is a partial function object, which is stored within the actor as its initial
behavior, see Become/Unbecome for further information on changing the behavior of an actor after its construction.
Props
Props is a configuration class to specify options for the creation of actors, think of it as an immutable and thus
freely shareable recipe for creating an actor including associated deployment information (e.g. which dispatcher
to use, see more below). Here are some examples of how to create a Props instance.
import akka.actor.Props;
Props props1 = Props.create(MyActor.class);
Props props2 = Props.create(ActorWithArgs.class,
() -> new ActorWithArgs("arg")); // careful, see below
Props props3 = Props.create(ActorWithArgs.class, "arg");
The second variant shows how to pass constructor arguments to the Actor being created, but it should only be
used outside of actors as explained below.
The last line shows a possibility to pass constructor arguments regardless of the context it is being used in.
The presence of a matching constructor is verified during construction of the Props object, resulting in an
IllegalArgumentEception if no or multiple matching constructors are found.
Dangerous Variants
// NOT RECOMMENDED within another actor:
// encourages to close over enclosing class
Props props7 = Props.create(ActorWithArgs.class,
() -> new ActorWithArgs("arg"));
This method is not recommended to be used within another actor because it encourages to close over the enclosing scope, resulting in non-serializable Props and possibly race conditions (breaking the actor encapsulation).
On the other hand using this variant in a Props factory in the actors companion object as documented under
Recommended Practices below is completely fine.
There were two use-cases for these methods: passing constructor arguments to the actorwhich is solved by the
newly introduced Props.create(clazz, args) method above or the recommended practice belowand
168
creating actors on the spot as anonymous classes. The latter should be solved by making these actors named
classes instead (if they are not declared within a top-level object then the enclosing instances this reference
needs to be passed as the first argument).
Warning: Declaring one actor within another is very dangerous and breaks actor encapsulation. Never pass
an actors this reference into Props!
Recommended Practices
It is a good idea to provide factory methods on the companion object of each Actor which help keeping the
creation of suitable Props as close to the actor definition as possible. This also avoids the pitfalls associated with
using the Props.create(...) method which takes a by-name argument, since within a companion object
the given code block will not retain a reference to its enclosing scope:
public class DemoActor extends AbstractActor {
/**
* Create Props for an actor of this type.
* @param magicNumber The magic number to be passed to this actors constructor.
* @return a Props for creating this actor, which can then be further configured
(e.g. calling .withDispatcher() on it)
*
*/
static Props props(Integer magicNumber) {
// You need to specify the actual type of the returned actor
// since Java 8 lambdas have some runtime type information erased
return Props.create(DemoActor.class, () -> new DemoActor(magicNumber));
}
private final Integer magicNumber;
DemoActor(Integer magicNumber) {
this.magicNumber = magicNumber;
receive(ReceiveBuilder.
match(Integer.class, i -> {
sender().tell(i + magicNumber, self());
}).build()
);
}
}
public class SomeOtherActor extends AbstractActor {
// Props(new DemoActor(42)) would not be safe
ActorRef demoActor = context().actorOf(DemoActor.props(42), "demo");
// ...
}
Using the ActorSystem will create top-level actors, supervised by the actor systems provided guardian actor,
while using an actors context will create a child actor.
169
It is recommended to create a hierarchy of children, grand-children and so on such that it fits the logical failurehandling structure of the application, see Actor Systems.
The call to actorOf returns an instance of ActorRef. This is a handle to the actor instance and the only way to
interact with it. The ActorRef is immutable and has a one to one relationship with the Actor it represents. The
ActorRef is also serializable and network-aware. This means that you can serialize it, send it over the wire and
use it on a remote host and it will still be representing the same Actor on the original node, across the network.
The name parameter is optional, but you should preferably name your actors, since that is used in log messages
and for identifying actors. The name must not be empty or start with $, but it may contain URL encoded characters (eg. %20 for a blank space). If the given name is already in use by another child to the same parent an
InvalidActorNameException is thrown.
Actors are automatically started asynchronously when created.
Dependency Injection
If your UntypedActor has a constructor that takes parameters then those need to be part of the Props as well, as
described above. But there are cases when a factory method must be used, for example when the actual constructor
arguments are determined by a dependency injection framework.
import akka.actor.Actor;
import akka.actor.IndirectActorProducer;
class DependencyInjector implements IndirectActorProducer {
final Object applicationContext;
final String beanName;
public DependencyInjector(Object applicationContext, String beanName) {
this.applicationContext = applicationContext;
this.beanName = beanName;
}
@Override
public Class<? extends Actor> actorClass() {
return MyActor.class;
}
@Override
public MyActor produce() {
MyActor result;
// obtain fresh Actor instance from DI framework ...
return result;
}
}
final ActorRef myActor = getContext().actorOf(
Props.create(DependencyInjector.class, applicationContext, "MyActor"),
"myactor3");
Warning: You might be tempted at times to offer an IndirectActorProducer which always returns
the same instance, e.g. by using a static field. This is not supported, as it goes against the meaning of an actor
restart, which is described here: What Restarting Means.
When using a dependency injection framework, actor beans MUST NOT have singleton scope.
Techniques for dependency injection and integration with dependency injection frameworks are described in more
170
depth in the Using Akka with Dependency Injection guideline and the Akka Java Spring tutorial in Typesafe
Activator.
The Inbox
When writing code outside of actors which shall communicate with actors, the ask pattern can be a solution (see
below), but there are two thing it cannot do: receiving multiple replies (e.g. by subscribing an ActorRef to a
notification service) and watching other actors lifecycle. For these purposes there is the Inbox class:
final Inbox inbox = Inbox.create(system);
inbox.send(target, "hello");
assert inbox.receive(Duration.create(1, TimeUnit.SECONDS)).equals("world");
The send method wraps a normal tell and supplies the internal actors reference as the sender. This allows the
reply to be received on the last line. Watching an actor is quite simple as well:
final Inbox inbox = Inbox.create(system);
inbox.watch(target);
target.tell(PoisonPill.getInstance(), ActorRef.noSender());
assert inbox.receive(Duration.create(1, TimeUnit.SECONDS)) instanceof Terminated;
171
The implementations shown above are the defaults provided by the AbstractActor class.
Actor Lifecycle
A path in an actor system represents a place which might be occupied by a living actor. Initially (apart from
system initialized actors) a path is empty. When actorOf() is called it assigns an incarnation of the actor
described by the passed Props to the given path. An actor incarnation is identified by the path and a UID. A
restart only swaps the Actor instance defined by the Props but the incarnation and hence the UID remains the
same.
The lifecycle of an incarnation ends when the actor is stopped. At that point the appropriate lifecycle events are
called and watching actors are notified of the termination. After the incarnation is stopped, the path can be reused
4.1. Actors (Java with Lambda Support)
172
again by creating an actor with actorOf(). In this case the name of the new incarnation will be the same as the
previous one but the UIDs will differ.
An ActorRef always represents an incarnation (path and UID) not just a given path. Therefore if an actor is
stopped and a new one with the same name is created an ActorRef of the old incarnation will not point to the
new one.
ActorSelection on the other hand points to the path (or multiple paths if wildcards are used) and is completely
oblivious to which incarnation is currently occupying it. ActorSelection cannot be watched for this reason.
It is possible to resolve the current incarnations ActorRef living under the path by sending an Identify
message to the ActorSelection which will be replied to with an ActorIdentity containing the correct
reference (see Identifying Actors via Actor Selection). This can also be done with the resolveOne method of
the ActorSelection, which returns a Future of the matching ActorRef.
Lifecycle Monitoring aka DeathWatch
In order to be notified when another actor terminates (i.e. stops permanently, not temporary failure and restart), an
actor may register itself for reception of the Terminated message dispatched by the other actor upon termination
(see Stopping Actors). This service is provided by the DeathWatch component of the actor system.
Registering a monitor is easy:
public class WatchActor extends AbstractActor {
private final ActorRef child = context().actorOf(Props.empty(), "target");
private ActorRef lastSender = system.deadLetters();
public WatchActor() {
context().watch(child); // <-- this is the only call needed for registration
receive(ReceiveBuilder.
matchEquals("kill", s -> {
context().stop(child);
lastSender = sender();
}).
match(Terminated.class, t -> t.actor().equals(child), t -> {
lastSender.tell("finished", self());
}).build()
);
}
}
It should be noted that the Terminated message is generated independent of the order in which registration and
termination occur. In particular, the watching actor will receive a Terminated message even if the watched
actor has already been terminated at the time of registration.
Registering multiple times does not necessarily lead to multiple messages being generated, but there is no guarantee that only exactly one such message is received: if termination of the watched actor has generated and queued
the message, and another registration is done before this message has been processed, then a second message will
be queued, because registering for monitoring of an already terminated actor leads to the immediate generation of
the Terminated message.
It is also possible to deregister from watching another actors liveliness using context.unwatch(target).
This works even if the Terminated message has already been enqueued in the mailbox; after calling unwatch
no Terminated message for that actor will be processed anymore.
Start Hook
Right after starting the actor, its preStart method is invoked.
@Override
public void preStart() {
173
This method is called when the actor is first created. During restarts it is called by the default implementation of
postRestart, which means that by overriding that method you can choose whether the initialization code in
this method is called only exactly once for this actor or for every restart. Initialization code which is part of the
actors constructor will always be called when an instance of the actor class is created, which happens at every
restart.
Restart Hooks
All actors are supervised, i.e. linked to another actor with a fault handling strategy. Actors may be restarted in
case an exception is thrown while processing a message (see Supervision and Monitoring). This restart involves
the hooks mentioned above:
1. The old actor is informed by calling preRestart with the exception which caused the restart and the
message which triggered that exception; the latter may be None if the restart was not caused by processing
a message, e.g. when a supervisor does not trap the exception and is restarted in turn by its supervisor, or if
an actor is restarted due to a siblings failure. If the message is available, then that messages sender is also
accessible in the usual way (i.e. by calling sender).
This method is the best place for cleaning up, preparing hand-over to the fresh actor instance, etc. By default
it stops all children and calls postStop.
2. The initial factory from the actorOf call is used to produce the fresh instance.
3. The new actors postRestart method is invoked with the exception which caused the restart. By default
the preStart is called, just as in the normal start-up case.
An actor restart replaces only the actual actor object; the contents of the mailbox is unaffected by the restart,
so processing of messages will resume after the postRestart hook returns. The message that triggered the
exception will not be received again. Any message sent to an actor while it is being restarted will be queued to its
mailbox as usual.
Warning: Be aware that the ordering of failure notifications relative to user messages is not deterministic. In
particular, a parent might restart its child before it has processed the last messages sent by the child before the
failure. See Discussion: Message Ordering for details.
Stop Hook
After stopping an actor, its postStop hook is called, which may be used e.g. for deregistering this actor from
other services. This hook is guaranteed to run after message queuing has been disabled for this actor, i.e. messages
sent to a stopped actor will be redirected to the deadLetters of the ActorSystem.
174
The supplied path is parsed as a java.net.URI, which basically means that it is split on / into path elements.
If the path starts with /, it is absolute and the look-up starts at the root guardian (which is the parent of "/user");
otherwise it starts at the current actor. If a path element equals .., the look-up will take a step up towards the
supervisor of the currently traversed actor, otherwise it will step down to the named child. It should be noted
that the .. in actor paths here always means the logical structure, i.e. the supervisor.
The path elements of an actor selection may contain wildcard patterns allowing for broadcasting of messages to
that section:
// will look all children to serviceB with names starting with worker
context().actorSelection("/user/serviceB/worker*");
// will look up all siblings beneath same supervisor
context().actorSelection("../*");
Messages can be sent via the ActorSelection and the path of the ActorSelection is looked up when
delivering each message. If the selection does not match any actors the message will be dropped.
To acquire an ActorRef for an ActorSelection you need to send a message to the selection and use the
sender() reference of the reply from the actor. There is a built-in Identify message that all Actors will
understand and automatically reply to with a ActorIdentity message containing the ActorRef. This message is handled specially by the actors which are traversed in the sense that if a concrete name lookup fails (i.e.
a non-wildcard path element does not correspond to a live actor) then a negative result is generated. Please note
that this does not mean that delivery of that reply is guaranteed, it still is a normal message.
import akka.actor.ActorIdentity;
import akka.actor.ActorSelection;
import akka.actor.Identify;
public class Follower extends AbstractActor {
final Integer identifyId = 1;
public Follower(){
ActorSelection selection = context().actorSelection("/user/another");
selection.tell(new Identify(identifyId), self());
receive(ReceiveBuilder.
match(ActorIdentity.class, id -> id.getRef() != null, id -> {
ActorRef ref = id.getRef();
context().watch(ref);
context().become(active(ref));
}).
match(ActorIdentity.class, id -> id.getRef() == null, id -> {
context().stop(self());
}).build()
);
}
final PartialFunction<Object, BoxedUnit> active(final ActorRef another) {
return ReceiveBuilder.
match(Terminated.class, t -> t.actor().equals(another), t -> {
context().stop(self());
}).build();
}
}
You can also acquire an ActorRef for an ActorSelection with the resolveOne method of the
ActorSelection. It returns a Future of the matching ActorRef if such an actor exists. It is completed
with failure [[akka.actor.ActorNotFound]] if no such actor exists or the identification didnt complete within the
supplied timeout.
Remote actor addresses may also be looked up, if remoting is enabled:
context().actorSelection("akka.tcp://app@otherhost:1234/user/serviceB");
175
176
The sender reference is passed along with the message and available within the receiving actor via its sender
method while processing this message. Inside of an actor it is usually self who shall be the sender, but there
can be cases where replies shall be routed to some other actore.g. the parentin which the second argument to
tell would be a different one. Outside of an actor and if no reply is needed the second argument can be null;
if a reply is needed outside of an actor you can use the ask-pattern described next..
Ask: Send-And-Receive-Future
The ask pattern involves actors as well as futures, hence it is offered as a use pattern rather than a method on
ActorRef:
import
import
import
import
import
import
import
static akka.pattern.Patterns.ask;
static akka.pattern.Patterns.pipe;
scala.concurrent.Future;
scala.concurrent.duration.Duration;
akka.dispatch.Futures;
akka.dispatch.Mapper;
akka.util.Timeout;
This example demonstrates ask together with the pipe pattern on futures, because this is likely to be a common
combination. Please note that all of the above is completely non-blocking and asynchronous: ask produces a
Future, two of which are composed into a new future using the Futures.sequence and map methods and
then pipe installs an onComplete-handler on the future to effect the submission of the aggregated Result to
another actor.
Using ask will send a message to the receiving Actor as with tell, and the receiving actor must reply with
sender().tell(reply, self()) in order to complete the returned Future with a value. The ask
operation involves creating an internal actor for handling this reply, which needs to have a timeout after which it
is destroyed in order not to leak resources; see more below.
Warning: To complete the future with an exception you need send a Failure message to the sender. This is
not done automatically when an actor throws an exception while processing a message.
try {
String result = operation();
177
sender().tell(result, self());
} catch (Exception e) {
sender().tell(new akka.actor.Status.Failure(e), self());
throw e;
}
If the actor does not complete the future, it will expire after the timeout period, specified as parameter to the ask
method; this will complete the Future with an AskTimeoutException.
See Futures for more information on how to await or query a future.
The onComplete, onSuccess, or onFailure methods of the Future can be used to register a callback to
get a notification when the Future completes. Gives you a way to avoid blocking.
Warning: When using future callbacks, inside actors you need to carefully avoid closing over the containing
actors reference, i.e. do not call methods or access mutable state on the enclosing actor from within the callback. This would break the actor encapsulation and may introduce synchronization bugs and race conditions
because the callback will be scheduled concurrently to the enclosing actor. Unfortunately there is not yet a
way to detect these illegal accesses at compile time. See also: Actors and shared mutable state
Forward message
You can forward a message from one actor to another. This means that the original sender address/reference is
maintained even though the message is going through a mediator. This can be useful when writing actors that
work as routers, load-balancers, replicators etc.
target.forward(result, context());
Both the argument to the AbstractActor receive method and the return type of the Actor receive
method is a PartialFunction<Object, BoxedUnit> that defines which messages your Actor can handle, along with the implementation of how the messages should be processed.
Dont let the type signature scare you. To allow you to easily build up a partial function there is a builder named
ReceiveBuilder that you can use.
Here is an example:
import
import
import
import
akka.actor.AbstractActor;
akka.event.Logging;
akka.event.LoggingAdapter;
akka.japi.pf.ReceiveBuilder;
178
receive(ReceiveBuilder.
match(String.class, s -> {
log.info("Received String message: {}", s);
}).
matchAny(o -> log.info("received unknown message")).build()
);
}
}
179
Termination of an actor proceeds in two steps: first the actor suspends its mailbox processing and sends a stop
command to all its children, then it keeps processing the internal termination notifications from its children until
the last one is gone, finally terminating itself (invoking postStop, dumping mailbox, publishing Terminated
on the DeathWatch, telling its supervisor). This procedure ensures that actor system sub-trees terminate in an
orderly fashion, propagating the stop command to the leaves and collecting their confirmation back to the stopped
supervisor. If one of the actors does not respond (i.e. processing a message for extended periods of time and
therefore not receiving the stop command), this whole process will be stuck.
Upon ActorSystem.shutdown, the system guardian actors will be stopped, and the aforementioned process
will ensure proper termination of the whole system.
The postStop hook is invoked after an actor is fully stopped. This enables cleaning up of resources:
@Override
public void postStop() {
// clean up some resources ...
}
Note: Since stopping an actor is asynchronous, you cannot immediately reuse the name of the child you just
stopped; this will result in an InvalidActorNameException. Instead, watch the terminating actor and
create its replacement in response to the Terminated message which will eventually arrive.
PoisonPill
You can also send an actor the akka.actor.PoisonPill message, which will stop the actor when the message is processed. PoisonPill is enqueued as ordinary messages and will be handled after messages that were
already queued in the mailbox.
Graceful Stop
gracefulStop is useful if you need to wait for termination or compose ordered termination of several actors:
import
import
import
import
import
static akka.pattern.Patterns.gracefulStop;
scala.concurrent.Await;
scala.concurrent.Future;
scala.concurrent.duration.Duration;
akka.pattern.AskTimeoutException;
try {
Future<Boolean> stopped =
gracefulStop(actorRef, Duration.create(5, TimeUnit.SECONDS), Manager.SHUTDOWN);
Await.result(stopped, Duration.create(6, TimeUnit.SECONDS));
// the actor has been stopped
} catch (AskTimeoutException e) {
// the actor wasnt stopped within 5 seconds
}
public class Manager extends AbstractActor {
private static enum Shutdown {
Shutdown
}
public static final Shutdown SHUTDOWN = Shutdown.Shutdown;
private ActorRef worker =
context().watch(context().actorOf(Props.create(Cruncher.class), "worker"));
public Manager() {
receive(ReceiveBuilder.
matchEquals("job", s -> {
180
worker.tell("crunch", self());
}).
matchEquals(SHUTDOWN, x -> {
worker.tell(PoisonPill.getInstance(), self());
context().become(shuttingDown);
}).build()
);
}
public PartialFunction<Object, BoxedUnit> shuttingDown =
ReceiveBuilder.
matchEquals("job", s -> {
sender().tell("service unavailable, shutting down", self());
}).
match(Terminated.class, t -> t.actor().equals(worker), t -> {
context().stop(self());
}).build();
}
When gracefulStop() returns successfully, the actors postStop() hook will have been executed: there
exists a happens-before edge between the end of postStop() and the return of gracefulStop().
In the above example a custom Manager.Shutdown message is sent to the target actor to initiate the process
of stopping the actor. You can use PoisonPill for this, but then you have limited possibilities to perform
interactions with other actors before stopping the target actor. Simple cleanup tasks can be handled in postStop.
Warning: Keep in mind that an actor stopping and its name being deregistered are separate events which
happen asynchronously from each other. Therefore it may be that you will find the name still in use after
gracefulStop() returned. In order to guarantee proper deregistration, only reuse names from within a
supervisor you control and only in response to a Terminated message, i.e. not for top-level actors.
4.1.10 Become/Unbecome
Upgrade
Akka supports hotswapping the Actors message loop (e.g. its implementation) at runtime: invoke the
context.become method from within the Actor. become takes a PartialFunction<Object,
BoxedUnit> that implements the new message handler. The hotswapped code is kept in a Stack which can
be pushed and popped.
Warning: Please note that the actor will revert to its original behavior when restarted by its Supervisor.
To hotswap the Actor behavior using become:
public class HotSwapActor extends AbstractActor {
private PartialFunction<Object, BoxedUnit> angry;
private PartialFunction<Object, BoxedUnit> happy;
public HotSwapActor() {
angry =
ReceiveBuilder.
matchEquals("foo", s -> {
sender().tell("I am already angry?", self());
}).
matchEquals("bar", s -> {
context().become(happy);
}).build();
happy = ReceiveBuilder.
181
matchEquals("bar", s -> {
sender().tell("I am already happy :-)", self());
}).
matchEquals("foo", s -> {
context().become(angry);
}).build();
receive(ReceiveBuilder.
matchEquals("foo", s -> {
context().become(angry);
}).
matchEquals("bar", s -> {
context().become(happy);
}).build()
);
}
}
This variant of the become method is useful for many different things, such as to implement a Finite State
Machine (FSM, for an example see Dining Hakkers). It will replace the current behavior (i.e. the top of the
behavior stack), which means that you do not use unbecome, instead always the next behavior is explicitly
installed.
The other way of using become does not replace but add to the top of the behavior stack. In this case care must
be taken to ensure that the number of pop operations (i.e. unbecome) matches the number of push ones in
the long run, otherwise this amounts to a memory leak (which is why this behavior is not the default).
public class Swapper extends AbstractLoggingActor {
public Swapper() {
receive(ReceiveBuilder.
matchEquals(Swap, s -> {
log().info("Hi");
context().become(ReceiveBuilder.
matchEquals(Swap, x -> {
log().info("Ho");
context().unbecome(); // resets the latest become (just for fun)
}).build(), false); // push on top instead of replace
}).build()
);
}
}
public class SwapperApp {
public static void main(String[] args) {
ActorSystem system = ActorSystem.create("SwapperSystem");
ActorRef swapper = system.actorOf(Props.create(Swapper.class), "swapper");
swapper.tell(Swap, ActorRef.noSender()); // logs Hi
swapper.tell(Swap, ActorRef.noSender()); // logs Ho
swapper.tell(Swap, ActorRef.noSender()); // logs Hi
swapper.tell(Swap, ActorRef.noSender()); // logs Ho
swapper.tell(Swap, ActorRef.noSender()); // logs Hi
swapper.tell(Swap, ActorRef.noSender()); // logs Ho
system.shutdown();
}
}
4.1.11 Stash
The AbstractActorWithStash class enables an actor to temporarily stash away messages that can not
or should not be handled using the actors current behavior. Upon changing the actors message handler, i.e.,
right before invoking context().become() or context().unbecome(), all stashed messages can be
182
unstashed, thereby prepending them to the actors mailbox. This way, the stashed messages can be processed in
the same order as they have been received originally. An actor that extends AbstractActorWithStash will
automatically get a deque-based mailbox.
Note:
The abstract class AbstractActorWithStash implements the marker interface
RequiresMessageQueue<DequeBasedMessageQueueSemantics> which requests the system
to automatically choose a deque based mailbox implementation for the actor. If you want more control over the
mailbox, see the documentation on mailboxes: Mailboxes.
Here is an example of the AbstractActorWithStash class in action:
public class ActorWithProtocol extends AbstractActorWithStash {
public ActorWithProtocol() {
receive(ReceiveBuilder.
matchEquals("open", s -> {
context().become(ReceiveBuilder.
matchEquals("write", ws -> { /* do writing */ }).
matchEquals("close", cs -> {
unstashAll();
context().unbecome();
}).
matchAny(msg -> stash()).build(), false);
}).
matchAny(msg -> stash()).build()
);
}
}
Invoking stash() adds the current message (the message that the actor received last) to the actors stash.
It is typically invoked when handling the default case in the actors message handler to stash messages that
arent handled by the other cases. It is illegal to stash the same message twice; to do so results in an
IllegalStateException being thrown. The stash may also be bounded in which case invoking stash()
may lead to a capacity violation, which results in a StashOverflowException. The capacity of the stash
can be configured using the stash-capacity setting (an Int) of the mailboxs configuration.
Invoking unstashAll() enqueues messages from the stash to the actors mailbox until the capacity of the mailbox (if any) has been reached (note that messages from the stash are prepended to the mailbox). In case a bounded
mailbox overflows, a MessageQueueAppendFailedException is thrown. The stash is guaranteed to be
empty after calling unstashAll().
The stash is backed by a scala.collection.immutable.Vector. As a result, even a very large number
of messages may be stashed without a major impact on performance.
Note that the stash is part of the ephemeral actor state, unlike the mailbox. Therefore, it should be managed like
other parts of the actors state which have the same property. The AbstractActorWithStash implementation of preRestart will call unstashAll(), which is usually the desired behavior.
Note: If you want to enforce that your actor can only work with an unbounded stash, then you should use the
AbstractActorWithUnboundedStash class instead.
183
victim.tell(akka.actor.Kill.getInstance(), ActorRef.noSender());
184
One useful usage of this pattern is to disable creation of new ActorRefs for children during restarts. This can
be achieved by overriding preRestart():
@Override
public void preStart() {
// Initialize children here
}
// Overriding postRestart to disable the call to preStart()
// after restarts
@Override
public void postRestart(Throwable reason) {
}
// The default implementation of preRestart() stops all the children
// of the actor. To opt-out from stopping the children, we
// have to override preRestart()
@Override
public void preRestart(Throwable reason, Option<Object> message)
throws Exception {
// Keep the call to postStop(), but no stopping of children
postStop();
}
Please note, that the child actors are still restarted, but no new ActorRef is created. One can recursively apply
the same principles for the children, ensuring that their preStart() method is called only at the creation of
their refs.
For more information see What Restarting Means.
Initialization via message passing
There are cases when it is impossible to pass all the information needed for actor initialization in the constructor,
for example in the presence of circular dependencies. In this case the actor should listen for an initialization
message, and use become() or a finite state-machine state transition to encode the initialized and uninitialized
states of the actor.
receive(ReceiveBuilder.
matchEquals("init", m1 -> {
initializeMe = "Up and running";
context().become(ReceiveBuilder.
matchEquals("U OK?", m2 -> {
sender().tell(initializeMe, self());
}).build());
}).build()
If the actor may receive messages before it has been initialized, a useful tool can be the Stash to save messages
until the initialization finishes, and replaying them after the actor became initialized.
Warning: This pattern should be used with care, and applied only when none of the patterns above are
applicable. One of the potential issues is that messages might be lost when sent to remote actors. Also,
publishing an ActorRef in an uninitialized state might lead to the condition that it receives a user message
before the initialization has been done.
185
This is something that the JVM can have problems optimizing and the resulting code might not be as performant
as the Scala equivalent or the corresponding untyped actor version.
186
Description
The progress Listener starts the work.
The Worker schedules work by sending Do messages periodically to itself
When receiving Do the Worker tells the CounterService to increment the counter, three times.
The Increment message is forwarded to the Counter, which updates its counter variable and sends
current value to the Storage.
The Worker asks the CounterService of current value of the counter and pipes the result back to
the Listener.
187
188
Step
1
2
3, 4,
5, 6
7
8
9, 10,
11
12
13,
14
15,
16
Description
The Storage throws StorageException.
The CounterService is supervisor of the Storage and restarts the Storage when
StorageException is thrown.
The Storage continues to fail and is restarted.
After 3 failures and restarts within 5 seconds the Storage is stopped by its supervisor, i.e. the
CounterService.
The CounterService is also watching the Storage for termination and receives the
Terminated message when the Storage has been stopped ...
and tells the Counter that there is no Storage.
The CounterService schedules a Reconnect message to itself.
When it receives the Reconnect message it creates a new Storage ...
and tells the Counter to use the new Storage
java.util.ArrayList;
java.util.HashMap;
java.util.List;
java.util.Map;
import
import
import
import
import
import
import
import
import
import
import
akka.actor.*;
akka.dispatch.Mapper;
akka.event.LoggingReceive;
akka.japi.pf.DeciderBuilder;
akka.japi.pf.ReceiveBuilder;
akka.util.Timeout;
com.typesafe.config.Config;
com.typesafe.config.ConfigFactory;
scala.concurrent.duration.Duration;
scala.PartialFunction;
scala.runtime.BoxedUnit;
import
import
import
import
import
static
static
static
static
static
akka.japi.Util.classTag;
akka.actor.SupervisorStrategy.resume;
akka.actor.SupervisorStrategy.restart;
akka.actor.SupervisorStrategy.stop;
akka.actor.SupervisorStrategy.escalate;
static
static
static
static
docs.actor.japi.FaultHandlingDocSample.WorkerApi.*;
docs.actor.japi.FaultHandlingDocSample.CounterServiceApi.*;
docs.actor.japi.FaultHandlingDocSample.CounterApi.*;
docs.actor.japi.FaultHandlingDocSample.StorageApi.*;
189
"akka.actor.debug {\n" +
" receive = on\n" +
" lifecycle = on\n" +
"}\n");
ActorSystem system = ActorSystem.create("FaultToleranceSample", config);
ActorRef worker = system.actorOf(Props.create(Worker.class), "worker");
ActorRef listener = system.actorOf(Props.create(Listener.class), "listener");
// start the work and listen on progress
// note that the listener is used as sender of the tell,
// i.e. it will receive replies from the worker
worker.tell(Start, listener);
}
/**
* Listens on progress from the worker and shuts down the system when enough
* work has been done.
*/
public static class Listener extends AbstractLoggingActor {
@Override
public void preStart() {
// If we dont get any progress within 15 seconds then the service
// is unavailable
context().setReceiveTimeout(Duration.create("15 seconds"));
}
public Listener() {
receive(LoggingReceive.create(ReceiveBuilder.
match(Progress.class, progress -> {
log().info("Current progress: {} %", progress.percent);
if (progress.percent >= 100.0) {
log().info("Thats all, shutting down");
context().system().shutdown();
}
}).
matchEquals(ReceiveTimeout.getInstance(), x -> {
// No progress within 15 seconds, ServiceUnavailable
log().error("Shutting down due to unavailable service");
context().system().shutdown();
}).build(), context()
));
}
}
public interface WorkerApi {
public static final Object Start = "Start";
public static final Object Do = "Do";
public static class Progress {
public final double percent;
public Progress(double percent) {
this.percent = percent;
}
public String toString() {
return String.format("%s(%s)", getClass().getSimpleName(), percent);
}
}
}
190
/**
* Worker performs some work when it receives the Start message. It will
* continuously notify the sender of the Start message of current Progress.
* The Worker supervise the CounterService.
*/
public static class Worker extends AbstractLoggingActor {
final Timeout askTimeout = new Timeout(Duration.create(5, "seconds"));
// The sender of the initial Start message will continuously be notified
// about progress
ActorRef progressListener;
final ActorRef counterService = context().actorOf(
Props.create(CounterService.class), "counter");
final int totalCount = 51;
// Stop the CounterService child if it throws ServiceUnavailable
private static final SupervisorStrategy strategy =
new OneForOneStrategy(DeciderBuilder.
match(ServiceUnavailable.class, e -> stop()).
matchAny(o -> escalate()).build());
@Override
public SupervisorStrategy supervisorStrategy() {
return strategy;
}
public Worker() {
receive(LoggingReceive.create(ReceiveBuilder.
matchEquals(Start, x -> progressListener == null, x -> {
progressListener = sender();
context().system().scheduler().schedule(
Duration.Zero(), Duration.create(1, "second"), self(), Do,
context().dispatcher(), null
);
}).
matchEquals(Do, x -> {
counterService.tell(new Increment(1), self());
counterService.tell(new Increment(1), self());
counterService.tell(new Increment(1), self());
// Send current progress to the initial sender
pipe(ask(counterService, GetCurrentCount, askTimeout)
.mapTo(classTag(CurrentCount.class))
.map(new Mapper<CurrentCount, Progress>() {
public Progress apply(CurrentCount c) {
return new Progress(100.0 * c.count / totalCount);
}
}, context().dispatcher()), context().dispatcher())
.to(progressListener);
}).build(), context())
);
}
}
public interface CounterServiceApi {
public static final Object GetCurrentCount = "GetCurrentCount";
public static class CurrentCount {
public final String key;
public final long count;
public CurrentCount(String key, long count) {
this.key = key;
191
this.count = count;
}
public String toString() {
return String.format("%s(%s, %s)", getClass().getSimpleName(), key, count);
}
}
public static class Increment {
public final long n;
public Increment(long n) {
this.n = n;
}
public String toString() {
return String.format("%s(%s)", getClass().getSimpleName(), n);
}
}
public static class ServiceUnavailable extends RuntimeException {
private static final long serialVersionUID = 1L;
public ServiceUnavailable(String msg) {
super(msg);
}
}
}
/**
* Adds the value received in Increment message to a persistent counter.
* Replies with CurrentCount when it is asked for CurrentCount. CounterService
* supervise Storage and Counter.
*/
public static class CounterService extends AbstractLoggingActor {
// Reconnect message
static final Object Reconnect = "Reconnect";
private static class SenderMsgPair {
final ActorRef sender;
final Object msg;
SenderMsgPair(ActorRef sender, Object msg) {
this.msg = msg;
this.sender = sender;
}
}
final String key = self().path().name();
ActorRef storage;
ActorRef counter;
final List<SenderMsgPair> backlog = new ArrayList<>();
final int MAX_BACKLOG = 10000;
// Restart the storage child when StorageException is thrown.
// After 3 restarts within 5 seconds it will be stopped.
private static final SupervisorStrategy strategy =
new OneForOneStrategy(3, Duration.create("5 seconds"), DeciderBuilder.
match(StorageException.class, e -> restart()).
matchAny(o -> escalate()).build());
192
@Override
public SupervisorStrategy supervisorStrategy() {
return strategy;
}
@Override
public void preStart() {
initStorage();
}
/**
* The child storage is restarted in case of failure, but after 3 restarts,
* and still failing it will be stopped. Better to back-off than
* continuously failing. When it has been stopped we will schedule a
* Reconnect after a delay. Watch the child so we receive Terminated message
* when it has been terminated.
*/
void initStorage() {
storage = context().watch(context().actorOf(
Props.create(Storage.class), "storage"));
// Tell the counter, if any, to use the new storage
if (counter != null)
counter.tell(new UseStorage(storage), self());
// We need the initial value to be able to operate
storage.tell(new Get(key), self());
}
public CounterService() {
receive(LoggingReceive.create(ReceiveBuilder.
match(Entry.class, entry -> entry.key.equals(key) && counter == null, entry -> {
// Reply from Storage of the initial value, now we can create the Counter
final long value = entry.value;
counter = context().actorOf(Props.create(Counter.class, key, value));
// Tell the counter to use current storage
counter.tell(new UseStorage(storage), self());
// and send the buffered backlog to the counter
for (SenderMsgPair each : backlog) {
counter.tell(each.msg, each.sender);
}
backlog.clear();
}).
match(Increment.class, increment -> {
forwardOrPlaceInBacklog(increment);
}).
matchEquals(GetCurrentCount, gcc -> {
forwardOrPlaceInBacklog(gcc);
}).
match(Terminated.class, o -> {
// After 3 restarts the storage child is stopped.
// We receive Terminated because we watch the child, see initStorage.
storage = null;
// Tell the counter that there is no storage for the moment
counter.tell(new UseStorage(null), self());
// Try to re-establish storage after while
context().system().scheduler().scheduleOnce(
Duration.create(10, "seconds"), self(), Reconnect,
context().dispatcher(), null);
}).
matchEquals(Reconnect, o -> {
// Re-establish storage after the scheduled delay
initStorage();
}).build(), context())
);
193
}
void
//
//
//
if
forwardOrPlaceInBacklog(Object msg) {
We need the initial value from storage before we can start delegate to
the counter. Before that we place the messages in a backlog, to be sent
to the counter when it is initialized.
(counter == null) {
if (backlog.size() >= MAX_BACKLOG)
throw new ServiceUnavailable("CounterService not available," +
" lack of initial value");
backlog.add(new SenderMsgPair(sender(), msg));
} else {
counter.forward(msg, context());
}
}
}
public interface CounterApi {
public static class UseStorage {
public final ActorRef storage;
public UseStorage(ActorRef storage) {
this.storage = storage;
}
public String toString() {
return String.format("%s(%s)", getClass().getSimpleName(), storage);
}
}
}
/**
* The in memory count variable that will send current value to the Storage,
* if there is any storage available at the moment.
*/
public static class Counter extends AbstractLoggingActor {
final String key;
long count;
ActorRef storage;
public Counter(String key, long initialValue) {
this.key = key;
this.count = initialValue;
receive(LoggingReceive.create(ReceiveBuilder.
match(UseStorage.class, useStorage -> {
storage = useStorage.storage;
storeCount();
}).
match(Increment.class, increment -> {
count += increment.n;
storeCount();
}).
matchEquals(GetCurrentCount, gcc -> {
sender().tell(new CurrentCount(key, count), self());
}).build(), context())
);
}
void storeCount() {
// Delegate dangerous work, to protect our valuable state.
// We can continue without storage.
194
if (storage != null) {
storage.tell(new Store(new Entry(key, count)), self());
}
}
}
public interface StorageApi {
public static class Store {
public final Entry entry;
public Store(Entry entry) {
this.entry = entry;
}
public String toString() {
return String.format("%s(%s)", getClass().getSimpleName(), entry);
}
}
public static class Entry {
public final String key;
public final long value;
public Entry(String key, long value) {
this.key = key;
this.value = value;
}
public String toString() {
return String.format("%s(%s, %s)", getClass().getSimpleName(), key, value);
}
}
public static class Get {
public final String key;
public Get(String key) {
this.key = key;
}
public String toString() {
return String.format("%s(%s)", getClass().getSimpleName(), key);
}
}
public static class StorageException extends RuntimeException {
private static final long serialVersionUID = 1L;
public StorageException(String msg) {
super(msg);
}
}
}
/**
* Saves key/value pairs to persistent storage when receiving Store message.
* Replies with current value when receiving Get message. Will throw
* StorageException if the underlying data store is out of order.
*/
public static class Storage extends AbstractLoggingActor {
final DummyDB db = DummyDB.instance;
195
public Storage() {
receive(LoggingReceive.create(ReceiveBuilder.
match(Store.class, store -> {
db.save(store.entry.key, store.entry.value);
}).
match(Get.class, get -> {
Long value = db.load(get.key);
sender().tell(new Entry(get.key, value == null ?
Long.valueOf(0L) : value), self());
}).build(), context())
);
}
}
public static class DummyDB {
public static final DummyDB instance = new DummyDB();
private final Map<String, Long> db = new HashMap<String, Long>();
private DummyDB() {
}
public synchronized void save(String key, Long value) throws StorageException {
if (11 <= value && value <= 14)
throw new StorageException("Simulated store failure " + value);
db.put(key, value);
}
public synchronized Long load(String key) throws StorageException {
return db.get(key);
}
}
}
I have chosen a few well-known exception types in order to demonstrate the application of the fault handling
directives described in Supervision and Monitoring. First off, it is a one-for-one strategy, meaning that each child
is treated separately (an all-for-one strategy works very similarly, the only difference is that any decision is applied
to all children of the supervisor, not only the failing one). There are limits set on the restart frequency, namely
maximum 10 restarts per minute. -1 and Duration.Inf() means that the respective limit does not apply,
leaving the possibility to specify an absolute upper limit on the restarts or to make the restarts work infinitely. The
child actor is stopped if the limit is exceeded.
Note: If the strategy is declared inside the supervising actor (as opposed to a separate class) its decider has access
196
to all internal state of the actor in a thread-safe fashion, including obtaining a reference to the currently failed child
(available as the getSender of the failure message).
197
@Override
public SupervisorStrategy supervisorStrategy() {
return strategy;
}
public Supervisor() {
receive(ReceiveBuilder.
match(Props.class, props -> {
sender().tell(context().actorOf(props), self());
}).build()
);
}
}
This supervisor will be used to create a child, with which we can experiment:
public class Child extends AbstractActor {
int state = 0;
public Child() {
receive(ReceiveBuilder.
match(Exception.class, exception -> { throw exception; }).
match(Integer.class, i -> state = i).
matchEquals("get", s -> sender().tell(state, self())).build()
);
}
}
The test is easier by using the utilities described in akka-testkit, where TestProbe provides an actor ref useful
for receiving and inspecting replies.
import akka.actor.*;
import
import
import
import
import
import
import
import
import
import
import
import
import
static akka.actor.SupervisorStrategy.resume;
static akka.actor.SupervisorStrategy.restart;
static akka.actor.SupervisorStrategy.stop;
static akka.actor.SupervisorStrategy.escalate;
akka.japi.pf.DeciderBuilder;
akka.japi.pf.ReceiveBuilder;
com.typesafe.config.Config;
com.typesafe.config.ConfigFactory;
scala.PartialFunction;
scala.concurrent.Await;
static akka.pattern.Patterns.ask;
scala.concurrent.duration.Duration;
akka.testkit.TestProbe;
198
@Test
public void mustEmploySupervisorStrategy() throws Exception {
// code here
}
}
The first test shall demonstrate the Resume directive, so we try it out by setting some non-initial state in the actor
and have it fail:
child.tell(42, ActorRef.noSender());
assert Await.result(ask(child, "get", 5000), timeout).equals(42);
child.tell(new ArithmeticException(), ActorRef.noSender());
assert Await.result(ask(child, "get", 5000), timeout).equals(42);
As you can see the value 42 survives the fault handling directive. Now, if we change the failure to a more serious
NullPointerException, that will no longer be the case:
child.tell(new NullPointerException(), ActorRef.noSender());
assert Await.result(ask(child, "get", 5000), timeout).equals(0);
And finally in case of the fatal IllegalArgumentException the child will be terminated by the supervisor:
final TestProbe probe = new TestProbe(system);
probe.watch(child);
child.tell(new IllegalArgumentException(), ActorRef.noSender());
probe.expectMsgClass(Terminated.class);
Up to now the supervisor was completely unaffected by the childs failure, because the directives set did handle it.
In case of an Exception, this is not true anymore and the supervisor escalates the failure.
child = (ActorRef) Await.result(ask(supervisor,
Props.create(Child.class), 5000), timeout);
probe.watch(child);
assert Await.result(ask(child, "get", 5000), timeout).equals(0);
child.tell(new Exception(), ActorRef.noSender());
probe.expectMsgClass(Terminated.class);
The supervisor itself is supervised by the top-level actor provided by the ActorSystem, which
has the default policy to restart in case of all Exception cases (with the notable exceptions of
ActorInitializationException and ActorKilledException). Since the default directive in case
of a restart is to kill all children, we expected our poor child not to survive this failure.
In case this is not desired (which depends on the use case), we need to use a different supervisor which overrides
this behavior.
public class Supervisor2 extends AbstractActor {
private static SupervisorStrategy strategy =
new OneForOneStrategy(10, Duration.create("1 minute"), DeciderBuilder.
match(ArithmeticException.class, e -> resume()).
match(NullPointerException.class, e -> restart()).
match(IllegalArgumentException.class, e -> stop()).
matchAny(o -> escalate()).build());
@Override
public SupervisorStrategy supervisorStrategy() {
199
return strategy;
}
public Supervisor2() {
receive(ReceiveBuilder.
match(Props.class, props -> {
sender().tell(context().actorOf(props), self());
}).build()
);
}
@Override
public void preRestart(Throwable cause, Option<Object> msg) {
// do not kill all children, which is the default here
}
}
With this parent, the child survives the escalated restart, as demonstrated in the last test:
superprops = Props.create(Supervisor2.class);
supervisor = system.actorOf(superprops);
child = (ActorRef) Await.result(ask(supervisor,
Props.create(Child.class), 5000), timeout);
child.tell(23, ActorRef.noSender());
assert Await.result(ask(child, "get", 5000), timeout).equals(23);
child.tell(new Exception(), ActorRef.noSender());
assert Await.result(ask(child, "get", 5000), timeout).equals(0);
200
import
import
import
import
import
akka.japi.pf.UnitMatch;
java.util.Arrays;
java.util.LinkedList;
java.util.List;
scala.concurrent.duration.Duration;
The contract of our Buncher actor is that it accepts or produces the following messages:
public final class SetTarget {
private final ActorRef ref;
public SetTarget(ActorRef ref) {
this.ref = ref;
}
public ActorRef getRef() {
return ref;
}
// boilerplate ...
}
public final class Queue {
private final Object obj;
public Queue(Object obj) {
this.obj = obj;
}
public Object getObj() {
return obj;
}
// boilerplate ...
}
public final class Batch {
private final List<Object> list;
public Batch(List<Object> list) {
this.list = list;
}
public List<Object> getList() {
return list;
}
// boilerplate ...
}
public enum Flush {
Flush
}
SetTarget is needed for starting it up, setting the destination for the Batches to be passed on; Queue will
add to the internal queue while Flush will mark the end of a burst.
The actor can be in two states: no message queued (aka Idle) or some message queued (aka Active). The
states and the state data is defined like this:
// states
enum State {
Idle, Active
}
// state data
interface Data {
201
}
enum Uninitialized implements Data {
Uninitialized
}
final class Todo implements Data {
private final ActorRef target;
private final List<Object> queue;
public Todo(ActorRef target, List<Object> queue) {
this.target = target;
this.queue = queue;
}
public ActorRef getTarget() {
return target;
}
public List<Object> getQueue() {
return queue;
}
// boilerplate ...
}
The actor starts out in the idle state. Once a message arrives it will go to the active state and stay there as long as
messages keep arriving and no flush is requested. The internal state data of the actor is made up of the target actor
reference to send the batches to and the actual queue of messages.
Now lets take a look at the skeleton for our FSM actor:
public class Buncher extends AbstractFSM<State, Data> {
{
startWith(Idle, Uninitialized);
when(Idle,
matchEvent(SetTarget.class, Uninitialized.class,
(setTarget, uninitialized) ->
stay().using(new Todo(setTarget.getRef(), new LinkedList<>()))));
// transition elided ...
when(Active, Duration.create(1, "second"),
matchEvent(Arrays.asList(Flush.class, StateTimeout()), Todo.class,
(event, todo) -> goTo(Idle).using(todo.copy(new LinkedList<>()))));
// unhandled elided ...
initialize();
}
}
The basic strategy is to declare the actor, by inheriting the AbstractFSM class and specifying the possible states
and data values as type parameters. Within the body of the actor a DSL is used for declaring the state machine:
startWith defines the initial state and initial data
then there is one when(<state>) { ... } declaration per state to be handled (could potentially be
multiple ones, the passed PartialFunction will be concatenated using orElse)
finally starting it up using initialize, which performs the transition into the initial state and sets up
timers (if required).
In this case, we start out in the Idle and Uninitialized state, where only the SetTarget() message
is handled; stay prepares to end this events processing for not leaving the current state, while the using
4.3. FSM (Java with Lambda Support)
202
modifier makes the FSM replace the internal state (which is Uninitialized at this point) with a fresh Todo()
object containing the target actor reference. The Active state has a state timeout declared, which means that
if no message is received for 1 second, a FSM.StateTimeout message will be generated. This has the same
effect as receiving the Flush command in this case, namely to transition back into the Idle state and resetting
the internal queue to the empty vector. But how do messages get queued? Since this shall work identically in
both states, we make use of the fact that any event which is not handled by the when() block is passed to the
whenUnhandled() block:
whenUnhandled(
matchEvent(Queue.class, Todo.class,
(queue, todo) -> goTo(Active).using(todo.addElement(queue.getObj()))).
anyEvent((event, state) -> {
log().warning("received unhandled request {} in state {}/{}",
event, stateName(), state);
return stay();
}));
The first case handled here is adding Queue() requests to the internal queue and going to the Active state
(this does the obvious thing of staying in the Active state if already there), but only if the FSM data are not
Uninitialized when the Queue() event is received. Otherwiseand in all other non-handled casesthe
second case just logs a warning and does not change the internal state.
The only missing piece is where the Batches are actually sent to the target, for which we use the
onTransition mechanism: you can declare multiple such blocks and all of them will be tried for matching
behavior in case a state transition occurs (i.e. only when the state actually changes).
onTransition(
matchState(Active, Idle, () -> {
// reuse this matcher
final UnitMatch<Data> m = UnitMatch.create(
matchData(Todo.class,
todo -> todo.getTarget().tell(new Batch(todo.getQueue()), self())));
m.match(stateData());
}).
state(Idle, Active, () -> {/* Do something here */}));
The transition callback is a partial function which takes as input a pair of statesthe current and the next state.
During the state change, the old state data is available via stateData as shown, and the new state data would
be available as nextStateData.
To verify that this buncher actually works, it is quite easy to write a test using the akka-testkit, here using JUnit as
an example:
public class BuncherTest {
static ActorSystem system;
@BeforeClass
public static void setup() {
system = ActorSystem.create("BuncherTest");
}
@AfterClass
public static void tearDown() {
JavaTestKit.shutdownActorSystem(system);
system = null;
}
@Test
public void testBuncherActorBatchesCorrectly() {
new JavaTestKit(system) {{
final ActorRef buncher =
system.actorOf(Props.create(Buncher.class));
final ActorRef probe = getRef();
203
4.3.3 Reference
The AbstractFSM Class
The AbstractFSM abstract class is the base class used to implement an FSM. It implements Actor since an
Actor is created to drive the FSM.
public class Buncher extends AbstractFSM<State, Data> {
{
// fsm body ...
}
}
Note: The AbstractFSM class defines a receive method which handles internal messages and passes everything
else through to the FSM logic (according to the current state). When overriding the receive method, keep in
mind that e.g. state timeout handling depends on actually passing the messages through the FSM logic.
The AbstractFSM class takes two type parameters:
1. the supertype of all state names, usually an enum,
2. the type of the state data which are tracked by the AbstractFSM module itself.
Note: The state data together with the state name describe the internal state of the state machine; if you stick to
this scheme and do not add mutable fields to the FSM class you have the advantage of making all changes of the
204
Defining States
A state is defined by one or more invocations of the method
when(<name>[, stateTimeout = <timeout>])(stateFunction).
The given name must be an object which is type-compatible with the first type parameter given to the
AbstractFSM class. This object is used as a hash key, so you must ensure that it properly implements equals
and hashCode; in particular it must not be mutable. The easiest fit for these requirements are case objects.
If the stateTimeout parameter is given, then all transitions into this state, including staying, receive this timeout by default. Initiating the transition with an explicit timeout may be used to override this default, see Initiating
Transitions for more information. The state timeout of any state may be changed during action processing with
setStateTimeout(state, duration). This enables runtime configuration e.g. via external message.
The stateFunction argument is a PartialFunction[Event, State], which is conveniently given
using the state function builder syntax as demonstrated below:
when(Idle,
matchEvent(SetTarget.class, Uninitialized.class,
(setTarget, uninitialized) ->
stay().using(new Todo(setTarget.getRef(), new LinkedList<>()))));
Warning: It is required that you define handlers for each of the possible FSM states, otherwise there will be
failures when trying to switch to undeclared states.
It is recommended practice to declare the states as an enum and then verify that there is a when clause for each of
the states. If you want to leave the handling of a state unhandled (more below), it still needs to be declared like
this:
when(SomeState, AbstractFSM.NullFunction());
205
Within this handler the state of the FSM may be queried using the stateName method.
IMPORTANT: This handler is not stacked, meaning that each invocation of whenUnhandled replaces the
previously installed handler.
Initiating Transitions
The result of any stateFunction must be a definition of the next state unless terminating the FSM, which is
described in Termination from Inside. The state definition can either be the current state, as described by the stay
directive, or it is a different state as given by goto(state). The resulting object allows further qualification by
way of the modifiers described in the following:
forMax(duration)
This modifier sets a state timeout on the next state. This means that a timer is started which upon expiry
sends a StateTimeout message to the FSM. This timer is canceled upon reception of any other message
in the meantime; you can rely on the fact that the StateTimeout message will not be processed after an
intervening message.
This modifier can also be used to override any default timeout which is specified for the target state. If you
want to cancel the default timeout, use Duration.Inf.
using(data)
This modifier replaces the old state data with the new data given. If you follow the advice above, this is the
only place where internal state data are ever modified.
replying(msg)
This modifier sends a reply to the currently processed message and otherwise does not modify the state
transition.
All modifiers can be chained to achieve a nice and concise description:
when(SomeState, matchAnyEvent((msg, data) -> {
return goTo(Processing).using(newData).
forMax(Duration.create(5, SECONDS)).replying(WillDo);
}));
The parentheses are not actually needed in all cases, but they visually distinguish between modifiers and their
arguments and therefore make the code even more pleasant to read for foreigners.
Note: Please note that the return statement may not be used in when blocks or similar; this is a Scala
restriction. Either refactor your code using if () ... else ... or move it into a method definition.
Monitoring Transitions
Transitions occur between states conceptually, which means after any actions you have put into the event handling block; this is obvious since the next state is only defined by the value returned by the event handling logic.
You do not need to worry about the exact order with respect to setting the internal state variable, as everything
within the FSM actor is running single-threaded anyway.
Internal Monitoring
Up to this point, the FSM DSL has been centered on states and events. The dual view is to describe it as a series
of transitions. This is enabled by the method
onTransition(handler)
which associates actions with a transition instead of with a state and event. The handler is a partial function which
takes a pair of states as input; no resulting state is needed as it is not possible to modify the transition in progress.
206
onTransition(
matchState(Active, Idle, () -> setTimer("timeout",
Tick, Duration.create(1, SECONDS), true)).
state(Active, null, () -> cancelTimer("timeout")).
state(null, Idle, (f, t) -> log().info("entering Idle from " + f)));
It is also possible to pass a function object accepting two states to onTransition, in case your transition
handling logic is implemented as a method:
public void handler(StateType from, StateType to) {
// handle transition here
}
onTransition(this::handler);
The handlers registered with this method are stacked, so you can intersperse onTransition blocks with when
blocks as suits your design. It should be noted, however, that all handlers will be invoked for each transition,
not only the first matching one. This is designed specifically so you can put all transition handling for a certain
aspect into one place without having to worry about earlier declarations shadowing later ones; the actions are still
executed in declaration order, though.
Note: This kind of internal monitoring may be used to structure your FSM according to transitions, so that for
example the cancellation of a timer upon leaving a certain state cannot be forgot when adding new target states.
External Monitoring
207
You can use onTermination(handler) to specify custom code that is executed when the FSM is stopped.
The handler is a partial function which takes a StopEvent(reason, stateName, stateData) as argument:
onTermination(
matchStop(Normal(),
(state, data) -> {/* Do something here */}).
stop(Shutdown(),
(state, data) -> {/* Do something here */}).
stop(Failure.class,
(reason, state, data) -> {/* Do something here */}));
As for the whenUnhandled case, this handler is not stacked, so each invocation of onTermination replaces
the previously installed handler.
Termination from Outside
When an ActorRef associated to a FSM is stopped using the stop method, its postStop hook will be
executed. The default implementation by the AbstractFSM class is to execute the onTermination handler
if that is prepared to handle a StopEvent(Shutdown, ...).
Warning: In case you override postStop and want to have your onTermination handler called, do not
forget to call super.postStop.
208
The logDepth defaults to zero, which turns off the event log.
Warning: The log buffer is allocated during actor creation, which is why the configuration is done using a
virtual method call. If you want to override with a val, make sure that its initialization happens before the
initializer of LoggingFSM runs, and do not change the value returned by logDepth after the buffer has
been allocated.
The contents of the event log are available using method getLog, which returns an IndexedSeq[LogEntry]
where the oldest entry is at index zero.
4.3.5 Examples
A bigger FSM example contrasted with Actors become/unbecome can be found in the Typesafe Activator
template named Akka FSM in Scala
209
Warning: This module is marked as experimental as of its introduction in Akka 2.3.0. We will continue to
improve this API based on our users feedback, which implies that while we try to keep incompatible changes
to a minimum the binary compatibility guarantee for maintenance releases does not apply to the contents of
the akka.persistence package.
Akka persistence is inspired by the eventsourced library. It follows the same concepts and architecture of
eventsourced but significantly differs on API and implementation level.
4.4.1 Dependencies
Akka persistence is a separate jar file. Make sure that you have the following dependency in your project:
<dependency>
<groupId>com.typesafe.akka</groupId>
<artifactId>akka-persistence-experimental_2.10</artifactId>
<version>2.3.1</version>
</dependency>
4.4.2 Architecture
Processor: A processor is a persistent, stateful actor. Messages sent to a processor are written to a journal
before its behavior is called. When a processor is started or restarted, journaled messages are replayed to
that processor, so that it can recover internal state from these messages.
View: A view is a persistent, stateful actor that receives journaled messages that have been written by
another processor. A view itself does not journal new messages, instead, it updates internal state only from
a processors replicated message stream.
Channel: Channels are used by processors and views to communicate with other actors. They prevent that
replayed messages are redundantly delivered to these actors and provide at-least-once message delivery
semantics, also in case of sender and receiver JVM crashes.
Journal: A journal stores the sequence of messages sent to a processor. An application can control which
messages are journaled and which are received by the processor without being journaled. The storage
backend of a journal is pluggable. The default journal storage plugin writes to the local filesystem, replicated
journals are available as Community plugins.
Snapshot store: A snapshot store persists snapshots of a processors or a views internal state. Snapshots
are used for optimizing recovery times. The storage backend of a snapshot store is pluggable. The default
snapshot storage plugin writes to the local filesystem.
Event sourcing. Based on the building blocks described above, Akka persistence provides abstractions for
the development of event sourced applications (see section Event sourcing)
4.4.3 Processors
A processor can be implemented by extending AbstractProcessor class and setting the initial behavior in
the constructor by calling the receive method
class MyProcessor extends AbstractProcessor {
public MyProcessor() {
receive(ReceiveBuilder.
match(Persistent.class, p -> {
// message successfully written to journal
Object payload = p.payload();
Long sequenceNr = p.sequenceNr();
// ...
}).
match(PersistenceFailure.class, failure -> {
210
Processors only write messages of type Persistent to the journal, others are received without being persisted. When a processors behavior is called with a Persistent message it can safely assume that this message has been successfully written to the journal. If a journal fails to write a Persistent message then the
processor is stopped, by default. If a processor should continue running on persistence failures it must handle
PersistenceFailure messages. In this case, a processor may want to inform the sender about the failure, so
that the sender can re-send the message, if needed.
An AbstractProcessor itself is an Actor and can therefore be instantiated with actorOf.
processor = context().actorOf(Props.create(MyProcessor.class), "myProcessor");
processor.tell(Persistent.create("foo"), null);
processor.tell("bar", null);
Recovery
By default, a processor is automatically recovered on start and on restart by replaying journaled messages. New
messages sent to a processor during recovery do not interfere with replayed messages. New messages will only
be received by a processor after recovery completes.
Recovery customization
Automated recovery on start can be disabled by overriding preStart with an empty implementation.
@Override
public void preStart() {}
If not overridden, preStart sends a Recover message to self(). Applications may also override
preStart to define further Recover parameters such as an upper sequence number bound, for example.
@Override
public void preStart() {
self().tell(Recover.create(457L), null);
}
Upper sequence number bounds can be used to recover a processor to past state instead of current state. Automated
recovery on restart can be disabled by overriding preRestart with an empty implementation.
@Override
public void preRestart(Throwable reason, Option<Object> message) {}
211
Recovery status
A processor can query its own recovery status via the methods
public boolean recoveryRunning();
public boolean recoveryFinished();
Sometimes there is a need for performing additional initialization when the recovery has completed, before processing any other message sent to the processor. The processor can send itself a message from preStart. It will
be stashed and received after recovery. The mailbox may contain other messages that are queued in front of that
message and therefore you need to stash until you receive that message.
class MyProcessor5 extends AbstractProcessor {
public MyProcessor5() {
receive(ReceiveBuilder.
matchEquals("FIRST", s -> {
recoveryCompleted();
getContext().become(active);
unstashAll();
}).
matchAny(message -> {
if (recoveryFinished()) {
stash();
} else {
active.apply(message);
}
}).
build()
);
}
@Override
public void preStart() throws Exception {
super.preStart();
self().tell("FIRST", self());
}
private void recoveryCompleted() {
// perform init after recovery, before any other messages
// ...
}
PartialFunction<Object, BoxedUnit> active =
ReceiveBuilder.
match(Persistent.class, message -> {/* ... */}).
build();
}
Failure handling
A persistent message that caused an exception will be received again by a processor after restart. To prevent a
replay of that message during recovery it can be deleted.
@Override
public void preRestart(Throwable reason, Option<Object> message) {
if (message.isDefined() && message.get() instanceof Persistent) {
deleteMessage(((Persistent) message.get()).sequenceNr());
}
super.preRestart(reason, message);
}
212
Message deletion
A processor can delete a single message by calling the deleteMessage method with the sequence number of
that message as argument. An optional permanent parameter specifies whether the message shall be permanently deleted from the journal or only marked as deleted. In both cases, the message wont be replayed. Later
extensions to Akka persistence will allow to replay messages that have been marked as deleted which can be useful
for debugging purposes, for example. To delete all messages (journaled by a single processor) up to a specified
sequence number, processors should call the deleteMessages method.
Identifiers
A processor must have an identifier that doesnt change across different actor incarnations. It defaults to the
String representation of processors path without the address part and can be obtained via the processorId
method.
public String processorId();
Applications can customize a processors id by specifying an actor name during processor creation as shown in
section Processors. This changes that processors name in its actor hierarchy and hence influences only part of the
processor id. To fully customize a processors id, the processorId method must be overridden.
@Override
public String processorId() {
return "my-stable-processor-id";
}
4.4.4 Views
Views can be implemented by extending the AbstractView abstract class, implement the processorId
method and setting the initial behavior in the constructor by calling the receive method.
class MyView extends AbstractView {
@Override
public String processorId() {
return "some-processor-id";
}
public MyView() {
receive(ReceiveBuilder.
match(Persistent.class, persistent -> {
// ...
}).build()
);
}
}
The processorId identifies the processor from which the view receives journaled messages. It is not necessary
the referenced processor is actually running. Views read messages from a processors journal directly. When a
processor is started later and begins to write new messages, the corresponding view is updated automatically, by
default.
Updates
The default update interval of all views of an actor system is configurable:
213
akka.persistence.view.auto-update-interval = 5s
View implementation classes may also override the autoUpdateInterval method to return a custom update
interval for a specific view class or view instance. Applications may also trigger additional updates at any time by
sending a view an Update message.
final ActorRef view = system.actorOf(Props.create(MyView.class));
view.tell(Update.create(true), null);
If the await parameter is set to true, messages that follow the Update request are processed when the incremental message replay, triggered by that update request, completed. If set to false (default), messages following
the update request may interleave with the replayed message stream. Automated updates always run with await
= false.
Automated updates of all views of an actor system can be turned off by configuration:
akka.persistence.view.auto-update = off
Implementation classes may override the configured default value by overriding the autoUpdate
method.
To limit the number of replayed messages per update request, applications can configure a custom akka.persistence.view.auto-update-replay-max value or override the
autoUpdateReplayMax method. The number of replayed messages for manual updates can be limited with
the replayMax parameter of the Update message.
Recovery
Initial recovery of views works in the very same way as for Processors (i.e. by sending a Recover
message to self). The maximum number of replayed messages during initial recovery is determined by
autoUpdateReplayMax. Further possibilities to customize initial recovery are explained in section Processors.
Identifiers
A view must have an identifier that doesnt change across different actor incarnations. It defaults to the String
representation of the actor path without the address part and can be obtained via the viewId method.
Applications can customize a views id by specifying an actor name during view creation. This changes that
views name in its actor hierarchy and hence influences only part of the view id. To fully customize a views
id, the viewId method must be overridden. Overriding viewId is the recommended way to generate stable
identifiers.
The viewId must differ from the referenced processorId, unless Snapshots of a view and its processor shall
be shared (which is what applications usually do not want).
4.4.5 Channels
Channels are special actors that are used by processors or views to communicate with other actors (channel destinations). The following discusses channels in context of processors but this is also applicable to views.
Channels prevent redundant delivery of replayed messages to destinations during processor recovery. A replayed
message is retained by a channel if its delivery has been confirmed by a destination.
class MyProcessor extends AbstractProcessor {
private final ActorRef destination;
private final ActorRef channel;
public MyProcessor() {
this.destination = context().actorOf(Props.create(MyDestination.class));
this.channel = context().actorOf(Channel.props(), "myChannel");
214
receive(ReceiveBuilder.
match(Persistent.class, p -> {
Persistent out = p.withPayload("done " + p.payload());
channel.tell(Deliver.create(out, destination.path()), self());
}).build()
);
}
}
class MyDestination extends AbstractActor {
public MyDestination() {
receive(ReceiveBuilder.
match(ConfirmablePersistent.class, p -> {
Object payload = p.payload();
Long sequenceNr = p.sequenceNr();
int redeliveries = p.redeliveries();
// ...
p.confirm();
}).build()
);
}
}
A channel is ready to use once it has been created, no recovery or further activation is needed. A Deliver request
instructs a channel to send a Persistent message to a destination. A destination is provided as ActorPath
and messages are sent by the channel via that paths ActorSelection. Sender references are preserved by a
channel, therefore, a destination can reply to the sender of a Deliver request.
If a processor wants to reply to a Persistent message sender it should use the sender() path as channel
destination.
channel.tell(Deliver.create(out, sender().path()), self());
A channel keeps messages in memory until their successful delivery has been confirmed or the maximum number
of re-deliveries is reached. To be notified about messages that have reached the maximum number of re-deliveries,
applications can register a listener at channel creation.
class MyListener extends AbstractActor {
public MyListener() {
receive(ReceiveBuilder.
215
match(RedeliverFailure.class, r -> {
Iterable<ConfirmablePersistent> messages = r.getMessages();
// ...
}).build()
);
}
}
final ActorRef myListener = context().actorOf(Props.create(MyListener.class));
context().actorOf(Channel.props(
ChannelSettings.create().withRedeliverFailureListener(null)));
A listener receives RedeliverFailure notifications containing all messages that could not be delivered. On
receiving a RedeliverFailure message, an application may decide to restart the sending processor to enforce
a re-send of these messages to the channel or confirm these messages to prevent further re-sends. The sending
processor can also be restarted any time later to re-send unconfirmed messages.
This combination of
message persistence by sending processors
message replays by sending processors
message re-deliveries by channels and
application-level confirmations (acknowledgements) by destinations
enables channels to provide at-least-once message delivery semantics. Possible duplicates can be detected by
destinations by tracking message sequence numbers. Message sequence numbers are generated per sending processor. Depending on how a processor routes outbound messages to destinations, they may either see a contiguous
message sequence or a sequence with gaps.
Warning: If a processor emits more than one outbound message per inbound Persistent message it
must use a separate channel for each outbound message to ensure that confirmations are uniquely identifiable,
otherwise, at-least-once message delivery semantics do not apply. This rule has been introduced to avoid
writing additional outbound message identifiers to the journal which would decrease the overall throughput.
It is furthermore recommended to collapse multiple outbound messages to the same destination into a single
outbound message, otherwise, if sent via multiple channels, their ordering is not defined.
If an application wants to have more control how sequence numbers are assigned to messages it should use an
application-specific sequence number generator and include the generated sequence numbers into the payload
of Persistent messages.
Persistent channels
Channels created with Channel.props do not persist messages. These channels are usually used in combination with a sending processor that takes care of persistence, hence, channel-specific persistence is not necessary in
this case. They are referred to as transient channels in the following.
Persistent channels are like transient channels but additionally persist messages before delivering them. Messages that have been persisted by a persistent channel are deleted when destinations confirm their delivery. A persistent channel can be created with PersistentChannel.props and configured with a
PersistentChannelSettings object.
final ActorRef channel = context().actorOf(
PersistentChannel.props(
PersistentChannelSettings.create()
.withRedeliverInterval(Duration.create(30, TimeUnit.SECONDS))
.withRedeliverMax(15)),
"myPersistentChannel");
channel.tell(Deliver.create(Persistent.create("example"), destination.path()), self());
216
A persistent channel is useful for delivery of messages to slow destinations or destinations that are unavailable for
a long time. It can constrain the number of pending confirmations based on the pendingConfirmationsMax
and pendingConfirmationsMin parameters of PersistentChannelSettings.
PersistentChannelSettings.create()
.withPendingConfirmationsMax(10000)
.withPendingConfirmationsMin(2000);
It suspends delivery when the number of pending confirmations reaches pendingConfirmationsMax and resumes delivery again when this number falls below pendingConfirmationsMin. This prevents both, flooding destinations with more messages than they can process and unlimited memory consumption by the channel.
A persistent channel continues to persist new messages even when message delivery is temporarily suspended.
Standalone usage
Applications may also use channels standalone. Transient channels can be used standalone if re-delivery attempts
to destinations are required but message loss in case of a sender JVM crash is not an issue. If message loss in
case of a sender JVM crash is an issue, persistent channels should be used. In this case, applications may want to
receive replies from the channel whether messages have been successfully persisted or not. This can be enabled
by creating the channel with the replyPersistent configuration parameter set to true:
PersistentChannelSettings.create().withReplyPersistent(true);
With this setting, either the successfully persisted message is replied to the sender or a PersistenceFailure
message. In case the latter case, the sender should re-send the message.
Identifiers
In the same way as Processors and Views, channels also have an identifier that defaults to a channels path.
A channel identifier can therefore be customized by using a custom actor name at channel creation. This
changes that channels name in its actor hierarchy and hence influences only part of the channel identifier.
To fully customize a channel identifier, it should be provided as argument Channel.props(String) or
PersistentChannel.props(String) (recommended to generate stable identifiers).
this.channel = context().actorOf(Channel.props("my-stable-channel-id"));
This is necessary for delivery confirmations to work properly. Both ways are equivalent but we recommend using
p.withPayload(...) for clarity. It is not allowed to send a message via a channel that has been created with
Persistent.create(...). This would redeliver the message on every replay even though its delivery was
confirmed by a destination.
Sequence number
The sequence number of a Persistent message can be obtained via its sequenceNr method. Persistent
messages are assigned sequence numbers on a per-processor basis (or per channel basis if used standalone). A
sequence starts at 1L and doesnt contain gaps unless a processor deletes messages.
217
4.4.7 Snapshots
Snapshots can dramatically reduce recovery times of processors and views. The following discusses snapshots in
context of processors but this is also applicable to views.
Processors can save snapshots of internal state by calling the saveSnapshot method. If saving of a snapshot
succeeds, the processor receives a SaveSnapshotSuccess message, otherwise a SaveSnapshotFailure
message
class MyProcessor extends AbstractProcessor {
private Object state;
public MyProcessor() {
receive(ReceiveBuilder.
match(String.class, s -> s.equals("snap"),
s -> saveSnapshot(state)).
match(SaveSnapshotSuccess.class, ss -> {
SnapshotMetadata metadata = ss.metadata();
// ...
}).
match(SaveSnapshotFailure.class, sf -> {
SnapshotMetadata metadata = sf.metadata();
// ...
}).build()
);
}
}
During recovery, the processor is offered a previously saved snapshot via a SnapshotOffer message from
which it can initialize internal state.
class MyProcessor extends AbstractProcessor {
private Object state;
public MyProcessor() {
receive(ReceiveBuilder.
match(SnapshotOffer.class, s -> {
state = s.snapshot();
// ...
}).
match(Persistent.class, p -> {/* ...*/}).build()
);
}
}
The replayed messages that follow the SnapshotOffer message, if any, are younger than the offered snapshot.
They finally recover the processor to its current (i.e. latest) state.
In general, a processor is only offered a snapshot if that processor has previously saved one or more snapshots
and at least one of these snapshots matches the SnapshotSelectionCriteria that can be specified for
recovery.
processor.tell(Recover.create(
SnapshotSelectionCriteria
.create(457L, System.currentTimeMillis())), null);
218
Snapshot deletion
A processor can delete individual snapshots by calling the deleteSnapshot method with the sequence number and the timestamp of a snapshot as argument.
To bulk-delete snapshots matching
SnapshotSelectionCriteria, processors should use the deleteSnapshots method.
akka.actor.ActorRef;
akka.actor.ActorSystem;
akka.actor.Props;
akka.japi.pf.ReceiveBuilder;
akka.persistence.AbstractEventsourcedProcessor;
akka.persistence.SnapshotOffer;
scala.PartialFunction;
scala.runtime.BoxedUnit;
import java.io.Serializable;
import java.util.ArrayList;
import static java.util.Arrays.asList;
class Cmd implements Serializable {
private final String data;
public Cmd(String data) {
this.data = data;
}
public String getData() {
return data;
}
}
class Evt implements Serializable {
private final String data;
public Evt(String data) {
this.data = data;
}
219
220
The example defines two data types, Cmd and Evt to represent commands and events, respectively. The state
of the ExampleProcessor is a list of persisted event data contained in ExampleState.
The processors receiveRecover method defines how state is updated during recovery by handling Evt
and SnapshotOffer messages. The processors receiveCommand method is a command handler. In this
example, a command is handled by generating two events which are then persisted and handled. Events are
persisted by calling persist with an event (or a sequence of events) as first argument and an event handler as
second argument.
The persist method persists events asynchronously and the event handler is executed for successfully persisted
events. Successfully persisted events are internally sent back to the processor as individual messages that trigger
event handler executions. An event handler may close over processor state and mutate it. The sender of a persisted
event is the sender of the corresponding command. This allows event handlers to reply to the sender of a command
(not shown).
The main responsibility of an event handler is changing processor state using event data and notifying others about
successful state changes by publishing events.
When persisting events with persist it is guaranteed that the processor will not receive further commands
between the persist call and the execution(s) of the associated event handler. This also holds for multiple
persist calls in context of a single command.
The easiest way to run this example yourself is to download Typesafe Activator and open the tutorial named Akka
Persistence Samples in Java with Lambdas. It contains instructions on how to run the EventsourcedExample.
Note: Its also possible to switch between different command handlers during normal processing and recovery
with context().become() and context().unbecome(). To get the actor into the same state after
recovery you need to take special care to perform the same state transitions with become and unbecome in the
receiveRecover method as you would have done in the command handler.
221
}
@Override public PartialFunction<Object, BoxedUnit> receiveCommand() {
return ReceiveBuilder.
match(String.class, s -> s.equals("cmd"),
s -> persist("evt", this::handleEvent)).build();
}
}
In larger integration scenarios, channel destinations may be actors that submit received events to an external
message broker, for example. After having successfully submitted an event, they should call confirm() on the
received ConfirmablePersistent message.
A new batch write is triggered by a processor as soon as a batch reaches the maximum size or if the journal
completed writing the previous batch. Batch writes are never timer-based which keeps latencies at a minimum.
Applications that want to have more explicit control over batch writes and batch sizes can send processors
PersistentBatch messages.
class MyProcessor extends AbstractProcessor {
public MyProcessor() {
receive(ReceiveBuilder.
match(Persistent.class, p -> p.payload().equals("a"),
p -> {/* ... */}).
match(Persistent.class, p -> p.payload().equals("b"),
p -> {/* ... */}).build()
);
}
}
class Example {
final ActorSystem system
= ActorSystem.create("example");
final ActorRef
processor = system.actorOf(Props.create(MyProcessor.class));
public void batchWrite() {
processor.tell(PersistentBatch
.create(asList(Persistent.create("a"),
Persistent.create("b"))), null);
}
// ...
}
Persistent messages contained in a PersistentBatch are always written atomically, even if the batch
size is greater than max-message-batch-size. Also, a PersistentBatch is written isolated from other
batches. Persistent messages contained in a PersistentBatch are received individually by a processor.
PersistentBatch
messages,
for
example,
are
used
internally
by
an
AbstractEventsourcedProcessor to ensure atomic writes of events. All events that are persisted
in context of a single command are written as a single batch to the journal (even if persist is called multiple
times per command). The recovery of an AbstractEventsourcedProcessor will therefore never be done
partially (with only a subset of events persisted by a single command).
Confirmation
and
deletion
operations
performed
by
Channels
are
also
222
batched.
The
maximum
confirmation
and
deletion
batch
sizes
are
with
akka.persistence.journal.max-confirmation-batch-size
akka.persistence.journal.max-deletion-batch-size, respectively.
configurable
and
akka.japi.pf.ReceiveBuilder;
scala.PartialFunction;
scala.concurrent.Future;
akka.japi.Option;
akka.japi.Procedure;
akka.persistence.*;
akka.persistence.journal.japi.*;
akka.persistence.snapshot.japi.*;
AsyncWriteJournal is an actor that should be extended if the storage backend API supports asynchronous,
non-blocking writes. In this case, the methods to be implemented are:
223
/**
* Java API, Plugin API: synchronously writes a batch of persistent messages to the
* journal. The batch write must be atomic i.e. either all persistent messages in the
* batch are written or none.
*/
Future<Void> doAsyncWriteMessages(Iterable<PersistentRepr> messages);
/**
* Java API, Plugin API: synchronously writes a batch of delivery confirmations to
* the journal.
*/
Future<Void> doAsyncWriteConfirmations(Iterable<PersistentConfirmation> confirmations);
/**
* Java API, Plugin API: synchronously deletes messages identified by messageIds
* from the journal. If permanent is set to false, the persistent messages are
* marked as deleted, otherwise they are permanently deleted.
*/
Future<Void> doAsyncDeleteMessages(Iterable<PersistentId> messageIds, boolean permanent);
/**
* Java API, Plugin API: synchronously deletes all persistent messages up to
* toSequenceNr. If permanent is set to false, the persistent messages are
* marked as deleted, otherwise they are permanently deleted.
*
* @see AsyncRecoveryPlugin
*/
Future<Void> doAsyncDeleteMessagesTo(String processorId, long toSequenceNr, boolean permanent);
Message replays and sequence number recovery are always asynchronous, therefore, any journal plugin must
implement:
/**
* Java API, Plugin API: asynchronously replays persistent messages.
* Implementations replay a message by calling replayCallback. The returned
* future must be completed when all messages (matching the sequence number
* bounds) have been replayed. The future must be completed with a failure if
* any of the persistent messages could not be replayed.
*
* The replayCallback must also be called with messages that have been marked
* as deleted. In this case a replayed messages deleted method must return
* true.
*
* The channel ids of delivery confirmations that are available for a replayed
* message must be contained in that messages confirms sequence.
*
* @param processorId processor id.
* @param fromSequenceNr sequence number where replay should start (inclusive).
* @param toSequenceNr sequence number where replay should end (inclusive).
* @param max maximum number of messages to be replayed.
* @param replayCallback called to replay a single message. Can be called from any
thread.
*
*/
Future<Void> doAsyncReplayMessages(String processorId, long fromSequenceNr, long toSequenceNr, lon
/**
* Java API, Plugin API: asynchronously reads the highest stored sequence number
* for the given processorId.
*
* @param processorId processor id.
* @param fromSequenceNr hint where to start searching for the highest sequence
number.
*
/
*
224
/**
* Java API, Plugin API: asynchronously loads a snapshot.
*
* @param processorId processor id.
* @param criteria selection criteria for loading.
*/
Future<Option<SelectedSnapshot>> doLoadAsync(String processorId, SnapshotSelectionCriteria criteri
/**
* Java API, Plugin API: asynchronously saves a snapshot.
*
* @param metadata snapshot metadata.
* @param snapshot snapshot.
*/
Future<Void> doSaveAsync(SnapshotMetadata metadata, Object snapshot);
/**
* Java API, Plugin API: called after successful saving of a snapshot.
*
* @param metadata snapshot metadata.
*/
void onSaved(SnapshotMetadata metadata) throws Exception;
/**
* Java API, Plugin API: deletes the snapshot identified by metadata.
*
* @param metadata snapshot metadata.
*/
void doDelete(SnapshotMetadata metadata) throws Exception;
/**
* Java API, Plugin API: deletes all snapshots matching criteria.
*
* @param processorId processor id.
* @param criteria selection criteria for deleting.
*/
void doDelete(String processorId, SnapshotSelectionCriteria criteria) throws Exception;
225
A snapshot store plugin can be activated with the following minimal configuration:
# Path to the snapshot store plugin to be used
akka.persistence.snapshot-store.plugin = "my-snapshot-store"
# My custom snapshot store plugin
my-snapshot-store {
# Class name of the plugin.
class = "docs.persistence.MySnapshotStore"
# Dispatcher for the plugin actor.
plugin-dispatcher = "akka.persistence.dispatchers.default-plugin-dispatcher"
}
With this plugin, each actor system runs its own private LevelDB instance.
Shared LevelDB journal
A LevelDB instance can also be shared by multiple actor systems (on the same or on different nodes). This, for
example, allows processors to failover to a backup node and continue using the shared journal instance from the
backup node.
Warning: A shared LevelDB instance is a single point of failure and should therefore only be used for testing
purposes. Highly-available, replicated journal are available as Community plugins.
A shared LevelDB instance is started by instantiating the SharedLeveldbStore actor.
final ActorRef store = system.actorOf(Props.create(SharedLeveldbStore.class), "store");
By default, the shared instance writes journaled messages to a local directory named journal in the current
working directory. The storage location can be changed by configuration:
akka.persistence.journal.leveldb-shared.store.dir = "target/shared"
Actor systems that use a shared LevelDB store must activate the akka.persistence.journal.leveldb-shared
plugin.
akka.persistence.journal.plugin = "akka.persistence.journal.leveldb-shared"
This plugin must be initialized by injecting the (remote) SharedLeveldbStore actor reference. Injection is
done by calling the SharedLeveldbJournal.setStore method with the actor reference as argument.
class SharedStorageUsage extends AbstractActor {
@Override
public void preStart() throws Exception {
String path = "akka.tcp://[email protected]:2552/user/store";
ActorSelection selection = context().actorSelection(path);
selection.tell(new Identify(1), self());
226
}
public SharedStorageUsage() {
receive(ReceiveBuilder.
match(ActorIdentity.class, ai -> {
if (ai.correlationId().equals(1)) {
ActorRef store = ai.getRef();
if (store != null) {
SharedLeveldbJournal.setStore(store, context().system());
}
}
}).build()
);
}
}
Internal journal commands (sent by processors) are buffered until injection completes. Injection is idempotent i.e.
only the first injection is used.
Local snapshot store
The default snapshot store plugin is akka.persistence.snapshot-store.local. It writes snapshot
files to the local filesystem. The default storage location is a directory named snapshots in the current working
directory. This can be changed by configuration where the specified path can be relative or absolute:
akka.persistence.snapshot-store.local.dir = "target/snapshots"
4.4.13 Testing
When running tests with LevelDB default settings in sbt, make sure to set fork := true in your sbt project
otherwise, youll see an UnsatisfiedLinkError. Alternatively, you can switch to a LevelDB Java port by
setting
akka.persistence.journal.leveldb.native = off
or
227
akka.persistence.journal.leveldb-shared.store.native = off
in your Akka configuration. The LevelDB Java port is for testing purposes only.
228
CHAPTER
FIVE
akka.dispatch.*;
scala.concurrent.ExecutionContext;
scala.concurrent.Future;
scala.concurrent.Await;
scala.concurrent.Promise;
akka.util.Timeout;
scala.concurrent.ExecutionContext;
scala.concurrent.ExecutionContext$;
ExecutionContext ec =
ExecutionContexts.fromExecutorService(yourExecutorServiceGoesHere);
//Use ec with your Futures
Future<String> f1 = Futures.successful("foo");
// Then you shut down the ExecutorService at the end of your application.
yourExecutorServiceGoesHere.shutdown();
229
import
import
import
import
import
import
akka.dispatch.*;
scala.concurrent.ExecutionContext;
scala.concurrent.Future;
scala.concurrent.Await;
scala.concurrent.Promise;
akka.util.Timeout;
This will cause the current thread to block and wait for the UntypedActor to complete the Future with its
reply. Blocking is discouraged though as it can cause performance problem. The blocking operations are located
in Await.result and Await.ready to make it easy to spot where blocking occurs. Alternatives to blocking
are discussed further within this documentation. Also note that the Future returned by an UntypedActor is a
Future<Object> since an UntypedActor is dynamic. That is why the cast to String is used in the above
sample.
Warning: Await.result and Await.ready are provided for exceptional situations where you must
block, a good rule of thumb is to only use them if you know why you must block. For all other cases, use
asynchronous composition as described below.
To send the result of a Future to an Actor, you can use the pipe construct:
akka.pattern.Patterns.pipe(future, system.dispatcher()).to(actor);
scala.concurrent.duration.Duration;
akka.japi.Function;
java.util.concurrent.Callable;
static akka.dispatch.Futures.future;
static java.util.concurrent.TimeUnit.SECONDS;
In the above code the block passed to future will be executed by the default Dispatcher, with the return
value of the block used to complete the Future (in this case, the result would be the string: HelloWorld).
Unlike a Future that is returned from an UntypedActor, this Future is properly typed, and we also avoid
the overhead of managing an UntypedActor.
You can also create already completed Futures using the Futures class, which can be either successes:
Future<String> future = Futures.successful("Yay!");
Or failures:
Future<String> otherFuture = Futures.failed(
new IllegalArgumentException("Bang!"));
It is also possible to create an empty Promise, to be filled later, and obtain the corresponding Future:
5.1. Futures
230
scala.concurrent.duration.Duration;
akka.japi.Function;
java.util.concurrent.Callable;
static akka.dispatch.Futures.future;
static java.util.concurrent.TimeUnit.SECONDS;
In this example we are joining two strings together within a Future. Instead of waiting for f1 to complete, we
apply our function that calculates the length of the string using the map method. Now we have a second Future,
f2, that will eventually contain an Integer. When our original Future, f1, completes, it will also apply our
function and complete the second Future with its result. When we finally get the result, it will contain the
number 10. Our original Future still contains the string HelloWorld and is unaffected by the map.
Something to note when using these methods: if the Future is still being processed when one of these methods
are called, it will be the completing thread that actually does the work. If the Future is already complete though,
it will be run in our current thread. For example:
final ExecutionContext ec = system.dispatcher();
Future<String> f1 = future(new Callable<String>() {
public String call() throws Exception {
Thread.sleep(100);
5.1. Futures
231
The original Future will take at least 0.1 second to execute now, which means it is still being processed at the
time we call map. The function we provide gets stored within the Future and later executed automatically by
the dispatcher when the result is ready.
If we do the opposite:
final ExecutionContext ec = system.dispatcher();
Future<String> f1 = future(new Callable<String>() {
public String call() {
return "Hello" + "World";
}
}, ec);
// Thread.sleep is only here to prove a point
Thread.sleep(100); // Do not use this in your code
Future<Integer> f2 = f1.map(new Mapper<String, Integer>() {
public Integer apply(String s) {
return s.length();
}
}, ec);
f2.onSuccess(new PrintResult<Integer>(), system.dispatcher());
Our little string has been processed long before our 0.1 second sleep has finished. Because of this, the dispatcher
has moved onto other messages that need processing and can no longer calculate the length of the string for us,
instead it gets calculated in the current thread just as if we werent using a Future.
Normally this works quite well as it means there is very little overhead to running a quick function. If there is
a possibility of the function taking a non-trivial amount of time to process it might be better to have this done
concurrently, and for that we use flatMap:
final ExecutionContext ec = system.dispatcher();
Future<String> f1 = future(new Callable<String>() {
public String call() {
return "Hello" + "World";
}
}, ec);
Future<Integer> f2 = f1.flatMap(new Mapper<String, Future<Integer>>() {
public Future<Integer> apply(final String s) {
return future(new Callable<Integer>() {
public Integer call() {
return s.length();
}
}, ec);
}
}, ec);
f2.onSuccess(new PrintResult<Integer>(), system.dispatcher());
5.1. Futures
232
Now our second Future is executed concurrently as well. This technique can also be used to combine the results
of several Futures into a single calculation, which will be better explained in the following sections.
If you need to do conditional propagation, you can use filter:
final ExecutionContext ec = system.dispatcher();
Future<Integer> future1 = Futures.successful(4);
Future<Integer> successfulFilter = future1.filter(Filter.filterOf(
new Function<Integer, Boolean>() {
public Boolean apply(Integer i) {
return i % 2 == 0;
}
}), ec);
Future<Integer> failedFilter = future1.filter(Filter.filterOf(
new Function<Integer, Boolean>() {
public Boolean apply(Integer i) {
return i % 2 != 0;
}
}), ec);
//When filter fails, the returned Future will be failed with a scala.MatchError
Composing Futures
It is very often desirable to be able to combine different Futures with each other, below are some examples on how
that can be done in a non-blocking fashion.
import static akka.dispatch.Futures.sequence;
final ExecutionContext ec = system.dispatcher();
//Some source generating a sequence of Future<Integer>:s
Iterable<Future<Integer>> listOfFutureInts = source;
// now we have a Future[Iterable[Integer]]
Future<Iterable<Integer>> futureListOfInts = sequence(listOfFutureInts, ec);
// Find the sum of the odd numbers
Future<Long> futureSum = futureListOfInts.map(
new Mapper<Iterable<Integer>, Long>() {
public Long apply(Iterable<Integer> ints) {
long sum = 0;
for (Integer i : ints)
sum += i;
return sum;
}
}, ec);
futureSum.onSuccess(new PrintResult<Long>(), system.dispatcher());
5.1. Futures
233
5.1. Futures
234
Same as with fold, the execution will be started when the last of the Futures is completed, you can also parallelize
it by chunking your futures into sub-sequences and reduce them, and then reduce the reduced results again.
This is just a sample of what can be done.
5.1.6 Callbacks
Sometimes you just want to listen to a Future being completed, and react to that not by creating a new Future,
but by side-effecting. For this Scala supports onComplete, onSuccess and onFailure, of which the latter
two are specializations of the first.
final ExecutionContext ec = system.dispatcher();
future.onSuccess(new OnSuccess<String>() {
public void onSuccess(String result) {
if ("bar" == result) {
//Do something if it resulted in "bar"
} else {
//Do something if it was some other String
}
}
}, ec);
final ExecutionContext ec = system.dispatcher();
future.onFailure(new OnFailure() {
public void onFailure(Throwable failure) {
if (failure instanceof IllegalStateException) {
//Do something if it was this particular failure
} else {
//Do something if it was some other failure
}
}
}, ec);
final ExecutionContext ec = system.dispatcher();
future.onComplete(new OnComplete<String>() {
public void onComplete(Throwable failure, String result) {
if (failure != null) {
//We got a failure, handle it here
} else {
// We got a result, do something with it
}
}
}, ec);
5.1.7 Ordering
Since callbacks are executed in any order and potentially in parallel, it can be tricky at the times when you need
sequential ordering of operations. But theres a solution! And its name is andThen, and it creates a new Future
with the specified callback, a Future that will have the same result as the Future its called on, which allows
for ordering like in the following sample:
final ExecutionContext ec = system.dispatcher();
Future<String> future1 = Futures.successful("value").andThen(
new OnComplete<String>() {
public void onComplete(Throwable failure, String result) {
if (failure != null)
sendToIssueTracker(failure);
}
5.1. Futures
235
}, ec).andThen(new OnComplete<String>() {
public void onComplete(Throwable failure, String result) {
if (result != null)
sendToTheInternetz(result);
}
}, ec);
You can also combine two Futures into a new Future that will hold a tuple of the two Futures successful results,
using the zip operation.
final ExecutionContext ec = system.dispatcher();
Future<String> future1 = Futures.successful("foo");
Future<String> future2 = Futures.successful("bar");
Future<String> future3 = future1.zip(future2).map(
new Mapper<scala.Tuple2<String, String>, String>() {
public String apply(scala.Tuple2<String, String> zipped) {
return zipped._1() + " " + zipped._2();
}
}, ec);
future3.onSuccess(new PrintResult<String>(), system.dispatcher());
5.1.9 Exceptions
Since the result of a Future is created concurrently to the rest of the program, exceptions must be handled differently. It doesnt matter if an UntypedActor or the dispatcher is completing the Future, if an Exception
is caught the Future will contain it instead of a valid result. If a Future does contain an Exception, calling
Await.result will cause it to be thrown again so it can be handled properly.
It is also possible to handle an Exception by returning a different result. This is done with the recover
method. For example:
final ExecutionContext ec = system.dispatcher();
Future<Integer> future = future(new Callable<Integer>() {
public Integer call() {
return 1 / 0;
}
}, ec).recover(new Recover<Integer>() {
public Integer recover(Throwable problem) throws Throwable {
if (problem instanceof ArithmeticException)
return 0;
else
throw problem;
}
}, ec);
future.onSuccess(new PrintResult<Integer>(), system.dispatcher());
5.1. Futures
236
5.1.10 After
akka.pattern.Patterns.after makes it easy to complete a Future with a value or exception after a
timeout.
import static akka.pattern.Patterns.after;
import java.util.Arrays;
final ExecutionContext ec = system.dispatcher();
Future<String> failExc = Futures.failed(new IllegalStateException("OHNOES1"));
Future<String> delayed = Patterns.after(Duration.create(200, "millis"),
system.scheduler(), ec, failExc);
Future<String> future = future(new Callable<String>() {
public String call() throws InterruptedException {
Thread.sleep(1000);
return "foo";
}
}, ec);
Future<String> result = Futures.firstCompletedOf(
Arrays.<Future<String>>asList(future, delayed), ec);
5.2 Agents
Agents in Akka are inspired by agents in Clojure.
Agents provide asynchronous change of individual locations. Agents are bound to a single storage location for
their lifetime, and only allow mutation of that location (to a new state) to occur as a result of an action. Update
actions are functions that are asynchronously applied to the Agents state and whose return value becomes the
Agents new state. The state of an Agent should be immutable.
5.2. Agents
237
While updates to Agents are asynchronous, the state of an Agent is always immediately available for reading by
any thread (using get) without any messages.
Agents are reactive.
The update actions of all Agents get interleaved amongst threads in an
ExecutionContext. At any point in time, at most one send action for each Agent is being executed. Actions
dispatched to an agent from another thread will occur in the order they were sent, potentially interleaved with
actions dispatched to the same agent from other threads.
Reading an Agents current value does not involve any message passing and happens immediately. So while
updates to an Agent are asynchronous, reading the state of an Agent is synchronous.
You can also get a Future to the Agents value, that will be completed after the currently queued updates have
completed:
import scala.concurrent.Future;
Future<Integer> future = agent.future();
You can also dispatch a function to update the internal state but on its own thread. This does not use the reactive
thread pool and can be used for long-running or blocking operations. You do this with the sendOff method.
Dispatches using either sendOff or send will still be executed in order.
5.2. Agents
238
import akka.dispatch.Mapper;
// sendOff a function
agent.sendOff(longRunningOrBlockingFunction,
theExecutionContextToExecuteItIn);
All send methods also have a corresponding alter method that returns a Future. See Futures for more
information on Futures.
import akka.dispatch.Mapper;
import scala.concurrent.Future;
// alter a value
Future<Integer> f1 = agent.alter(7);
// alter a function (Mapper)
Future<Integer> f2 = agent.alter(new Mapper<Integer, Integer>() {
public Integer apply(Integer i) {
return i * 2;
}
});
import akka.dispatch.Mapper;
import scala.concurrent.Future;
// alterOff a function (Mapper)
Future<Integer> f3 = agent.alterOff(longRunningOrBlockingFunction,
theExecutionContextToExecuteItIn);
5.2.4 Configuration
There are several configuration properties for the agents module, please refer to the reference configuration.
5.2. Agents
239
CHAPTER
SIX
NETWORKING
6.1 Cluster Specification
Note: This document describes the design concepts of the clustering. It is divided into two parts, where the first
part describes what is currently implemented and the second part describes what is planned as future enhancements/additions. References to unimplemented parts have been marked with the footnote [*]
240
Gossip
The cluster membership used in Akka is based on Amazons Dynamo system and particularly the approach taken
in Bashos Riak distributed database. Cluster membership is communicated using a Gossip Protocol, where the
current state of the cluster is gossiped randomly through the cluster, with preference to members that have not
seen the latest version.
Vector Clocks Vector clocks are a type of data structure and algorithm for generating a partial ordering of events
in a distributed system and detecting causality violations.
We use vector clocks to reconcile and merge differences in cluster state during gossiping. A vector clock is a set
of (node, counter) pairs. Each update to the cluster state has an accompanying update to the vector clock.
Gossip Convergence Information about the cluster converges locally at a node at certain points in time. This is
when a node can prove that the cluster state he is observing has been observed by all other nodes in the cluster.
Convergence is implemented by passing a set of nodes that have seen current state version during gossip. This
information is referred to as the seen set in the gossip overview. When all nodes are included in the seen set there
is convergence.
Gossip convergence cannot occur while any nodes are unreachable. The nodes need to become reachable
again, or moved to the down and removed states (see the Membership Lifecycle section below). This only
blocks the leader from performing its cluster membership management and does not influence the application
running on top of the cluster. For example this means that during a network partition it is not possible to add more
nodes to the cluster. The nodes can join, but they will not be moved to the up state until the partition has healed
or the unreachable nodes have been downed.
Failure Detector The failure detector is responsible for trying to detect if a node is unreachable from the
rest of the cluster. For this we are using an implementation of The Phi Accrual Failure Detector by Hayashibara
et al.
An accrual failure detector decouple monitoring and interpretation. That makes them applicable to a wider area
of scenarios and more adequate to build generic failure detection services. The idea is that it is keeping a history
of failure statistics, calculated from heartbeats received from other nodes, and is trying to do educated guesses by
taking multiple factors, and how they accumulate over time, into account in order to come up with a better guess
if a specific node is up or down. Rather than just answering yes or no to the question is the node down? it
returns a phi value representing the likelihood that the node is down.
The threshold that is the basis for the calculation is configurable by the user. A low threshold is prone
to generate many wrong suspicions but ensures a quick detection in the event of a real crash. Conversely, a high
threshold generates fewer mistakes but needs more time to detect actual crashes. The default threshold is
8 and is appropriate for most situations. However in cloud environments, such as Amazon EC2, the value could
be increased to 12 in order to account for network issues that sometimes occur on such platforms.
In a cluster each node is monitored by a few (default maximum 5) other nodes, and when any of these detects the
node as unreachable that information will spread to the rest of the cluster through the gossip. In other words,
only one node needs to mark a node unreachable to have the rest of the cluster mark that node unreachable.
The nodes to monitor are picked out of neighbors in a hashed ordered node ring. This is to increase the likelihood
to monitor across racks and data centers, but the order is the same on all nodes, which ensures full coverage.
Heartbeats are sent out every second and every heartbeat is performed in a request/reply handshake with the replies
used as input to the failure detector.
The failure detector will also detect if the node becomes reachable again. When all nodes that monitored the
unreachable node detects it as reachable again the cluster, after gossip dissemination, will consider it as
reachable.
If system messages cannot be delivered to a node it will be quarantined and then it cannot come back from
unreachable. This can happen if the there are too many unacknowledged system messages (e.g. watch,
Terminated, remote actor deployment, failures of actors supervised by remote parent). Then the node needs to be
241
moved to the down or removed states (see the Membership Lifecycle section below) and the actor system must
be restarted before it can join the cluster again.
Leader After gossip convergence a leader for the cluster can be determined. There is no leader election
process, the leader can always be recognised deterministically by any node whenever there is gossip convergence. The leader is just a role, any node can be the leader and it can change between convergence rounds. The
leader is simply the first node in sorted order that is able to take the leadership role, where the preferred member
states for a leader are up and leaving (see the Membership Lifecycle section below for more information
about member states).
The role of the leader is to shift members in and out of the cluster, changing joining members to the up
state or exiting members to the removed state. Currently leader actions are only triggered by receiving a
new cluster state with gossip convergence.
The leader also has the power, if configured so, to auto-down a node that according to the Failure Detector
is considered unreachable. This means setting the unreachable node status to down automatically after a
configured time of unreachability.
Seed Nodes The seed nodes are configured contact points for new nodes joining the cluster. When a new node
is started it sends a message to all seed nodes and then sends a join command to the seed node that answers first.
The seed nodes configuration value does not have any influence on the running cluster itself, it is only relevant for
new nodes joining the cluster as it helps them to find contact points to send the join command to; a new member
can send this command to any current member of the cluster, not only to the seed nodes.
Gossip Protocol A variation of push-pull gossip is used to reduce the amount of gossip information sent around
the cluster. In push-pull gossip a digest is sent representing current versions but not actual values; the recipient of
the gossip can then send back any values for which it has newer versions and also request values for which it has
outdated versions. Akka uses a single shared state with a vector clock for versioning, so the variant of push-pull
gossip used in Akka makes use of this version to only push the actual state as needed.
Periodically, the default is every 1 second, each node chooses another random node to initiate a round of gossip
with. If less than of the nodes resides in the seen set (have seen the new state) then the cluster gossips 3 times
instead of once every second. This adjusted gossip interval is a way to speed up the convergence process in the
early dissemination phase after a state change.
The choice of node to gossip with is random but it is biased to towards nodes that might not have seen the
current state version. During each round of gossip exchange when no convergence it uses a probability of 0.8
(configurable) to gossip to a node not part of the seen set, i.e. that probably has an older version of the state.
Otherwise gossip to any random live node.
This biased selection is a way to speed up the convergence process in the late dissemination phase after a state
change.
For clusters larger than 400 nodes (configurable, and suggested by empirical evidence) the 0.8 probability is
gradually reduced to avoid overwhelming single stragglers with too many concurrent gossip requests. The gossip
receiver also has a mechanism to protect itself from too many simultaneous gossip messages by dropping messages
that have been enqueued in the mailbox for too long time.
While the cluster is in a converged state the gossiper only sends a small gossip status message containing the
gossip version to the chosen node. As soon as there is a change to the cluster (meaning non-convergence) then it
goes back to biased gossip again.
The recipient of the gossip state or the gossip status can use the gossip version (vector clock) to determine whether:
1. it has a newer version of the gossip state, in which case it sends that back to the gossiper
2. it has an outdated version of the state, in which case the recipient requests the current state from the gossiper
by sending back its version of the gossip state
3. it has conflicting gossip versions, in which case the different versions are merged and sent back
242
If the recipient and the gossip have the same version then the gossip state is not sent or requested.
The periodic nature of the gossip has a nice batching effect of state changes, e.g. joining several nodes quickly
after each other to one node will result in only one state change to be spread to other members in the cluster.
The gossip messages are serialized with protobuf and also gzipped to reduce payload size.
Membership Lifecycle
A node begins in the joining state. Once all nodes have seen that the new node is joining (through gossip
convergence) the leader will set the member state to up.
If a node is leaving the cluster in a safe, expected manner then it switches to the leaving state. Once the leader
sees the convergence on the node in the leaving state, the leader will then move it to exiting. Once all
nodes have seen the exiting state (convergence) the leader will remove the node from the cluster, marking it as
removed.
If a node is unreachable then gossip convergence is not possible and therefore any leader actions are also
not possible (for instance, allowing a node to become a part of the cluster). To be able to move forward the state
of the unreachable nodes must be changed. It must become reachable again or marked as down. If the
node is to join the cluster again the actor system must be restarted and go through the joining process again. The
cluster can, through the leader, also auto-down a node after a configured time of unreachability..
Note: If you have auto-down enabled and the failure detector triggers, you can over time end up with a lot of
single node clusters if you dont put measures in place to shut down nodes that have become unreachable.
This follows from the fact that the unreachable node will likely see the rest of the cluster as unreachable,
become its own leader and form its own cluster.
243
244
Partitioning [*]
If 5 new nodes join the cluster and in sorted order these nodes appear after the current nodes 2, 4, 5, 7, and 8, then
the partition table could be updated to the following, with all instances on the same physical nodes as before:
A -> { 1, 2, 4, 5 }
B -> { 7, 9, 10 }
C -> { 12, 14, 15, 1, 2 }
When rebalancing is required the leader will schedule handoffs, gossiping a set of pending changes, and when
each change is complete the leader will update the partition table.
Additional Leader Responsibilities
After moving a member from joining to up, the leader can start assigning partitions [*] to the new node, and when
a node is leaving the leader will reassign partitions [*] across the cluster (it is possible for a leaving node to
itself be the leader). When all partition handoff [*] has completed then the node will change to the exiting
state.
On convergence the leader can schedule rebalancing across the cluster, but it may also be possible for the user to
explicitly rebalance the cluster by specifying migrations [*], or to rebalance [*] the cluster automatically based
on metrics from member nodes. Metrics may be spread using the gossip protocol or possibly more efficiently
6.1. Cluster Specification
245
using a random chord method, where the leader contacts several random nodes around the cluster ring and
each contacted node gathers information from their immediate neighbours, giving a random sampling of load
information.
Handoff
Handoff for an actor-based system is different than for a data-based system. The most important point is that
message ordering (from a given node to a given actor instance) may need to be maintained. If an actor is a
singleton actor (only one instance possible throughout the cluster) then the cluster may also need to assure that
there is only one such actor active at any one time. Both of these situations can be handled by forwarding and
buffering messages during transitions.
A graceful handoff (one where the previous host node is up and running during the handoff), given a previous host
node N1, a new host node N2, and an actor partition A to be migrated from N1 to N2, has this general structure:
1. the leader sets a pending change for N1 to handoff A to N2
2. N1 notices the pending change and sends an initialization message to N2
3. in response N2 creates A and sends back a ready message
4. after receiving the ready message N1 marks the change as complete and shuts down A
5. the leader sees the migration is complete and updates the partition table
6. all nodes eventually see the new partitioning and use N2
Transitions There are transition times in the handoff process where different approaches can be used to give
different guarantees.
Migration Transition The first transition starts when N1 initiates the moving of A and ends when N1 receives
the ready message, and is referred to as the migration transition.
The first question is; during the migration transition, should:
N1 continue to process messages for A?
Or is it important that no messages for A are processed on N1 once migration begins?
If it is okay for the previous host node N1 to process messages during migration then there is nothing that needs
to be done at this point.
If no messages are to be processed on the previous host node during migration then there are two possibilities: the
messages are forwarded to the new host and buffered until the actor is ready, or the messages are simply dropped
by terminating the actor and allowing the normal dead letter process to be used.
Update Transition The second transition begins when the migration is marked as complete and ends when all
nodes have the updated partition table (when all nodes will use N2 as the host for A, i.e. we have convergence)
and is referred to as the update transition.
Once the update transition begins N1 can forward any messages it receives for A to the new host N2. The question
is whether or not message ordering needs to be preserved. If messages sent to the previous host node N1 are being
forwarded, then it is possible that a message sent to N1 could be forwarded after a direct message to the new host
N2, breaking message ordering from a client to actor A.
In this situation N2 can keep a buffer for messages per sending node. Each buffer is flushed and removed when
an acknowledgement (ack) message has been received. When each node in the cluster sees the partition update
it first sends an ack message to the previous host node N1 before beginning to use N2 as the new host for A. Any
messages sent from the client node directly to N2 will be buffered. N1 can count down the number of acks to
determine when no more forwarding is needed. The ack message from any node will always follow any other
messages sent to N1. When N1 receives the ack message it also forwards it to N2 and again this ack message will
follow any other messages already forwarded for A. When N2 receives an ack message, the buffer for the sending
246
node can be flushed and removed. Any subsequent messages from this sending node can be queued normally.
Once all nodes in the cluster have acknowledged the partition change and N2 has cleared all buffers, the handoff is
complete and message ordering has been preserved. In practice the buffers should remain small as it is only those
messages sent directly to N2 before the acknowledgement has been forwarded that will be buffered.
Graceful Handoff A more complete process for graceful handoff would be:
1. the leader sets a pending change for N1 to handoff A to N2
2. N1 notices the pending change and sends an initialization message to N2. Options:
(a) keep A on N1 active and continuing processing messages as normal
(b) N1 forwards all messages for A to N2
(c) N1 drops all messages for A (terminate A with messages becoming dead letters)
3. in response N2 creates A and sends back a ready message. Options:
(a) N2 simply processes messages for A as normal
(b) N2 creates a buffer per sending node for A. Each buffer is opened (flushed and removed) when an
acknowledgement for the sending node has been received (via N1)
4. after receiving the ready message N1 marks the change as complete. Options:
(a) N1 forwards all messages for A to N2 during the update transition
(b) N1 drops all messages for A (terminate A with messages becoming dead letters)
5. the leader sees the migration is complete and updates the partition table
6. all nodes eventually see the new partitioning and use N2
(a) each node sends an acknowledgement message to N1
(b) when N1 receives the acknowledgement it can count down the pending acknowledgements and remove
forwarding when complete
(c) when N2 receives the acknowledgement it can open the buffer for the sending node (if buffers are
used)
The default approach is to take options 2a, 3a, and 4a - allowing A on N1 to continue processing messages during
migration and then forwarding any messages during the update transition. This assumes stateless actors that do
not have a dependency on message ordering from any given source.
If an actor has persistent (durable) state then nothing needs to be done, other than migrating the actor.
If message ordering needs to be maintained during the update transition then option 3b can be used, creating
buffers per sending node.
If the actors are robust to message send failures then the dropping messages approach can be used (with no
forwarding or buffering needed).
If an actor is a singleton (only one instance possible throughout the cluster) and state is transferred during
the migration initialization, then options 2b and 3b would be required.
Stateful Actor Replication [*]
247
To enable cluster capabilities in your Akka project you should, at a minimum, add the Remoting settings, but with
akka.cluster.ClusterActorRefProvider. The akka.cluster.seed-nodes should normally
also be added to your application.conf file.
The seed nodes are configured contact points for initial, automatic, join of the cluster.
Note that if you are going to start the nodes on different machines you need to specify the ip-addresses or host
names of the machines in application.conf instead of 127.0.0.1
248
An actor that uses the cluster extension may look like this:
package sample.cluster.simple;
import
import
import
import
import
import
import
import
import
akka.actor.UntypedActor;
akka.cluster.Cluster;
akka.cluster.ClusterEvent;
akka.cluster.ClusterEvent.MemberEvent;
akka.cluster.ClusterEvent.MemberUp;
akka.cluster.ClusterEvent.MemberRemoved;
akka.cluster.ClusterEvent.UnreachableMember;
akka.event.Logging;
akka.event.LoggingAdapter;
The actor registers itself as subscriber of certain cluster events. It receives events corresponding to the current
state of the cluster when the subscription starts and then it receives events for changes that happen in the cluster.
The easiest way to run this example yourself is to download Typesafe Activator and open the tutorial named Akka
Cluster Samples with Java. It contains instructions of how to run the SimpleClusterApp.
249
This can also be defined as Java system properties when starting the JVM using the following syntax:
-Dakka.cluster.seed-nodes.0=akka.tcp://ClusterSystem@host1:2552
-Dakka.cluster.seed-nodes.1=akka.tcp://ClusterSystem@host2:2552
The seed nodes can be started in any order and it is not necessary to have all seed nodes running, but the node
configured as the first element in the seed-nodes configuration list must be started when initially starting a
cluster, otherwise the other seed-nodes will not become initialized and no other node can join the cluster. The
reason for the special first seed node is to avoid forming separated islands when starting from an empty cluster. It
is quickest to start all configured seed nodes at the same time (order doesnt matter), otherwise it can take up to
the configured seed-node-timeout until the nodes can join.
Once more than two seed nodes have been started it is no problem to shut down the first seed node. If the first
seed node is restarted it will first try join the other seed nodes in the existing cluster.
If you dont configure the seed nodes you need to join manually, using JMX or Command Line Management. You
can join to any node in the cluster. It doesnt have to be configured as a seed node.
Joining can also be performed programatically with Cluster.get(system).join. Note that you can only
join to an existing cluster member, which means that for bootstrapping some node must join itself.
You may also use Cluster.get(system).joinSeedNodes, which is attractive when dynamically discovering other nodes at startup by using some external tool or API. When using joinSeedNodes you should not
include the node itself except for the node that is supposed to be the first seed node, and that should be placed first
in parameter to joinSeedNodes.
Unsuccessful join attempts are automatically retried after the time period defined in configuration property
retry-unsuccessful-join-after. When using seed-nodes this means that a new seed node is
picked. When joining manually or programatically this means that the last join request is retried. Retries can
be disabled by setting the property to off.
An actor system can only join a cluster once. Additional attempts will be ignored. When it has successfully joined
it must be restarted to be able to join another cluster or to join the same cluster again. It can use the same host
name and port after the restart, but it must have been removed from the cluster before the join request is accepted.
This means that the cluster leader member will change the unreachable node status to down automatically
after the configured time of unreachability.
250
Be aware of that using auto-down implies that two separate clusters will automatically be formed in case of
network partition. That might be desired by some applications but not by others.
Note: If you have auto-down enabled and the failure detector triggers, you can over time end up with a lot of
single node clusters if you dont put measures in place to shut down nodes that have become unreachable.
This follows from the fact that the unreachable node will likely see the rest of the cluster as unreachable,
become its own leader and form its own cluster.
6.2.5 Leaving
There are two ways to remove a member from the cluster.
You can just stop the actor system (or the JVM process). It will be detected as unreachable and removed after the
automatic or manual downing as described above.
A more graceful exit can be performed if you tell the cluster that a node shall leave. This can be
performed using JMX or Command Line Management. It can also be performed programatically with
Cluster.get(system).leave(address).
Note that this command can be issued to any member in the cluster, not necessarily the one that is leaving. The
cluster extension, but not the actor system or JVM, of the leaving member will be shutdown after the leader has
changed status of the member to Exiting. Thereafter the member will be removed from the cluster. Normally this
is handled automatically, but in case of network failures during this process it might still be necessary to set the
nodes status to Down in order to complete the removal.
of
the
cluster
membership
by
using
251
252
}
public String getReason() {
return reason;
}
public TransformationJob getJob() {
return job;
}
@Override
public String toString() {
return "JobFailed(" + reason + ")";
}
}
public static final String BACKEND_REGISTRATION = "BackendRegistration";
}
253
if (member.hasRole("frontend"))
getContext().actorSelection(member.address() + "/user/frontend").tell(
BACKEND_REGISTRATION, getSelf());
}
}
Note that the TransformationBackend actor subscribes to cluster events to detect new, potential, frontend
nodes, and send them a registration message so that they know that they can use the backend worker.
The frontend that receives user jobs and delegates to one of the registered backend workers:
public class TransformationFrontend extends UntypedActor {
List<ActorRef> backends = new ArrayList<ActorRef>();
int jobCounter = 0;
@Override
public void onReceive(Object message) {
if ((message instanceof TransformationJob) && backends.isEmpty()) {
TransformationJob job = (TransformationJob) message;
getSender().tell(
new JobFailed("Service unavailable, try again later", job),
getSender());
} else if (message instanceof TransformationJob) {
TransformationJob job = (TransformationJob) message;
jobCounter++;
backends.get(jobCounter % backends.size())
.forward(job, getContext());
} else if (message.equals(BACKEND_REGISTRATION)) {
getContext().watch(getSender());
backends.add(getSender());
} else if (message instanceof Terminated) {
Terminated terminated = (Terminated) message;
backends.remove(terminated.getActor());
} else {
unhandled(message);
}
}
}
Note that the TransformationFrontend actor watch the registered backend to be able to remove it from
its list of available backend workers. Death watch uses the cluster failure detector for nodes in the cluster, i.e.
it detects network failures and JVM crashes, in addition to graceful termination of watched actor. Death watch
generates the Terminated message to the watching actor when the unreachable cluster node has been downed
and removed.
The Typesafe Activator tutorial named Akka Cluster Samples with Java. contains the full source code and instructions of how to run the Worker Dial-in Example.
254
The roles of the nodes is part of the membership information in MemberEvent that you can subscribe to.
In a similar way you can define required number of members of a certain role before the leader changes member
status of Joining members to Up.
akka.cluster.role {
frontend.min-nr-of-members = 1
backend.min-nr-of-members = 2
}
You can start the actors in a registerOnMemberUp callback, which will be invoked when the current member
status is changed tp Up, i.e. the cluster has at least the defined number of members.
Cluster.get(system).registerOnMemberUp(new Runnable() {
@Override
public void run() {
system.actorOf(Props.create(FactorialFrontend.class, upToN, true),
"factorialFrontend");
}
});
This callback can be used for other things than starting actors.
255
where F is the cumulative distribution function of a normal distribution with mean and standard deviation estimated
from historical heartbeat inter-arrival times.
In the Configuration you can adjust the akka.cluster.failure-detector.threshold to define when
a phi value is considered to be a failure.
A low threshold is prone to generate many false positives but ensures a quick detection in the event of a real
crash. Conversely, a high threshold generates fewer mistakes but needs more time to detect actual crashes.
The default threshold is 8 and is appropriate for most situations. However in cloud environments, such as
Amazon EC2, the value could be increased to 12 in order to account for network issues that sometimes occur on
such platforms.
The following chart illustrates how phi increase with increasing time since the previous heartbeat.
256
Phi is calculated from the mean and standard deviation of historical inter arrival times. The previous chart is an
example for standard deviation of 200 ms. If the heartbeats arrive with less deviation the curve becomes steeper,
i.e. it is possible to determine failure more quickly. The curve looks like this for a standard deviation of 100 ms.
To
be
able
to
survive
sudden
abnormalities,
such
as
garbage
collection
pauses
and
transient
network
failures
the
failure
detector
is
configured
with
a
margin,
akka.cluster.failure-detector.acceptable-heartbeat-pause.
You may want to
adjust the Configuration of this depending on you environment. This is how the curve looks like for
acceptable-heartbeat-pause configured to 3 seconds.
6.2. Cluster Usage
257
Death watch uses the cluster failure detector for nodes in the cluster, i.e. it detects network failures and JVM
crashes, in addition to graceful termination of watched actor. Death watch generates the Terminated message
to the watching actor when the unreachable cluster node has been downed and removed.
If you encounter suspicious false positives when the system is under load you should define a separate dispatcher
for the cluster actors as described in Cluster Dispatcher.
258
akka.actor.deployment {
/statsService/workerRouter {
router = consistent-hashing-group
nr-of-instances = 100
routees.paths = ["/user/statsWorker"]
cluster {
enabled = on
allow-local-routees = on
use-role = compute
}
}
}
Note: The routee actors should be started as early as possible when starting the actor system, because the router
will try to use them as soon as the member status is changed to Up. If it is not available at that point it will be
removed from the router and it will only re-try when the cluster members are changed.
It is the relative actor paths defined in routees.paths that identify what actor to lookup. It is possible to limit
the lookup of routees to member nodes tagged with a certain role by specifying use-role.
nr-of-instances defines total number of routees in the cluster. Setting nr-of-instances to a high value
will result in new routees added to the router when nodes join the cluster.
The same type of router could also have been defined in code:
int totalInstances = 100;
Iterable<String> routeesPaths = Collections
.singletonList("/user/statsWorker");
boolean allowLocalRoutees = true;
String useRole = "compute";
ActorRef workerRouter = getContext().actorOf(
new ClusterRouterGroup(new ConsistentHashingGroup(routeesPaths),
new ClusterRouterGroupSettings(totalInstances, routeesPaths,
allowLocalRoutees, useRole)).props(), "workerRouter2");
259
The service that receives text from users and splits it up into words, delegates to workers and aggregates:
260
261
getContext().stop(getSelf());
}
} else if (message == ReceiveTimeout.getInstance()) {
replyTo.tell(new JobFailed("Service unavailable, try again later"),
getSelf());
getContext().stop(getSelf());
} else {
unhandled(message);
}
}
}
This means that user requests can be sent to StatsService on any node and it will use StatsWorker on all
nodes.
The Typesafe Activator tutorial named Akka Cluster Samples with Java. contains the full source code and instructions of how to run the Router Example with Group of Routees.
Router with Pool of Remote Deployed Routees
When using a Pool with routees created and deployed on the cluster member nodes the configuration for a router
looks like this:
akka.actor.deployment {
/singleton/statsService/workerRouter {
router = consistent-hashing-pool
nr-of-instances = 100
cluster {
enabled = on
max-nr-of-instances-per-node = 3
allow-local-routees = on
use-role = compute
}
}
}
It is possible to limit the deployment of routees to member nodes tagged with a certain role by specifying
use-role.
nr-of-instances defines total number of routees in the cluster, but the number of routees per node,
max-nr-of-instances-per-node, will not be exceeded. Setting nr-of-instances to a high value
will result in creating and deploying additional routees when new nodes join the cluster.
262
The same type of router could also have been defined in code:
int totalInstances = 100;
int maxInstancesPerNode = 3;
boolean allowLocalRoutees = false;
String useRole = "compute";
ActorRef workerRouter = getContext().actorOf(
new ClusterRouterPool(new ConsistentHashingPool(0),
new ClusterRouterPoolSettings(totalInstances, maxInstancesPerNode,
allowLocalRoutees, useRole)).props(Props
.create(StatsWorker.class)), "workerRouter3");
We also need an actor on each node that keeps track of where current single master exists and delegates jobs to
the StatsService. That is provided by the ClusterSingletonProxy.
system.actorOf(ClusterSingletonProxy.defaultProps("/user/singleton/statsService",
"compute"), "statsServiceProxy");
The ClusterSingletonProxy receives text from users and delegates to the current StatsService, the
single master. It listens to cluster events to lookup the StatsService on the oldest node.
All nodes start ClusterSingletonProxy and the ClusterSingletonManager. The router is now
configured like this:
akka.actor.deployment {
/singleton/statsService/workerRouter {
router = consistent-hashing-pool
nr-of-instances = 100
cluster {
enabled = on
max-nr-of-instances-per-node = 3
allow-local-routees = on
use-role = compute
}
}
}
The Typesafe Activator tutorial named Akka Cluster Samples with Java. contains the full source code and instructions of how to run the Router Example with Pool of Remote Deployed Routees.
263
ing a native OS library. To enable usage of Sigar you need to add the directory of the native library to
-Djava.libarary.path=<path_of_sigar_libs> add the following dependency:
<dependency>
<groupId>org.fusesource</groupId>
<artifactId>sigar</artifactId>
<version>1.6.4</version>
</dependency>
264
}
}
BigInteger factorial(int n) {
BigInteger acc = BigInteger.ONE;
for (int i = 1; i <= n; ++i) {
acc = acc.multiply(BigInteger.valueOf(i));
}
return acc;
}
}
The frontend that receives user jobs and delegates to the backends via the router:
public class FactorialFrontend extends UntypedActor {
final int upToN;
final boolean repeat;
LoggingAdapter log = Logging.getLogger(getContext().system(), this);
ActorRef backend = getContext().actorOf(FromConfig.getInstance().props(),
"factorialBackendRouter");
public FactorialFrontend(int upToN, boolean repeat) {
this.upToN = upToN;
this.repeat = repeat;
}
@Override
public void preStart() {
sendJobs();
getContext().setReceiveTimeout(Duration.create(10, TimeUnit.SECONDS));
}
@Override
public void onReceive(Object message) {
if (message instanceof FactorialResult) {
FactorialResult result = (FactorialResult) message;
if (result.n == upToN) {
log.debug("{}! = {}", result.n, result.factorial);
if (repeat)
sendJobs();
else
getContext().stop(getSelf());
}
} else if (message instanceof ReceiveTimeout) {
log.info("Timeout");
sendJobs();
} else {
unhandled(message);
}
}
void sendJobs() {
log.info("Starting batch of factorials up to [{}]", upToN);
for (int n = 1; n <= upToN; n++) {
backend.tell(n, getSelf());
}
}
}
265
As you can see, the router is defined in the same way as other routers, and in this case it is configured as follows:
akka.actor.deployment {
/factorialFrontend/factorialBackendRouter = {
router = adaptive-group
# metrics-selector = heap
# metrics-selector = load
# metrics-selector = cpu
metrics-selector = mix
nr-of-instances = 100
routees.paths = ["/user/factorialBackend"]
cluster {
enabled = on
use-role = backend
allow-local-routees = off
}
}
}
It is only router type adaptive and the metrics-selector that is specific to this router, other things work
in the same way as other routers.
The same type of router could also have been defined in code:
int totalInstances = 100;
Iterable<String> routeesPaths = Arrays.asList("/user/factorialBackend", "");
boolean allowLocalRoutees = true;
String useRole = "backend";
ActorRef backend = getContext().actorOf(
new ClusterRouterGroup(new AdaptiveLoadBalancingGroup(
HeapMetricsSelector.getInstance(), Collections.<String> emptyList()),
new ClusterRouterGroupSettings(totalInstances, routeesPaths,
allowLocalRoutees, useRole)).props(), "factorialBackendRouter2");
int totalInstances = 100;
int maxInstancesPerNode = 3;
boolean allowLocalRoutees = false;
String useRole = "backend";
ActorRef backend = getContext().actorOf(
new ClusterRouterPool(new AdaptiveLoadBalancingPool(
SystemLoadAverageMetricsSelector.getInstance(), 0),
new ClusterRouterPoolSettings(totalInstances, maxInstancesPerNode,
allowLocalRoutees, useRole)).props(Props
.create(FactorialBackend.class)), "factorialBackendRouter3");
The Typesafe Activator tutorial named Akka Cluster Samples with Java. contains the full source code and instructions of how to run the Adaptive Load Balancing sample.
Subscribe to Metrics Events
It is possible to subscribe to the metrics events directly to implement other functionality.
import
import
import
import
import
import
import
import
import
import
akka.actor.UntypedActor;
akka.cluster.Cluster;
akka.cluster.ClusterEvent.ClusterMetricsChanged;
akka.cluster.ClusterEvent.CurrentClusterState;
akka.cluster.NodeMetrics;
akka.cluster.StandardMetrics;
akka.cluster.StandardMetrics.HeapMemory;
akka.cluster.StandardMetrics.Cpu;
akka.event.Logging;
akka.event.LoggingAdapter;
266
@Override
public void onReceive(Object message) {
if (message instanceof ClusterMetricsChanged) {
ClusterMetricsChanged clusterMetrics = (ClusterMetricsChanged) message;
for (NodeMetrics nodeMetrics : clusterMetrics.getNodeMetrics()) {
if (nodeMetrics.address().equals(cluster.selfAddress())) {
logHeap(nodeMetrics);
logCpu(nodeMetrics);
}
}
} else if (message instanceof CurrentClusterState) {
// ignore
} else {
unhandled(message);
}
}
void logHeap(NodeMetrics nodeMetrics) {
HeapMemory heap = StandardMetrics.extractHeapMemory(nodeMetrics);
if (heap != null) {
log.info("Used heap: {} MB", ((double) heap.used()) / 1024 / 1024);
}
}
void logCpu(NodeMetrics nodeMetrics) {
Cpu cpu = StandardMetrics.extractCpu(nodeMetrics);
if (cpu != null && cpu.systemLoadAverage().isDefined()) {
log.info("Load: {} ({} processors)", cpu.systemLoadAverage().get(),
cpu.processors());
}
}
}
267
6.2.16 JMX
Information and management of the cluster is available as JMX MBeans with the root name akka.Cluster.
The JMX information can be displayed with an ordinary JMX console such as JConsole or JVisualVM.
From JMX you can:
see what members that are part of the cluster
see status of this node
join this node to another node in cluster
mark any node in the cluster as down
tell any node in the cluster to leave
Member nodes are identified
name>@<hostname>:<port>.
by
their
address,
in
format
akka.<protocol>://<actor-system-
To be able to use the script you must enable remote monitoring and management when starting the JVMs of the
cluster nodes, as described in Monitoring and Management Using JMX Technology
Example of system properties to enable remote monitoring and management:
java -Dcom.sun.management.jmxremote.port=9999 \
-Dcom.sun.management.jmxremote.authenticate=false \
-Dcom.sun.management.jmxremote.ssl=false
6.2.18 Configuration
There are several configuration properties for the cluster. We refer to the reference configuration for more information.
268
Cluster Dispatcher
Under the hood the cluster extension is implemented with actors and it can be necessary to create a bulkhead
for those actors to avoid disturbance from other actors. Especially the heartbeating actors that is used for failure
detection can generate false positives if they are not given a chance to run at regular intervals. For this purpose
you can define a separate dispatcher to be used for the cluster actors:
akka.cluster.use-dispatcher = cluster-dispatcher
cluster-dispatcher {
type = "Dispatcher"
executor = "fork-join-executor"
fork-join-executor {
parallelism-min = 2
parallelism-max = 4
}
}
6.3 Remoting
For an introduction of remoting capabilities of Akka please see Location Transparency.
Note: As explained in that chapter Akka remoting is designed for communication in a peer-to-peer fashion and
it has limitations for client-server setups. In particular Akka Remoting does not work with Network Address
Translation and Load Balancers, among others.
To enable remote capabilities in your Akka project you should, at a minimum, add the following changes to your
application.conf file:
akka {
actor {
provider = "akka.remote.RemoteActorRefProvider"
}
remote {
enabled-transports = ["akka.remote.netty.tcp"]
netty.tcp {
hostname = "127.0.0.1"
port = 2552
}
}
}
6.3. Remoting
269
As you can see in the example above there are four things you need to add to get started:
Change
provider
from
akka.actor.LocalActorRefProvider
akka.remote.RemoteActorRefProvider
to
Add host name - the machine you want to run the actor system on; this host name is exactly what is passed
to remote systems in order to identify this system and consequently used for connecting back to this system
if need be, hence set it to a reachable IP address or resolvable name in case you want to communicate across
the network.
Add port number - the port the actor system should listen on, set to 0 to have it chosen automatically
Note: The port number needs to be unique for each actor system on the same machine even if the actor systems
have different names. This is because each actor system has its own network subsystem listening for connections
and handling messages as not to interfere with other actor systems.
The example above only illustrates the bare minimum of properties you have to add to enable remoting. All
settings are described in Remote Configuration.
As you can see from the example above the following pattern is used to find an actor on a remote node:
akka.<protocol>://<actorsystemname>@<hostname>:<port>/<actor path>
Once you obtained a selection to the actor you can interact with it they same way you would with a local actor,
e.g.:
selection.tell("Pretty awesome feature", getSelf());
To acquire an ActorRef for an ActorSelection you need to send a message to the selection and use the
getSender reference of the reply from the actor. There is a built-in Identify message that all Actors will understand and automatically reply to with a ActorIdentity message containing the ActorRef. This can also
be done with the resolveOne method of the ActorSelection, which returns a Future of the matching
ActorRef.
Note: For more details on how actor addresses and paths are formed and used, please refer to Actor References,
Paths and Addresses.
6.3. Remoting
270
The configuration above instructs Akka to react when an actor with path /sampleActor is created, i.e. using system.actorOf(new Props(...), "sampleActor"). This specific actor will not be directly
instantiated, but instead the remote daemon of the remote system will be asked to create the actor, which in this
sample corresponds to [email protected]:2553.
Once you have configured the properties above you would do the following in code:
ActorRef actor = system.actorOf(Props.create(SampleActor.class), "sampleActor");
actor.tell("Pretty slick", ActorRef.noSender());
The actor class SampleActor has to be available to the runtimes using it, i.e. the classloader of the actor systems
has to have a JAR containing the class.
Note: In order to ensure serializability of Props when passing constructor arguments to the actor being created,
do not make the factory a non-static inner class: this will inherently capture a reference to its enclosing object,
which in most cases is not serializable. It is best to make a static inner class which implements Creator.
Serializability
of
all
Props
can
be
tested
by
setting
the
configuration
item
akka.actor.serialize-creators=on. Only Props whose deploy has LocalScope are exempt
from this check.
Note:
You can use asterisks as wildcard matches for the actor path sections, so you could specify:
/*/sampleActor and that would match all sampleActor on that level in the hierarchy. You can also use
wildcard in the last position to match all actors at a certain level: /someParent/*. Non-wildcard matches
always have higher priority to match than wildcards, so: /foo/bar is considered more specific than /foo/*
and only the highest priority match is used. Please note that it cannot be used to partially match section, like this:
/foo*/bar, /f*o/bar etc.
akka.actor.ActorRef;
akka.actor.Address;
akka.actor.AddressFromURIString;
akka.actor.Deploy;
akka.actor.Props;
akka.actor.ActorSystem;
akka.remote.RemoteScope;
you can advise the system to create a child on that remote node like so:
ActorRef ref = system.actorOf(Props.create(SampleActor.class).withDeploy(
new Deploy(new RemoteScope(addr))));
6.3. Remoting
271
Each link with a remote system can be in one of the four states as illustrated above. Before any communication
happens with a remote system at a given Address the state of the association is Idle. The first time a message
is attempted to be sent to the remote system or an inbound connection is accepted the state of the link transitions
to Active denoting that the two systems has messages to send or receive and no failures were encountered so
far. When a communication failure happens and the connection is lost between the two systems the link becomes
Gated.
In this state the system will not attempt to connect to the remote host and all outbound messages will be dropped. The time while the link is in the Gated state is controlled by the setting
akka.remote.retry-gate-closed-for: after this time elapses the link state transitions to Idle again.
Gate is one-sided in the sense that whenever a successful inbound connection is accepted from a remote system
during Gate it automatically transitions to Active and communication resumes immediately.
In the face of communication failures that are unrecoverable because the state of the participating systems are
inconsistent, the remote system becomes Quarantined. Unlike Gate, quarantining is permanent and lasts
until one of the systems is restarted. After a restart communication can be resumed again and the link can become
Active again.
6.3. Remoting
272
Failure Detector
Under the hood remote death watch uses heartbeat messages and a failure detector to generate Terminated
message from network failures and JVM crashes, in addition to graceful termination of watched actor.
The heartbeat arrival times is interpreted by an implementation of The Phi Accrual Failure Detector.
The suspicion level of failure is given by a value called phi. The basic idea of the phi failure detector is to express
the value of phi on a scale that is dynamically adjusted to reflect current network conditions.
The value of phi is calculated as:
phi = -log10(1 - F(timeSinceLastHeartbeat))
where F is the cumulative distribution function of a normal distribution with mean and standard deviation estimated
from historical heartbeat inter-arrival times.
In the Remote Configuration you can adjust the akka.remote.watch-failure-detector.threshold
to define when a phi value is considered to be a failure.
A low threshold is prone to generate many false positives but ensures a quick detection in the event of a real
crash. Conversely, a high threshold generates fewer mistakes but needs more time to detect actual crashes.
The default threshold is 10 and is appropriate for most situations. However in cloud environments, such as
Amazon EC2, the value could be increased to 12 in order to account for network issues that sometimes occur on
such platforms.
The following chart illustrates how phi increase with increasing time since the previous heartbeat.
6.3. Remoting
273
Phi is calculated from the mean and standard deviation of historical inter arrival times. The previous chart is an
example for standard deviation of 200 ms. If the heartbeats arrive with less deviation the curve becomes steeper,
i.e. it is possible to determine failure more quickly. The curve looks like this for a standard deviation of 100 ms.
To
be
able
to
survive
sudden
abnormalities,
such
as
garbage
collection
pauses
and
transient
network
failures
the
failure
detector
is
configured
with
a
margin,
akka.remote.watch-failure-detector.acceptable-heartbeat-pause.
You may want
to adjust the Remote Configuration of this depending on you environment. This is how the curve looks like for
acceptable-heartbeat-pause configured to 3 seconds.
6.3. Remoting
274
6.3.6 Serialization
When using remoting for actors you must ensure that the props and messages used for those actors are serializable. Failing to do so will cause the system to behave in an unintended way.
For more information please see Serialization
This configuration setting will clone the actor defined in the Props of the remotePool 10 times and deploy it
evenly distributed across the two given target nodes.
A group of remote actors can be configured as:
akka.actor.deployment {
/parent/remoteGroup {
router = round-robin-group
routees.paths = [
"akka.tcp://[email protected]:2552/user/workers/w1",
"akka.tcp://[email protected]:2552/user/workers/w1",
"akka.tcp://[email protected]:2552/user/workers/w1"]
}
}
This configuration setting will send messages to the defined remote actor paths. It requires that you create the
destination actors on the remote nodes with matching paths. That is not done by the router.
6.3. Remoting
275
key-password = "changeme"
trust-store-password = "changeme"
protocol = "TLSv1"
random-number-generator = "AES128CounterSecureRNG"
enabled-algorithms = [TLS_RSA_WITH_AES_128_CBC_SHA]
}
}
}
Remote Events
It is possible to listen to events that occur in Akka Remote, and to subscribe/unsubscribe to these events you
simply register as listener to the below described types in on the ActorSystem.eventStream.
Note: To subscribe to any remote event, subscribe to RemotingLifecycleEvent. To subscribe to events
related only to the lifecycle of associations, subscribe to akka.remote.AssociationEvent.
Note: The use of term Association instead of Connection reflects that the remoting subsystem may use connectionless transports, but an association similar to transport layer connections is maintained between endpoints
by the Akka protocol.
By default an event listener is registered which logs all of the events described below. This default was chosen to
help setting up a system, but it is quite common to switch this logging off once that phase of the project is finished.
Note: In order to switch off the logging, set akka.remote.log-remote-lifecycle-events = off
in your application.conf.
To be notified when an association is over (disconnected) listen to DisassociatedEvent which holds the
direction of the association (inbound or outbound) and the addresses of the involved parties.
To be notified when an association is successfully established (connected) listen to AssociatedEvent which
holds the direction of the association (inbound or outbound) and the addresses of the involved parties.
To intercept errors directly related to associations, listen to AssociationErrorEvent which holds the direction of the association (inbound or outbound), the addresses of the involved parties and the Throwable cause.
6.3. Remoting
276
To be notified when the remoting subsystem is ready to accept associations, listen to RemotingListenEvent
which contains the addresses the remoting listens on.
To be notified when the remoting subsystem has been shut down, listen to RemotingShutdownEvent.
To intercept generic remoting related errors, listen to RemotingErrorEvent which holds the Throwable
cause.
This disallows sending of system messages (actor life-cycle commands, DeathWatch, etc.) and any message
extending PossiblyHarmful to the system on which this flag is set. Should a client send them nonetheless
they are dropped and logged (at DEBUG level in order to reduce the possibilities for a denial of service attack).
PossiblyHarmful covers the predefined messages like PoisonPill and Kill, but it can also be added as
a marker trait to user-defined messages.
Messages sent with actor selection are by default discarded in untrusted mode, but permission to receive actor
selection messages can be granted to specific actors defined in configuration:
akka.remote.trusted-selection-paths = ["/user/receptionist", "/user/namingService"]
6.3. Remoting
277
SSL
SSL can be used as the remote transport by adding akka.remote.netty.ssl to the enabled-transport
configuration section. See a description of the settings in the Remote Configuration section.
The SSL support is implemented with Java Secure Socket Extension, please consult the offical Java Secure Socket
Extension documentation and related resources for troubleshooting.
Note:
When
using
SHA1PRNG
on
Linux
its
recommended
specify
-Djava.security.egd=file:/dev/./urandom as argument to the JVM to prevent blocking. It
is NOT as secure because it reuses the seed. Use /dev/./urandom, not /dev/urandom as that doesnt work
according to Bug ID: 6202721.
6.4 Serialization
Akka has a built-in Extension for serialization, and it is both possible to use the built-in serializers and to write
your own.
The serialization mechanism is both used by Akka internally to serialize messages, and available for ad-hoc
serialization of whatever you might need it for.
6.4. Serialization
278
6.4.1 Usage
Configuration
For Akka to know which Serializer to use for what, you need edit your Configuration, in the akka.actor.serializers-section you bind names to implementations of the
akka.serialization.Serializer you wish to use, like this:
akka {
actor {
serializers {
java = "akka.serialization.JavaSerializer"
proto = "akka.remote.serialization.ProtobufSerializer"
myown = "docs.serialization.MyOwnSerializer"
}
}
}
After youve bound names to different implementations of Serializer you need to wire which classes should
be serialized using which Serializer, this is done in the akka.actor.serialization-bindings-section:
akka {
actor {
serializers {
java = "akka.serialization.JavaSerializer"
proto = "akka.remote.serialization.ProtobufSerializer"
myown = "docs.serialization.MyOwnSerializer"
}
serialization-bindings {
"java.lang.String" = java
"docs.serialization.Customer" = java
"com.google.protobuf.Message" = proto
"docs.serialization.MyOwnSerializable" = myown
"java.lang.Boolean" = myown
}
}
}
You only need to specify the name of an interface or abstract base class of the messages. In case of ambiguity, i.e. the message implements several of the configured classes, the most specific configured class will be
used, i.e. the one of which all other candidates are superclasses. If this condition cannot be met, because e.g.
java.io.Serializable and MyOwnSerializable both apply and neither is a subtype of the other, a
warning will be issued.
Akka
provides
serializers
for
java.io.Serializable
and
protobuf
com.google.protobuf.GeneratedMessage by default (the latter only if depending on
the akka-remote module), so normally you dont need to add configuration for that; since
com.google.protobuf.GeneratedMessage implements java.io.Serializable, protobuf
messages will always by serialized using the protobuf protocol unless specifically overridden. In order to disable
a default serializer, map its marker type to none:
akka.actor.serialization-bindings {
"java.io.Serializable" = none
}
Verification
If you want to verify that your messages are serializable you can enable the following config option:
6.4. Serialization
279
akka {
actor {
serialize-messages = on
}
}
Warning: We only recommend using the config option turned on when youre running tests. It is completely
pointless to have it turned on in other scenarios.
If you want to verify that your Props are serializable you can enable the following config option:
akka {
actor {
serialize-creators = on
}
}
Warning: We only recommend using the config option turned on when youre running tests. It is completely
pointless to have it turned on in other scenarios.
Programmatic
If you want to programmatically serialize/deserialize using Akka Serialization, heres some examples:
import akka.actor.*;
import akka.serialization.*;
ActorSystem system = ActorSystem.create("example");
// Get the Serialization Extension
Serialization serialization = SerializationExtension.get(system);
// Have something to serialize
String original = "woohoo";
// Find the Serializer for it
Serializer serializer = serialization.findSerializerFor(original);
// Turn it into bytes
byte[] bytes = serializer.toBinary(original);
// Turn it back into an object,
// the nulls are for the class manifest and for the classloader
String back = (String) serializer.fromBinary(bytes);
// Voil!
assertEquals(original, back);
6.4.2 Customization
So,
lets say that you want to create your own Serializer,
docs.serialization.MyOwnSerializer in the config example above?
6.4. Serialization
you
saw
the
280
Then you only need to fill in the blanks, bind it to a name in your Configuration and then list which classes that
should be serialized using it.
Serializing ActorRefs
All ActorRefs are serializable using JavaSerializer, but in case you are writing your own serializer, you might
want to know how to serialize and deserialize them properly. In the general case, the local address to be
used depends on the type of remote address which shall be the recipient of the serialized information. Use
Serialization.serializedActorPath(actorRef) like this:
import akka.actor.*;
import akka.serialization.*;
// Serialize
// (beneath toBinary)
String identifier = Serialization.serializedActorPath(theActorRef);
// Then just serialize the identifier however you like
// Deserialize
// (beneath fromBinary)
final ActorRef deserializedActorRef = extendedSystem.provider().resolveActorRef(
identifier);
// Then just use the ActorRef
6.4. Serialization
281
This assumes that serialization happens in the context of sending a message through the remote transport. There
are other uses of serialization, though, e.g. storing actor references outside of an actor application (database, etc.).
In this case, it is important to keep in mind that the address part of an actors path determines how that actor is
communicated with. Storing a local actor path might be the right choice if the retrieval happens in the same logical
context, but it is not enough when deserializing it on a different network host: for that it would need to include
the systems remote transport address. An actor system is not limited to having just one remote transport per
se, which makes this question a bit more interesting. To find out the appropriate address to use when sending to
remoteAddr you can use ActorRefProvider.getExternalAddressFor(remoteAddr) like this:
public class ExternalAddressExt implements Extension {
private final ExtendedActorSystem system;
public ExternalAddressExt(ExtendedActorSystem system) {
this.system = system;
}
public Address getAddressFor(Address remoteAddress) {
final scala.Option<Address> optAddr = system.provider()
.getExternalAddressFor(remoteAddress);
if (optAddr.isDefined()) {
return optAddr.get();
} else {
throw new UnsupportedOperationException(
"cannot send to remote address " + remoteAddress);
}
}
}
public class ExternalAddress extends
AbstractExtensionId<ExternalAddressExt> implements ExtensionIdProvider {
public static final ExternalAddress ID = new ExternalAddress();
public ExternalAddress lookup() {
return ID;
}
public ExternalAddressExt createExtension(ExtendedActorSystem system) {
return new ExternalAddressExt(system);
}
}
public class ExternalAddressExample {
public String serializeTo(ActorRef ref, Address remote) {
return ref.path().toSerializationFormatWithAddress(
ExternalAddress.ID.get(system).getAddressFor(remote));
}
}
282
6.5 I/O
6.5.1 Introduction
The akka.io package has been developed in collaboration between the Akka and spray.io teams. Its design
combines experiences from the spray-io module with improvements that were jointly developed for more
general consumption as an actor-based service.
6.5. I/O
283
The guiding design goal for this I/O implementation was to reach extreme scalability, make no compromises
in providing an API correctly matching the underlying transport mechanism and to be fully event-driven, nonblocking and asynchronous. The API is meant to be a solid foundation for the implementation of network protocols
and building higher abstractions; it is not meant to be a full-service high-level NIO wrapper for end users.
The manager receives I/O command messages and instantiates worker actors in response. The worker actors
present themselves to the API user in the reply to the command that was sent. For example after a Connect
command sent to the TCP manager the manager creates an actor representing the TCP connection. All operations
related to the given TCP connections can be invoked by sending messages to the connection actor which announces
itself by sending a Connected message.
DeathWatch and Resource Management
I/O worker actors receive commands and also send out events. They usually need a user-side counterpart actor
listening for these events (such events could be inbound connections, incoming bytes or acknowledgements for
writes). These worker actors watch their listener counterparts. If the listener stops then the worker will automatically release any resources that it holds. This design makes the API more robust against resource leaks.
Thanks to the completely actor based approach of the I/O API the opposite direction works as well: a user actor
responsible for handling a connection can watch the connection actor to be notified if it unexpectedly terminates.
Write models (Ack, Nack)
I/O devices have a maximum throughput which limits the frequency and size of writes. When an application tries
to push more data than a device can handle, the driver has to buffer bytes until the device is able to write them.
With buffering it is possible to handle short bursts of intensive writes but no buffer is infinite. Flow control
is needed to avoid overwhelming device buffers.
Akka supports two types of flow control:
Ack-based, where the driver notifies the writer when writes have succeeded.
Nack-based, where the driver notifies the writer when writes have failed.
Each of these models is available in both the TCP and the UDP implementations of Akka I/O.
Individual writes can be acknowledged by providing an ack object in the write message (Write in the case of
TCP and Send for UDP). When the write is complete the worker will send the ack object to the writing actor. This
can be used to implement ack-based flow control; sending new data only when old data has been acknowledged.
If a write (or any other command) fails, the driver notifies the actor that sent the command with a special message
(CommandFailed in the case of UDP and TCP). This message will also notify the writer of a failed write,
serving as a nack for that write. Please note, that in a nack-based flow-control setting the writer has to be prepared
for the fact that the failed write might not be the most recent write it sent. For example, the failure notification for
a write W1 might arrive after additional write commands W2 and W3 have been sent. If the writer wants to resend
any nacked messages it may need to keep a buffer of pending messages.
6.5. I/O
284
Warning: An acknowledged write does not mean acknowledged delivery or storage; receiving an ack for a
write simply signals that the I/O driver has successfully processed the write. The Ack/Nack protocol described
here is a means of flow control not error handling. In other words, data may still be lost, even if every write is
acknowledged.
ByteString
To maintain isolation, actors should communicate with immutable objects only. ByteString is an immutable
container for bytes. It is used by Akkas I/O system as an efficient, immutable alternative the traditional byte
containers used for I/O on the JVM, such as byte[] and ByteBuffer.
ByteString is a rope-like data structure that is immutable and provides fast concatenation and slicing operations (perfect for I/O). When two ByteStrings are concatenated together they are both stored within the
resulting ByteString instead of copying both to a new array. Operations such as drop and take return
ByteStrings that still reference the original array, but just change the offset and length that is visible. Great
care has also been taken to make sure that the internal array cannot be modified. Whenever a potentially unsafe
array is used to create a new ByteString a defensive copy is created. If you require a ByteString that only
blocks a much memory as necessary for its content, use the compact method to get a CompactByteString
instance. If the ByteString represented only a slice of the original array, this will result in copying all bytes in
that slice.
ByteString inherits all methods from IndexedSeq, and it also has some new ones. For more information,
look up the akka.util.ByteString class and its companion object in the ScalaDoc.
ByteString also comes with its own optimized builder and iterator classes ByteStringBuilder and
ByteIterator which provide extra features in addition to those of normal builders and iterators.
Compatibility with java.io
java.net.InetSocketAddress;
akka.actor.ActorRef;
akka.actor.ActorSystem;
akka.actor.Props;
akka.actor.UntypedActor;
akka.io.Tcp;
akka.io.Tcp.Bound;
akka.io.Tcp.CommandFailed;
akka.io.Tcp.Connected;
akka.io.Tcp.ConnectionClosed;
akka.io.Tcp.Received;
akka.io.TcpMessage;
akka.japi.Procedure;
akka.util.ByteString;
285
All of the Akka I/O APIs are accessed through manager objects. When using an I/O API, the first step is to acquire
a reference to the appropriate manager. The code below shows how to acquire a reference to the Tcp manager.
final ActorRef tcpManager = Tcp.get(getContext().system()).manager();
The manager is an actor that handles the underlying low level I/O resources (selectors, channels) and instantiates
workers for specific tasks, such as listening to incoming connections.
6.6.1 Connecting
public class Client extends UntypedActor {
final InetSocketAddress remote;
final ActorRef listener;
public Client(InetSocketAddress remote, ActorRef listener) {
this.remote = remote;
this.listener = listener;
final ActorRef tcp = Tcp.get(getContext().system()).manager();
tcp.tell(TcpMessage.connect(remote), getSelf());
}
@Override
public void onReceive(Object msg) throws Exception {
if (msg instanceof CommandFailed) {
listener.tell("failed", getSelf());
getContext().stop(getSelf());
} else if (msg instanceof Connected) {
listener.tell(msg, getSelf());
getSender().tell(TcpMessage.register(getSelf()), getSelf());
getContext().become(connected(getSender()));
}
}
private Procedure<Object> connected(final ActorRef connection) {
return new Procedure<Object>() {
@Override
public void apply(Object msg) throws Exception {
if (msg instanceof ByteString) {
connection.tell(TcpMessage.write((ByteString) msg), getSelf());
} else if (msg instanceof CommandFailed) {
// OS kernel socket buffer was full
} else if (msg instanceof Received) {
listener.tell(((Received) msg).data(), getSelf());
} else if (msg.equals("close")) {
connection.tell(TcpMessage.close(), getSelf());
} else if (msg instanceof ConnectionClosed) {
getContext().stop(getSelf());
}
}
};
}
}
286
The first step of connecting to a remote address is sending a Connect message to the TCP manager; in addition
to the simplest form shown above there is also the possibility to specify a local InetSocketAddress to bind
to and a list of socket options to apply.
Note: The SO_NODELAY (TCP_NODELAY on Windows) socket option defaults to true in Akka, independently of the OS default settings. This setting disables Nagles algorithm, considerably improving latency for
most applications. This setting could be overridden by passing SO.TcpNoDelay(false) in the list of socket
options of the Connect message.
The TCP manager will then reply either with a CommandFailed or it will spawn an internal actor representing
the new connection. This new actor will then send a Connected message to the original sender of the Connect
message.
In order to activate the new connection a Register message must be sent to the connection actor, informing
that one about who shall receive data from the socket. Before this step is done the connection cannot be used, and
there is an internal timeout after which the connection actor will shut itself down if no Register message is
received.
The connection actor watches the registered handler and closes the connection when that one terminates, thereby
cleaning up all internal resources associated with that connection.
The actor in the example above uses become to switch from unconnected to connected operation, demonstrating
the commands and events which are observed in that state. For a discussion on CommandFailed see Throttling
Reads and Writes below. ConnectionClosed is a trait, which marks the different connection close events.
The last line handles all connection close events in the same way. It is possible to listen for more fine-grained
connection close events, see Closing Connections below.
287
To create a TCP server and listen for inbound connections, a Bind command has to be sent to the TCP manager.
This will instruct the TCP manager to listen for TCP connections on a particular InetSocketAddress; the
port may be specified as 0 in order to bind to a random port.
The actor sending the Bind message will receive a Bound message signalling that the server is ready to accept
incoming connections; this message also contains the InetSocketAddress to which the socket was actually
bound (i.e. resolved IP address and correct port number).
From this point forward the process of handling connections is the same as for outgoing connections. The example
demonstrates that handling the reads from a certain connection can be delegated to another actor by naming it as
the handler when sending the Register message. Writes can be sent from any actor in the system to the
connection actor (i.e. the actor which sent the Connected message). The simplistic handler is defined as:
public class SimplisticHandler extends UntypedActor {
@Override
public void onReceive(Object msg) throws Exception {
if (msg instanceof Received) {
final ByteString data = ((Received) msg).data();
System.out.println(data);
getSender().tell(TcpMessage.write(data), getSelf());
} else if (msg instanceof ConnectionClosed) {
getContext().stop(getSelf());
}
}
}
For a more complete sample which also takes into account the possibility of failures when sending please see
Throttling Reads and Writes below.
The only difference to outgoing connections is that the internal actor managing the listen portthe sender of the
Bound messagewatches the actor which was named as the recipient for Connected messages in the Bind
message. When that actor terminates the listen port will be closed and all resources associated with it will be
released; existing connections will not be terminated at this point.
288
289
290
}
private final Procedure<Object> buffering = new Procedure<Object>() {
@Override
public void apply(Object msg) throws Exception {
if (msg instanceof Received) {
buffer(((Received) msg).data());
} else if (msg == ACK) {
acknowledge();
} else if (msg instanceof ConnectionClosed) {
if (((ConnectionClosed) msg).isPeerClosed()) {
closing = true;
} else {
// could also be ErrorClosed, in which case we just give up
getContext().stop(getSelf());
}
}
}
};
// storage omitted ...
}
The principle is simple: when having written a chunk always wait for the Ack to come back before sending the
next chunk. While waiting we switch behavior such that new incoming data are buffered. The helper functions
used are a bit lengthy but not complicated:
protected void buffer(ByteString data) {
storage.add(data);
stored += data.size();
if (stored > maxStored) {
log.warning("drop connection to [{}] (buffer overrun)", remote);
getContext().stop(getSelf());
} else if (stored > highWatermark) {
log.debug("suspending reading");
connection.tell(TcpMessage.suspendReading(), getSelf());
suspended = true;
}
}
protected void acknowledge() {
final ByteString acked = storage.remove();
stored -= acked.size();
transferred += acked.size();
if (suspended && stored < lowWatermark) {
log.debug("resuming reading");
connection.tell(TcpMessage.resumeReading(), getSelf());
suspended = false;
}
if (storage.isEmpty()) {
if (closing) {
getContext().stop(getSelf());
} else {
getContext().unbecome();
}
} else {
connection.tell(TcpMessage.write(storage.peek(), ACK), getSelf());
291
}
}
The most interesting part is probably the last: an Ack removes the oldest data chunk from the buffer, and if that
was the last chunk then we either close the connection (if the peer closed its half already) or return to the idle
behavior; otherwise we just send the next buffered chunk and stay waiting for the next Ack.
Back-pressure can be propagated also across the reading side back to the writer on the other end of the connection
by sending the SuspendReading command to the connection actor. This will lead to no data being read from
the socket anymore (although this does happen after a delay because it takes some time until the connection actor
processes this command, hence appropriate head-room in the buffer should be present), which in turn will lead
to the O/S kernel buffer filling up on our end, then the TCP window mechanism will stop the remote side from
writing, filling up its write buffer, until finally the writer on the other side cannot push any data into the socket
anymore. This is how end-to-end back-pressure is realized across a TCP connection.
292
getContext().become(buffering((Ack) w.ack()));
} else if (msg instanceof ConnectionClosed) {
final ConnectionClosed cl = (ConnectionClosed) msg;
if (cl.isPeerClosed()) {
if (storage.isEmpty()) {
getContext().stop(getSelf());
} else {
getContext().become(closing);
}
}
}
}
};
// buffering ...
// closing ...
// storage omitted ...
}
The principle here is to keep writing until a CommandFailed is received, using acknowledgements only to
prune the resend buffer. When a such a failure was received, transition into a different state for handling and
handle resending of all queued data:
protected Procedure<Object> buffering(final Ack nack) {
return new Procedure<Object>() {
private int toAck = 10;
private boolean peerClosed = false;
@Override
public void apply(Object msg) throws Exception {
if (msg instanceof Received) {
buffer(((Received) msg).data());
} else if (msg instanceof WritingResumed) {
writeFirst();
} else if (msg instanceof ConnectionClosed) {
if (((ConnectionClosed) msg).isPeerClosed())
peerClosed = true;
else
getContext().stop(getSelf());
} else if (msg instanceof Integer) {
final int ack = (Integer) msg;
acknowledge(ack);
if (ack >= nack.ack) {
// otherwise it was the ack of the last successful write
if (storage.isEmpty()) {
if (peerClosed)
getContext().stop(getSelf());
else
getContext().become(writing);
} else {
if (toAck > 0) {
// stay in ACK-based mode for a short while
writeFirst();
293
--toAck;
} else {
// then return to NACK-based again
writeAll();
if (peerClosed)
getContext().become(closing);
else
getContext().become(writing);
}
}
}
}
}
};
}
It should be noted that all writes which are currently buffered have also been sent to the connection actor upon
entering this state, which means that the ResumeWriting message is enqueued after those writes, leading to
the reception of all outstanding CommandFailed messages (which are ignored in this state) before receiving
the WritingResumed signal. That latter message is sent by the connection actor only once the internally
queued write has been fully completed, meaning that a subsequent write will not fail. This is exploited by the
EchoHandler to switch to an ACK-based approach for the first ten writes after a failure before resuming the
optimistic write-through behavior.
protected Procedure<Object> closing = new Procedure<Object>() {
@Override
public void apply(Object msg) throws Exception {
if (msg instanceof CommandFailed) {
// the command can only have been a Write
connection.tell(TcpMessage.resumeWriting(), getSelf());
getContext().become(closeResend, false);
} else if (msg instanceof Integer) {
acknowledge((Integer) msg);
if (storage.isEmpty())
getContext().stop(getSelf());
}
}
};
protected Procedure<Object> closeResend = new Procedure<Object>() {
@Override
public void apply(Object msg) throws Exception {
if (msg instanceof WritingResumed) {
writeAll();
getContext().unbecome();
} else if (msg instanceof Integer) {
acknowledge((Integer) msg);
}
}
};
Closing the connection while still sending all data is a bit more involved than in the ACK-based approach: the
idea is to always send all outstanding messages and acknowledge all successful writes, and if a failure happens
then switch behavior to await the WritingResumed event and start over.
The helper functions are very similar to the ACK-based case:
protected void buffer(ByteString data) {
storage.add(data);
stored += data.size();
if (stored > MAX_STORED) {
log.warning("drop connection to [{}] (buffer overrun)", remote);
294
getContext().stop(getSelf());
} else if (stored > HIGH_WATERMARK) {
log.debug("suspending reading at {}", currentOffset());
connection.tell(TcpMessage.suspendReading(), getSelf());
suspended = true;
}
}
protected void acknowledge(int ack) {
assert ack == storageOffset;
assert !storage.isEmpty();
final ByteString acked = storage.remove();
stored -= acked.size();
transferred += acked.size();
storageOffset += 1;
if (suspended && stored < LOW_WATERMARK) {
log.debug("resuming reading");
connection.tell(TcpMessage.resumeReading(), getSelf());
suspended = false;
}
}
The idea here is that reading is not resumed until the previous write has been completely acknowledged by the
connection actor. Every pull mode connection actor starts from suspended state. To start the flow of data we send
a ResumeReading in the preStart method to tell the connection actor that we are ready to receive the first
chunk of data. Since we only resume reading when the previous data chunk has been completely written there is
no need for maintaining a buffer.
To enable pull reading on an outbound connection the pullMode parameter of the Connect should be set to
true:
final List<Inet.SocketOption> options = new ArrayList<Inet.SocketOption>();
tcp.tell(
TcpMessage.connect(new InetSocketAddress("localhost", 3000), null, options, null, true),
getSelf()
);
295
One of the effects of this setting is that all connections accepted by this listener actor will use pull mode reading.
Another effect of this setting is that in addition of setting all inbound connections to pull mode, accepting connections becomes pull based, too. This means that after handling one (or more) Connected events the listener actor
has to be resumed by sending it a ResumeAccepting message.
Listener actors with pull mode start suspended so to start accepting connections a ResumeAccepting command
has to be sent to the listener actor after binding was successful:
public void onReceive(Object message) throws Exception {
if (message instanceof Tcp.Bound) {
listener = getSender();
// Accept connections one by one
listener.tell(TcpMessage.resumeAccepting(1), getSelf());
} else if (message instanceof Tcp.Connected) {
ActorRef handler = getContext().actorOf(Props.create(PullEcho.class, getSender()));
getSender().tell(TcpMessage.register(handler), getSelf());
// Resume accepting connections
listener.tell(TcpMessage.resumeAccepting(1), getSelf());
}
}
As shown in the example after handling an incoming connection we need to resume accepting again. The
ResumeAccepting message accepts a batchSize parameter that specifies how many new connections are
accepted before a next ResumeAccepting message is needed to resume handling of new connections.
296
The simplest form of UDP usage is to just send datagrams without the need of getting a reply. To this end a simple
sender facility is provided as demonstrated above. The UDP extension is queried using the simpleSender
message, which is answered by a SimpleSenderReady notification. The sender of this message is the newly
created sender actor which from this point onward can be used to send datagrams to arbitrary destinations; in this
example it will just send any UTF-8 encoded String it receives to a predefined remote address.
Note: The simple sender will not shut itself down because it cannot know when you are done with it. You will
need to send it a PoisonPill when you want to close the ephemeral port the sender is bound to.
297
@Override
public void apply(Object msg) throws Exception {
if (msg instanceof Udp.Received) {
final Udp.Received r = (Udp.Received) msg;
// echo server example: send back the data
socket.tell(UdpMessage.send(r.data(), r.sender()), getSelf());
// or do some processing and forward it on
final Object processed = // parse data etc., e.g. using PipelineStage
nextActor.tell(processed, getSelf());
} else if (msg.equals(UdpMessage.unbind())) {
socket.tell(msg, getSelf());
} else if (msg instanceof Udp.Unbound) {
getContext().stop(getSelf());
} else unhandled(msg);
}
};
}
}
If you want to implement a UDP server which listens on a socket for incoming datagrams then you need to use the
bind command as shown above. The local address specified may have a zero port in which case the operating
system will automatically choose a free port and assign it to the new socket. Which port was actually bound can
be found out by inspecting the Bound message.
The sender of the Bound message is the actor which manages the new socket. Sending datagrams is achieved by
using the send message type and the socket can be closed by sending a unbind command, in which case the
socket actor will reply with a Unbound notification.
Received datagrams are sent to the actor designated in the bind message, whereas the Bound message will be
sent to the sender of the bind.
298
Consequently the example shown here looks quite similar to the previous one, the biggest difference is the absence
of remote address information in send and Received messages.
Note: There is a small performance benefit in using connection based UDP API over the connectionless one. If
there is a SecurityManager enabled on the system, every connectionless message send has to go through a security
check, while in the case of connection-based UDP the security check is cached after connect, thus writes do not
suffer an additional performance penalty.
6.8 ZeroMQ
Akka provides a ZeroMQ module which abstracts a ZeroMQ connection and therefore allows interaction between
Akka actors to take place over ZeroMQ connections. The messages can be of a proprietary format or they can be
defined using Protobuf. The socket actor is fault-tolerant by default and when you use the newSocket method to
create new sockets it will properly reinitialize the socket.
ZeroMQ is very opinionated when it comes to multi-threading so configuration option akka.zeromq.socketdispatcher always needs to be configured to a PinnedDispatcher, because the actual ZeroMQ socket can only
be accessed by the thread that created it.
The ZeroMQ module for Akka is written against an API introduced in JZMQ, which uses JNI to interact with
the native ZeroMQ library. Instead of using JZMQ, the module uses ZeroMQ binding for Scala that uses the
native ZeroMQ library through JNA. In other words, the only native library that this module requires is the native
ZeroMQ library. The benefit of the scala library is that you dont need to compile and manage native dependencies
at the cost of some runtime performance. The scala-bindings are compatible with the JNI bindings so they are a
drop-in replacement, in case you really need to get that extra bit of performance out.
Note: The currently used version of zeromq-scala-bindings is only compatible with zeromq 2; zeromq
3 is not supported.
6.8. ZeroMQ
299
6.8.1 Connection
ZeroMQ supports multiple connectivity patterns, each aimed to meet a different set of requirements. Currently, this module supports publisher-subscriber connections and connections based on dealers and routers.
For connecting or accepting connections, a socket must be created. Sockets are always created using the
akka.zeromq.ZeroMQExtension, for example:
import akka.zeromq.Bind;
import akka.zeromq.ZeroMQExtension;
ActorRef pubSocket = ZeroMQExtension.get(system).newPubSocket(
new Bind("tcp://127.0.0.1:1233"));
Above examples will create a ZeroMQ Publisher socket that is Bound to the port 21231 on localhost.
Similarly you can create a subscription socket, with a listener, that subscribes to all messages from the publisher
using:
import akka.zeromq.Connect;
import akka.zeromq.Listener;
import akka.zeromq.Subscribe;
ActorRef listener = system.actorOf(Props.create(ListenerActor.class));
ActorRef subSocket = ZeroMQExtension.get(system).newSubSocket(
new Connect("tcp://127.0.0.1:1233"),
new Listener(listener), Subscribe.all());
public class ListenerActor extends UntypedActor {
public void onReceive(Object message) throws Exception {
//...
}
}
The following sub-sections describe the supported connection patterns and how they can be used in an Akka
environment. However, for a comprehensive discussion of connection patterns, please refer to ZeroMQ The
Guide.
Publisher-Subscriber Connection
In a publisher-subscriber (pub-sub) connection, the publisher accepts one or more subscribers. Each subscriber
shall subscribe to one or more topics, whereas the publisher publishes messages to a set of topics. Also, a subscriber can subscribe to all available topics. In an Akka environment, pub-sub connections shall be used when an
actor sends messages to one or more actors that do not interact with the actor that sent the message.
When youre using zeromq pub/sub you should be aware that it needs multicast - check your cloud - to work
properly and that the filtering of events for topics happens client side, so all events are always broadcasted to every
subscriber.
An actor is subscribed to a topic as follows:
ActorRef subTopicSocket = ZeroMQExtension.get(system).newSubSocket(
new Connect("tcp://127.0.0.1:1233"),
new Listener(listener), new Subscribe("foo.bar"));
It is a prefix match so it is subscribed to all topics starting with foo.bar. Note that if the given string is empty
or Subscribe.all() is used, the actor is subscribed to all topics.
To unsubscribe from a topic you do the following:
import akka.zeromq.Unsubscribe;
subTopicSocket.tell(new Unsubscribe("foo.bar"), ActorRef.noSender());
6.8. ZeroMQ
300
To publish messages to a topic you must use two Frames with the topic in the first frame.
import akka.util.ByteString;
import akka.zeromq.ZMQMessage;
pubSocket.tell(ZMQMessage.withFrames(ByteString.fromString("foo.bar"),
ByteString.fromArray(payload)), ActorRef.noSender());
Pub-Sub in Action
akka.actor.ActorRef;
akka.actor.UntypedActor;
akka.actor.Props;
akka.event.Logging;
akka.event.LoggingAdapter;
org.junit.*;
scala.concurrent.duration.Duration;
akka.serialization.SerializationExtension;
akka.serialization.Serialization;
java.io.Serializable;
java.lang.management.ManagementFactory;
6.8. ZeroMQ
301
Lets add one subscriber that logs the information. It subscribes to all topics starting with "health", i.e. both
Heap and Load events.
public class Logger extends UntypedActor {
ActorRef subSocket = ZeroMQExtension.get(getContext().system()).newSubSocket(
new Connect("tcp://127.0.0.1:1237"),
new Listener(getSelf()), new Subscribe("health"));
Serialization ser = SerializationExtension.get(getContext().system());
SimpleDateFormat timestampFormat = new SimpleDateFormat("HH:mm:ss.SSS");
LoggingAdapter log = Logging.getLogger(getContext().system(), this);
@Override
public void onReceive(Object message) {
if (message instanceof ZMQMessage) {
ZMQMessage m = (ZMQMessage) message;
String topic = m.frame(0).utf8String();
// the first frame is the topic, second is the message
if ("health.heap".equals(topic)) {
6.8. ZeroMQ
302
Another subscriber keep track of used heap and warns if too much heap is used. It only subscribes to Heap events.
public class HeapAlerter extends UntypedActor {
ActorRef subSocket = ZeroMQExtension.get(getContext().system()).newSubSocket(
new Connect("tcp://127.0.0.1:1237"),
new Listener(getSelf()), new Subscribe("health.heap"));
Serialization ser = SerializationExtension.get(getContext().system());
LoggingAdapter log = Logging.getLogger(getContext().system(), this);
int count = 0;
@Override
public void onReceive(Object message) {
if (message instanceof ZMQMessage) {
ZMQMessage m = (ZMQMessage) message;
String topic = m.frame(0).utf8String();
// the first frame is the topic, second is the message
if ("health.heap".equals(topic)) {
Heap heap = ser.<Heap>deserialize(m.frame(1).toArray(), Heap.class).get();
if (((double) heap.used / heap.max) > 0.9) {
count += 1;
} else {
count = 0;
}
if (count > 10) {
log.warning("Need more memory, using {} %",
(100.0 * heap.used / heap.max));
}
}
} else {
unhandled(message);
}
}
}
system.actorOf(Props.create(HeapAlerter.class), "alerter");
Router-Dealer Connection
While Pub/Sub is nice the real advantage of zeromq is that it is a lego-box for reliable messaging. And because
there are so many integrations the multi-language support is fantastic. When youre using ZeroMQ to integrate
many systems youll probably need to build your own ZeroMQ devices. This is where the router and dealer socket
types come in handy. With those socket types you can build your own reliable pub sub broker that uses TCP/IP
and does publisher side filtering of events.
6.8. ZeroMQ
303
To create a Router socket that has a high watermark configured, you would do:
ActorRef highWatermarkSocket = ZeroMQExtension.get(system).newRouterSocket(
new SocketOption[] { new Listener(listener),
new Bind("tcp://127.0.0.1:1233"), new HighWatermark(50000) });
The akka-zeromq module accepts most if not all the available configuration options for a zeromq socket.
Push-Pull Connection
Akka ZeroMQ module supports Push-Pull connections.
You can create a Push connection through the:
ActorRef newPushSocket(SocketOption[] socketParameters);
6.8.2 Configuration
There are several configuration properties for the zeromq module, please refer to the reference configuration.
6.9 Camel
6.9.1 Introduction
The akka-camel module allows Untyped Actors to receive and send messages over a great variety of protocols and
APIs. In addition to the native Scala and Java actor API, actors can now exchange messages with other systems
over large number of protocols and APIs such as HTTP, SOAP, TCP, FTP, SMTP or JMS, to mention a few. At
the moment, approximately 80 protocols and APIs are supported.
Apache Camel
The akka-camel module is based on Apache Camel, a powerful and light-weight integration framework for the
JVM. For an introduction to Apache Camel you may want to read this Apache Camel article. Camel comes with
a large number of components that provide bindings to different protocols and APIs. The camel-extra project
provides further components.
6.9. Camel
304
Consumer
Heres an example of using Camels integration components in Akka.
import akka.camel.CamelMessage;
import akka.camel.javaapi.UntypedConsumerActor;
public class MyEndpoint extends UntypedConsumerActor{
private String uri;
public String getEndpointUri() {
return uri;
}
public void onReceive(Object message) throws Exception {
if (message instanceof CamelMessage) {
/* ... */
} else
unhandled(message);
}
// Extra constructor to change the default uri,
// for instance to "jetty:http://localhost:8877/example"
public MyEndpoint(String uri) {
this.uri = uri;
}
public MyEndpoint() {
this.uri = "mina2:tcp://localhost:6200?textline=true";
}
}
The above example exposes an actor over a TCP endpoint via Apache Camels Mina component. The actor
implements the getEndpointUri method to define an endpoint from which it can receive messages. After starting
the actor, TCP clients can immediately send messages to and receive responses from that actor. If the message
exchange should go over HTTP (via Camels Jetty component), the actors getEndpointUri method should return
a different URI, for instance jetty:http://localhost:8877/example. In the above case an extra constructor is added
that can set the endpoint URI, which would result in the getEndpointUri returning the URI that was set using this
constructor.
Producer
Actors can also trigger message exchanges with external systems i.e. produce to Camel endpoints.
import akka.camel.javaapi.UntypedProducerActor;
public class Orders extends UntypedProducerActor {
public String getEndpointUri() {
return "jms:queue:Orders";
}
}
In the above example, any message sent to this actor will be sent to the JMS queue Orders. Producer actors
may choose from the same set of Camel components as Consumer actors do. Below an example of how to send a
message to the Orders producer.
ActorSystem system = ActorSystem.create("some-system");
Props props = Props.create(Orders.class);
ActorRef producer = system.actorOf(props, "jmsproducer");
producer.tell("<order amount=\"100\" currency=\"PLN\" itemId=\"12345\"/>",
ActorRef.noSender());
6.9. Camel
305
CamelMessage
The number of Camel components is constantly increasing. The akka-camel module can support these in a plugand-play manner. Just add them to your applications classpath, define a component-specific endpoint URI and
use it to exchange messages over the component-specific protocols or APIs. This is possible because Camel components bind protocol-specific message formats to a Camel-specific normalized message format. The normalized
message format hides protocol-specific details from Akka and makes it therefore very easy to support a large
number of protocols through a uniform Camel component interface. The akka-camel module further converts
mutable Camel messages into immutable representations which are used by Consumer and Producer actors for
pattern matching, transformation, serialization or storage. In the above example of the Orders Producer, the XML
message is put in the body of a newly created Camel Message with an empty set of headers. You can also create a
CamelMessage yourself with the appropriate body and headers as you see fit.
CamelExtension
The akka-camel module is implemented as an Akka Extension, the CamelExtension object. Extensions will
only be loaded once per ActorSystem, which will be managed by Akka. The CamelExtension object
provides access to the Camel interface. The Camel interface in turn provides access to two important Apache
Camel objects, the CamelContext and the ProducerTemplate. Below you can see how you can get access to these
Apache Camel objects.
ActorSystem system = ActorSystem.create("some-system");
Camel camel = CamelExtension.get(system);
CamelContext camelContext = camel.context();
ProducerTemplate producerTemplate = camel.template();
One CamelExtension is only loaded once for every one ActorSystem, which makes it safe to call the
CamelExtension at any point in your code to get to the Apache Camel objects associated with it. There is
one CamelContext and one ProducerTemplate for every one ActorSystem that uses a CamelExtension.
By Default, a new CamelContext is created when the CamelExtension starts. If you want to inject your own
context instead, you can implement the ContextProvider interface and add the FQCN of your implementation in the
config, as the value of the akka.camel.context-provider. This interface define a single method getContext()
used to load the CamelContext.
Below an example on how to add the ActiveMQ component to the CamelContext, which is required when you
would like to use the ActiveMQ component.
ActorSystem system = ActorSystem.create("some-system");
Camel camel = CamelExtension.get(system);
CamelContext camelContext = camel.context();
// camelContext.addComponent("activemq", ActiveMQComponent.activeMQComponent(
//
"vm://localhost?broker.persistent=false"));
The CamelContext joins the lifecycle of the ActorSystem and CamelExtension it is associated with;
the CamelContext is started when the CamelExtension is created, and it is shut down when the associated
ActorSystem is shut down. The same is true for the ProducerTemplate.
The CamelExtension is used by both Producer and Consumer actors to interact with Apache Camel internally. You can access the CamelExtension inside a Producer or a Consumer using the camel method, or
get straight at the CamelContext using the getCamelContext method or to the ProducerTemplate using the
getProducerTemplate method. Actors are created and started asynchronously. When a Consumer actor is created,
the Consumer is published at its Camel endpoint (more precisely, the route is added to the CamelContext from
the Endpoint to the actor). When a Producer actor is created, a SendProcessor and Endpoint are created so that
the Producer can send messages to it. Publication is done asynchronously; setting up an endpoint may still be in
progress after you have requested the actor to be created. Some Camel components can take a while to startup,
and in some cases you might want to know when the endpoints are activated and ready to be used. The Camel
interface allows you to find out when the endpoint is activated or deactivated.
import akka.actor.ActorRef;
import akka.actor.ActorSystem;
import akka.actor.Props;
6.9. Camel
306
import
import
import
import
import
import
import
import
import
akka.camel.Camel;
akka.camel.CamelExtension;
akka.camel.javaapi.UntypedConsumerActor;
akka.testkit.JavaTestKit;
akka.testkit.TestKit;
akka.util.Timeout;
scala.concurrent.Future;
scala.concurrent.duration.Duration;
static java.util.concurrent.TimeUnit.SECONDS;
// ..
ActorSystem system = ActorSystem.create("some-system");
Props props = Props.create(MyConsumer.class);
ActorRef producer = system.actorOf(props,"myproducer");
Camel camel = CamelExtension.get(system);
// get a future reference to the activation of the endpoint of the Consumer Actor
Timeout timeout = new Timeout(Duration.create(10, SECONDS));
Future<ActorRef> activationFuture = camel.activationFutureFor(producer,
timeout, system.dispatcher());
The above code shows that you can get a Future to the activation of the route from the endpoint to the actor,
or you can wait in a blocking fashion on the activation of the route. An ActivationTimeoutException is
thrown if the endpoint could not be activated within the specified timeout. Deactivation works in a similar fashion:
// ..
system.stop(producer);
// get a future reference to the deactivation of the endpoint of the Consumer Actor
Future<ActorRef> deactivationFuture = camel.deactivationFutureFor(producer,
timeout, system.dispatcher());
Deactivation of a Consumer or a Producer actor happens when the actor is terminated. For a Consumer, the route
to the actor is stopped. For a Producer, the SendProcessor is stopped. A DeActivationTimeoutException
is thrown if the associated camel objects could not be deactivated within the specified timeout.
akka.camel.CamelMessage;
akka.camel.javaapi.UntypedConsumerActor;
akka.event.Logging;
akka.event.LoggingAdapter;
6.9. Camel
307
Whenever a file is put into the data/input/actor directory, its content is picked up by the Camel file component and
sent as message to the actor. Messages consumed by actors from Camel endpoints are of type CamelMessage.
These are immutable representations of Camel messages.
Heres another example that sets the endpointUri to jetty:http://localhost:8877/camel/default.
It causes Camels Jetty component to start an embedded Jetty server, accepting HTTP connections from localhost
on port 8877.
import akka.camel.CamelMessage;
import akka.camel.javaapi.UntypedConsumerActor;
public class Consumer2 extends UntypedConsumerActor {
public String getEndpointUri() {
return "jetty:http://localhost:8877/camel/default";
}
public void onReceive(Object message) {
if (message instanceof CamelMessage) {
CamelMessage camelMessage = (CamelMessage) message;
String body = camelMessage.getBodyAs(String.class, getCamelContext());
getSender().tell(String.format("Received message: %s",body), getSelf());
} else
unhandled(message);
}
}
akka.actor.Status;
akka.camel.Ack;
akka.camel.CamelMessage;
akka.camel.javaapi.UntypedConsumerActor;
6.9. Camel
308
Consumer timeout
Camel Exchanges (and their corresponding endpoints) that support two-way communications need to wait for a
response from an actor before returning it to the initiating client. For some endpoint types, timeout values can be
defined in an endpoint-specific way which is described in the documentation of the individual Camel components.
Another option is to configure timeouts on the level of consumer actors.
Two-way communications between a Camel endpoint and an actor are initiated by sending the request message to
the actor with the ask pattern and the actor replies to the endpoint when the response is ready. The ask request to
the actor can timeout, which will result in the Exchange failing with a TimeoutException set on the failure of the
Exchange. The timeout on the consumer actor can be overridden with the replyTimeout, as shown below.
import
import
import
import
akka.camel.CamelMessage;
akka.camel.javaapi.UntypedConsumerActor;
scala.concurrent.duration.Duration;
scala.concurrent.duration.FiniteDuration;
import java.util.concurrent.TimeUnit;
public class Consumer4 extends UntypedConsumerActor {
private final static FiniteDuration timeout =
Duration.create(500, TimeUnit.MILLISECONDS);
@Override
public FiniteDuration replyTimeout() {
return timeout;
}
public String getEndpointUri() {
return "jetty:http://localhost:8877/camel/default";
}
public void onReceive(Object message) {
if (message instanceof CamelMessage) {
CamelMessage camelMessage = (CamelMessage) message;
String body = camelMessage.getBodyAs(String.class, getCamelContext());
getSender().tell(String.format("Hello %s",body), getSelf());
} else
unhandled(message);
}
}
6.9. Camel
309
import akka.camel.javaapi.UntypedProducerActor;
public class Producer1 extends UntypedProducerActor {
public String getEndpointUri() {
return "http://localhost:8080/news";
}
}
Producer1 inherits a default implementation of the onReceive method from the UntypedProducerActor
class. To customize a producer actors default behavior you must override the UntypedProducerActor.onTransformResponse and UntypedProducerActor.onTransformOutgoingMessage methods. This is explained
later in more detail. Producer Actors cannot override the UntypedProducerActor.onReceive method.
Any message sent to a Producer actor will be sent to the associated Camel endpoint, in the above example to
http://localhost:8080/news. The UntypedProducerActor always sends messages asynchronously. Response messages (if supported by the configured endpoint) will, by default, be returned to the original sender. The
following example uses the ask pattern to send a message to a Producer actor and waits for a response.
ActorSystem system = ActorSystem.create("some-system");
Props props = Props.create(FirstProducer.class);
ActorRef producer = system.actorOf(props,"myproducer");
Future<Object> future = Patterns.ask(producer, "some request", 1000);
The future contains the response CamelMessage, or an AkkaCamelException when an error occurred, which
contains the headers of the response.
Custom Processing
Instead of replying to the initial sender, producer actors can implement custom response processing by overriding
the onRouteResponse method. In the following example, the response message is forwarded to a target actor
instead of being replied to the original sender.
import akka.actor.UntypedActor;
import akka.camel.CamelMessage;
public class ResponseReceiver extends UntypedActor{
public void onReceive(Object message) {
if(message instanceof CamelMessage) {
// do something with the forwarded response
}
}
}
import akka.actor.ActorRef;
import akka.camel.javaapi.UntypedProducerActor;
public class Forwarder extends UntypedProducerActor {
private String uri;
private ActorRef target;
public Forwarder(String uri, ActorRef target) {
this.uri = uri;
this.target = target;
}
public String getEndpointUri() {
return uri;
}
@Override
public void onRouteResponse(Object message) {
target.forward(message, getContext());
6.9. Camel
310
}
}
ActorSystem system = ActorSystem.create("some-system");
Props receiverProps = Props.create(ResponseReceiver.class);
final ActorRef receiver = system.actorOf(receiverProps,"responseReceiver");
ActorRef forwardResponse = system.actorOf(Props.create(
Forwarder.class, "http://localhost:8080/news/akka", receiver));
// the Forwarder sends out a request to the web page and forwards the response to
// the ResponseReceiver
forwardResponse.tell("some request", ActorRef.noSender());
Before producing messages to endpoints, producer actors can pre-process them by overriding the UntypedProducerActor.onTransformOutgoingMessage method.
import
import
import
import
akka.camel.CamelMessage;
akka.camel.javaapi.UntypedProducerActor;
akka.dispatch.Mapper;
akka.japi.Function;
6.9. Camel
311
Message correlation
To correlate request with response messages, applications can set the Message.MessageExchangeId message
header.
ActorSystem system = ActorSystem.create("some-system");
Props props = Props.create(Orders.class);
ActorRef producer = system.actorOf(props,"jmsproducer");
Map<String,Object> headers = new HashMap<String, Object>();
headers.put(CamelMessage.MessageExchangeId(),"123");
producer.tell(new CamelMessage("<order amount=\"100\" currency=\"PLN\" " +
"itemId=\"12345\"/>",headers), ActorRef.noSender());
ProducerTemplate
The UntypedProducerActor class is a very convenient way for actors to produce messages to Camel endpoints.
Actors may also use a Camel ProducerTemplate for producing messages to endpoints.
import
import
import
import
akka.actor.UntypedActor;
akka.camel.Camel;
akka.camel.CamelExtension;
org.apache.camel.ProducerTemplate;
For initiating a a two-way message exchange, one of the ProducerTemplate.request* methods must be
used.
import
import
import
import
akka.actor.UntypedActor;
akka.camel.Camel;
akka.camel.CamelExtension;
org.apache.camel.ProducerTemplate;
6.9. Camel
312
6.9. Camel
313
where <actor-path> is the ActorPath to the actor. The <options> are name-value pairs separated by &
(i.e. name1=value1&name2=value2&...).
URI options
Type
Duration
Default
false
autoAck
Boolean
true
Description
The reply timeout, specified in the same way that you use the
duration in akka, for instance 10 seconds except that in the url it
is handy to use a + between the amount and the unit, like for
example 200+millis
See also Consumer timeout.
If set to true, in-only message exchanges are auto-acknowledged
when the message is added to the actors mailbox. If set to false,
actors must acknowledge the receipt of the message.
See also Delivery acknowledgements.
In the following example, a custom route to an actor is created, using the actors path.
import
import
import
import
akka.actor.UntypedActor;
akka.camel.CamelMessage;
akka.dispatch.Mapper;
akka.japi.Function;
6.9. Camel
314
}
}
ActorSystem system = ActorSystem.create("some-system");
Camel camel = CamelExtension.get(system);
ActorRef responder = system.actorOf(Props.create(Responder.class), "TestResponder");
camel.context().addRoutes(new CustomRouteBuilder(responder));
The CamelPath.toCamelUri converts the ActorRef to the Camel actor component URI format which points to the
actor endpoint as described above. When a message is received on the jetty endpoint, it is routed to the Responder
actor, which in return replies back to the client of the HTTP request.
Intercepting route construction
The previous section, Akka Camel components, explained how to setup a route to an actor manually. It was the
applications responsibility to define the route and add it to the current CamelContext. This section explains a
more convenient way to define custom routes: akka-camel is still setting up the routes to consumer actors (and
adds these routes to the current CamelContext) but applications can define extensions to these routes. Extensions
can be defined with Camels Java DSL or Scala DSL. For example, an extension could be a custom error handler
that redelivers messages from an endpoint to an actors bounded mailbox when the mailbox was full.
The following examples demonstrate how to extend a route to a consumer actor for handling exceptions thrown
by that actor.
import
import
import
import
import
import
import
import
akka.actor.Status;
akka.camel.CamelMessage;
akka.camel.javaapi.UntypedConsumerActor;
akka.dispatch.Mapper;
org.apache.camel.builder.Builder;
org.apache.camel.model.ProcessorDefinition;
org.apache.camel.model.RouteDefinition;
scala.Option;
6.9. Camel
315
@Override
public Mapper<RouteDefinition,
ProcessorDefinition<?>> getRouteDefinitionHandler() {
return mapper;
}
@Override
public void preRestart(Throwable reason, Option<Object> message) {
getSender().tell(new Status.Failure(reason), getSelf());
}
}
The above ErrorThrowingConsumer sends the Failure back to the sender in preRestart because the Exception that
is thrown in the actor would otherwise just crash the actor, by default the actor would be restarted, and the response
would never reach the client of the Consumer.
The akka-camel module creates a RouteDefinition instance by calling from(endpointUri) on a Camel RouteBuilder
(where endpointUri is the endpoint URI of the consumer actor) and passes that instance as argument to the route
definition handler *). The route definition handler then extends the route and returns a ProcessorDefinition (in the
above example, the ProcessorDefinition returned by the end method. See the org.apache.camel.model package for
details). After executing the route definition handler, akka-camel finally calls a to(targetActorUri) on the returned
ProcessorDefinition to complete the route to the consumer actor (where targetActorUri is the actor component
URI as described in Access to actors). If the actor cannot be found, a ActorNotRegisteredException is thrown.
*) Before passing the RouteDefinition instance to the route definition handler, akka-camel may make some further
modifications to it.
6.9.6 Examples
The Typesafe Activator tutorial named Akka Camel Samples with Java contains 3 samples:
Asynchronous routing and transformation - This example demonstrates how to implement consumer and
producer actors that support Asynchronous routing with their Camel endpoints.
Custom Camel route - Demonstrates the combined usage of a Producer and a Consumer actor as well
as the inclusion of a custom Camel route.
Quartz Scheduler Example - Showing how simple is to implement a cron-style scheduler by using the Camel
Quartz component
6.9.7 Configuration
There are several configuration properties for the Camel module, please refer to the reference configuration.
6.9. Camel
316
CHAPTER
SEVEN
UTILITIES
7.1 Event Bus
Originally conceived as a way to send messages to groups of actors, the EventBus has been generalized into a
set of abstract base classes implementing a simple interface:
/**
* Attempts to register the subscriber to the specified Classifier
* @return true if successful and false if not (because it was already
subscribed to that Classifier, or otherwise)
*
*/
public boolean subscribe(Subscriber subscriber, Classifier to);
/**
* Attempts to deregister the subscriber from the specified Classifier
* @return true if successful and false if not (because it wasnt subscribed
to that Classifier, or otherwise)
*
*/
public boolean unsubscribe(Subscriber subscriber, Classifier from);
/**
* Attempts to deregister the subscriber from all Classifiers it may be subscribed to
*/
public void unsubscribe(Subscriber subscriber);
/**
* Publishes the specified Event to this bus
*/
public void publish(Event event);
Note: Please note that the EventBus does not preserve the sender of the published messages. If you need a
reference to the original sender you have to provide it inside the message.
This mechanism is used in different places within Akka, e.g. the Event Stream. Implementations can make use of
the specific building blocks presented below.
An event bus must define the following three type parameters:
Event (E) is the type of all events published on that bus
Subscriber (S) is the type of subscribers allowed to register on that event bus
Classifier (C) defines the classifier to be used in selecting subscribers for dispatching events
The traits below are still generic in these types, but they need to be defined for any concrete implementation.
317
7.1.1 Classifiers
The classifiers presented here are part of the Akka distribution, but rolling your own in case you do not find a
perfect match is not difficult, check the implementation of the existing ones on github
Lookup Classification
The simplest classification is just to extract an arbitrary classifier from each event and maintaining a set of
subscribers for each possible classifier. This can be compared to tuning in on a radio station. The trait
LookupClassification is still generic in that it abstracts over how to compare subscribers and how exactly to classify.
The necessary methods to be implemented are illustrated with the following example:
import akka.event.japi.LookupEventBus;
public class MsgEnvelope {
public final String topic;
public final Object payload;
public MsgEnvelope(String topic, Object payload) {
this.topic = topic;
this.payload = payload;
}
}
/**
* Publishes the payload of the MsgEnvelope when the topic of the
* MsgEnvelope equals the String specified when subscribing.
*/
public class LookupBusImpl extends LookupEventBus<MsgEnvelope, ActorRef, String> {
// is used for extracting the classifier from the incoming events
@Override public String classify(MsgEnvelope event) {
return event.topic;
}
// will be invoked for each event for all subscribers which registered themselves
// for the events classifier
@Override public void publish(MsgEnvelope event, ActorRef subscriber) {
subscriber.tell(event.payload, ActorRef.noSender());
}
// must define a full order over the subscribers, expressed as expected from
// java.lang.Comparable.compare
@Override public int compareSubscribers(ActorRef a, ActorRef b) {
return a.compareTo(b);
}
// determines the initial size of the index data structure
// used internally (i.e. the expected number of different classifiers)
@Override public int mapSize() {
return 128;
}
}
318
319
This classifier is also efficient in case no subscribers are found for an event, but it uses conventional locking to
synchronize an internal classifier cache, hence it is not well-suited to use cases in which subscriptions change
with very high frequency (keep in mind that opening a classifier by sending the first message will also have to
re-check all previous subscriptions).
Scanning Classification
The previous classifier was built for multi-classifier subscriptions which are strictly hierarchical, this classifier is
useful if there are overlapping classifiers which cover various parts of the event space without forming a hierarchy.
It can be compared to tuning in on (possibly multiple) radio stations by geographical reachability (for old-school
radio-wave transmission).
The necessary methods to be implemented are illustrated with the following example:
import akka.event.japi.ScanningEventBus;
/**
* Publishes String messages with length less than or equal to the length
* specified when subscribing.
*/
public class ScanningBusImpl extends ScanningEventBus<String, ActorRef, Integer> {
// is needed for determining matching classifiers and storing them in an
// ordered collection
@Override public int compareClassifiers(Integer a, Integer b) {
return a.compareTo(b);
}
// is needed for storing subscribers in an ordered collection
@Override public int compareSubscribers(ActorRef a, ActorRef b) {
return a.compareTo(b);
}
// determines whether a given classifier shall match a given event; it is invoked
// for each subscription for all received events, hence the name of the classifier
@Override public boolean matches(Integer classifier, String event) {
return event.length() <= classifier;
}
// will be invoked for each event for all subscribers which registered themselves
// for the events classifier
@Override public void publish(String event, ActorRef subscriber) {
subscriber.tell(event, ActorRef.noSender());
}
}
This classifier takes always a time which is proportional to the number of subscriptions, independent of how many
actually match.
320
Actor Classification
This classification was originally developed specifically for implementing DeathWatch: subscribers as well as
classifiers are of type ActorRef.
The necessary methods to be implemented are illustrated with the following example:
import akka.event.japi.ActorEventBus;
public class Notification {
public final ActorRef ref;
public final int id;
public Notification(ActorRef ref, int id) {
this.ref = ref;
this.id = id;
}
}
public class ActorBusImpl extends ActorEventBus<Notification> {
// is used for extracting the classifier from the incoming events
@Override public ActorRef classify(Notification event) {
return event.ref;
}
// determines the initial size of the index data structure
// used internally (i.e. the expected number of different classifiers)
@Override public int mapSize() {
return 128;
}
}
This classifier is still is generic in the event type, and it is efficient for all use cases.
321
import
import
import
import
import
akka.actor.Props;
akka.actor.ActorRef;
akka.actor.ActorSystem;
akka.actor.UntypedActor;
akka.actor.DeadLetter;
Default Handlers
Upon start-up the actor system creates and subscribes actors to the event stream for logging: these are the handlers
which are configured for example in application.conf:
akka {
loggers = ["akka.event.Logging$DefaultLogger"]
}
The handlers listed here by fully-qualified class name will be subscribed to all log event classes with priority higher
than or equal to the configured log-level and their subscriptions are kept in sync when changing the log-level at
runtime:
system.eventStream.setLogLevel(Logging.DebugLevel());
This means that log events for a level which will not be logged are typically not dispatched at all (unless manual
subscriptions to the respective event class have been done)
Dead Letters
As described at Stopping actors, messages queued when an actor terminates or sent after its death are re-routed
to the dead letter mailbox, which by default will publish the messages wrapped in DeadLetter. This wrapper
holds the original sender, receiver and message of the envelope which was redirected.
Other Uses
The event stream is always there and ready to be used, just publish your own events (it accepts Object) and
subscribe listeners to the corresponding JVM classes.
7.2 Logging
Logging in Akka is not tied to a specific logging backend. By default log messages are printed to STDOUT, but
you can plug-in a SLF4J logger or your own logger. Logging is performed asynchronously to ensure that logging
has minimal performance impact. Logging generally means IO and locks, which can slow down the operations of
your code if it was performed synchronously.
7.2. Logging
322
The Java Class of the log source is also included in the generated LogEvent. In case of a simple string this
is replaced with a marker class akka.event.DummyClassForStringSources in order to allow special
treatment of this case, e.g. in the SLF4J event listener which will then use the string instead of the class name for
looking up the logger instance to use.
Logging of Dead Letters
By default messages sent to dead letters are logged at info level. Existence of dead letters does not necessarily
indicate a problem, but it might be, and therefore they are logged by default. After a few messages this logging
is turned off, to avoid flooding the logs. You can disable this logging completely or adjust how many dead letters
7.2. Logging
323
that are logged. During system shutdown it is likely that you see dead letters, since pending messages in the actor
mailboxes are sent to dead letters. You can also disable logging of dead letters during shutdown.
akka {
log-dead-letters = 10
log-dead-letters-during-shutdown = on
}
To customize the logging further or take other actions for dead letters you can subscribe to the Event Stream.
Auxiliary logging options
Akka has a couple of configuration options for very low level debugging, that makes most sense in for developers
and not for operations.
You almost definitely need to have logging set to DEBUG to use any of the options below:
akka {
loglevel = "DEBUG"
}
This config option is very good if you want to know what config settings are loaded by Akka:
akka {
# Log the complete configuration at INFO level when the actor system is started.
# This is useful when you are uncertain of what configuration is used.
log-config-on-start = on
}
If you want very detailed logging of all automatically received messages that are processed by Actors:
akka {
actor {
debug {
# enable DEBUG logging of all AutoReceiveMessages (Kill, PoisonPill et.c.)
autoreceive = on
}
}
}
If you want very detailed logging of all lifecycle changes of Actors (restarts, deaths etc):
akka {
actor {
debug {
# enable DEBUG logging of actor lifecycle changes
lifecycle = on
}
}
}
If you want very detailed logging of all events, transitions and timers of FSM Actors that extend LoggingFSM:
akka {
actor {
debug {
# enable DEBUG logging of all LoggingFSMs for events, transitions and timers
fsm = on
}
}
}
7.2. Logging
324
akka {
actor {
debug {
# enable DEBUG logging of subscription changes on the eventStream
event-stream = on
}
}
}
If you want to see all messages that are received through remoting at DEBUG log level: (This is logged as they
are received by the transport layer, not by any Actor)
akka {
remote {
# If this is "on", Akka will log all inbound messages at DEBUG level,
# if off then they are not logged
log-received-messages = on
}
}
If you want to see message types with payload size in bytes larger than a specified limit at INFO log level:
akka {
remote {
# Logging of message types with payload size in bytes larger than
# this value. Maximum detected size per message type is logged once,
# with an increase threshold of 10%.
# By default this feature is turned off. Activate it by setting the property to
# a value in bytes, such as 1000b. Note that for all messages larger than this
# limit there will be extra performance and scalability cost.
log-frame-size-exceeding = 1000b
}
}
Also see the logging options for TestKit: Tracing Actor Invocations.
Turn Off Logging
To turn off logging you can configure the log levels to be OFF like this.
akka {
stdout-loglevel = "OFF"
loglevel = "OFF"
}
The stdout-loglevel is only in effect during system startup and shutdown, and setting it to OFF as well,
ensures that nothing gets logged during system startup or shutdown.
7.2. Logging
325
7.2.2 Loggers
Logging is performed asynchronously through an event bus. Log events are processed by an event handler actor
and it will receive the log events in the same order as they were emitted.
You can configure which event handlers are created at system start-up and listen to logging events. That is done
using the loggers element in the Configuration. Here you can also define the log level.
akka {
# Loggers to register at boot time (akka.event.Logging$DefaultLogger logs
# to STDOUT)
loggers = ["akka.event.Logging$DefaultLogger"]
# Options: OFF, ERROR, WARNING, INFO, DEBUG
loglevel = "DEBUG"
}
The default one logs to STDOUT and is registered by default. It is not intended to be used for production. There
is also an Logging to stdout during startup and shutdown logger available in the akka-slf4j module.
Example of creating a listener:
import akka.event.Logging;
import akka.event.LoggingAdapter;
import
import
import
import
import
akka.event.Logging.InitializeLogger;
akka.event.Logging.Error;
akka.event.Logging.Warning;
akka.event.Logging.Info;
akka.event.Logging.Debug;
7.2.4 SLF4J
Akka provides a logger for SL4FJ. This module is available in the akka-slf4j.jar. It has one single dependency;
the slf4j-api jar. In runtime you also need a SLF4J backend, we recommend Logback:
<dependency>
<groupId>ch.qos.logback</groupId>
<artifactId>logback-classic</artifactId>
7.2. Logging
326
<version>1.0.13</version>
</dependency>
You need to enable the Slf4jLogger in the loggers element in the Configuration. Here you can also define the
log level of the event bus. More fine grained log levels can be defined in the configuration of the SLF4J backend
(e.g. logback.xml).
akka {
loggers = ["akka.event.slf4j.Slf4jLogger"]
loglevel = "DEBUG"
}
One gotcha is that the timestamp is attributed in the event handler, not when actually doing the logging.
The SLF4J logger selected for each log event is chosen based on the Class of the log source specified when creating the LoggingAdapter, unless that was given directly as a string in which case
that string is used (i.e.
LoggerFactory.getLogger(Class c) is used in the first case and
LoggerFactory.getLogger(String s) in the second).
Note: Beware that the actor systems name is appended to a String log source if the LoggingAdapter was
created giving an ActorSystem to the factory. If this is not intended, give a LoggingBus instead as shown
below:
final LoggingAdapter log = Logging.getLogger(system.eventStream(), "my.string");
Note: It will probably be a good idea to use the sourceThread MDC value also in non-Akka parts of the
application in order to have this property consistently available in the logs.
Another helpful facility is that Akka captures the actors address when instantiating a logger within it, meaning
that the full instance identification is available for associating log messages e.g. with members of a router. This
information is available in the MDC with attribute name akkaSource:
<appender name="STDOUT" class="ch.qos.logback.core.ConsoleAppender">
<encoder>
<pattern>%date{ISO8601} %-5level %logger{36} %X{akkaSource} - %msg%n</pattern>
</encoder>
</appender>
For more details on what this attribute containsalso for non-actorsplease see How to Log.
More accurate timestamps for log output in MDC
Akkas logging is asynchronous which means that the timestamp of a log entry is taken from when the underlying
logger implementation is called, which can be surprising at first. If you want to more accurately output the
timestamp, use the MDC attribute akkaTimestamp:
7.2. Logging
327
Once you have the logger, you just need to add the custom values before you log something. This way, the values
will be put in the SLF4J MDC right before appending the log and removed after.
Note: The cleanup (removal) should be done in the actor at the end, otherwise, next message will log with same
mdc values, if it is not set to a new map. Use log.clearMDC().
import
import
import
import
akka.event.Logging;
akka.event.DiagnosticLoggingAdapter;
java.util.HashMap;
java.util.Map;
Now, the values will be available in the MDC, so you can use them in the layout pattern:
<appender name="STDOUT" class="ch.qos.logback.core.ConsoleAppender">
<encoder>
<pattern>
%-5level %logger{36} [req: %X{requestId}, visitor: %X{visitorId}] - %msg%n
</pattern>
</encoder>
</appender>
7.3 Scheduler
Sometimes the need for making things happen in the future arises, and where do you go look then? Look
no further than ActorSystem! There you find the scheduler method that returns an instance of
7.3. Scheduler
328
akka.actor.Scheduler, this instance is unique per ActorSystem and is used internally for scheduling things
to happen at specific points in time.
You can schedule sending of messages to actors and execution of tasks (functions or Runnable). You will get a
Cancellable back that you can call cancel on to cancel the execution of the scheduled operation.
Warning: The default implementation of Scheduler used by Akka is based on job buckets which
are emptied according to a fixed schedule. It does not execute tasks at the exact time, but on every tick,
it will run everything that is (over)due. The accuracy of the default Scheduler can be modified by the
akka.scheduler.tick-duration configuration property.
Schedule a Runnable, that sends the current time to the testActor, to be executed after 50ms:
system.scheduler().scheduleOnce(Duration.create(50, TimeUnit.MILLISECONDS),
new Runnable() {
@Override
public void run() {
testActor.tell(System.currentTimeMillis(), ActorRef.noSender());
}
}, system.dispatcher());
Warning: If you schedule Runnable instances you should be extra careful to not pass or close over unstable
references. In practice this means that you should not call methods on the enclosing Actor from within the
Runnable. If you need to schedule an invocation it is better to use the schedule() variant accepting a
message and an ActorRef to schedule a message to self (containing the necessary parameters) and then call
the method when the message is received.
Schedule to send the Tick-message to the tickActor after 0ms repeating every 50ms:
import
import
import
import
import
akka.actor.Props;
scala.concurrent.duration.Duration;
java.util.concurrent.TimeUnit;
akka.actor.UntypedActor;
akka.actor.Cancellable;
7.3. Scheduler
329
7.3. Scheduler
330
*/
@Override
public abstract double maxFrequency();
}
7.4 Duration
Durations are used throughout the Akka library, wherefore this concept is represented by a special data type,
scala.concurrent.duration.Duration. Values of this type may represent infinite (Duration.Inf,
Duration.MinusInf) or finite durations, or be Duration.Undefined.
7.4.2 Scala
In Scala durations are constructable using a mini-DSL and support all expected arithmetic operations:
7.4. Duration
331
import scala.concurrent.duration._
val fivesec = 5.seconds
val threemillis = 3.millis
val diff = fivesec - threemillis
assert(diff < fivesec)
val fourmillis = threemillis * 4 / 3 // you cannot write it the other way around
val n = threemillis / (1 millisecond)
Note: You may leave out the dot if the expression is clearly delimited (e.g. within parentheses or in an argument
list), but it is recommended to use it if the time unit is the last token on a line, otherwise semi-colon inference
might go wrong, depending on what starts the next line.
7.4.3 Java
Java provides less syntactic sugar, so you have to spell out the operations as method calls instead:
import scala.concurrent.duration.Duration;
import scala.concurrent.duration.Deadline;
final Duration fivesec = Duration.create(5, "seconds");
final Duration threemillis = Duration.create("3 millis");
final Duration diff = fivesec.minus(threemillis);
assert diff.lt(fivesec);
assert Duration.Zero().lt(Duration.Inf());
7.4.4 Deadline
Durations have a brother named Deadline, which is a class holding a representation of an absolute point in
time, and support deriving a duration from this by calculating the difference between now and the deadline. This
is useful when you want to keep one overall deadline without having to take care of the book-keeping wrt. the
passing of time yourself:
val deadline = 10.seconds.fromNow
// do something
val rest = deadline.timeLeft
332
resource exhaustion. This will affect all users, even those who are not using functionality dependent on this third
party web service.
Introducing circuit breakers on the web service call would cause the requests to begin to fail-fast, letting the user
know that something is wrong and that they need not refresh their request. This also confines the failure behavior
to only those users that are using functionality dependent on the third party, other users are no longer affected as
there is no resource exhaustion. Circuit breakers can also allow savvy developers to mark portions of the site that
use the functionality unavailable, or perhaps show some cached content as appropriate while the breaker is open.
The Akka library provides an implementation of a circuit breaker called akka.pattern.CircuitBreaker
which has the behavior described below.
7.5.3 Examples
Initialization
Heres how a CircuitBreaker would be configured for:
5 maximum failures
a call timeout of 10 seconds
333
scala.concurrent.duration._
akka.pattern.CircuitBreaker
akka.pattern.pipe
akka.actor.Actor
akka.actor.ActorLogging
scala.concurrent.Future
akka.event.Logging
Java
import
import
import
import
import
import
akka.actor.UntypedActor;
scala.concurrent.Future;
akka.event.LoggingAdapter;
scala.concurrent.duration.Duration;
akka.pattern.CircuitBreaker;
akka.event.Logging;
334
Call Protection
Heres how the CircuitBreaker would be used to protect an asynchronous call as well as a synchronous one:
Scala
def dangerousCall: String = "This really isnt that dangerous of a call after all"
def receive = {
case "is my middle name" =>
breaker.withCircuitBreaker(Future(dangerousCall)) pipeTo sender()
case "block for me" =>
sender() ! breaker.withSyncCircuitBreaker(dangerousCall)
}
Java
public String dangerousCall() {
return "This really isnt that dangerous of a call after all";
}
@Override
public void onReceive(Object message) {
if (message instanceof String) {
String m = (String) message;
if ("is my middle name".equals(m)) {
final Future<String> f = future(
new Callable<String>() {
public String call() {
return dangerousCall();
}
}, getContext().dispatcher());
pipe(breaker.callWithCircuitBreaker(
new Callable<Future<String>>() {
public Future<String> call() throws Exception {
return f;
}
}), getContext().dispatcher()).to(getSender());
}
if ("block for me".equals(m)) {
getSender().tell(breaker
.callWithSyncCircuitBreaker(
new Callable<String>() {
@Override
public String call() throws Exception {
return dangerousCall();
}
}), getSelf());
}
}
}
Note:
Using the CircuitBreaker companion objects apply or create methods will return a
CircuitBreaker where callbacks are executed in the callers thread. This can be useful if the asynchronous
Future behavior is unnecessary, for example invoking a synchronous-only API.
335
Then we need to create an ExtensionId for our extension so we can grab ahold of it.
import akka.actor.*;
import java.util.concurrent.atomic.AtomicLong;
public class CountExtension extends AbstractExtensionId<CountExtensionImpl>
implements ExtensionIdProvider {
//This will be the identifier of our CountExtension
public final static CountExtension CountExtensionProvider = new CountExtension();
private CountExtension() {}
//The lookup method is required by ExtensionIdProvider,
// so we return ourselves here, this allows us
// to configure our extension to be loaded when
// the ActorSystem starts up
public CountExtension lookup() {
return CountExtension.CountExtensionProvider; //The public static final
}
//This method will be called by Akka
// to instantiate our Extension
public CountExtensionImpl createExtension(ExtendedActorSystem system) {
return new CountExtensionImpl();
}
}
336
7.6.3 Applicability
The sky is the limit! By the way, did you know that Akkas Typed Actors, Serialization and other
features are implemented as Akka Extensions?
Application specific settings
The Configuration can be used for application specific settings. A good practice is to place those settings in an
Extension.
Sample configuration:
myapp {
db {
uri = "mongodb://example1.com:27017,example2.com:27017"
}
circuit-breaker {
timeout = 30 seconds
}
}
The Extension:
import
import
import
import
import
import
import
import
akka.actor.Extension;
akka.actor.AbstractExtensionId;
akka.actor.ExtensionIdProvider;
akka.actor.ActorSystem;
akka.actor.ExtendedActorSystem;
scala.concurrent.duration.Duration;
com.typesafe.config.Config;
java.util.concurrent.TimeUnit;
337
Use it:
public class MyActor extends UntypedActor {
// typically you would use static import of the Settings.SettingsProvider field
final SettingsImpl settings =
Settings.SettingsProvider.get(getContext().system());
Connection connection =
connect(settings.DB_URI, settings.CIRCUIT_BREAKER_TIMEOUT);
}
7.7 Microkernel
The purpose of the Akka Microkernel is to offer a bundling mechanism so that you can distribute an Akka application as a single payload, without the need to run in a Java Application Server or manually having to create a
launcher script.
The Akka Microkernel is included in the Akka download found at downloads.
To run an application with the microkernel you need to create a Bootable class that handles the startup and shutdown the application. An example is included below.
Put your application jar in the deploy directory and additional dependencies in the lib directory to have them
automatically loaded and placed on the classpath.
To start the kernel use the scripts in the bin directory, passing the boot classes for your application.
The start script adds config directory first in the classpath, followed by lib/*. It runs java with main class
akka.kernel.Main and the supplied Bootable class as argument.
Example command (on a unix-based system):
7.7. Microkernel
338
bin/akka sample.kernel.hello.HelloKernel
akka.actor.ActorRef;
akka.actor.UntypedActor;
akka.actor.ActorSystem;
akka.actor.Props;
akka.kernel.Bootable;
7.7. Microkernel
339
CHAPTER
EIGHT
340
The second variant sets up an initial one shot message send in the preStart method of the actor, and the
then the actor when it receives this message sets up a new one shot message send. You also have to override
postRestart so we dont call preStart and schedule the initial message send again.
Note: With this approach we wont fill up the mailbox with tick messages if the actor is under pressure, but only
schedule a new tick message when we have seen the previous one.
public class ScheduleInReceive extends UntypedActor {
@Override
public void preStart() {
getContext().system().scheduler().scheduleOnce(
Duration.create(500, TimeUnit.MILLISECONDS),
getSelf(), "tick", getContext().dispatcher(), null);
}
// override postRestart so we dont call preStart and schedule a new message
@Override
public void postRestart(Throwable reason) {
}
@Override
public void onReceive(Object message) throws Exception {
if (message.equals("tick")) {
// send another periodic tick after the specified delay
getContext().system().scheduler().scheduleOnce(
Duration.create(1000, TimeUnit.MILLISECONDS),
getSelf(), "tick", getContext().dispatcher(), null);
// do something useful here
}
else {
unhandled(message);
}
}
}
341
import scala.concurrent.duration.Duration;
import
import
import
import
import
import
import
import
import
import
import
import
import
import
import
import
akka.actor.Actor;
akka.actor.ActorKilledException;
akka.actor.ActorRef;
akka.actor.ActorRefFactory;
akka.actor.Cancellable;
akka.actor.OneForOneStrategy;
akka.actor.Props;
akka.actor.Scheduler;
akka.actor.Status;
akka.actor.SupervisorStrategy;
akka.actor.SupervisorStrategy.Directive;
akka.actor.Terminated;
akka.actor.UntypedActor;
akka.japi.Function;
akka.pattern.Patterns;
akka.util.Timeout;
}
private static class AskTimeout {
}
public static class AskSupervisorCreator extends UntypedActor {
@Override
public void onReceive(Object message) throws Exception {
if (message instanceof AskParam) {
ActorRef supervisor = getContext().actorOf(
Props.create(AskSupervisor.class));
supervisor.forward(message, getContext());
} else {
unhandled(message);
}
}
}
public static class AskSupervisor extends UntypedActor {
private ActorRef targetActor;
private ActorRef caller;
private AskParam askParam;
private Cancellable timeoutMessage;
@Override
public SupervisorStrategy supervisorStrategy() {
return new OneForOneStrategy(0, Duration.Zero(),
new Function<Throwable, Directive>() {
public Directive apply(Throwable cause) {
caller.tell(new Status.Failure(cause), self());
return SupervisorStrategy.stop();
342
}
});
}
@Override
public void onReceive(Object message) throws Exception {
if (message instanceof AskParam) {
askParam = (AskParam) message;
caller = getSender();
targetActor = getContext().actorOf(askParam.props);
getContext().watch(targetActor);
targetActor.forward(askParam.message, getContext());
Scheduler scheduler = getContext().system().scheduler();
timeoutMessage = scheduler.scheduleOnce(askParam.timeout.duration(),
self(), new AskTimeout(), context().dispatcher(), null);
} else if (message instanceof Terminated) {
Throwable ex = new ActorKilledException("Target actor terminated.");
caller.tell(new Status.Failure(ex), self());
timeoutMessage.cancel();
getContext().stop(self());
} else if (message instanceof AskTimeout) {
Throwable ex = new TimeoutException("Target actor timed out after "
+ askParam.timeout.toString());
caller.tell(new Status.Failure(ex), self());
getContext().stop(self());
} else
unhandled(message);
}
}
public static Future<Object> askOf(ActorRef supervisorCreator, Props props,
Object message, Timeout timeout) {
AskParam param = new AskParam(props, message, timeout);
return Patterns.ask(supervisorCreator, param, timeout);
}
synchronized public static ActorRef createSupervisorCreator(
ActorRefFactory factory) {
return factory.actorOf(Props.create(AskSupervisorCreator.class));
}
}
In the askOf method the SupervisorCreator is sent the user message. The SupervisorCreator creates a SupervisorActor and forwards the message. This prevents the actor system from overloading due to actor creations.
The SupervisorActor is responsible to create the user actor, forwards the message, handles actor termination and
supervision. Additionally the SupervisorActor stops the user actor if execution time expired.
In case of an exception the supervisor tells the temporary actor which exception was thrown. Afterwards the actor
hierarchy is stopped.
Finally we are able to execute an actor and receive the results or exceptions.
package docs.pattern;
import
import
import
import
import
import
import
scala.concurrent.Await;
scala.concurrent.Future;
akka.actor.ActorRef;
akka.actor.ActorRefFactory;
akka.actor.Props;
akka.actor.UntypedActor;
akka.util.Timeout;
343
Note: Spread the word: this is the easiest way to get famous!
Please keep this pattern at the end of this file.
344
CHAPTER
NINE
EXPERIMENTAL MODULES
The following modules of Akka are marked as experimental, which means that they are in early access mode,
which also means that they are not covered by commercial support. The purpose of releasing them early, as
experimental, is to make them easily available and improve based on feedback, or even discover that the module
wasnt useful.
An experimental module doesnt have to obey the rule of staying binary compatible between micro releases.
Breaking API changes may be introduced in minor releases without notice as we refine and simplify based on
your feedback. An experimental module may be dropped in minor releases without prior deprecation.
9.1 Persistence
Akka persistence enables stateful actors to persist their internal state so that it can be recovered when an actor is
started, restarted after a JVM crash or by a supervisor, or migrated in a cluster. The key concept behind Akka
persistence is that only changes to an actors internal state are persisted but never its current state directly (except
for optional snapshots). These changes are only ever appended to storage, nothing is ever mutated, which allows
for very high transaction rates and efficient replication. Stateful actors are recovered by replaying stored changes
to these actors from which they can rebuild internal state. This can be either the full history of changes or starting
from a snapshot which can dramatically reduce recovery times. Akka persistence also provides point-to-point
communication channels with at-least-once message delivery semantics.
Warning: This module is marked as experimental as of its introduction in Akka 2.3.0. We will continue to
improve this API based on our users feedback, which implies that while we try to keep incompatible changes
to a minimum the binary compatibility guarantee for maintenance releases does not apply to the contents of
the akka.persistence package.
Akka persistence is inspired by and the official replacement of the eventsourced library. It follows the same
concepts and architecture of eventsourced but significantly differs on API and implementation level. See also
Migration Guide Eventsourced to Akka Persistence 2.3.x
9.1.1 Dependencies
Akka persistence is a separate jar file. Make sure that you have the following dependency in your project:
"com.typesafe.akka" %% "akka-persistence-experimental" % "2.3.1"
9.1.2 Architecture
Processor: A processor is a persistent, stateful actor. Messages sent to a processor are written to a journal
before its receive method is called. When a processor is started or restarted, journaled messages are
replayed to that processor, so that it can recover internal state from these messages.
345
View: A view is a persistent, stateful actor that receives journaled messages that have been written by
another processor. A view itself does not journal new messages, instead, it updates internal state only from
a processors replicated message stream.
Channel: Channels are used by processors and views to communicate with other actors. They prevent that
replayed messages are redundantly delivered to these actors and provide at-least-once message delivery
semantics, also in case of sender and receiver JVM crashes.
Journal: A journal stores the sequence of messages sent to a processor. An application can control which
messages are journaled and which are received by the processor without being journaled. The storage
backend of a journal is pluggable. The default journal storage plugin writes to the local filesystem, replicated
journals are available as Community plugins.
Snapshot store: A snapshot store persists snapshots of a processors or a views internal state. Snapshots
are used for optimizing recovery times. The storage backend of a snapshot store is pluggable. The default
snapshot storage plugin writes to the local filesystem.
Event sourcing. Based on the building blocks described above, Akka persistence provides abstractions for
the development of event sourced applications (see section Event sourcing)
9.1.3 Processors
A processor can be implemented by extending the Processor trait and implementing the receive method.
import akka.persistence.{ Persistent, PersistenceFailure, Processor }
class MyProcessor extends Processor {
def receive = {
case Persistent(payload, sequenceNr) =>
// message successfully written to journal
case PersistenceFailure(payload, sequenceNr, cause) =>
// message failed to be written to journal
case other =>
// message not written to journal
}
}
Processors only write messages of type Persistent to the journal, others are received without being persisted.
When a processors receive method is called with a Persistent message it can safely assume that this
message has been successfully written to the journal. If a journal fails to write a Persistent message then
the processor is stopped, by default. If a processor should continue running on persistence failures it must handle
PersistenceFailure messages. In this case, a processor may want to inform the sender about the failure, so
that the sender can re-send the message, if needed.
A Processor itself is an Actor and can therefore be instantiated with actorOf.
import akka.actor.Props
val processor = actorOf(Props[MyProcessor], name = "myProcessor")
processor ! Persistent("foo") // will be journaled
processor ! "bar" // will not be journaled
Recovery
By default, a processor is automatically recovered on start and on restart by replaying journaled messages. New
messages sent to a processor during recovery do not interfere with replayed messages. New messages will only
be received by a processor after recovery completes.
9.1. Persistence
346
Recovery customization
Automated recovery on start can be disabled by overriding preStart with an empty implementation.
override def preStart() = ()
If not overridden, preStart sends a Recover() message to self. Applications may also override
preStart to define further Recover() parameters such as an upper sequence number bound, for example.
override def preStart() {
self ! Recover(toSequenceNr = 457L)
}
Upper sequence number bounds can be used to recover a processor to past state instead of current state. Automated
recovery on restart can be disabled by overriding preRestart with an empty implementation.
override def preRestart(reason: Throwable, message: Option[Any]) = ()
Recovery status
A processor can query its own recovery status via the methods
def recoveryRunning: Boolean
def recoveryFinished: Boolean
Sometimes there is a need for performing additional initialization when the recovery has completed, before processing any other message sent to the processor. The processor can send itself a message from preStart. It will
be stashed and received after recovery. The mailbox may contain other messages that are queued in front of that
message and therefore you need to stash until you receive that message.
override def preStart(): Unit = {
super.preStart()
self ! "FIRST"
}
def receive = initializing.orElse(active)
def initializing: Receive = {
case "FIRST" =>
recoveryCompleted()
context.become(active)
unstashAll()
case other if recoveryFinished =>
stash()
}
def recoveryCompleted(): Unit = {
// perform init after recovery, before any other messages
// ...
}
def active: Receive = {
case Persistent(msg, _) => //...
}
9.1. Persistence
347
Failure handling
A persistent message that caused an exception will be received again by a processor after restart. To prevent a
replay of that message during recovery it can be deleted.
override def preRestart(reason: Throwable, message: Option[Any]) {
message match {
case Some(p: Persistent) => deleteMessage(p.sequenceNr)
case _
=>
}
super.preRestart(reason, message)
}
Message deletion
A processor can delete a single message by calling the deleteMessage method with the sequence number of
that message as argument. An optional permanent parameter specifies whether the message shall be permanently deleted from the journal or only marked as deleted. In both cases, the message wont be replayed. Later
extensions to Akka persistence will allow to replay messages that have been marked as deleted which can be useful
for debugging purposes, for example. To delete all messages (journaled by a single processor) up to a specified
sequence number, processors should call the deleteMessages method.
Identifiers
A processor must have an identifier that doesnt change across different actor incarnations. It defaults to the
String representation of processors path without the address part and can be obtained via the processorId
method.
def processorId: String
Applications can customize a processors id by specifying an actor name during processor creation as shown in
section Processors. This changes that processors name in its actor hierarchy and hence influences only part of the
processor id. To fully customize a processors id, the processorId method must be overridden.
override def processorId = "my-stable-processor-id"
9.1.4 Views
Views can be implemented by extending the View trait and implementing the receive and the processorId
methods.
class MyView extends View {
def processorId: String = "some-processor-id"
def receive: Actor.Receive = {
case Persistent(payload, sequenceNr) => // ...
}
}
The processorId identifies the processor from which the view receives journaled messages. It is not necessary
the referenced processor is actually running. Views read messages from a processors journal directly. When a
processor is started later and begins to write new messages, the corresponding view is updated automatically, by
default.
9.1. Persistence
348
Updates
The default update interval of all views of an actor system is configurable:
akka.persistence.view.auto-update-interval = 5s
View implementation classes may also override the autoUpdateInterval method to return a custom update
interval for a specific view class or view instance. Applications may also trigger additional updates at any time by
sending a view an Update message.
val view = system.actorOf(Props[MyView])
view ! Update(await = true)
If the await parameter is set to true, messages that follow the Update request are processed when the incremental message replay, triggered by that update request, completed. If set to false (default), messages following
the update request may interleave with the replayed message stream. Automated updates always run with await
= false.
Automated updates of all views of an actor system can be turned off by configuration:
akka.persistence.view.auto-update = off
Implementation classes may override the configured default value by overriding the autoUpdate
method.
To limit the number of replayed messages per update request, applications can configure a custom akka.persistence.view.auto-update-replay-max value or override the
autoUpdateReplayMax method. The number of replayed messages for manual updates can be limited with
the replayMax parameter of the Update message.
Recovery
Initial recovery of views works in the very same way as for Processors (i.e. by sending a Recover
message to self). The maximum number of replayed messages during initial recovery is determined by
autoUpdateReplayMax. Further possibilities to customize initial recovery are explained in section Processors.
Identifiers
A view must have an identifier that doesnt change across different actor incarnations. It defaults to the String
representation of the actor path without the address part and can be obtained via the viewId method.
Applications can customize a views id by specifying an actor name during view creation. This changes that
views name in its actor hierarchy and hence influences only part of the view id. To fully customize a views
id, the viewId method must be overridden. Overriding viewId is the recommended way to generate stable
identifiers.
The viewId must differ from the referenced processorId, unless Snapshots of a view and its processor shall
be shared (which is what applications usually do not want).
9.1.5 Channels
Channels are special actors that are used by processors or views to communicate with other actors (channel destinations). The following discusses channels in context of processors but this is also applicable to views.
Channels prevent redundant delivery of replayed messages to destinations during processor recovery. A replayed
message is retained by a channel if its delivery has been confirmed by a destination.
import akka.actor.{ Actor, Props }
import akka.persistence.{ Channel, Deliver, Persistent, Processor }
class MyProcessor extends Processor {
9.1. Persistence
349
A channel is ready to use once it has been created, no recovery or further activation is needed. A Deliver request
instructs a channel to send a Persistent message to a destination. A destination is provided as ActorPath
and messages are sent by the channel via that paths ActorSelection. Sender references are preserved by a
channel, therefore, a destination can reply to the sender of a Deliver request.
Note: Sending via a channel has at-least-once delivery semanticsby virtue of either the sending actor or
the channel being persistentwhich means that the semantics do not match those of a normal ActorRef send
operation:
it is not at-most-once delivery
message order for the same senderreceiver pair is not retained due to possible resends
after a crash and restart of the destination messages are still deliveredto the new actor incarnation
These semantics match precisely what an ActorPath represents (see actor-lifecycle-scala), therefore you need
to supply a path and not a reference when constructing Deliver messages.
If a processor wants to reply to a Persistent message sender it should use the sender path as channel
destination.
channel ! Deliver(p.withPayload(s"processed ${payload}"), sender.path)
9.1. Persistence
350
A channel keeps messages in memory until their successful delivery has been confirmed or the maximum number
of re-deliveries is reached. To be notified about messages that have reached the maximum number of re-deliveries,
applications can register a listener at channel creation.
class MyListener extends Actor {
def receive = {
case RedeliverFailure(messages) => // ...
}
}
val myListener = context.actorOf(Props[MyListener])
val myChannel = context.actorOf(Channel.props(
ChannelSettings(redeliverFailureListener = Some(myListener))))
A listener receives RedeliverFailure notifications containing all messages that could not be delivered. On
receiving a RedeliverFailure message, an application may decide to restart the sending processor to enforce
a re-send of these messages to the channel or confirm these messages to prevent further re-sends. The sending
processor can also be restarted any time later to re-send unconfirmed messages.
This combination of
message persistence by sending processors
message replays by sending processors
message re-deliveries by channels and
application-level confirmations (acknowledgements) by destinations
enables channels to provide at-least-once message delivery semantics. Possible duplicates can be detected by
destinations by tracking message sequence numbers. Message sequence numbers are generated per sending processor. Depending on how a processor routes outbound messages to destinations, they may either see a contiguous
message sequence or a sequence with gaps.
Warning: If a processor emits more than one outbound message per inbound Persistent message it
must use a separate channel for each outbound message to ensure that confirmations are uniquely identifiable,
otherwise, at-least-once message delivery semantics do not apply. This rule has been introduced to avoid
writing additional outbound message identifiers to the journal which would decrease the overall throughput.
It is furthermore recommended to collapse multiple outbound messages to the same destination into a single
outbound message, otherwise, if sent via multiple channels, their ordering is not defined.
If an application wants to have more control how sequence numbers are assigned to messages it should use an
application-specific sequence number generator and include the generated sequence numbers into the payload
of Persistent messages.
Persistent channels
Channels created with Channel.props do not persist messages. These channels are usually used in combination with a sending processor that takes care of persistence, hence, channel-specific persistence is not necessary in
this case. They are referred to as transient channels in the following.
Persistent channels are like transient channels but additionally persist messages before delivering them. Messages that have been persisted by a persistent channel are deleted when destinations confirm their delivery. A persistent channel can be created with PersistentChannel.props and configured with a
PersistentChannelSettings object.
val channel = context.actorOf(PersistentChannel.props(
PersistentChannelSettings(redeliverInterval = 30 seconds, redeliverMax = 15)),
name = "myPersistentChannel")
channel ! Deliver(Persistent("example"), destination.path)
9.1. Persistence
351
A persistent channel is useful for delivery of messages to slow destinations or destinations that are unavailable for
a long time. It can constrain the number of pending confirmations based on the pendingConfirmationsMax
and pendingConfirmationsMin parameters of PersistentChannelSettings.
PersistentChannelSettings(
pendingConfirmationsMax = 10000,
pendingConfirmationsMin = 2000)
It suspends delivery when the number of pending confirmations reaches pendingConfirmationsMax and resumes delivery again when this number falls below pendingConfirmationsMin. This prevents both, flooding destinations with more messages than they can process and unlimited memory consumption by the channel.
A persistent channel continues to persist new messages even when message delivery is temporarily suspended.
Standalone usage
Applications may also use channels standalone. Transient channels can be used standalone if re-delivery attempts
to destinations are required but message loss in case of a sender JVM crash is not an issue. If message loss in
case of a sender JVM crash is an issue, persistent channels should be used. In this case, applications may want to
receive replies from the channel whether messages have been successfully persisted or not. This can be enabled
by creating the channel with the replyPersistent configuration parameter set to true:
PersistentChannelSettings(replyPersistent = true)
With this setting, either the successfully persisted message is replied to the sender or a PersistenceFailure
message. In case the latter case, the sender should re-send the message.
Identifiers
In the same way as Processors and Views, channels also have an identifier that defaults to a channels path.
A channel identifier can therefore be customized by using a custom actor name at channel creation. This
changes that channels name in its actor hierarchy and hence influences only part of the channel identifier.
To fully customize a channel identifier, it should be provided as argument Channel.props(String) or
PersistentChannel.props(String) (recommended to generate stable identifiers).
context.actorOf(Channel.props("my-stable-channel-id"))
Inside processors, new persistent messages are derived from the current persistent message before sending them
via a channel, either by calling p.withPayload(...) or Persistent(...) where the latter uses the
implicit currentPersistentMessage made available by Processor.
implicit def currentPersistentMessage: Option[Persistent]
This is necessary for delivery confirmations to work properly. Both ways are equivalent but we recommend using
p.withPayload(...) for clarity.
9.1. Persistence
352
Sequence number
The sequence number of a Persistent message can be obtained via its
def sequenceNr: Long
Persistent messages are assigned sequence numbers on a per-processor basis (or per channel basis if used standalone). A sequence starts at 1L and doesnt contain gaps unless a processor deletes messages.
9.1.7 Snapshots
Snapshots can dramatically reduce recovery times of processors and views. The following discusses snapshots in
context of processors but this is also applicable to views.
Processors can save snapshots of internal state by calling the saveSnapshot method. If saving of a snapshot
succeeds, the processor receives a SaveSnapshotSuccess message, otherwise a SaveSnapshotFailure
message
class MyProcessor extends Processor {
var state: Any = _
def receive = {
case "snap"
=> saveSnapshot(state)
case SaveSnapshotSuccess(metadata)
=> // ...
case SaveSnapshotFailure(metadata, reason) => // ...
}
}
During recovery, the processor is offered a previously saved snapshot via a SnapshotOffer message from
which it can initialize internal state.
class MyProcessor extends Processor {
var state: Any = _
def receive = {
case SnapshotOffer(metadata, offeredSnapshot) => state = offeredSnapshot
case Persistent(payload, sequenceNr)
=> // ...
}
}
The replayed messages that follow the SnapshotOffer message, if any, are younger than the offered snapshot.
They finally recover the processor to its current (i.e. latest) state.
In general, a processor is only offered a snapshot if that processor has previously saved one or more snapshots
and at least one of these snapshots matches the SnapshotSelectionCriteria that can be specified for
recovery.
processor ! Recover(fromSnapshot = SnapshotSelectionCriteria(
maxSequenceNr = 457L,
maxTimestamp = System.currentTimeMillis))
9.1. Persistence
353
Snapshot deletion
A processor can delete individual snapshots by calling the deleteSnapshot method with the sequence number and the timestamp of a snapshot as argument.
To bulk-delete snapshots matching
SnapshotSelectionCriteria, processors should use the deleteSnapshots method.
9.1. Persistence
354
}
case "snap" => saveSnapshot(state)
case "print" => println(state)
}
}
The example defines two data types, Cmd and Evt to represent commands and events, respectively. The state
of the ExampleProcessor is a list of persisted event data contained in ExampleState.
The processors receiveRecover method defines how state is updated during recovery by handling Evt
and SnapshotOffer messages. The processors receiveCommand method is a command handler. In this
example, a command is handled by generating two events which are then persisted and handled. Events are
persisted by calling persist with an event (or a sequence of events) as first argument and an event handler as
second argument.
The persist method persists events asynchronously and the event handler is executed for successfully persisted
events. Successfully persisted events are internally sent back to the processor as individual messages that trigger
event handler executions. An event handler may close over processor state and mutate it. The sender of a persisted
event is the sender of the corresponding command. This allows event handlers to reply to the sender of a command
(not shown).
The main responsibility of an event handler is changing processor state using event data and notifying others about
successful state changes by publishing events.
When persisting events with persist it is guaranteed that the processor will not receive further commands
between the persist call and the execution(s) of the associated event handler. This also holds for multiple
persist calls in context of a single command.
The easiest way to run this example yourself is to download Typesafe Activator and open the tutorial named Akka
Persistence Samples with Scala. It contains instructions on how to run the EventsourcedExample.
Note: Its also possible to switch between different command handlers during normal processing and recovery
with context.become() and context.unbecome(). To get the actor into the same state after recovery you need to take special care to perform the same state transitions with become and unbecome in the
receiveRecover method as you would have done in the command handler.
9.1. Persistence
355
}
}
}
In larger integration scenarios, channel destinations may be actors that submit received events to an external
message broker, for example. After having successfully submitted an event, they should call confirm() on the
received ConfirmablePersistent message.
A new batch write is triggered by a processor as soon as a batch reaches the maximum size or if the journal
completed writing the previous batch. Batch writes are never timer-based which keeps latencies at a minimum.
Applications that want to have more explicit control over batch writes and batch sizes can send processors
PersistentBatch messages.
class MyProcessor extends Processor {
def receive = {
case Persistent("a", _) => // ...
case Persistent("b", _) => // ...
}
}
val system = ActorSystem("example")
val processor = system.actorOf(Props[MyProcessor])
processor ! PersistentBatch(List(Persistent("a"), Persistent("b")))
Persistent messages contained in a PersistentBatch are always written atomically, even if the batch
size is greater than max-message-batch-size. Also, a PersistentBatch is written isolated from other
batches. Persistent messages contained in a PersistentBatch are received individually by a processor.
PersistentBatch messages, for example, are used internally by an EventsourcedProcessor to ensure atomic writes of events. All events that are persisted in context of a single command are written as a
single batch to the journal (even if persist is called multiple times per command). The recovery of an
EventsourcedProcessor will therefore never be done partially (with only a subset of events persisted by a
single command).
Confirmation
and
deletion
operations
performed
by
Channels
batched.
The
maximum
confirmation
and
deletion
batch
sizes
are
with
akka.persistence.journal.max-confirmation-batch-size
akka.persistence.journal.max-deletion-batch-size, respectively.
are
also
configurable
and
9.1. Persistence
356
import akka.persistence.journal._
import akka.persistence.snapshot._
AsyncWriteJournal is an actor that should be extended if the storage backend API supports asynchronous,
non-blocking writes. In this case, the methods to be implemented are:
/**
* Plugin API: asynchronously writes a batch of persistent messages to the journal.
* The batch write must be atomic i.e. either all persistent messages in the batch
* are written or none.
*/
def asyncWriteMessages(messages: immutable.Seq[PersistentRepr]): Future[Unit]
/**
* Plugin API: asynchronously writes a batch of delivery confirmations to the journal.
*/
def asyncWriteConfirmations(confirmations: immutable.Seq[PersistentConfirmation]): Future[Unit]
/**
* Plugin API: asynchronously deletes messages identified by messageIds from the
* journal. If permanent is set to false, the persistent messages are marked as
* deleted, otherwise they are permanently deleted.
*/
def asyncDeleteMessages(messageIds: immutable.Seq[PersistentId], permanent: Boolean): Future[Unit]
/**
* Plugin API: asynchronously deletes all persistent messages up to toSequenceNr
* (inclusive). If permanent is set to false, the persistent messages are marked
* as deleted, otherwise they are permanently deleted.
9.1. Persistence
357
*/
def asyncDeleteMessagesTo(processorId: String, toSequenceNr: Long, permanent: Boolean): Future[Uni
Message replays and sequence number recovery are always asynchronous, therefore, any journal plugin must
implement:
/**
* Plugin API: asynchronously replays persistent messages. Implementations replay
* a message by calling replayCallback. The returned future must be completed
* when all messages (matching the sequence number bounds) have been replayed.
* The future must be completed with a failure if any of the persistent messages
* could not be replayed.
*
* The replayCallback must also be called with messages that have been marked
* as deleted. In this case a replayed messages deleted method must return
* true.
*
* The channel ids of delivery confirmations that are available for a replayed
* message must be contained in that messages confirms sequence.
*
* @param processorId processor id.
* @param fromSequenceNr sequence number where replay should start (inclusive).
* @param toSequenceNr sequence number where replay should end (inclusive).
* @param max maximum number of messages to be replayed.
* @param replayCallback called to replay a single message. Can be called from any
thread.
*
*
* @see [[AsyncWriteJournal]]
* @see [[SyncWriteJournal]]
*/
def asyncReplayMessages(processorId: String, fromSequenceNr: Long, toSequenceNr: Long, max: Long)(
/**
* Plugin API: asynchronously reads the highest stored sequence number for the
* given processorId.
*
* @param processorId processor id.
* @param fromSequenceNr hint where to start searching for the highest sequence
number.
*
/
*
def asyncReadHighestSequenceNr(processorId: String, fromSequenceNr: Long): Future[Long]
9.1. Persistence
358
/**
* Plugin API: asynchronously loads a snapshot.
*
* @param processorId processor id.
* @param criteria selection criteria for loading.
*/
def loadAsync(processorId: String, criteria: SnapshotSelectionCriteria): Future[Option[SelectedSna
/**
* Plugin API: asynchronously saves a snapshot.
*
* @param metadata snapshot metadata.
* @param snapshot snapshot.
*/
def saveAsync(metadata: SnapshotMetadata, snapshot: Any): Future[Unit]
/**
* Plugin API: called after successful saving of a snapshot.
*
* @param metadata snapshot metadata.
*/
def saved(metadata: SnapshotMetadata)
/**
* Plugin API: deletes the snapshot identified by metadata.
*
* @param metadata snapshot metadata.
*/
def delete(metadata: SnapshotMetadata)
/**
* Plugin API: deletes all snapshots matching criteria.
*
* @param processorId processor id.
* @param criteria selection criteria for deleting.
*/
def delete(processorId: String, criteria: SnapshotSelectionCriteria)
A snapshot store plugin can be activated with the following minimal configuration:
# Path to the snapshot store plugin to be used
akka.persistence.snapshot-store.plugin = "my-snapshot-store"
# My custom snapshot store plugin
my-snapshot-store {
# Class name of the plugin.
class = "docs.persistence.MySnapshotStore"
# Dispatcher for the plugin actor.
plugin-dispatcher = "akka.persistence.dispatchers.default-plugin-dispatcher"
}
9.1. Persistence
359
With this plugin, each actor system runs its own private LevelDB instance.
Shared LevelDB journal
A LevelDB instance can also be shared by multiple actor systems (on the same or on different nodes). This, for
example, allows processors to failover to a backup node and continue using the shared journal instance from the
backup node.
Warning: A shared LevelDB instance is a single point of failure and should therefore only be used for testing
purposes. Highly-available, replicated journal are available as Community plugins.
A shared LevelDB instance is started by instantiating the SharedLeveldbStore actor.
import akka.persistence.journal.leveldb.SharedLeveldbStore
val store = system.actorOf(Props[SharedLeveldbStore], "store")
By default, the shared instance writes journaled messages to a local directory named journal in the current
working directory. The storage location can be changed by configuration:
akka.persistence.journal.leveldb-shared.store.dir = "target/shared"
Actor systems that use a shared LevelDB store must activate the akka.persistence.journal.leveldb-shared
plugin.
akka.persistence.journal.plugin = "akka.persistence.journal.leveldb-shared"
This plugin must be initialized by injecting the (remote) SharedLeveldbStore actor reference. Injection is
done by calling the SharedLeveldbJournal.setStore method with the actor reference as argument.
trait SharedStoreUsage extends Actor {
override def preStart(): Unit = {
context.actorSelection("akka.tcp://[email protected]:2552/user/store") ! Identify(1)
}
def receive = {
case ActorIdentity(1, Some(store)) =>
SharedLeveldbJournal.setStore(store, context.system)
}
}
Internal journal commands (sent by processors) are buffered until injection completes. Injection is idempotent i.e.
only the first injection is used.
Local snapshot store
The default snapshot store plugin is akka.persistence.snapshot-store.local. It writes snapshot
files to the local filesystem. The default storage location is a directory named snapshots in the current working
directory. This can be changed by configuration where the specified path can be relative or absolute:
9.1. Persistence
360
akka.persistence.snapshot-store.local.dir = "target/snapshots"
9.1.13 Testing
When running tests with LevelDB default settings in sbt, make sure to set fork := true in your sbt project
otherwise, youll see an UnsatisfiedLinkError. Alternatively, you can switch to a LevelDB Java port by
setting
akka.persistence.journal.leveldb.native = off
or
akka.persistence.journal.leveldb-shared.store.native = off
in your Akka configuration. The LevelDB Java port is for testing purposes only.
9.1.14 Miscellaneous
State machines
State machines can be persisted by mixing in the FSM trait into processors.
import akka.actor.FSM
import akka.persistence.{ Processor, Persistent }
class PersistentDoor extends Processor with FSM[String, Int] {
startWith("closed", 0)
when("closed") {
case Event(Persistent("open", _), counter) =>
goto("open") using (counter + 1) replying (counter)
}
when("open") {
case Event(Persistent("close", _), counter) =>
goto("closed") using (counter + 1) replying (counter)
9.1. Persistence
361
}
}
9.1.15 Configuration
There are several configuration properties for the persistence module, please refer to the reference configuration.
362
The test conductor server is responsible for coordinating barriers and sending commands to the test conductor
clients that act upon them, e.g. throttling network traffic to/from another client. More information on the possible
operations is availible in the akka.remote.testconductor.Conductor API documentation.
363
Defaults to
multiNodeJavaName
The name of the default Java executable on the target machines. Defaults to java.
Here are some examples of how you define hosts:
localhost
The current user on localhost using the default java.
user1@host1
User user1 on host host1 with the default java.
user2@host2:/usr/lib/jvm/java-7-openjdk-amd64/bin/java
User user2 on host host2 using java 7.
host3:/usr/lib/jvm/java-6-openjdk-amd64/bin/java
The current user on host host3 using java 6.
Running the Multi Node Tests
To run all the multi node test in multi-node mode (i.e. distributing the jar files and kicking off the tests remotely)
from inside sbt, use the multi-node-test task:
multi-node-test
To run all of them in multi-jvm mode (i.e. all JVMs on the local machine) do:
364
multi-jvm:test
More than one test name can be listed to run multiple specific tests. Tab completion in sbt makes it easy to
complete the test names.
If you are using the latest nightly build you should pick a timestamped Akka version from
http://repo.typesafe.com/typesafe/snapshots/com/typesafe/akka/akka-multi-node-testkit_2.10/. We recommend
against using SNAPSHOT in order to obtain stable builds.
package sample.multinode
import akka.remote.testkit.MultiNodeConfig
object MultiNodeSampleConfig extends MultiNodeConfig {
val node1 = role("node1")
val node2 = role("node2")
}
And then finally to the node test code. That starts the two nodes, and demonstrates a barrier, and a remote actor
message send/receive.
package sample.multinode
import akka.remote.testkit.MultiNodeSpec
365
import akka.testkit.ImplicitSender
import akka.actor.{ Props, Actor }
class MultiNodeSampleSpecMultiJvmNode1 extends MultiNodeSample
class MultiNodeSampleSpecMultiJvmNode2 extends MultiNodeSample
object MultiNodeSample {
class Ponger extends Actor {
def receive = {
case "ping" => sender() ! "pong"
}
}
}
class MultiNodeSample extends MultiNodeSpec(MultiNodeSampleConfig)
with STMultiNodeSpec with ImplicitSender {
import MultiNodeSampleConfig._
import MultiNodeSample._
def initialParticipants = roles.size
"A MultiNodeSample" must {
"wait for all nodes to enter a barrier" in {
enterBarrier("startup")
}
"send to and receive from a remote node" in {
runOn(node1) {
enterBarrier("deployed")
val ponger = system.actorSelection(node(node2) / "user" / "ponger")
ponger ! "ping"
expectMsg("pong")
}
runOn(node2) {
system.actorOf(Props[Ponger], "ponger")
enterBarrier("deployed")
}
enterBarrier("finished")
}
}
}
The easiest way to run this example yourself is to download Typesafe Activator and open the tutorial named Akka
Multi-Node Testing Sample with Scala.
366
Dont ask for the address of a node using node(address) after the node has been shut down. Grab the
address before shutting down the node.
Dont use MultiNodeSpec methods like address lookup, barrier entry et.c. from other threads than the main
test thread. This also means that you shouldnt use them from inside an actor, a future, or a scheduled task.
9.2.8 Configuration
There are several configuration properties for the Multi-Node Testing module, please refer to the reference configuration.
akka.actor.AbstractActor;
akka.event.Logging;
akka.event.LoggingAdapter;
akka.japi.pf.ReceiveBuilder;
367
receive(ReceiveBuilder.
match(String.class, s -> {
log.info("Received String message: {}", s);
}).
matchAny(o -> log.info("received unknown message")).build()
);
}
}
Please note that the Akka Actor receive message loop is exhaustive, which is different compared to Erlang and
the late Scala Actors. This means that you need to provide a pattern match for all messages that it can accept
and if you want to be able to handle unknown messages then you need to have a default case as in the example
above. Otherwise an akka.actor.UnhandledMessage(message, sender, recipient) will be
published to the ActorSystems EventStream.
Note further that the return type of the behavior defined above is Unit; if the actor shall reply to the received
message then this must be done explicitly as explained below.
The argument to the receive method is a partial function object, which is stored within the actor as its initial
behavior, see Become/Unbecome for further information on changing the behavior of an actor after its construction.
Props
Props is a configuration class to specify options for the creation of actors, think of it as an immutable and thus
freely shareable recipe for creating an actor including associated deployment information (e.g. which dispatcher
to use, see more below). Here are some examples of how to create a Props instance.
import akka.actor.Props;
Props props1 = Props.create(MyActor.class);
Props props2 = Props.create(ActorWithArgs.class,
() -> new ActorWithArgs("arg")); // careful, see below
Props props3 = Props.create(ActorWithArgs.class, "arg");
The second variant shows how to pass constructor arguments to the Actor being created, but it should only be
used outside of actors as explained below.
The last line shows a possibility to pass constructor arguments regardless of the context it is being used in.
The presence of a matching constructor is verified during construction of the Props object, resulting in an
IllegalArgumentEception if no or multiple matching constructors are found.
Dangerous Variants
// NOT RECOMMENDED within another actor:
// encourages to close over enclosing class
Props props7 = Props.create(ActorWithArgs.class,
() -> new ActorWithArgs("arg"));
This method is not recommended to be used within another actor because it encourages to close over the enclosing scope, resulting in non-serializable Props and possibly race conditions (breaking the actor encapsulation).
On the other hand using this variant in a Props factory in the actors companion object as documented under
Recommended Practices below is completely fine.
There were two use-cases for these methods: passing constructor arguments to the actorwhich is solved by the
newly introduced Props.create(clazz, args) method above or the recommended practice belowand
creating actors on the spot as anonymous classes. The latter should be solved by making these actors named
classes instead (if they are not declared within a top-level object then the enclosing instances this reference
needs to be passed as the first argument).
368
Warning: Declaring one actor within another is very dangerous and breaks actor encapsulation. Never pass
an actors this reference into Props!
Recommended Practices
It is a good idea to provide factory methods on the companion object of each Actor which help keeping the
creation of suitable Props as close to the actor definition as possible. This also avoids the pitfalls associated with
using the Props.create(...) method which takes a by-name argument, since within a companion object
the given code block will not retain a reference to its enclosing scope:
public class DemoActor extends AbstractActor {
/**
* Create Props for an actor of this type.
* @param magicNumber The magic number to be passed to this actors constructor.
* @return a Props for creating this actor, which can then be further configured
(e.g. calling .withDispatcher() on it)
*
*/
static Props props(Integer magicNumber) {
// You need to specify the actual type of the returned actor
// since Java 8 lambdas have some runtime type information erased
return Props.create(DemoActor.class, () -> new DemoActor(magicNumber));
}
private final Integer magicNumber;
DemoActor(Integer magicNumber) {
this.magicNumber = magicNumber;
receive(ReceiveBuilder.
match(Integer.class, i -> {
sender().tell(i + magicNumber, self());
}).build()
);
}
}
public class SomeOtherActor extends AbstractActor {
// Props(new DemoActor(42)) would not be safe
ActorRef demoActor = context().actorOf(DemoActor.props(42), "demo");
// ...
}
Using the ActorSystem will create top-level actors, supervised by the actor systems provided guardian actor,
while using an actors context will create a child actor.
public class FirstActor extends AbstractActor {
final ActorRef child = context().actorOf(Props.create(MyActor.class), "myChild");
// plus some behavior ...
}
369
It is recommended to create a hierarchy of children, grand-children and so on such that it fits the logical failurehandling structure of the application, see Actor Systems.
The call to actorOf returns an instance of ActorRef. This is a handle to the actor instance and the only way to
interact with it. The ActorRef is immutable and has a one to one relationship with the Actor it represents. The
ActorRef is also serializable and network-aware. This means that you can serialize it, send it over the wire and
use it on a remote host and it will still be representing the same Actor on the original node, across the network.
The name parameter is optional, but you should preferably name your actors, since that is used in log messages
and for identifying actors. The name must not be empty or start with $, but it may contain URL encoded characters (eg. %20 for a blank space). If the given name is already in use by another child to the same parent an
InvalidActorNameException is thrown.
Actors are automatically started asynchronously when created.
Dependency Injection
If your UntypedActor has a constructor that takes parameters then those need to be part of the Props as well, as
described above. But there are cases when a factory method must be used, for example when the actual constructor
arguments are determined by a dependency injection framework.
import akka.actor.Actor;
import akka.actor.IndirectActorProducer;
class DependencyInjector implements IndirectActorProducer {
final Object applicationContext;
final String beanName;
public DependencyInjector(Object applicationContext, String beanName) {
this.applicationContext = applicationContext;
this.beanName = beanName;
}
@Override
public Class<? extends Actor> actorClass() {
return MyActor.class;
}
@Override
public MyActor produce() {
MyActor result;
// obtain fresh Actor instance from DI framework ...
return result;
}
}
final ActorRef myActor = getContext().actorOf(
Props.create(DependencyInjector.class, applicationContext, "MyActor"),
"myactor3");
Warning: You might be tempted at times to offer an IndirectActorProducer which always returns
the same instance, e.g. by using a static field. This is not supported, as it goes against the meaning of an actor
restart, which is described here: What Restarting Means.
When using a dependency injection framework, actor beans MUST NOT have singleton scope.
Techniques for dependency injection and integration with dependency injection frameworks are described in more
depth in the Using Akka with Dependency Injection guideline and the Akka Java Spring tutorial in Typesafe
Activator.
370
The Inbox
When writing code outside of actors which shall communicate with actors, the ask pattern can be a solution (see
below), but there are two thing it cannot do: receiving multiple replies (e.g. by subscribing an ActorRef to a
notification service) and watching other actors lifecycle. For these purposes there is the Inbox class:
final Inbox inbox = Inbox.create(system);
inbox.send(target, "hello");
assert inbox.receive(Duration.create(1, TimeUnit.SECONDS)).equals("world");
The send method wraps a normal tell and supplies the internal actors reference as the sender. This allows the
reply to be received on the last line. Watching an actor is quite simple as well:
final Inbox inbox = Inbox.create(system);
inbox.watch(target);
target.tell(PoisonPill.getInstance(), ActorRef.noSender());
assert inbox.receive(Duration.create(1, TimeUnit.SECONDS)) instanceof Terminated;
371
}
postStop();
}
public void postRestart(Throwable reason) {
preStart();
}
public void postStop() {
}
The implementations shown above are the defaults provided by the AbstractActor class.
Actor Lifecycle
A path in an actor system represents a place which might be occupied by a living actor. Initially (apart from
system initialized actors) a path is empty. When actorOf() is called it assigns an incarnation of the actor
described by the passed Props to the given path. An actor incarnation is identified by the path and a UID. A
restart only swaps the Actor instance defined by the Props but the incarnation and hence the UID remains the
same.
The lifecycle of an incarnation ends when the actor is stopped. At that point the appropriate lifecycle events are
called and watching actors are notified of the termination. After the incarnation is stopped, the path can be reused
again by creating an actor with actorOf(). In this case the name of the new incarnation will be the same as the
previous one but the UIDs will differ.
An ActorRef always represents an incarnation (path and UID) not just a given path. Therefore if an actor is
stopped and a new one with the same name is created an ActorRef of the old incarnation will not point to the
new one.
372
ActorSelection on the other hand points to the path (or multiple paths if wildcards are used) and is completely
oblivious to which incarnation is currently occupying it. ActorSelection cannot be watched for this reason.
It is possible to resolve the current incarnations ActorRef living under the path by sending an Identify
message to the ActorSelection which will be replied to with an ActorIdentity containing the correct
reference (see Identifying Actors via Actor Selection). This can also be done with the resolveOne method of
the ActorSelection, which returns a Future of the matching ActorRef.
Lifecycle Monitoring aka DeathWatch
In order to be notified when another actor terminates (i.e. stops permanently, not temporary failure and restart), an
actor may register itself for reception of the Terminated message dispatched by the other actor upon termination
(see Stopping Actors). This service is provided by the DeathWatch component of the actor system.
Registering a monitor is easy:
public class WatchActor extends AbstractActor {
private final ActorRef child = context().actorOf(Props.empty(), "target");
private ActorRef lastSender = system.deadLetters();
public WatchActor() {
context().watch(child); // <-- this is the only call needed for registration
receive(ReceiveBuilder.
matchEquals("kill", s -> {
context().stop(child);
lastSender = sender();
}).
match(Terminated.class, t -> t.actor().equals(child), t -> {
lastSender.tell("finished", self());
}).build()
);
}
}
It should be noted that the Terminated message is generated independent of the order in which registration and
termination occur. In particular, the watching actor will receive a Terminated message even if the watched
actor has already been terminated at the time of registration.
Registering multiple times does not necessarily lead to multiple messages being generated, but there is no guarantee that only exactly one such message is received: if termination of the watched actor has generated and queued
the message, and another registration is done before this message has been processed, then a second message will
be queued, because registering for monitoring of an already terminated actor leads to the immediate generation of
the Terminated message.
It is also possible to deregister from watching another actors liveliness using context.unwatch(target).
This works even if the Terminated message has already been enqueued in the mailbox; after calling unwatch
no Terminated message for that actor will be processed anymore.
Start Hook
Right after starting the actor, its preStart method is invoked.
@Override
public void preStart() {
target = context().actorOf(Props.create(MyActor.class, "target"));
}
This method is called when the actor is first created. During restarts it is called by the default implementation of
postRestart, which means that by overriding that method you can choose whether the initialization code in
this method is called only exactly once for this actor or for every restart. Initialization code which is part of the
373
actors constructor will always be called when an instance of the actor class is created, which happens at every
restart.
Restart Hooks
All actors are supervised, i.e. linked to another actor with a fault handling strategy. Actors may be restarted in
case an exception is thrown while processing a message (see Supervision and Monitoring). This restart involves
the hooks mentioned above:
1. The old actor is informed by calling preRestart with the exception which caused the restart and the
message which triggered that exception; the latter may be None if the restart was not caused by processing
a message, e.g. when a supervisor does not trap the exception and is restarted in turn by its supervisor, or if
an actor is restarted due to a siblings failure. If the message is available, then that messages sender is also
accessible in the usual way (i.e. by calling sender).
This method is the best place for cleaning up, preparing hand-over to the fresh actor instance, etc. By default
it stops all children and calls postStop.
2. The initial factory from the actorOf call is used to produce the fresh instance.
3. The new actors postRestart method is invoked with the exception which caused the restart. By default
the preStart is called, just as in the normal start-up case.
An actor restart replaces only the actual actor object; the contents of the mailbox is unaffected by the restart,
so processing of messages will resume after the postRestart hook returns. The message that triggered the
exception will not be received again. Any message sent to an actor while it is being restarted will be queued to its
mailbox as usual.
Warning: Be aware that the ordering of failure notifications relative to user messages is not deterministic. In
particular, a parent might restart its child before it has processed the last messages sent by the child before the
failure. See Discussion: Message Ordering for details.
Stop Hook
After stopping an actor, its postStop hook is called, which may be used e.g. for deregistering this actor from
other services. This hook is guaranteed to run after message queuing has been disabled for this actor, i.e. messages
sent to a stopped actor will be redirected to the deadLetters of the ActorSystem.
The supplied path is parsed as a java.net.URI, which basically means that it is split on / into path elements.
If the path starts with /, it is absolute and the look-up starts at the root guardian (which is the parent of "/user");
otherwise it starts at the current actor. If a path element equals .., the look-up will take a step up towards the
supervisor of the currently traversed actor, otherwise it will step down to the named child. It should be noted
that the .. in actor paths here always means the logical structure, i.e. the supervisor.
374
The path elements of an actor selection may contain wildcard patterns allowing for broadcasting of messages to
that section:
// will look all children to serviceB with names starting with worker
context().actorSelection("/user/serviceB/worker*");
// will look up all siblings beneath same supervisor
context().actorSelection("../*");
Messages can be sent via the ActorSelection and the path of the ActorSelection is looked up when
delivering each message. If the selection does not match any actors the message will be dropped.
To acquire an ActorRef for an ActorSelection you need to send a message to the selection and use the
sender() reference of the reply from the actor. There is a built-in Identify message that all Actors will
understand and automatically reply to with a ActorIdentity message containing the ActorRef. This message is handled specially by the actors which are traversed in the sense that if a concrete name lookup fails (i.e.
a non-wildcard path element does not correspond to a live actor) then a negative result is generated. Please note
that this does not mean that delivery of that reply is guaranteed, it still is a normal message.
import akka.actor.ActorIdentity;
import akka.actor.ActorSelection;
import akka.actor.Identify;
public class Follower extends AbstractActor {
final Integer identifyId = 1;
public Follower(){
ActorSelection selection = context().actorSelection("/user/another");
selection.tell(new Identify(identifyId), self());
receive(ReceiveBuilder.
match(ActorIdentity.class, id -> id.getRef() != null, id -> {
ActorRef ref = id.getRef();
context().watch(ref);
context().become(active(ref));
}).
match(ActorIdentity.class, id -> id.getRef() == null, id -> {
context().stop(self());
}).build()
);
}
final PartialFunction<Object, BoxedUnit> active(final ActorRef another) {
return ReceiveBuilder.
match(Terminated.class, t -> t.actor().equals(another), t -> {
context().stop(self());
}).build();
}
}
You can also acquire an ActorRef for an ActorSelection with the resolveOne method of the
ActorSelection. It returns a Future of the matching ActorRef if such an actor exists. It is completed
with failure [[akka.actor.ActorNotFound]] if no such actor exists or the identification didnt complete within the
supplied timeout.
Remote actor addresses may also be looked up, if remoting is enabled:
context().actorSelection("akka.tcp://app@otherhost:1234/user/serviceB");
375
will be true even if an actor with that exact path is created after acquiring the actor reference. For remote actor
references acquired with actorFor the behaviour is different and sending messages to such a reference will under
the hood look up the actor by path on the remote system for every message send.
The sender reference is passed along with the message and available within the receiving actor via its sender
method while processing this message. Inside of an actor it is usually self who shall be the sender, but there
can be cases where replies shall be routed to some other actore.g. the parentin which the second argument to
376
tell would be a different one. Outside of an actor and if no reply is needed the second argument can be null;
if a reply is needed outside of an actor you can use the ask-pattern described next..
Ask: Send-And-Receive-Future
The ask pattern involves actors as well as futures, hence it is offered as a use pattern rather than a method on
ActorRef:
import
import
import
import
import
import
import
static akka.pattern.Patterns.ask;
static akka.pattern.Patterns.pipe;
scala.concurrent.Future;
scala.concurrent.duration.Duration;
akka.dispatch.Futures;
akka.dispatch.Mapper;
akka.util.Timeout;
This example demonstrates ask together with the pipe pattern on futures, because this is likely to be a common
combination. Please note that all of the above is completely non-blocking and asynchronous: ask produces a
Future, two of which are composed into a new future using the Futures.sequence and map methods and
then pipe installs an onComplete-handler on the future to effect the submission of the aggregated Result to
another actor.
Using ask will send a message to the receiving Actor as with tell, and the receiving actor must reply with
sender().tell(reply, self()) in order to complete the returned Future with a value. The ask
operation involves creating an internal actor for handling this reply, which needs to have a timeout after which it
is destroyed in order not to leak resources; see more below.
Warning: To complete the future with an exception you need send a Failure message to the sender. This is
not done automatically when an actor throws an exception while processing a message.
try {
String result = operation();
sender().tell(result, self());
} catch (Exception e) {
sender().tell(new akka.actor.Status.Failure(e), self());
throw e;
}
377
If the actor does not complete the future, it will expire after the timeout period, specified as parameter to the ask
method; this will complete the Future with an AskTimeoutException.
See Futures for more information on how to await or query a future.
The onComplete, onSuccess, or onFailure methods of the Future can be used to register a callback to
get a notification when the Future completes. Gives you a way to avoid blocking.
Warning: When using future callbacks, inside actors you need to carefully avoid closing over the containing
actors reference, i.e. do not call methods or access mutable state on the enclosing actor from within the callback. This would break the actor encapsulation and may introduce synchronization bugs and race conditions
because the callback will be scheduled concurrently to the enclosing actor. Unfortunately there is not yet a
way to detect these illegal accesses at compile time. See also: Actors and shared mutable state
Forward message
You can forward a message from one actor to another. This means that the original sender address/reference is
maintained even though the message is going through a mediator. This can be useful when writing actors that
work as routers, load-balancers, replicators etc.
target.forward(result, context());
Both the argument to the AbstractActor receive method and the return type of the Actor receive
method is a PartialFunction<Object, BoxedUnit> that defines which messages your Actor can handle, along with the implementation of how the messages should be processed.
Dont let the type signature scare you. To allow you to easily build up a partial function there is a builder named
ReceiveBuilder that you can use.
Here is an example:
import
import
import
import
akka.actor.AbstractActor;
akka.event.Logging;
akka.event.LoggingAdapter;
akka.japi.pf.ReceiveBuilder;
378
}
}
379
supervisor. If one of the actors does not respond (i.e. processing a message for extended periods of time and
therefore not receiving the stop command), this whole process will be stuck.
Upon ActorSystem.shutdown, the system guardian actors will be stopped, and the aforementioned process
will ensure proper termination of the whole system.
The postStop hook is invoked after an actor is fully stopped. This enables cleaning up of resources:
@Override
public void postStop() {
// clean up some resources ...
}
Note: Since stopping an actor is asynchronous, you cannot immediately reuse the name of the child you just
stopped; this will result in an InvalidActorNameException. Instead, watch the terminating actor and
create its replacement in response to the Terminated message which will eventually arrive.
PoisonPill
You can also send an actor the akka.actor.PoisonPill message, which will stop the actor when the message is processed. PoisonPill is enqueued as ordinary messages and will be handled after messages that were
already queued in the mailbox.
Graceful Stop
gracefulStop is useful if you need to wait for termination or compose ordered termination of several actors:
import
import
import
import
import
static akka.pattern.Patterns.gracefulStop;
scala.concurrent.Await;
scala.concurrent.Future;
scala.concurrent.duration.Duration;
akka.pattern.AskTimeoutException;
try {
Future<Boolean> stopped =
gracefulStop(actorRef, Duration.create(5, TimeUnit.SECONDS), Manager.SHUTDOWN);
Await.result(stopped, Duration.create(6, TimeUnit.SECONDS));
// the actor has been stopped
} catch (AskTimeoutException e) {
// the actor wasnt stopped within 5 seconds
}
public class Manager extends AbstractActor {
private static enum Shutdown {
Shutdown
}
public static final Shutdown SHUTDOWN = Shutdown.Shutdown;
private ActorRef worker =
context().watch(context().actorOf(Props.create(Cruncher.class), "worker"));
public Manager() {
receive(ReceiveBuilder.
matchEquals("job", s -> {
worker.tell("crunch", self());
}).
matchEquals(SHUTDOWN, x -> {
worker.tell(PoisonPill.getInstance(), self());
context().become(shuttingDown);
}).build()
380
);
}
public PartialFunction<Object, BoxedUnit> shuttingDown =
ReceiveBuilder.
matchEquals("job", s -> {
sender().tell("service unavailable, shutting down", self());
}).
match(Terminated.class, t -> t.actor().equals(worker), t -> {
context().stop(self());
}).build();
}
When gracefulStop() returns successfully, the actors postStop() hook will have been executed: there
exists a happens-before edge between the end of postStop() and the return of gracefulStop().
In the above example a custom Manager.Shutdown message is sent to the target actor to initiate the process
of stopping the actor. You can use PoisonPill for this, but then you have limited possibilities to perform
interactions with other actors before stopping the target actor. Simple cleanup tasks can be handled in postStop.
Warning: Keep in mind that an actor stopping and its name being deregistered are separate events which
happen asynchronously from each other. Therefore it may be that you will find the name still in use after
gracefulStop() returned. In order to guarantee proper deregistration, only reuse names from within a
supervisor you control and only in response to a Terminated message, i.e. not for top-level actors.
9.3.10 Become/Unbecome
Upgrade
Akka supports hotswapping the Actors message loop (e.g. its implementation) at runtime: invoke the
context.become method from within the Actor. become takes a PartialFunction<Object,
BoxedUnit> that implements the new message handler. The hotswapped code is kept in a Stack which can
be pushed and popped.
Warning: Please note that the actor will revert to its original behavior when restarted by its Supervisor.
To hotswap the Actor behavior using become:
public class HotSwapActor extends AbstractActor {
private PartialFunction<Object, BoxedUnit> angry;
private PartialFunction<Object, BoxedUnit> happy;
public HotSwapActor() {
angry =
ReceiveBuilder.
matchEquals("foo", s -> {
sender().tell("I am already angry?", self());
}).
matchEquals("bar", s -> {
context().become(happy);
}).build();
happy = ReceiveBuilder.
matchEquals("bar", s -> {
sender().tell("I am already happy :-)", self());
}).
matchEquals("foo", s -> {
context().become(angry);
}).build();
381
receive(ReceiveBuilder.
matchEquals("foo", s -> {
context().become(angry);
}).
matchEquals("bar", s -> {
context().become(happy);
}).build()
);
}
}
This variant of the become method is useful for many different things, such as to implement a Finite State
Machine (FSM, for an example see Dining Hakkers). It will replace the current behavior (i.e. the top of the
behavior stack), which means that you do not use unbecome, instead always the next behavior is explicitly
installed.
The other way of using become does not replace but add to the top of the behavior stack. In this case care must
be taken to ensure that the number of pop operations (i.e. unbecome) matches the number of push ones in
the long run, otherwise this amounts to a memory leak (which is why this behavior is not the default).
public class Swapper extends AbstractLoggingActor {
public Swapper() {
receive(ReceiveBuilder.
matchEquals(Swap, s -> {
log().info("Hi");
context().become(ReceiveBuilder.
matchEquals(Swap, x -> {
log().info("Ho");
context().unbecome(); // resets the latest become (just for fun)
}).build(), false); // push on top instead of replace
}).build()
);
}
}
public class SwapperApp {
public static void main(String[] args) {
ActorSystem system = ActorSystem.create("SwapperSystem");
ActorRef swapper = system.actorOf(Props.create(Swapper.class), "swapper");
swapper.tell(Swap, ActorRef.noSender()); // logs Hi
swapper.tell(Swap, ActorRef.noSender()); // logs Ho
swapper.tell(Swap, ActorRef.noSender()); // logs Hi
swapper.tell(Swap, ActorRef.noSender()); // logs Ho
swapper.tell(Swap, ActorRef.noSender()); // logs Hi
swapper.tell(Swap, ActorRef.noSender()); // logs Ho
system.shutdown();
}
}
9.3.11 Stash
The AbstractActorWithStash class enables an actor to temporarily stash away messages that can not
or should not be handled using the actors current behavior. Upon changing the actors message handler, i.e.,
right before invoking context().become() or context().unbecome(), all stashed messages can be
unstashed, thereby prepending them to the actors mailbox. This way, the stashed messages can be processed in
the same order as they have been received originally. An actor that extends AbstractActorWithStash will
automatically get a deque-based mailbox.
Note:
The abstract class AbstractActorWithStash implements the marker interface
RequiresMessageQueue<DequeBasedMessageQueueSemantics> which requests the system
382
to automatically choose a deque based mailbox implementation for the actor. If you want more control over the
mailbox, see the documentation on mailboxes: Mailboxes.
Here is an example of the AbstractActorWithStash class in action:
public class ActorWithProtocol extends AbstractActorWithStash {
public ActorWithProtocol() {
receive(ReceiveBuilder.
matchEquals("open", s -> {
context().become(ReceiveBuilder.
matchEquals("write", ws -> { /* do writing */ }).
matchEquals("close", cs -> {
unstashAll();
context().unbecome();
}).
matchAny(msg -> stash()).build(), false);
}).
matchAny(msg -> stash()).build()
);
}
}
Invoking stash() adds the current message (the message that the actor received last) to the actors stash.
It is typically invoked when handling the default case in the actors message handler to stash messages that
arent handled by the other cases. It is illegal to stash the same message twice; to do so results in an
IllegalStateException being thrown. The stash may also be bounded in which case invoking stash()
may lead to a capacity violation, which results in a StashOverflowException. The capacity of the stash
can be configured using the stash-capacity setting (an Int) of the mailboxs configuration.
Invoking unstashAll() enqueues messages from the stash to the actors mailbox until the capacity of the mailbox (if any) has been reached (note that messages from the stash are prepended to the mailbox). In case a bounded
mailbox overflows, a MessageQueueAppendFailedException is thrown. The stash is guaranteed to be
empty after calling unstashAll().
The stash is backed by a scala.collection.immutable.Vector. As a result, even a very large number
of messages may be stashed without a major impact on performance.
Note that the stash is part of the ephemeral actor state, unlike the mailbox. Therefore, it should be managed like
other parts of the actors state which have the same property. The AbstractActorWithStash implementation of preRestart will call unstashAll(), which is usually the desired behavior.
Note: If you want to enforce that your actor can only work with an unbounded stash, then you should use the
AbstractActorWithUnboundedStash class instead.
383
384
@Override
public void postRestart(Throwable reason) {
}
// The default implementation of preRestart() stops all the children
// of the actor. To opt-out from stopping the children, we
// have to override preRestart()
@Override
public void preRestart(Throwable reason, Option<Object> message)
throws Exception {
// Keep the call to postStop(), but no stopping of children
postStop();
}
Please note, that the child actors are still restarted, but no new ActorRef is created. One can recursively apply
the same principles for the children, ensuring that their preStart() method is called only at the creation of
their refs.
For more information see What Restarting Means.
Initialization via message passing
There are cases when it is impossible to pass all the information needed for actor initialization in the constructor,
for example in the presence of circular dependencies. In this case the actor should listen for an initialization
message, and use become() or a finite state-machine state transition to encode the initialized and uninitialized
states of the actor.
receive(ReceiveBuilder.
matchEquals("init", m1 -> {
initializeMe = "Up and running";
context().become(ReceiveBuilder.
matchEquals("U OK?", m2 -> {
sender().tell(initializeMe, self());
}).build());
}).build()
If the actor may receive messages before it has been initialized, a useful tool can be the Stash to save messages
until the initialization finishes, and replaying them after the actor became initialized.
Warning: This pattern should be used with care, and applied only when none of the patterns above are
applicable. One of the potential issues is that messages might be lost when sent to remote actors. Also,
publishing an ActorRef in an uninitialized state might lead to the condition that it receives a user message
before the initialization has been done.
385
akka.actor.AbstractFSM;
akka.actor.ActorRef;
akka.japi.pf.UnitMatch;
java.util.Arrays;
java.util.LinkedList;
java.util.List;
scala.concurrent.duration.Duration;
The contract of our Buncher actor is that it accepts or produces the following messages:
public final class SetTarget {
private final ActorRef ref;
public SetTarget(ActorRef ref) {
this.ref = ref;
}
public ActorRef getRef() {
return ref;
}
// boilerplate ...
}
public final class Queue {
private final Object obj;
public Queue(Object obj) {
this.obj = obj;
}
public Object getObj() {
return obj;
}
386
// boilerplate ...
}
public final class Batch {
private final List<Object> list;
public Batch(List<Object> list) {
this.list = list;
}
public List<Object> getList() {
return list;
}
// boilerplate ...
}
public enum Flush {
Flush
}
SetTarget is needed for starting it up, setting the destination for the Batches to be passed on; Queue will
add to the internal queue while Flush will mark the end of a burst.
The actor can be in two states: no message queued (aka Idle) or some message queued (aka Active). The
states and the state data is defined like this:
// states
enum State {
Idle, Active
}
// state data
interface Data {
}
enum Uninitialized implements Data {
Uninitialized
}
final class Todo implements Data {
private final ActorRef target;
private final List<Object> queue;
public Todo(ActorRef target, List<Object> queue) {
this.target = target;
this.queue = queue;
}
public ActorRef getTarget() {
return target;
}
public List<Object> getQueue() {
return queue;
}
// boilerplate ...
}
The actor starts out in the idle state. Once a message arrives it will go to the active state and stay there as long as
messages keep arriving and no flush is requested. The internal state data of the actor is made up of the target actor
reference to send the batches to and the actual queue of messages.
Now lets take a look at the skeleton for our FSM actor:
387
The basic strategy is to declare the actor, by inheriting the AbstractFSM class and specifying the possible states
and data values as type parameters. Within the body of the actor a DSL is used for declaring the state machine:
startWith defines the initial state and initial data
then there is one when(<state>) { ... } declaration per state to be handled (could potentially be
multiple ones, the passed PartialFunction will be concatenated using orElse)
finally starting it up using initialize, which performs the transition into the initial state and sets up
timers (if required).
In this case, we start out in the Idle and Uninitialized state, where only the SetTarget() message
is handled; stay prepares to end this events processing for not leaving the current state, while the using
modifier makes the FSM replace the internal state (which is Uninitialized at this point) with a fresh Todo()
object containing the target actor reference. The Active state has a state timeout declared, which means that
if no message is received for 1 second, a FSM.StateTimeout message will be generated. This has the same
effect as receiving the Flush command in this case, namely to transition back into the Idle state and resetting
the internal queue to the empty vector. But how do messages get queued? Since this shall work identically in
both states, we make use of the fact that any event which is not handled by the when() block is passed to the
whenUnhandled() block:
whenUnhandled(
matchEvent(Queue.class, Todo.class,
(queue, todo) -> goTo(Active).using(todo.addElement(queue.getObj()))).
anyEvent((event, state) -> {
log().warning("received unhandled request {} in state {}/{}",
event, stateName(), state);
return stay();
}));
The first case handled here is adding Queue() requests to the internal queue and going to the Active state
(this does the obvious thing of staying in the Active state if already there), but only if the FSM data are not
Uninitialized when the Queue() event is received. Otherwiseand in all other non-handled casesthe
second case just logs a warning and does not change the internal state.
The only missing piece is where the Batches are actually sent to the target, for which we use the
onTransition mechanism: you can declare multiple such blocks and all of them will be tried for matching
behavior in case a state transition occurs (i.e. only when the state actually changes).
onTransition(
matchState(Active, Idle, () -> {
// reuse this matcher
final UnitMatch<Data> m = UnitMatch.create(
388
matchData(Todo.class,
todo -> todo.getTarget().tell(new Batch(todo.getQueue()), self())));
m.match(stateData());
}).
state(Idle, Active, () -> {/* Do something here */}));
The transition callback is a partial function which takes as input a pair of statesthe current and the next state.
During the state change, the old state data is available via stateData as shown, and the new state data would
be available as nextStateData.
To verify that this buncher actually works, it is quite easy to write a test using the akka-testkit, here using JUnit as
an example:
public class BuncherTest {
static ActorSystem system;
@BeforeClass
public static void setup() {
system = ActorSystem.create("BuncherTest");
}
@AfterClass
public static void tearDown() {
JavaTestKit.shutdownActorSystem(system);
system = null;
}
@Test
public void testBuncherActorBatchesCorrectly() {
new JavaTestKit(system) {{
final ActorRef buncher =
system.actorOf(Props.create(Buncher.class));
final ActorRef probe = getRef();
buncher.tell(new SetTarget(probe), probe);
buncher.tell(new Queue(42), probe);
buncher.tell(new Queue(43), probe);
LinkedList<Object> list1 = new LinkedList<>();
list1.add(42);
list1.add(43);
expectMsgEquals(new Batch(list1));
buncher.tell(new Queue(44), probe);
buncher.tell(Flush, probe);
buncher.tell(new Queue(45), probe);
LinkedList<Object> list2 = new LinkedList<>();
list2.add(44);
expectMsgEquals(new Batch(list2));
LinkedList<Object> list3 = new LinkedList<>();
list3.add(45);
expectMsgEquals(new Batch(list3));
system.stop(buncher);
}};
}
@Test
public void testBuncherActorDoesntBatchUninitialized() {
new JavaTestKit(system) {{
final ActorRef buncher =
system.actorOf(Props.create(Buncher.class));
final ActorRef probe = getRef();
buncher.tell(new Queue(42), probe);
389
expectNoMsg();
system.stop(buncher);
}};
}
}
9.4.3 Reference
The AbstractFSM Class
The AbstractFSM abstract class is the base class used to implement an FSM. It implements Actor since an
Actor is created to drive the FSM.
public class Buncher extends AbstractFSM<State, Data> {
{
// fsm body ...
}
}
Note: The AbstractFSM class defines a receive method which handles internal messages and passes everything
else through to the FSM logic (according to the current state). When overriding the receive method, keep in
mind that e.g. state timeout handling depends on actually passing the messages through the FSM logic.
The AbstractFSM class takes two type parameters:
1. the supertype of all state names, usually an enum,
2. the type of the state data which are tracked by the AbstractFSM module itself.
Note: The state data together with the state name describe the internal state of the state machine; if you stick to
this scheme and do not add mutable fields to the FSM class you have the advantage of making all changes of the
internal state explicit in a few well-known places.
Defining States
A state is defined by one or more invocations of the method
when(<name>[, stateTimeout = <timeout>])(stateFunction).
The given name must be an object which is type-compatible with the first type parameter given to the
AbstractFSM class. This object is used as a hash key, so you must ensure that it properly implements equals
and hashCode; in particular it must not be mutable. The easiest fit for these requirements are case objects.
If the stateTimeout parameter is given, then all transitions into this state, including staying, receive this timeout by default. Initiating the transition with an explicit timeout may be used to override this default, see Initiating
Transitions for more information. The state timeout of any state may be changed during action processing with
setStateTimeout(state, duration). This enables runtime configuration e.g. via external message.
The stateFunction argument is a PartialFunction[Event, State], which is conveniently given
using the state function builder syntax as demonstrated below:
when(Idle,
matchEvent(SetTarget.class, Uninitialized.class,
(setTarget, uninitialized) ->
stay().using(new Todo(setTarget.getRef(), new LinkedList<>()))));
Warning: It is required that you define handlers for each of the possible FSM states, otherwise there will be
failures when trying to switch to undeclared states.
390
It is recommended practice to declare the states as an enum and then verify that there is a when clause for each of
the states. If you want to leave the handling of a state unhandled (more below), it still needs to be declared like
this:
when(SomeState, AbstractFSM.NullFunction());
Within this handler the state of the FSM may be queried using the stateName method.
IMPORTANT: This handler is not stacked, meaning that each invocation of whenUnhandled replaces the
previously installed handler.
Initiating Transitions
The result of any stateFunction must be a definition of the next state unless terminating the FSM, which is
described in Termination from Inside. The state definition can either be the current state, as described by the stay
directive, or it is a different state as given by goto(state). The resulting object allows further qualification by
way of the modifiers described in the following:
forMax(duration)
This modifier sets a state timeout on the next state. This means that a timer is started which upon expiry
sends a StateTimeout message to the FSM. This timer is canceled upon reception of any other message
in the meantime; you can rely on the fact that the StateTimeout message will not be processed after an
intervening message.
This modifier can also be used to override any default timeout which is specified for the target state. If you
want to cancel the default timeout, use Duration.Inf.
using(data)
This modifier replaces the old state data with the new data given. If you follow the advice above, this is the
only place where internal state data are ever modified.
replying(msg)
This modifier sends a reply to the currently processed message and otherwise does not modify the state
transition.
391
The parentheses are not actually needed in all cases, but they visually distinguish between modifiers and their
arguments and therefore make the code even more pleasant to read for foreigners.
Note: Please note that the return statement may not be used in when blocks or similar; this is a Scala
restriction. Either refactor your code using if () ... else ... or move it into a method definition.
Monitoring Transitions
Transitions occur between states conceptually, which means after any actions you have put into the event handling block; this is obvious since the next state is only defined by the value returned by the event handling logic.
You do not need to worry about the exact order with respect to setting the internal state variable, as everything
within the FSM actor is running single-threaded anyway.
Internal Monitoring
Up to this point, the FSM DSL has been centered on states and events. The dual view is to describe it as a series
of transitions. This is enabled by the method
onTransition(handler)
which associates actions with a transition instead of with a state and event. The handler is a partial function which
takes a pair of states as input; no resulting state is needed as it is not possible to modify the transition in progress.
onTransition(
matchState(Active, Idle, () -> setTimer("timeout",
Tick, Duration.create(1, SECONDS), true)).
state(Active, null, () -> cancelTimer("timeout")).
state(null, Idle, (f, t) -> log().info("entering Idle from " + f)));
It is also possible to pass a function object accepting two states to onTransition, in case your transition
handling logic is implemented as a method:
public void handler(StateType from, StateType to) {
// handle transition here
}
onTransition(this::handler);
The handlers registered with this method are stacked, so you can intersperse onTransition blocks with when
blocks as suits your design. It should be noted, however, that all handlers will be invoked for each transition,
not only the first matching one. This is designed specifically so you can put all transition handling for a certain
aspect into one place without having to worry about earlier declarations shadowing later ones; the actions are still
executed in declaration order, though.
Note: This kind of internal monitoring may be used to structure your FSM according to transitions, so that for
example the cancellation of a timer upon leaving a certain state cannot be forgot when adding new target states.
392
External Monitoring
You can use onTermination(handler) to specify custom code that is executed when the FSM is stopped.
The handler is a partial function which takes a StopEvent(reason, stateName, stateData) as argument:
onTermination(
matchStop(Normal(),
(state, data) -> {/* Do something here */}).
stop(Shutdown(),
(state, data) -> {/* Do something here */}).
393
stop(Failure.class,
(reason, state, data) -> {/* Do something here */}));
As for the whenUnhandled case, this handler is not stacked, so each invocation of onTermination replaces
the previously installed handler.
Termination from Outside
When an ActorRef associated to a FSM is stopped using the stop method, its postStop hook will be
executed. The default implementation by the AbstractFSM class is to execute the onTermination handler
if that is prepared to handle a StopEvent(Shutdown, ...).
Warning: In case you override postStop and want to have your onTermination handler called, do not
forget to call super.postStop.
394
The logDepth defaults to zero, which turns off the event log.
Warning: The log buffer is allocated during actor creation, which is why the configuration is done using a
virtual method call. If you want to override with a val, make sure that its initialization happens before the
initializer of LoggingFSM runs, and do not change the value returned by logDepth after the buffer has
been allocated.
The contents of the event log are available using method getLog, which returns an IndexedSeq[LogEntry]
where the oldest entry is at index zero.
9.4.5 Examples
A bigger FSM example contrasted with Actors become/unbecome can be found in the Typesafe Activator
template named Akka FSM in Scala
Another reason for marking a module as experimental is that its too early to tell if the module has a maintainer
that can take the responsibility of the module over time. These modules live in the akka-contrib subproject:
395
It is a lot more likely for an unintended message delivery failure to occur when a message send crosses JVM
boundaries, i.e. an intermediate unreliable network is involved. If someone unplugs an ethernet cable, or a power
failure shuts down a router, messages will be lost while the actors would be able to process them just fine.
Note: This does not mean that message send semantics are different between local and remote operations, it just
means that in practice there is a difference between how good the best effort is.
To bridge the disparity between local and remote sends is the goal of this pattern. When sending from A to
B must be as reliable as in-JVM, regardless of the deployment, then you can interject a reliable tunnel and send
through that instead. The tunnel consists of two end-points, where the ingress point P (the proxy) is a child of A
and the egress point E is a child of P, deployed onto the same network node where B lives. Messages sent to P will
be wrapped in an envelope, tagged with a sequence number and sent to E, who verifies that the received envelope
has the right sequence number (the next expected one) and forwards the contained message to B. When B receives
this message, the sender() will be a reference to the sender() of the original message to P. Reliability is added
by E replying to orderly received messages with an ACK, so that P can tick those messages off its resend list. If
ACKs do not come in a timely fashion, P will try to resend until successful.
Exactly what does it guarantee?
Sending via a ReliableProxy makes the message send exactly as reliable as if the represented target were to
live within the same JVM, provided that the remote actor system does not terminate. In effect, both ends (i.e. JVM
and actor system) must be considered as one when evaluating the reliability of this communication channel. The
benefit is that the network in-between is taken out of that equation.
Connecting to the target The proxy tries to connect to the target using the mechanism outlined in
actorSelection-scala. Once connected, if the tunnel terminates the proxy will optionally try to reconnect
to the target using using the same process.
Note that during the reconnection process there is a possibility that a message could be delivered to the target
more than once. Consider the case where a message is delivered to the target and the target system crashes
before the ACK is sent to the sender. After the proxy reconnects to the target it will start resending all of
the messages that it has not received an ACK for, and the message that it never got an ACK for will be redelivered.
Either this possibility should be considered in the design of the target or reconnection should be disabled.
396
How to use it
Since this implementation does not offer much in the way of configuration, simply instantiate a proxy wrapping a
target ActorPath. From Java it looks like this:
import akka.contrib.pattern.ReliableProxy;
Since the ReliableProxy actor is an fsm-scala, it also offers the capability to subscribe to state transitions.
If you need to know when all enqueued messages have been received by the remote end-point (and consequently
been forwarded to the target), you can subscribe to the FSM notifications and observe a transition from state
ReliableProxy.Active to state ReliableProxy.Idle.
public class ProxyTransitionParent extends UntypedActor {
private final ActorRef proxy;
private ActorRef client = null;
public ProxyTransitionParent(ActorPath targetPath) {
proxy = getContext().actorOf(
ReliableProxy.props(targetPath,
Duration.create(100, TimeUnit.MILLISECONDS)));
proxy.tell(new FSM.SubscribeTransitionCallBack(getSelf()), getSelf());
}
public void onReceive(Object msg) {
if ("hello".equals(msg)) {
proxy.tell("world!", getSelf());
client = getSender();
} else if (msg instanceof FSM.CurrentState<?>) {
// get initial state
} else if (msg instanceof FSM.Transition<?>) {
@SuppressWarnings("unchecked")
final FSM.Transition<ReliableProxy.State> transition =
(FSM.Transition<ReliableProxy.State>) msg;
assert transition.fsmRef().equals(proxy);
397
if (transition.from().equals(ReliableProxy.active()) &&
transition.to().equals(ReliableProxy.idle())) {
client.tell("done", getSelf());
}
}
}
}
Configuration
Set akka.reliable-proxy.debug to on to turn on extra debug logging for your ReliableProxy
actors.
akka.reliable-proxy.default-connect-interval is used only if you create a
ReliableProxy with no reconnections (that is, reconnectAfter == None). The default
value is the value of the configuration property akka.remote.retry-gate-closed-for. For
example, if akka.remote.retry-gate-closed-for is 5 s case the ReliableProxy will send
an Identify message to the target every 5 seconds to try to resolve the ActorPath to an ActorRef
so that messages can be sent to the target.
The Actor Contract
Message it Processes
FSM.SubscribeTransitionCallBack and FSM.UnsubscribeTransitionCallBack, see
fsm-scala
ReliableProxy.Unsent, see the API documentation for details.
any other message is transferred through the reliable tunnel and forwarded to the designated target actor
Messages it Sends
FSM.CurrentState and FSM.Transition, see fsm-scala
ReliableProxy.TargetChanged is sent to the FSM transition subscribers if the proxy reconnects to
a new target.
ReliableProxy.ProxyTerminated is sent to the FSM transition subscribers if the proxy is stopped.
398
Exceptions it Escalates
no specific exception types
any exception encountered by either the local or remote end-point are escalated (only fatal VM errors)
Arguments it Takes
target is the ActorPath to the actor to which the tunnel shall reliably deliver messages, B in the above
illustration.
retryAfter is the timeout for receiving ACK messages from the remote end-point; once it fires, all outstanding
message sends will be retried.
reconnectAfter is an optional interval between connection attempts. It is also used as the interval between
receiving a Terminated for the tunnel and attempting to reconnect to the target actor.
maxConnectAttempts is an optional maximum number of attempts to connect to the target while in the
Connecting state.
Throttling Actor Messages
Introduction
Suppose you are writing an application that makes HTTP requests to an external web service and that this web
service has a restriction in place: you may not make more than 10 requests in 1 minute. You will get blocked or
need to pay if you dont stay under this limit. In such a scenario you will want to employ a message throttler.
This extension module provides a simple implementation of a throttling actor, the TimerBasedThrottler.
How to use it
You can use a TimerBasedThrottler as follows. From Java it looks like this:
// A simple actor that prints whatever it receives
ActorRef printer = system.actorOf(Props.create(Printer.class));
// The throttler for this example, setting the rate
ActorRef throttler = system.actorOf(Props.create(TimerBasedThrottler.class,
new Throttler.Rate(3, Duration.create(1, TimeUnit.SECONDS))));
// Set the target
throttler.tell(new Throttler.SetTarget(printer), null);
// These three messages will be sent to the target immediately
throttler.tell("1", null);
throttler.tell("2", null);
throttler.tell("3", null);
// These two will wait until a second has passed
throttler.tell("4", null);
throttler.tell("5", null);
//A simple actor that prints whatever it receives
public class Printer extends UntypedActor {
@Override
public void onReceive(Object msg) {
System.out.println(msg);
}
}
399
case x println(x)
}
}
val printer = system.actorOf(Props[PrintActor])
// The throttler for this example, setting the rate
val throttler = system.actorOf(Props(classOf[TimerBasedThrottler],
3 msgsPer 1.second))
// Set the target
throttler ! SetTarget(Some(printer))
// These three messages will be sent to the target immediately
throttler ! "1"
throttler ! "2"
throttler ! "3"
// These two will wait until a second has passed
throttler ! "4"
throttler ! "5"
TimerBasedThrottler uses a timer internally. When the throttlers rate is 3 msg/s, for example, the throttler
will start a timer that triggers every second and each time will give the throttler exactly three vouchers; each
voucher gives the throttler a right to deliver a message. In this way, at most 3 messages will be sent out by the
throttler in each interval.
It should be noted that such timer-based throttlers provide relatively weak guarantees:
Only start times are taken into account. This may be a problem if, for example, the throttler is used to
throttle requests to an external web service. If a web request takes very long on the server then the rate
observed on the server may be higher.
A timer-based throttler only makes guarantees for the intervals of its own timer. In our example, no more
than 3 messages are delivered within such intervals. Other intervals on the timeline, however, may contain
more calls.
The two cases are illustrated in the two figures below, each showing a timeline and three intervals of the timer. The
message delivery times chosen by the throttler are indicated by dots, and as you can see, each interval contains
at most 3 point, so the throttler works correctly. Still, there is in each example an interval (the red one) that is
problematic. In the first scenario, this is because the delivery times are merely the start times of longer requests
(indicated by the four bars above the timeline that start at the dots), so that the server observes four requests during
the red interval. In the second scenario, the messages are centered around one of the points in time where the timer
triggers, causing the red interval to contain too many messages.
For some application scenarios, the guarantees provided by a timer-based throttler might be too weak. Charles
Cordingleys blog post discusses a throttler with stronger guarantees (it solves problem 2 from above). Future
versions of this module may feature throttlers with better guarantees.
400
401
Running this application (try it in the Akka sources by saying sbt akka-contrib/test:run) may produce
the following output (note the processing of World on lines 2 and 16):
Hello
World
[ERROR] [12/17/2012 16:28:36.581] [MySystem-peek-dispatcher-5] [akka://MySystem/user/myActor] DONT
java.lang.Exception: DONTWANNA
at akka.contrib.mailbox.MyActor.doStuff(PeekMailbox.scala:105)
at akka.contrib.mailbox.MyActor$$anonfun$receive$1.applyOrElse(PeekMailbox.scala:98)
at akka.actor.ActorCell.receiveMessage(ActorCell.scala:425)
at akka.actor.ActorCell.invoke(ActorCell.scala:386)
at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:230)
at akka.dispatch.Mailbox.run(Mailbox.scala:212)
at akka.dispatch.ForkJoinExecutorConfigurator$MailboxExecutionTask.exec(AbstractDispatcher.sc
at scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:262)
at scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:975)
at scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1478)
at scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:104)
World
Normally one would want to make processing idempotent (i.e. it does not matter if a message is processed twice)
or context.become a different behavior upon restart; the above example included the println(msg) call
just to demonstrate the re-processing.
Cluster Singleton
For some use cases it is convenient and sometimes also mandatory to ensure that you have exactly one actor of a
certain type running somewhere in the cluster.
Some examples:
single point of responsibility for certain cluster-wide consistent decisions, or coordination of actions across
the cluster system
single entry point to an external system
single master, many workers
centralized naming service, or routing logic
Using a singleton should not be the first design choice. It has several drawbacks, such as single-point of bottleneck.
Single-point of failure is also a relevant concern, but for some cases this feature takes care of that by making sure
that another singleton instance will eventually be started.
The cluster singleton pattern is implemented by akka.contrib.pattern.ClusterSingletonManager.
It manages one singleton actor instance among all cluster nodes or a group of nodes tagged with a specific role.
ClusterSingletonManager is an actor that is supposed to be started on all nodes, or all nodes with
specified role, in the cluster. The actual singleton actor is started by the ClusterSingletonManager on the
oldest node by creating a child actor from supplied Props. ClusterSingletonManager makes sure that at
most one singleton instance is running at any point in time.
The singleton actor is always running on the oldest member, which can be determined by
Member#isOlderThan. This can change when removing that member from the cluster. Be aware that
there is a short time period when there is no active singleton during the hand-over process.
The cluster failure detector will notice when oldest node becomes unreachable due to things like JVM crash, hard
shut down, or network failure. Then a new oldest node will take over and a new singleton actor is created. For
these failure scenarios there will not be a graceful hand-over, but more than one active singletons is prevented by
all reasonable means. Some corner cases are eventually resolved by configurable timeouts.
You can access the singleton actor by using the provided akka.contrib.pattern.ClusterSingletonProxy,
which will route all messages to the current instance of the singleton. The proxy will keep track of the oldest node
in the cluster and resolve the singletons ActorRef by explicitly sending the singletons actorSelection
the akka.actor.Identify message and waiting for it to reply. This is performed periodically if the
402
singleton doesnt reply within a certain (configurable) time. Given the implementation, there might be periods of
time during which the ActorRef is unavailable, e.g., when a node leaves the cluster. In these cases, the proxy
will stash away all messages until it is able to identify the singleton. Its worth noting that messages can always
be lost because of the distributed nature of these actors. As always, additional logic should be implemented in the
singleton (acknowledgement) and in the client (retry) actors to ensure at-least-once message delivery.
An Example
Assume that we need one single entry point to an external system. An actor that receives messages from a JMS
queue with the strict requirement that only one JMS consumer must exist to be make sure that the messages are
processed in order. That is perhaps not how one would like to design things, but a typical real-world scenario
when integrating with external systems.
On each node in the cluster you need to start the ClusterSingletonManager and supply the Props of the
singleton actor, in this case the JMS queue consumer.
In Scala:
system.actorOf(ClusterSingletonManager.props(
singletonProps = Props(classOf[Consumer], queue, testActor),
singletonName = "consumer",
terminationMessage = End,
role = Some("worker")),
name = "singleton")
Here we limit the singleton to nodes tagged with the "worker" role, but all nodes, independent of role, can be
used by specifying None as role parameter.
The
corresponding
Java
API
for
the
singeltonProps
function
is
akka.contrib.pattern.ClusterSingletonPropsFactory. The Java API takes a plain String for
the role parameter and null means that all nodes, independent of role, are used.
In Java:
Note: The singletonProps/singletonPropsFactory is invoked when creating the singleton actor and
it must not use members that are not thread safe, e.g. mutable state in enclosing actor.
Here we use an application specific terminationMessage to be able to close the resources before actually
stopping the singleton actor. Note that PoisonPill is a perfectly fine terminationMessage if you only
need to stop the actor.
Here is how the singleton actor handles the terminationMessage in this example.
case End
queue ! UnregisterConsumer
case UnregistrationOk
context stop self
case Ping
sender ! Pong
Note that you can send back current state to the ClusterSingletonManager before terminating. This message will be sent over to the ClusterSingletonManager at the new oldest node and it will be passed to the
singletonProps factory when creating the new singleton instance.
With the names given above, access to the singleton can be obtained from any cluster node using a properly
configured proxy.
In Scala:
403
system.actorOf(ClusterSingletonProxy.props(
singletonPath = "/user/singleton/consumer",
role = Some("worker")),
name = "consumerProxy")
In Java:
A more comprehensive sample is available in the Typesafe Activator tutorial named Distributed workers with
Akka and Scala! and Distributed workers with Akka and Java!.
Cluster Sharding
Cluster sharding is useful when you need to distribute actors across several nodes in the cluster and want to be
able to interact with them using their logical identifier, but without having to care about their physical location in
the cluster, which might also change over time.
It could for example be actors representing Aggregate Roots in Domain-Driven Design terminology. Here we call
these actors entries. These actors typically have persistent (durable) state, but this feature is not limited to actors
with persistent state.
Cluster sharding is typically used when you have many stateful actors that together consume more resources (e.g.
memory) than fit on one machine. If you only have a few stateful actors it might be easier to run them on a Cluster
Singleton node.
In this context sharding means that actors with an identifier, so called entries, can be automatically distributed
across multiple nodes in the cluster. Each entry actor runs only at one place, and messages can be sent to the
entry without requiring the sender() to know the location of the destination actor. This is achieved by sending the
messages via a ShardRegion actor provided by this extension, which knows how to route the message with the
entry id to the final destination.
An Example in Java
404
The above actor uses event sourcing and the support provided in UntypedEventsourcedProcessor to store
its state. It does not have to be a processor, but in case of failure or migration of entries between nodes it must be
able to recover its state if it is valuable.
When using the sharding extension you are first, typically at system startup on each node in the cluster, supposed
to register the supported entry types with the ClusterSharding.start method.
405
ClusterSharding.get(system).start("Counter", Props.create(Counter.class),
messageExtractor);
The messageExtractor defines application specific methods to extract the entry identifier and the shard identifier from incoming messages.
ShardRegion.MessageExtractor messageExtractor = new ShardRegion.MessageExtractor() {
@Override
public String entryId(Object message) {
if (message instanceof Counter.EntryEnvelope)
return String.valueOf(((Counter.EntryEnvelope) message).id);
else if (message instanceof Counter.Get)
return String.valueOf(((Counter.Get) message).counterId);
else
return null;
}
@Override
public Object entryMessage(Object message) {
if (message instanceof Counter.EntryEnvelope)
return ((Counter.EntryEnvelope) message).payload;
else
return message;
}
@Override
public String shardId(Object message) {
if (message instanceof Counter.EntryEnvelope) {
long id = ((Counter.EntryEnvelope) message).id;
return String.valueOf(id % 10);
} else if (message instanceof Counter.Get) {
long id = ((Counter.Get) message).counterId;
return String.valueOf(id % 10);
} else {
return null;
}
}
};
This example illustrates two different ways to define the entry identifier in the messages:
The Get message includes the identifier itself.
The EntryEnvelope holds the identifier, and the actual message that is sent to the entry actor is wrapped
in the envelope.
Note how these two messages types are handled in the entryId and entryMessage methods shown above.
A shard is a group of entries that will be managed together. The grouping is defined by the shardResolver
function shown above. Creating a good sharding algorithm is an interesting challenge in itself. Try to produce a
uniform distribution, i.e. same amount of entries in each shard. As a rule of thumb, the number of shards should
be a factor ten greater than the planned maximum number of cluster nodes.
Messages to the entries are always sent via the local ShardRegion. The ShardRegion actor for a named
entry type can be retrieved with ClusterSharding.shardRegion. The ShardRegion will lookup the
location of the shard for the entry if it does not already know its location. It will delegate the message to the right
node and it will create the entry actor on demand, i.e. when the first message for a specific entry is delivered.
ActorRef counterRegion = ClusterSharding.get(system).shardRegion("Counter");
counterRegion.tell(new Counter.Get(100), getSelf());
counterRegion.tell(new Counter.EntryEnvelope(100,
406
Counter.CounterOp.INCREMENT), getSelf());
counterRegion.tell(new Counter.Get(100), getSelf());
An Example in Scala
object Increment
object Decrement
class Get(counterId: Long)
class EntryEnvelope(id: Long, payload: Any)
The above actor uses event sourcing and the support provided in EventsourcedProcessor to store its state.
It does not have to be a processor, but in case of failure or migration of entries between nodes it must be able to
recover its state if it is valuable.
When using the sharding extension you are first, typically at system startup on each node in the cluster, supposed
to register the supported entry types with the ClusterSharding.start method.
ClusterSharding(system).start(
typeName = "Counter",
entryProps = Some(Props[Counter]),
idExtractor = idExtractor,
shardResolver = shardResolver)
The idExtractor and shardResolver are two application specific functions to extract the entry identifier
and the shard identifier from incoming messages.
val idExtractor: ShardRegion.IdExtractor = {
case EntryEnvelope(id, payload) (id.toString, payload)
case msg @ Get(id)
(id.toString, msg)
}
val shardResolver: ShardRegion.ShardResolver = msg msg match {
case EntryEnvelope(id, _) (id % 12).toString
407
case Get(id)
(id % 12).toString
This example illustrates two different ways to define the entry identifier in the messages:
The Get message includes the identifier itself.
The EntryEnvelope holds the identifier, and the actual message that is sent to the entry actor is wrapped
in the envelope.
Note how these two messages types are handled in the idExtractor function shown above.
A shard is a group of entries that will be managed together. The grouping is defined by the shardResolver
function shown above. Creating a good sharding algorithm is an interesting challenge in itself. Try to produce a
uniform distribution, i.e. same amount of entries in each shard. As a rule of thumb, the number of shards should
be a factor ten greater than the planned maximum number of cluster nodes.
Messages to the entries are always sent via the local ShardRegion. The ShardRegion actor for a named
entry type can be retrieved with ClusterSharding.shardRegion. The ShardRegion will lookup the
location of the shard for the entry if it does not already know its location. It will delegate the message to the right
node and it will create the entry actor on demand, i.e. when the first message for a specific entry is delivered.
val counterRegion: ActorRef = ClusterSharding(system).shardRegion("Counter")
counterRegion ! Get(100)
expectMsg(0)
counterRegion ! EntryEnvelope(100, Increment)
counterRegion ! Get(100)
expectMsg(1)
A more comprehensive sample is available in the Typesafe Activator tutorial named Akka Cluster Sharding with
Scala!.
How it works
The ShardRegion actor is started on each node in the cluster, or group of nodes tagged with a specific role.
The ShardRegion is created with two application specific functions to extract the entry identifier and the shard
identifier from incoming messages. A shard is a group of entries that will be managed together. For the first
message in a specific shard the ShardRegion request the location of the shard from a central coordinator, the
ShardCoordinator.
The ShardCoordinator decides which ShardRegion that owns the shard. The ShardRegion receives
the decided home of the shard and if that is the ShardRegion instance itself it will create a local child actor
representing the entry and direct all messages for that entry to it. If the shard home is another ShardRegion
instance messages will be forwarded to that ShardRegion instance instead. While resolving the location of a
shard incoming messages for that shard are buffered and later delivered when the shard home is known. Subsequent messages to the resolved shard can be delivered to the target destination immediately without involving the
ShardCoordinator.
Scenario 1:
1. Incoming message M1 to ShardRegion instance R1.
2. M1 is mapped to shard S1. R1 doesnt know about S1, so it asks the coordinator C for the location of S1.
3. C answers that the home of S1 is R1.
4. R1 creates child actor for the entry E1 and sends buffered messages for S1 to E1 child
5. All incoming messages for S1 which arrive at R1 can be handled by R1 without C. It creates entry children
as needed, and forwards messages to them.
Scenario 2:
1. Incoming message M2 to R1.
408
2. M2 is mapped to S2. R1 doesnt know about S2, so it asks C for the location of S2.
3. C answers that the home of S2 is R2.
4. R1 sends buffered messages for S2 to R2
5. All incoming messages for S2 which arrive at R1 can be handled by R1 without C. It forwards messages to
R2.
6. R2 receives message for S2, ask C, which answers that the home of S2 is R2, and we are in Scenario 1 (but
for R2).
To make sure that at most one instance of a specific entry actor is running somewhere in the cluster it is important
that all nodes have the same view of where the shards are located. Therefore the shard allocation decisions are
taken by the central ShardCoordinator, which is running as a cluster singleton, i.e. one instance on the oldest
member among all cluster nodes or a group of nodes tagged with a specific role.
The logic that decides where a shard is to be located is defined in a pluggable shard allocation strategy. The
default implementation ShardCoordinator.LeastShardAllocationStrategy allocates new shards
to the ShardRegion with least number of previously allocated shards. This strategy can be replaced by an
application specific implementation.
To be able to use newly added members in the cluster the coordinator facilitates rebalancing of shards, i.e. migrate
entries from one node to another. In the rebalance process the coordinator first notifies all ShardRegion actors
that a handoff for a shard has started. That means they will start buffering incoming messages for that shard,
in the same way as if the shard location is unknown. During the rebalance process the coordinator will not
answer any requests for the location of shards that are being rebalanced, i.e. local buffering will continue until
the handoff is completed. The ShardRegion responsible for the rebalanced shard will stop all entries in that
shard by sending PoisonPill to them. When all entries have been terminated the ShardRegion owning the
entries will acknowledge the handoff as completed to the coordinator. Thereafter the coordinator will reply to
requests for the location of the shard and thereby allocate a new home for the shard and then buffered messages
in the ShardRegion actors are delivered to the new location. This means that the state of the entries are not
transferred or migrated. If the state of the entries are of importance it should be persistent (durable), e.g. with
akka-persistence, so that it can be recovered at the new location.
The logic that decides which shards to rebalance is defined in a pluggable shard allocation strategy. The default implementation ShardCoordinator.LeastShardAllocationStrategy picks shards for handoff from the ShardRegion with most number of previously allocated shards. They will then be allocated to the
ShardRegion with least number of previously allocated shards, i.e. new members in the cluster. There is a
configurable threshold of how large the difference must be to begin the rebalancing. This strategy can be replaced
by an application specific implementation.
The state of shard locations in the ShardCoordinator is persistent (durable) with akka-persistence
to survive failures. Since it is running in a cluster akka-persistence must be configured with a distributed
journal. When a crashed or unreachable coordinator node has been removed (via down) from the cluster a new
ShardCoordinator singleton actor will take over and the state is recovered. During such a failure period
shards with known location are still available, while messages for new (unknown) shards are buffered until the
new ShardCoordinator becomes available.
As long as a sender() uses the same ShardRegion actor to deliver messages to an entry actor the order of the
messages is preserved. As long as the buffer limit is not reached messages are delivered on a best effort basis, with
at-most once delivery semantics, in the same way as ordinary message sending. Reliable end-to-end messaging,
with at-least-once semantics can be added by using channels in akka-persistence.
Some additional latency is introduced for messages targeted to new or previously unused shards due to the roundtrip to the coordinator. Rebalancing of shards may also add latency. This should be considered when designing
the application specific shard resolution, e.g. to avoid too fine grained shards.
Proxy Only Mode
The ShardRegion actor can also be started in proxy only mode, i.e. it will not host any entries itself, but
knows how to delegate messages to the right location. A ShardRegion starts in proxy only mode if the roles
409
of the node does not include the node role specified in akka.contrib.cluster.sharding.role config
property or if the specified entryProps is None / null.
Passivation
If the state of the entries are persistent you may stop entries that are not used to reduce memory consumption. This
is done by the application specific implementation of the entry actors for example by defining receive timeout
(context.setReceiveTimeout). If a message is already enqueued to the entry when it stops itself the
enqueued message in the mailbox will be dropped. To support graceful passivation without loosing such messages
the entry actor can send ShardRegion.Passivate to its parent ShardRegion. The specified wrapped
message in Passivate will be sent back to the entry, which is then supposed to stop itself. Incoming messages
will be buffered by the ShardRegion between reception of Passivate and termination of the entry. Such
buffered messages are thereafter delivered to a new incarnation of the entry.
Configuration
Custom
shard
allocation
strategy
can
be
defined
in
an
optional
parameter
to
ClusterSharding.start. See the API documentation of ShardAllocationStrategy (Scala)
or AbstractShardAllocationStrategy (Java) for details of how to implement a custom shard allocation
strategy.
410
A subscriber actor:
public class Subscriber extends UntypedActor {
LoggingAdapter log = Logging.getLogger(getContext().system(), this);
411
public Subscriber() {
ActorRef mediator =
DistributedPubSubExtension.get(getContext().system()).mediator();
// subscribe to the topic named "content"
mediator.tell(new DistributedPubSubMediator.Subscribe("content", getSelf()),
getSelf());
}
public void onReceive(Object msg) {
if (msg instanceof String)
log.info("Got: {}", msg);
else if (msg instanceof DistributedPubSubMediator.SubscribeAck)
log.info("subscribing");
else
unhandled(msg);
}
}
Subscriber actors can be started on several nodes in the cluster, and all will receive messages published to the
content topic.
system.actorOf(Props.create(Subscriber.class), "subscriber1");
//another node
system.actorOf(Props.create(Subscriber.class), "subscriber2");
system.actorOf(Props.create(Subscriber.class), "subscriber3");
A subscriber actor:
class Subscriber extends Actor with ActorLogging {
import DistributedPubSubMediator.{ Subscribe, SubscribeAck }
val mediator = DistributedPubSubExtension(context.system).mediator
// subscribe to the topic named "content"
mediator ! Subscribe("content", self)
412
def receive = {
case SubscribeAck(Subscribe("content", self))
context become ready
}
def ready: Actor.Receive = {
case s: String
log.info("Got {}", s)
}
}
Subscriber actors can be started on several nodes in the cluster, and all will receive messages published to the
content topic.
runOn(first) {
system.actorOf(Props[Subscriber], "subscriber1")
}
runOn(second) {
system.actorOf(Props[Subscriber], "subscriber2")
system.actorOf(Props[Subscriber], "subscriber3")
}
A more comprehensive sample is available in the Typesafe Activator tutorial named Akka Clustered PubSub with
Scala!.
DistributedPubSubExtension
In
the
example
above
the
mediator
is
started
and
accessed
with
the
akka.contrib.pattern.DistributedPubSubExtension.
That is convenient and perfectly
fine in most cases, but it can be good to know that it is possible to start the mediator actor as an ordinary actor and
you can have several different mediators at the same time to be able to divide a large number of actors/topics to
different mediators. For example you might want to use different cluster roles for different mediators.
The DistributedPubSubExtension can be configured with the following properties:
# Settings for the DistributedPubSubExtension
akka.contrib.cluster.pub-sub {
# Actor name of the mediator actor, /user/distributedPubSubMediator
name = distributedPubSubMediator
413
It is recommended to load the extension when the actor system is started by defining it in akka.extensions
configuration property. Otherwise it will be activated when first used and then it takes a while for it to be populated.
akka.extensions = ["akka.contrib.pattern.DistributedPubSubExtension"]
Cluster Client
An actor system that is not part of the cluster can communicate with actors somewhere in the cluster via this
ClusterClient. The client can of course be part of another cluster. It only needs to know the location of one
(or more) nodes to use as initial contact points. It will establish a connection to a ClusterReceptionist
somewhere in the cluster. It will monitor the connection to the receptionist and establish a new connection
if the link goes down. When looking for a new receptionist it uses fresh contact points retrieved from previous establishment, or periodically refreshed contacts, i.e. not necessarily the initial contact points. Also,
note its necessary to change akka.actor.provider from akka.actor.LocalActorRefProvider
to akka.remote.RemoteActorRefProvider or akka.cluster.ClusterActorRefProvider
when using the cluster client.
The receptionist is supposed to be started on all nodes, or all nodes with specified role, in the cluster. The
receptionist can be started with the ClusterReceptionistExtension or as an ordinary actor.
You can send messages via the ClusterClient to any actor in the cluster that is registered in the DistributedPubSubMediator used by the ClusterReceptionist.
The
ClusterReceptionistExtension provides methods for registration of actors that should be reachable from the client. Messages are wrapped in ClusterClient.Send, ClusterClient.SendToAll or
ClusterClient.Publish.
1. ClusterClient.Send
The message will be delivered to one recipient with a matching path, if any such exists. If several entries match
the path the message will be delivered to one random destination. The sender() of the message can specify that
local affinity is preferred, i.e. the message is sent to an actor in the same local actor system as the used receptionist
actor, if any such exists, otherwise random to any other matching entry.
2. ClusterClient.SendToAll
The message will be delivered to all recipients with a matching path.
3. ClusterClient.Publish
The message will be delivered to all recipients Actors that have been registered as subscribers to the named topic.
Response messages from the destination actor are tunneled via the receptionist to avoid inbound connections from
other cluster nodes to the client, i.e. the sender(), as seen by the destination actor, is not the client itself. The
9.5. External Contributions
414
sender() of the response messages, as seen by the client, is preserved as the original sender(), so the client can
choose to send subsequent messages directly to the actor in the cluster.
While establishing a connection to a receptionist the ClusterClient will buffer messages and send
them when the connection is established.
If the buffer is full the ClusterClient will throw
akka.actor.StashOverflowException, which can be handled in by the supervision strategy of the parent actor. The size of the buffer can be configured by the following stash-capacity setting of the mailbox
that is used by the ClusterClient actor.
akka.contrib.cluster.client {
mailbox {
mailbox-type = "akka.dispatch.UnboundedDequeBasedMailbox"
stash-capacity = 1000
}
}
An Example
On the cluster nodes first start the receptionist. Note, it is recommended to load the extension when the actor
system is started by defining it in the akka.extensions configuration property:
akka.extensions = ["akka.contrib.pattern.ClusterReceptionistExtension"]
Next, register the actors that should be available for the client.
runOn(host1) {
val serviceA = system.actorOf(Props[Service], "serviceA")
ClusterReceptionistExtension(system).registerService(serviceA)
}
runOn(host2, host3) {
val serviceB = system.actorOf(Props[Service], "serviceB")
ClusterReceptionistExtension(system).registerService(serviceB)
}
On the client you create the ClusterClient actor and use it as a gateway for sending messages to the actors
identified by their path (without address information) somewhere in the cluster.
runOn(client) {
val c = system.actorOf(ClusterClient.props(initialContacts))
c ! ClusterClient.Send("/user/serviceA", "hello", localAffinity = true)
c ! ClusterClient.SendToAll("/user/serviceB", "hi")
}
You will probably define the address information of the initial contact points in configuration or system property.
A more comprehensive sample is available in the Typesafe Activator tutorial named Distributed workers with
Akka and Scala! and Distributed workers with Akka and Java!.
ClusterReceptionistExtension
In
the
example
above
the
receptionist
is
started
and
accessed
with
the
akka.contrib.pattern.ClusterReceptionistExtension.
That is convenient and
perfectly fine in most cases, but it can be good to know that it is possible to start the
akka.contrib.pattern.ClusterReceptionist actor as an ordinary actor and you can have
several different receptionists at the same time, serving different types of clients.
9.5. External Contributions
415
Aggregator Pattern
The aggregator pattern supports writing actors that aggregate data from multiple other actors and updates its state
based on those responses. It is even harder to optionally aggregate more data based on the runtime state of the
actor or take certain actions (sending another message and get another response) based on two or more previous
responses.
A common thought is to use the ask pattern to request information from other actors. However, ask creates another
actor specifically for the ask. We cannot use a callback from the future to update the state as the thread executing
the callback is not defined. This will likely close-over the current actor.
The aggregator pattern solves such scenarios. It makes sure were acting from the same actor in the scope of the
actor receive.
Introduction
The aggregator pattern allows match patterns to be dynamically added to and removed from an actor from inside
the message handling logic. All match patterns are called from the receive loop and run in the thread handling
the incoming message. These dynamically added patterns and logic can safely read and/or modify this actors
mutable state without risking integrity or concurrency issues.
Usage
To use the aggregator pattern, you need to extend the Aggregator trait. The trait takes care of receive and
actors extending this trait should not override receive. The trait provides the expect, expectOnce, and
unexpect calls. The expect and expectOnce calls return a handle that can be used for later de-registration
by passing the handle to unexpect.
expect is often used for standing matches such as catching error messages or timeouts.
expect {
case TimedOut collectBalances(force = true)
}
416
expectOnce is used for matching the initial message as well as other requested messages
expectOnce {
case GetCustomerAccountBalances(id, types)
new AccountAggregator(sender(), id, types)
case _
sender() ! CantUnderstand
context.stop(self)
}
def fetchCheckingAccountsBalance() {
context.actorOf(Props[CheckingAccountProxy]) ! GetAccountBalances(id)
expectOnce {
case CheckingAccountBalances(balances)
results += (Checking -> balances)
collectBalances()
}
}
unexpect can be used for expecting multiple responses until a timeout or when the logic dictates such an
expect no longer applies.
val handle = expect {
case Response(name, value)
values += value
if (values.size > 3) processList()
case TimedOut processList()
}
def processList() {
unexpect(handle)
if (values.size > 0) {
context.actorSelection("/user/evaluator") ! values.toList
expectOnce {
case EvaluationResults(name, eval) processFinal(eval)
}
} else processFinal(List.empty[Int])
}
As the name eludes, expect keeps the partial function matching any received messages until unexpect is
called or the actor terminates, whichever comes first. On the other hand, expectOnce removes the partial
function once a match has been established.
It is a common pattern to register the initial expectOnce from the construction of the actor to accept the initial
message. Once that message is received, the actor starts doing all aggregations and sends the response back to the
original requester. The aggregator should terminate after the response is sent (or timed out). A different original
request should use a different actor instance.
As you can see, aggregator actors are generally stateful, short lived actors.
Sample Use Case - AccountBalanceRetriever
This example below shows a typical and intended use of the aggregator pattern.
import scala.collection._
import scala.concurrent.duration._
import scala.math.BigDecimal.int2bigDecimal
import akka.actor._
/**
* Sample and test code for the aggregator patter.
* This is based on Jamie Allens tutorial at
417
* http://jaxenter.com/tutorial-asynchronous-programming-with-akka-actors-46220.html
*/
sealed trait AccountType
case object Checking extends AccountType
case object Savings extends AccountType
case object MoneyMarket extends AccountType
case class GetCustomerAccountBalances(id: Long, accountTypes: Set[AccountType])
case class GetAccountBalances(id: Long)
case class AccountBalances(accountType: AccountType,
balance: Option[List[(Long, BigDecimal)]])
case class CheckingAccountBalances(balances: Option[List[(Long, BigDecimal)]])
case class SavingsAccountBalances(balances: Option[List[(Long, BigDecimal)]])
case class MoneyMarketAccountBalances(balances: Option[List[(Long, BigDecimal)]])
case object TimedOut
case object CantUnderstand
class SavingsAccountProxy extends Actor {
def receive = {
case GetAccountBalances(id: Long)
sender() ! SavingsAccountBalances(Some(List((1, 150000), (2, 29000))))
}
}
class CheckingAccountProxy extends Actor {
def receive = {
case GetAccountBalances(id: Long)
sender() ! CheckingAccountBalances(Some(List((3, 15000))))
}
}
class MoneyMarketAccountProxy extends Actor {
def receive = {
case GetAccountBalances(id: Long)
sender() ! MoneyMarketAccountBalances(None)
}
}
class AccountBalanceRetriever extends Actor with Aggregator {
import context._
expectOnce {
case GetCustomerAccountBalances(id, types)
new AccountAggregator(sender(), id, types)
case _
sender() ! CantUnderstand
context.stop(self)
}
class AccountAggregator(originalSender: ActorRef,
id: Long, types: Set[AccountType]) {
val results =
mutable.ArrayBuffer.empty[(AccountType, Option[List[(Long, BigDecimal)]])]
if (types.size > 0)
types foreach {
case Checking
fetchCheckingAccountsBalance()
case Savings
fetchSavingsAccountsBalance()
case MoneyMarket fetchMoneyMarketAccountsBalance()
418
}
else collectBalances() // Empty type list yields empty response
context.system.scheduler.scheduleOnce(1.second, self, TimedOut)
expect {
case TimedOut collectBalances(force = true)
}
def fetchCheckingAccountsBalance() {
context.actorOf(Props[CheckingAccountProxy]) ! GetAccountBalances(id)
expectOnce {
case CheckingAccountBalances(balances)
results += (Checking -> balances)
collectBalances()
}
}
def fetchSavingsAccountsBalance() {
context.actorOf(Props[SavingsAccountProxy]) ! GetAccountBalances(id)
expectOnce {
case SavingsAccountBalances(balances)
results += (Savings -> balances)
collectBalances()
}
}
def fetchMoneyMarketAccountsBalance() {
context.actorOf(Props[MoneyMarketAccountProxy]) ! GetAccountBalances(id)
expectOnce {
case MoneyMarketAccountBalances(balances)
results += (MoneyMarket -> balances)
collectBalances()
}
}
def collectBalances(force: Boolean = false) {
if (results.size == types.size || force) {
originalSender ! results.toList // Make sure it becomes immutable
context.stop(self)
}
}
}
}
class
class
class
class
class
InitialRequest(name: String)
Request(name: String)
Response(name: String, value: String)
EvaluationResults(name: String, eval: List[Int])
FinalResponse(qualifiedValues: List[String])
/**
* An actor sample demonstrating use of unexpect and chaining.
* This is just an example and not a complete test case.
*/
class ChainingSample extends Actor with Aggregator {
expectOnce {
419
Pitfalls
The current implementation does not match the sender of the message. This is designed to work with
ActorSelection as well as ActorRef. Without the sender(), there is a chance a received message
can be matched by more than one partial function. The partial function that was registered via expect
or expectOnce first (chronologically) and is not yet de-registered by unexpect takes precedence in
this case. Developers should make sure the messages can be uniquely matched or the wrong logic can be
executed for a certain message.
The sender referenced in any expect or expectOnce logic refers to the sender() of that particular
message and not the sender() of the original message. The original sender() still needs to be saved so a final
response can be sent back.
context.become is not supported when extending the Aggregator trait.
We strongly recommend against overriding receive. If your use case really dictates, you may do so with
extreme caution. Always provide a pattern match handling aggregator messages among your receive
pattern matches, as follows:
case msg if handleMessage(msg) // noop
// side effects of handleMessage does the actual match
Sorry, there is not yet a Java implementation of the aggregator pattern available.
9.5. External Contributions
420
421
CHAPTER
TEN
For example:
git clone git://github.com/akka/akka.git
If you have already cloned the repository previously then you can update the code with git pull:
git pull origin master
Building
To compile all the Akka core modules use the compile command:
sbt compile
422
If compiling and testing are successful then you have everything working for the latest Akka development version.
Parallel Execution
By default the tests are executed sequentially. They can be executed in parallel to reduce build times, if hardware
can handle the increased memory and cpu usage. Add the following system property to sbt launch script to activate
parallel execution:
-Dakka.parallelExecution=true
Note: Akka generates class diagrams for the API documentation using ScalaDoc. This needs the dot command
from the Graphviz software package to be installed to avoid errors. You can disable the diagram generation by
adding the flag -Dakka.scaladoc.diagrams=false
423
10.1.4 Dependencies
You
can
look
at
the
Ivy
dependency
resolution
information
that
is
created
on
sbt update
and
found
in
~/.ivy2/cache.
For
example,
the
~/.ivy2/cache/com.typesafe.akka-akka-remote-compile.xml file contains the resolution information for the akka-remote module compile dependencies. If you open this file in a web browser you
will get an easy to navigate view of dependencies.
10.2.1 Setup
The multi-JVM testing is an sbt plugin that you can find at http://github.com/typesafehub/sbt-multi-jvm.
You can add it as a plugin by adding the following to your project/plugins.sbt:
addSbtPlugin("com.typesafe.sbt" % "sbt-multi-jvm" % "0.3.8")
You can then add multi-JVM testing to build.sbt or project/Build.scala by including the MultiJvm
settings and config. Please note that MultiJvm test sources are located in src/multi-jvm/..., and not in
src/test/....
Here is an example build.sbt file for sbt 0.13 that uses the MultiJvm plugin:
import com.typesafe.sbt.SbtMultiJvm
import com.typesafe.sbt.SbtMultiJvm.MultiJvmKeys.MultiJvm
val akkaVersion = "2.3.1"
val project = Project(
id = "akka-sample-multi-node-scala",
base = file("."),
settings = Project.defaultSettings ++ SbtMultiJvm.multiJvmSettings ++ Seq(
name := "akka-sample-multi-node-scala",
version := "1.0",
scalaVersion := "2.10.3",
libraryDependencies ++= Seq(
"com.typesafe.akka" %% "akka-remote" % akkaVersion,
"com.typesafe.akka" %% "akka-multi-node-testkit" % akkaVersion,
"org.scalatest" %% "scalatest" % "2.0" % "test"),
// make sure that MultiJvm test are compiled by the default test compilation
compile in MultiJvm <<= (compile in MultiJvm) triggeredBy (compile in Test),
// disable parallel tests
parallelExecution in Test := false,
// make sure that MultiJvm tests are executed by the default test target,
// and combine the results from ordinary test and multi-jvm tests
executeTests in Test <<= (executeTests in Test, executeTests in MultiJvm) map {
case (testResults, multiNodeResults) =>
val overall =
424
Or one can change to the akka-remote-tests project first, and then run the tests:
project akka-remote-tests
multi-jvm:test
More than one test name can be listed to run multiple specific tests. Tab-completion in sbt makes it easy to
complete the test names.
Its also possible to specify JVM options with test-only by including those options after the test names and
--. For example:
multi-jvm:test-only akka.remote.RandomRoutedRemoteActor -- -Dsome.option=something
That is, each test has MultiJvm in the middle of its name. The part before it groups together tests/applications
under a single TestName that will run together. The part after, the NodeName, is a distinguishing name for each
forked JVM.
So to create a 3-node test called Sample, you can create three applications like the following:
package sample
object SampleMultiJvmNode1 {
def main(args: Array[String]) {
println("Hello from node 1")
}
}
425
object SampleMultiJvmNode2 {
def main(args: Array[String]) {
println("Hello from node 2")
}
}
object SampleMultiJvmNode3 {
def main(args: Array[String]) {
println("Hello from node 3")
}
}
When you call multi-jvm:run sample.Sample at the sbt prompt, three JVMs will be spawned, one for
each node. It will look like this:
> multi-jvm:run sample.Sample
...
[info] * sample.Sample
[JVM-1] Hello from node 1
[JVM-2] Hello from node 2
[JVM-3] Hello from node 3
[success] Total time: ...
You can change what the MultiJvm identifier is. For example, to change it to ClusterTest use the
multiJvmMarker setting:
multiJvmMarker in MultiJvm := "ClusterTest"
SampleMultiJvmNode2.opts:
-Dakka.remote.port=9992 -Xmx256m
SampleMultiJvmNode3.opts:
-Dakka.remote.port=9993 -Xmx256m
426
10.2.6 ScalaTest
There is also support for creating ScalaTest tests rather than applications. To do this use the same naming convention as above, but create ScalaTest suites rather than objects with main methods. You need to have ScalaTest on
the classpath. Here is a similar example to the one above but using ScalaTest:
package sample
import org.scalatest.WordSpec
import org.scalatest.matchers.MustMatchers
class SpecMultiJvmNode1 extends WordSpec with MustMatchers {
"A node" should {
"be able to say hello" in {
val message = "Hello from node 1"
message must be("Hello from node 1")
}
}
}
class SpecMultiJvmNode2 extends WordSpec with MustMatchers {
"A node" should {
"be able to say hello" in {
val message = "Hello from node 2"
message must be("Hello from node 2")
}
}
}
To run just these tests you would call multi-jvm:test-only sample.Spec at the sbt prompt.
10.3.1 Requirements
In order to form a general and extensible IO layer basis for a wide range of applications, with Akka remoting and
spray HTTP being the initial ones, the following requirements were established as key drivers for the design:
scalability to millions of concurrent connections
lowest possible latency in getting data from an input channel into the target actors mailbox
maximal throughput
optional back-pressure in both directions (i.e. throttling local senders as well as allowing local readers to
throttle remote senders, where allowed by the protocol)
a purely actor-based API with immutable data representation
extensibility for integrating new transports by way of a very lean SPI; the goal is to not force I/O mechanisms
into a lowest common denominator but instead allow completely protocol-specific user-level APIs.
427
428
Akka is using Scalariform to format the source code as part of the build. So just hack away and then run sbt
compile and it will reformat the code according to Akka standards.
10.4.2 Process
Make sure you have signed the Akka CLA, if not, sign it online.
Pick a ticket, if there is no ticket for your work then create one first.
Start working in a feature branch. Name it something like wip-<ticket number>-<descriptive
name>-<your username>.
When you are done, create a GitHub Pull-Request towards the targeted branch and email the Akka Mailing
List that you want it reviewed
When theres consensus on the review, someone from the Akka Core Team will merge it.
429
10.4.4 Testing
All code that is checked in should have tests. All testing is done with ScalaTest and ScalaCheck.
Name tests as Test.scala if they do not depend on any external stuff. That keeps surefire happy.
Name tests as Spec.scala if they have external dependencies.
Actor TestKit
There is a useful test kit for testing actors: akka.util.TestKit. It enables assertions concerning replies received and
their timing, there is more documentation in the akka-testkit module.
Multi-JVM Testing
Included in the example is an sbt trait for multi-JVM testing which will fork JVMs for multi-node testing. There
is support for running applications (objects with main methods) and running ScalaTest tests.
NetworkFailureTest
You can use the NetworkFailureTest trait to test network failure.
10.5.1 Sphinx
For more details see The Sphinx Documentation
10.5.2 reStructuredText
For more details see The reST Quickref
Sections
Section headings are very flexible in reST. We use the following convention in the Akka documentation:
# (over and under) for module headings
= for sections
- for subsections
^ for subsubsections
~ for subsubsubsections
430
Cross-referencing
Sections that may be cross-referenced across the documentation should be marked with a reference. To
mark a section use .. _ref-name: before the section heading. The section can then be linked with
:ref:ref-name. These are unique references across the entire documentation.
For example:
.. _akka-module:
#############
Akka Module
#############
This is the module documentation.
.. _akka-section:
Akka Section
============
Akka Subsection
--------------Here is a reference to "akka section": :ref:akka-section which will have the
name "Akka Section".
Installing Sphinx on OS X
Install Homebrew
Install Python and pip:
brew install python
/usr/local/share/python/easy_install pip
431
tlmgr
tlmgr
tlmgr
tlmgr
tlmgr
tlmgr
tlmgr
tlmgr
update --self
install titlesec
install framed
install threeparttable
install wrapfig
install helvetic
install courier
install multirow
If you get the error unknown locale: UTF-8 when generating the documentation the solution is to define the
following environment variables:
export LANG=en_US.UTF-8
export LC_ALL=en_US.UTF-8
432
10.6 Team
Name
Jonas Bonr
Viktor Klang
Roland Kuhn
Patrik Nordwall
Bjrn Antonsson
Endre Varga
Mathias Doenitz
Johannes Rudolph
Raymond Roestenburg
Piotr Gabryanczyk
Helena Edelson
Martin Krasser
Henrik Engstrm
Peter Vlugter
Derek Williams
Debasish Ghosh
Ross McDonald
Eckhart Hertzler
Mikael Hgqvist
Tim Perrett
Jeanfrancois Arcand
Jan Van Besien
Michael Kober
Peter Veentjer
Irmo Manie
Heiko Seeberger
Hiram Chirino
Scott Clasen
10.6. Team
Role
Founder, Despot, Committer
Honorary Member
Project Lead
Core Team
Core Team
Core Team
Committer
Committer
Committer
Committer
Committer
Committer
Alumnus
Alumnus
Alumnus
Alumnus
Alumnus
Alumnus
Alumnus
Alumnus
Alumnus
Alumnus
Alumnus
Alumnus
Alumnus
Alumnus
Alumnus
Alumnus
433
CHAPTER
ELEVEN
PROJECT INFORMATION
11.1 Migration Guides
11.1.1 Migration Guide 1.3.x to 2.0.x
Migration from 1.3.x to 2.0.x is described in the documentation of 2.0.
434
Routers
The routers have been cleaned up and enhanced. The routing logic has been extracted to be usable within normal
actors as well. Some usability problems have been have been solved, such as properly reject invalid configuration
combinations. Routees can be dynamically added and removed by sending special management messages to the
router.
The two types of routers have been named Pool and Group to make them more distinguishable and reduce
confusion of their subtle differences:
Pool - The router creates routees as child actors and removes them from the router if they terminate.
Group - The routee actors are created externally to the router and the router sends messages to the specified
path using actor selection, without watching for termination.
Configuration of routers is compatible with 2.2.x, but the router type should preferably be specified with -pool
or -group suffix.
Some classes used for programmatic definition of routers have been renamed, but the old classes remain as deprecated. The compiler will guide you with deprecation warning. For example RoundRobinRouter has been
renamed to RoundRobinPool or RoundRobinGroup depending on which type you are actually using.
There is no replacement for SmallestMailboxRouter combined with routee paths, i.e. a group, because that
combination is not useful.
An optional API enhancement that makes the code read better is to use the props method instead of
withRouter. withRouter has not been deprecated and you can continue to use that if you prefer that way of
defining a router.
Example in Scala:
context.actorOf(FromConfig.props(Props[Worker]), "router1")
context.actorOf(RoundRobinPool(5).props(Props[Worker]), "router2")
Example in Java:
getContext().actorOf(FromConfig.getInstance().props(Props.create(Worker.class)),
"router1");
getContext().actorOf(new RoundRobinPool(5).props(Props.create(Worker.class)),
"router2");
To support multiple routee paths for a cluster aware router sending to paths the deployment configuration
property cluster.routees-path has been changed to string list routees.paths property. The old
cluster.routees-path is deprecated, but still working during the deprecation phase.
Example:
/router4 {
router = round-robin-group
nr-of-instances = 10
routees.paths = ["/user/myserviceA", "/user/myserviceB"]
cluster.enabled = on
}
The API for creating custom routers and resizers have changed without keeping the old API as deprecated. That
should be a an API used by only a few users and they should be able to migrate to the new API without much
trouble.
Read more about the new routers in the documentation for Scala and documentation for Java.
Akka IO is no longer experimental
The core IO layer introduced in Akka 2.2 is now a fully supported module of Akka.
435
436
437
akka-sbt-plugin is Removed
akka-sbt-plugin for packaging of application binaries has been removed. Version 2.2.3 can still be used
independent of Akka version of the application. Version 2.2.3 can be used with both sbt 0.12 and 0.13.
sbt-native-packager is the recommended tool for creating distributions of Akka applications when using sbt.
Parens Added to sender
Parens were added to the sender() method of the Actor Scala API to highlight that the sender() reference is
not referentially transparent and must not be exposed to other threads, for example by closing over it when using
future callbacks.
It is recommended to use this new convention:
sender() ! "reply"
However, it is not mandatory to use parens and you do not have to change anything.
ReliableProxy Constructor Changed
The constructor of ReliableProxy in akka-contrib has been changed to take an ActorPath instead
of an ActorRef. Also it takes new parameters to support reconnection. Use the new props factory methods,
ReliableProxy.props.
OSGi Changes
akka-osgi no longer contains the akka-actor classes, instead akka-actor is a bundle now. akka-osgi
only contains a few OSGi helpers, most notably the BundleDelegatingClassLoader which resolves e.g.
reference.conf files (hence these are not copied into the akka-osgi bundle any longer either).
akka-osgi-aries has been removed. Similar can be implemented outside of Akka if needed.
TestKit: reworked time dilation
TestDuration has been changed into an implicit value class plus a Java API in JavaTestKit. Please change:
import akka.testkit.duration2TestDuration
into:
import akka.testkit.TestDuration
438
Extensions
Eventsourced and Akka Persistence are both extending-akka-scala.
Eventsourced: EventsourcingExtension
Must be explicitly created with an actor system and an application-defined journal actor as arguments. (see
example usage).
Processors and Channels must be created with the factory methods processorOf and channelOf defined on EventsourcingExtension.
Is a central registry of created processors and channels.
Akka Persistence: Persistence extension
Must not be explicitly created by an application. A Persistence extension is implicitly created upon first
processor or channel creation. Journal actors are automatically created from a journal plugin configuration
(see Journal plugin API).
Processors and Channels can be created like any other actor with actorOf without using the
Persistence extension.
Is not a central registry of processors and channels.
Processors
Eventsourced: Eventsourced
Stackable trait that adds journaling (write-ahead-logging) to actors (see processor definition and creation).
Name Eventsourced caused some confusion in the past as many examples used Eventsourced processors for command sourcing. See also this FAQ entry for clarification.
Must be explicitly recovered using recovery methods on EventsourcingExtension. Applications are
responsible for avoiding an interference of replayed messages and new messages i.e. applications have to
explicitly wait for recovery to complete. Recovery on processor re-start is not supported out-of-the box.
Journaling-preserving behavior changes are only possible with special-purpose methods become
and unbecome, in addition to non-journaling-preserving behavior changes with default methods
context.become and context.unbecome.
Writes messages of type Message to the journal (see processor usage). Sender references are not stored in
the journal i.e. the sender reference of a replayed message is always system.deadLetters.
Supports snapshots.
Identifiers are of type Int and must be application-defined.
Does not support batch-writes of messages to the journal.
Does not support stashing of messages.
Akka Persistence: Processor
Trait that adds journaling (write-ahead-logging) to actors (see Processors) and used by applications for
command sourcing. Corresponds to Eventsourced processors in Eventsourced but is not a stackable
trait.
Automatically recovers on start and re-start, by default. Recovery can be customized or turned off by overriding actor life cycle hooks preStart and preRestart. Processor takes care that new messages
never interfere with replayed messages. New messages are internally buffered until recovery completes.
No special-purpose behavior change methods. Default behavior change methods context.become and
context.unbecome can be used and are journaling-preserving.
Writes messages of type Persistent to the journal (see Persistent messages). Corresponds to Message
in Eventsourced. Sender references are written to the journal. A reply to senders must therefore be done via
439
a channel in order to avoid redundant replies during replay. Sender references of type PromiseActorRef
are not journaled, they are system.deadLetters on replay.
Supports Snapshots.
Identifiers are of type String, have a default value and can be overridden by applications.
Supports Batch writes.
Supports stashing of messages.
Akka Persistence: EventsourcedProcessor
Extension trait and pattern on top of Processor to support Event sourcing. Has no direct counterpart in
Eventsourced. Can be considered as a replacement of two processors in Eventsourced where one processor
processes commands and the other processes events that have been emitted by the command processor.
Channels
Eventsourced: DefaultChannel
Prevents redundant delivery of messages to a destination (see usage example and default channel).
Is associated with a single destination actor reference. A new incarnation of the destination is not automatically resolved by the channel. In this case a new channel must be created.
Must be explicitly activated
EventsourcingExtension.
using
methods
deliver
or
recover
defined
on
Must be activated after all processors have been activated. Depending on the recovery method, this is either
done automatically or must be followed by the application (see non-blocking recovery). This is necessary
for a network of processors and channels to recover consistently.
Does not redeliver messages on missing or negative delivery confirmation.
Cannot be used standalone.
Akka Persistence: Channel
Prevents redundant delivery of messages to a destination (see Channels) i.e. serves the same primary purpose as in Eventsourced.
Is not associated with a single destination. A destination can be specified with each Deliver request and
is referred to by an actor path. A destination path is resolved to the current destination incarnation during
delivery (via actorSelection).
Must not be explicitly activated. Also, a network of processors and channels automatically recover consistently, even if they are distributed. This enhancement, together with improved processor recovery, makes
recovery of complex Akka Persistence applications trivial. No special recovery procedures must be run by
applications.
Redelivers messages on missing delivery confirmation (see Message re-delivery). In contrast to
Eventsourced, Akka Persistence doesnt distinguish between missing and negative confirmations. It only
has a notion of missing confirmations using timeouts (which are closely related to negative confirmations
as both trigger message redelivery).
Can be used standalone.
Persistent channels
Eventsourced: ReliableChannel
Provides DefaultChannel functionality plus persistence and recovery from sender JVM crashes (see
ReliableChannel). Also provides message redelivery in case of missing or negative delivery confirmations.
Delivers next message to a destination only if previous message has been successfully delivered (flow control is done by destination).
11.1. Migration Guides
440
Stops itself when the maximum number of redelivery attempts has been reached.
Cannot reply on persistence.
Can be used standalone.
Akka Persistence: PersistentChannel
Provides Channel functionality plus persistence and recovery from sender JVM crashes (see Persistent
channels). Same message redelivery features as Channel.
Redelivers unconfirmed messages concurrently to newly delivered messages. Flow control is done by channel using a configurable minimum and maximum number of pending confirmations.
Optionally notifies applications about messages for which the maximum number of delivery attempts has
been reached (also offered by Channel).
Can reply on persistence (= accept acknowledgement).
Can be used standalone.
Views
Eventsourced:
No direct support for views. Only way to maintain a view is to use a channel and a processor as destination.
Akka Persistence: View
Receives the message stream written by a Processor or EventsourcedProcessor by reading it
directly from the journal (see Views). Alternative to using channels. Useful in situations where actors shall
receive a persistent message stream in correct order without duplicates.
Can be used in combination with Channels for sending messages.
Supports Snapshots.
Serializers
Eventsourced:
Uses a protobuf serializer for serializing Message objects.
Delegates to a configurable Akka serializer for serializing Message payloads.
Delegates to a configurable, proprietary (stream) serializer for serializing snapshots.
See Serialization.
Akka Persistence:
Uses a protobuf serializer for serializing Persistent objects.
Delegates to a configurable Akka serializer for serializing Persistent payloads.
Delegates to a configurable Akka serializer for serializing snapshots.
See Custom serialization.
Sequence numbers
Eventsourced:
Generated on a per-journal basis.
Akka Persistence:
Generated on a per-processor basis.
441
Storage plugins
Eventsourced:
Plugin API: SynchronousWriteReplaySupport and AsynchronousWriteReplaySupport
No separation between journal and snapshot storage plugins.
All plugins pre-packaged with project (see journals and snapshot configuration)
Akka Persistence:
Plugin API: SyncWriteJournal, AsyncWriteJournal, AsyncRecovery, SnapshotStore
(see Storage plugins).
Clear separation between journal and snapshot storage plugins.
Limited number of Pre-packaged plugins (LevelDB journal and local snapshot store).
Impressive list of community plugins.
11.2.1 Browsing
Tickets
You can find the Akka tickets here
Roadmaps
The roadmap for each Akka milestone is here
11.3 Licenses
11.3.1 Akka License
This software is licensed under the Apache 2 license, quoted below.
Copyright 2009-2014 Typesafe Inc. <http://www.typesafe.com>
Licensed under the Apache License, Version 2.0 (the "License"); you may not
use this file except in compliance with the License. You may obtain a copy of
442
the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
License for the specific language governing permissions and limitations under
the License.
11.4 Sponsors
11.4.1 Typesafe
Typesafe is the company behind the Akka Project, Scala Programming Language, Play Web Framework, Scala
IDE, Simple Build Tool and many other open source projects. It also provides the Typesafe Stack, a full-featured
development stack consisting of AKka, Play and Scala. Learn more at typesafe.com.
11.4.2 YourKit
YourKit is kindly supporting open source projects with its full-featured Java Profiler.
YourKit, LLC is the creator of innovative and intelligent tools for profiling Java and .NET applications. Take a
look at YourKits leading software products: YourKit Java Profiler and YourKit .NET Profiler
11.5 Project
11.5.1 Commercial Support
Commercial support is provided by Typesafe. Akka is part of the Typesafe Reactive Platform.
11.5.3 Downloads
http://akka.io/downloads
11.4. Sponsors
443
Define the library dependencies with the timestamp as version. For example:
libraryDependencies += "com.typesafe.akka" % "akka-remote_2.10" %
"2.1-20121016-001042"
Define the library dependencies with the timestamp as version. For example:
<dependencies>
<dependency>
<groupId>com.typesafe.akka</groupId>
<artifactId>akka-remote_2.10</artifactId>
<version>2.1-20121016-001042</version>
</dependency>
</dependencies>
11.5. Project
444
CHAPTER
TWELVE
ADDITIONAL INFORMATION
12.1 Frequently Asked Questions
12.1.1 Akka Project
Where does the name Akka come from?
It is the name of a beautiful Swedish mountain up in the northern part of Sweden called Laponia. The mountain is
also sometimes called The Queen of Laponia.
Akka is also the name of a goddess in the Smi (the native Swedish population) mythology. She is the goddess
that stands for all the beauty and good in the world. The mountain can be seen as the symbol of this goddess.
Also, the name AKKA is the a palindrome of letters A and K as in Actor Kernel.
Akka is also:
the name of the goose that Nils traveled across Sweden on in The Wonderful Adventures of Nils by the
Swedish writer Selma Lagerlf.
the Finnish word for nasty elderly woman and the word for elder sister in the Indian languages Tamil,
Telugu, Kannada and Marathi.
a font
a town in Morocco
a near-earth asteroid
12.1.4 Remoting
I want to send to a remote system but it does not do anything
Make sure that you have remoting enabled on both ends: client and server. Both do need hostname and port
configured, and you will need to know the port of the server; the client can use an automatic port in most cases
(i.e. configure port zero). If both systems are running on the same network host, their ports must be different
If you still do not see anything, look at what the logging of remote life-cycle events tells you (normally logged
at INFO level) or switch on Auxiliary remote logging options to see all sent and received messages (logged at
DEBUG level).
Which options shall I enable when debugging remoting issues?
Have a look at the Remote Configuration, the typical candidates are:
akka.remote.log-sent-messages
akka.remote.log-received-messages
akka.remote.log-remote-lifecycle-events (this also includes deserialization errors)
446
12.1.5 Debugging
How do I turn on debug logging?
To turn on debug logging in your actor system add the following to your configuration:
akka.loglevel = DEBUG
To enable different types of debug logging add the following to your configuration:
akka.actor.debug.receive will log all messages sent to an actor if that actors receive method is a
LoggingReceive
akka.actor.debug.autoreceive will log all special messages like Kill, PoisonPill e.t.c.
sent to all actors
akka.actor.debug.lifecycle will log all actor lifecycle events of all actors
Read more about it in the docs for Logging and actor.logging-scala.
447
12.2 Books
Akka in Action, by Raymond Roestenburg and Rob Bakker, Manning Publications Co., ISBN:
9781617291012, est fall 2013
Akka Concurrency, by Derek Wyatt, artima developer, ISBN: 0981531660, est April 2013
Akka Essentials, by Munish K. Gupta, PACKT Publishing, ISBN: 1849518289, October 2012
12.3.2 Groovy/Groovy++
Read more here: https://gist.github.com/620439.
12.3.3 Clojure
Read more here: http://blog.darevay.com/2011/06/clojure-and-akka-a-match-made-in/.
12.4.2 Activator
To bootstrap Akka inside an OSGi environment, you can use the akka.osgi.ActorSystemActivator
class to conveniently set up the ActorSystem.
import akka.actor.{ Props, ActorSystem }
import org.osgi.framework.BundleContext
import akka.osgi.ActorSystemActivator
class Activator extends ActorSystemActivator {
def configure(context: BundleContext, system: ActorSystem) {
// optionally register the ActorSystem in the OSGi Service Registry
registerService(context, system)
val someActor = system.actorOf(Props[SomeActor], name = "someName")
someActor ! SomeMessage
}
}
12.2. Books
448
The ActorSystemActivator creates the actor system with a class loader that finds resources
(reference.conf files) and classes from the application bundle and all transitive dependencies.
The ActorSystemActivator class is included in the akka-osgi artifact:
<dependency>
<groupId>com.typesafe.akka</groupId>
<artifactId>akka-osgi_2.10</artifactId>
<version>2.3.1</version>
</dependency>
12.4.3 Sample
A complete sample project is provided in akka-sample-osgi-dining-hakkers.
12.5.2 Spray
The Spray toolkit is built using Akka, and is a minimalistic HTTP/REST layer.
449