Documentation

You are viewing the documentation for the 3.0.5 release. The latest stable release series is 3.0.x.

§Understanding Play thread pools

Play Framework is, from the bottom up, an asynchronous web framework. Thread pools in Play are tuned to use fewer threads than in traditional web frameworks, since IO in play-core never blocks.

Because of this, if you plan to write blocking IO code, or code that could potentially do a lot of CPU intensive work, you need to know exactly which thread pool is bearing that workload, and you need to tune it accordingly. Doing blocking IO without taking this into account is likely to result in very poor performance from Play Framework, for example, you may see only a few requests per second being handled, while CPU usage sits at 5%. In comparison, benchmarks on typical development hardware (eg, a MacBook Pro) have shown Play to be able to handle workloads in the hundreds or even thousands of requests per second without a sweat when tuned correctly.

§Knowing when you are blocking

The most common place where a typical Play application will block is when it’s talking to a database. Unfortunately, none of the major databases provide asynchronous database drivers for the JVM, so for most databases, your only option is to using blocking IO. A notable exception to this is ReactiveMongo, a driver for MongoDB that uses Play’s Iteratee library to talk to MongoDB.

Other cases when your code may block include:

In general, if the API you are using returns Futures, it is non-blocking, otherwise it is blocking.

Note that you may be tempted to therefore wrap your blocking code in Futures. This does not make it non-blocking, it just means the blocking will happen in a different thread. You still need to make sure that the thread pool that you are using has enough threads to handle the blocking. Please see Play’s example templates on https://playframework.com/download#examples for how to configure your application for a blocking API.

In contrast, the following types of IO do not block:

§Play’s thread pools

Play uses a number of different thread pools for different purposes:

§Using the default thread pool

All actions in Play Framework use the default thread pool. When doing certain asynchronous operations, for example, calling map or flatMap on a future, you may need to provide an implicit execution context to execute the given functions in. An execution context is basically another name for a ThreadPool.

In most situations, the appropriate execution context to use will be the Play default thread pool. This is accessible through @Inject()(implicit ec: ExecutionContext) This can be used by injecting it into your Scala source file:

class Samples @Inject() (components: ControllerComponents)(implicit ec: ExecutionContext)
    extends AbstractController(components) {
  def someAsyncAction = Action.async {
    someCalculation()
      .map { result => Ok(s"The answer is $result") }
      .recover {
        case e: TimeoutException =>
          InternalServerError("Calculation timed out!")
      }
  }

  def someCalculation(): Future[Int] = {
    Future.successful(42)
  }
}

or using CompletionStage with an ClassLoaderExecutionContext in Java code:

import java.util.concurrent.CompletableFuture;
import java.util.concurrent.CompletionStage;
import javax.inject.Inject;
import play.libs.concurrent.ClassLoaderExecutionContext;
import play.mvc.*;

public class MyController extends Controller {

  private ClassLoaderExecutionContext clExecutionContext;

  @Inject
  public MyController(ClassLoaderExecutionContext ec) {
    this.clExecutionContext = ec;
  }

  public CompletionStage<Result> index() {
    // Use a different task with explicit EC
    return calculateResponse()
        .thenApplyAsync(
            answer -> {
              return ok("answer was " + answer).flashing("info", "Response updated!");
            },
            clExecutionContext.current());
  }

  private static CompletionStage<String> calculateResponse() {
    return CompletableFuture.completedFuture("42");
  }
}

This execution context connects directly to the Application’s ActorSystem and uses Pekko’s default dispatcher.

§Configuring the default thread pool

The default thread pool can be configured using standard Pekko configuration in application.conf under the pekko namespace.

If you want to configure the default dispatcher, use another dispatcher, or define a new dispatcher to use, see the Types of dispatchers section of Pekko’s reference documentation for full details.

The full configuration options available to you can be found in the Configuration section.

§Using other thread pools

In certain circumstances, you may wish to dispatch work to other thread pools. This may include CPU heavy work, or IO work, such as database access. To do this, you should first create a ThreadPool, this can be done easily in Scala:

val myExecutionContext: ExecutionContext = actorSystem.dispatchers.lookup("my-context")

In this case, we are using Pekko to create the ExecutionContext, but you could also easily create your own ExecutionContexts using Java executors, or the Scala fork join thread pool, for example. Play provides play.libs.concurrent.CustomExecutionContext and play.api.libs.concurrent.CustomExecutionContext that can be used to create your own execution contexts. Please see ScalaAsync or JavaAsync for further details.

To configure this Pekko execution context, you can add the following configuration to your application.conf:

my-context {
  fork-join-executor {
    parallelism-factor = 20.0
    parallelism-max = 200
  }
}

To use this execution context in Scala, you would simply use the scala Future companion object function:

Future {
  // Some blocking or expensive code here
}(myExecutionContext)

or you could just use it implicitly:

implicit val ec = myExecutionContext

Future {
  // Some blocking or expensive code here
}

In addition, please see the example templates on https://playframework.com/download#examples for examples of how to configure your application for a blocking API.

§Class loaders

Class loaders need special handling in a multithreaded environment such as a Play program.

§Application class loader

In a Play application the thread context class loader may not always be able to load application classes. You should explicitly use the application class loader to load classes.

Java
Class myClass = app.classloader().loadClass(myClassName);
Scala
val myClass = app.classloader.loadClass(myClassName)

Being explicit about loading classes is most important when running Play in development mode (using run) rather than production mode. That’s because Play’s development mode uses multiple class loaders so that it can support automatic application reloading. Some of Play’s threads might be bound to a class loader that only knows about a subset of your application’s classes.

In some cases you may not be able to explicitly use the application classloader. This is sometimes the case when using third party libraries. In this case you may need to set the thread context class loader explicitly before you call the third party code. If you do, remember to restore the context class loader back to its previous value once you’ve finished calling the third party code.

§Switching threads

The problem with class loaders however is that as soon as control switches to another thread, you lose access to the original class loader. So if you were to map a CompletionStage using thenApplyAsync, or using thenApply at a point in time after the Future associated with that CompletionStage had completed, and you then try to access the original class loader, it probably won’t work . To address this, Play provides an ClassLoaderExecutionContext. This allows you to capture the current class loader in an Executor, which you can then pass to the CompletionStage *Async methods such as thenApplyAsync(), and when the executor executes your callback, it will ensure the class loader remains in scope.

To use the ClassLoaderExecutionContext, inject it into your component, and then pass the current execution context anytime a CompletionStage is interacted with. For example:

import java.util.concurrent.CompletableFuture;
import java.util.concurrent.CompletionStage;
import javax.inject.Inject;
import play.libs.concurrent.ClassLoaderExecutionContext;
import play.mvc.*;

public class MyController extends Controller {

  private ClassLoaderExecutionContext clExecutionContext;

  @Inject
  public MyController(ClassLoaderExecutionContext ec) {
    this.clExecutionContext = ec;
  }

  public CompletionStage<Result> index() {
    // Use a different task with explicit EC
    return calculateResponse()
        .thenApplyAsync(
            answer -> {
              return ok("answer was " + answer).flashing("info", "Response updated!");
            },
            clExecutionContext.current());
  }

  private static CompletionStage<String> calculateResponse() {
    return CompletableFuture.completedFuture("42");
  }
}

If you have a custom executor, you can wrap it in an ClassLoaderExecutionContext simply by passing it to the ClassLoaderExecutionContexts constructor.

§Best practices

How you should best divide work in your application between different thread pools greatly depends on the types of work that your application is doing, and the control you want to have over how much of which work can be done in parallel. There is no one size fits all solution to the problem, and the best decision for you will come from understanding the blocking-IO requirements of your application and the implications they have on your thread pools. It may help to do load testing on your application to tune and verify your configuration.

Note: In a blocking environment, thread-pool-executor is better than fork-join because no work-stealing is possible, and a fixed-pool-size size should be used and set to the maximum size of the underlying resource.

Given the fact that JDBC is blocking, thread pools can be sized to the number of connections available to a database pool, assuming that the thread pool is used exclusively for database access. Fewer threads will not consume the number of connections available. Any more threads than the number of connections available could be wasteful given contention for the connections.

Below we outline a few common profiles that people may want to use in Play Framework:

§Pure asynchronous

In this case, you are doing no blocking IO in your application. Since you are never blocking, the default configuration of one thread per processor suits your use case perfectly, so no extra configuration needs to be done. The Play default execution context can be used in all cases.

§Highly synchronous

This profile matches that of a traditional synchronous IO based web framework, such as a Java servlet container. It uses large thread pools to handle blocking IO. It is useful for applications where most actions are doing database synchronous IO calls, such as accessing a database, and you don’t want or need control over concurrency for different types of work. This profile is the simplest for handling blocking IO.

In this profile, you would use the default execution context everywhere, but configure it to have a very large number of threads in its pool. Because the default thread pool is used for both servicing Play requests and database requests, the fixed pool size should be the maximum size of database connection pool, plus the number of cores, plus a couple extra for housekeeping, like so:

pekko {
  actor {
    default-dispatcher {
      executor = "thread-pool-executor"
      throughput = 1
      thread-pool-executor {
        fixed-pool-size = 55 # db conn pool (50) + number of cores (4) + housekeeping (1)
      }
    }
  }
}

This profile is recommended for Java applications that do synchronous IO, since it is harder in Java to dispatch work to other threads.

In addition, please see the example templates on https://playframework.com/download#examples for examples of how to configure your application for a blocking API.

§Many specific thread pools

This profile is for when you want to do a lot of synchronous IO, but you also want to control exactly how much of which types of operations your application does at once. In this profile, you would only do non blocking operations in the default execution context, and then dispatch blocking operations to different execution contexts for those specific operations.

In this case, you might create a number of different execution contexts for different types of operations, like this:

object Contexts {
  implicit val simpleDbLookups: ExecutionContext = actorSystem.dispatchers.lookup("contexts.simple-db-lookups")
  implicit val expensiveDbLookups: ExecutionContext =
    actorSystem.dispatchers.lookup("contexts.expensive-db-lookups")
  implicit val dbWriteOperations: ExecutionContext =
    actorSystem.dispatchers.lookup("contexts.db-write-operations")
  implicit val expensiveCpuOperations: ExecutionContext =
    actorSystem.dispatchers.lookup("contexts.expensive-cpu-operations")
}

These might then be configured like so:

contexts {
  simple-db-lookups {
    executor = "thread-pool-executor"
    throughput = 1
    thread-pool-executor {
      fixed-pool-size = 20
    }
  }
  expensive-db-lookups {
    executor = "thread-pool-executor"
    throughput = 1
    thread-pool-executor {
      fixed-pool-size = 20
    }
  }
  db-write-operations {
    executor = "thread-pool-executor"
    throughput = 1
    thread-pool-executor {
      fixed-pool-size = 10
    }
  }
  expensive-cpu-operations {
    fork-join-executor {
      parallelism-max = 2
    }
  }
}

Then in your code, you would create Futures and pass the relevant ExecutionContext for the type of work that Future was doing.

Note: The configuration namespace can be chosen freely, as long as it matches the dispatcher ID passed to app.actorSystem.dispatchers.lookup. The CustomExecutionContext class will do this for you automatically.

§Few specific thread pools

This is a combination between the many specific thread pools and the highly synchronized profile. You would do most simple IO in the default execution context and set the number of threads there to be reasonably high (say 100), but then dispatch certain expensive operations to specific contexts, where you can limit the number of them that are done at one time.

§Debugging Thread Pools

There are many possible settings for a dispatcher, and it can be hard to see which ones have been applied and what the defaults are, particularly when overriding the default dispatcher. The pekko.log-config-on-start configuration option shows the entire applied configuration when the application is loaded:

pekko.log-config-on-start = on

Note that you must have Pekko logging set to a debug level to see output, so you should add the following to logback.xml:

<logger name="org.apache.pekko" level="DEBUG" />

Once you see the logged HOCON output, you can copy and paste it into an “example.conf” file and view it in IntelliJ IDEA, which supports HOCON syntax. You should see your changes merged in with Pekko’s dispatcher, so if you override thread-pool-executor you will see it merged:

{ 
  # Elided HOCON... 
  "actor" : {
    "default-dispatcher" : {
      # application.conf @ file:/Users/wsargent/work/catapi/target/universal/stage/conf/application.conf: 19
      "executor" : "thread-pool-executor"
    }
  }
}

Note also that Play has different configuration settings for development mode than it does for production. To ensure that the thread pool settings are correct, you should run Play in a production configuration.

Next: Configuring Pekko Http Server Backend