This is where you define the load you want to inject to your server.
You can configure assertions and protocols with these two methods:
assertions: set assertions on the simulation, see the dedicated section here
protocols: set protocols definitions, see the dedicated section here.
The definition of the injection profile of users is done with the
inject method. This method takes as argument a sequence of injection steps that will be processed sequentially.
Open vs Closed Workload Models¶
When it comes to load model, systems behave in 2 different ways:
- Closed systems, where you control the concurrent number of users
- Open systems, where you control the arrival rate of users
Make sure to use the proper load model that matches the load your live system experiences.
Closed system are system where the number of concurrent users is capped. At full capacity, a new user can effectively enter the system only once another exits.
Typical systems that behave this way are:
- call center when all operators are busy
- ticketing websites where users get placed into a queue when system is at full capacity
On the contrary, open system have no control over the number concurrent users: users keep on arriving even though applications has some trouble serving them. Most websites behave this way.
If you’re using a closed workload model in your load tests while your system actually is a open one, your test is broken and you’re testing some different imaginary behavior. In such case, when system under test starts to has some trouble, response time will increase, journey time will become longer, so number of concurrent users will increase and injector will slow down to match the imaginary cap you’ve set.
Open and closed workload models are antinomical and you can’t mix them in the same injection profile.
setUp( scn.inject( nothingFor(4 seconds), // 1 atOnceUsers(10), // 2 rampUsers(10) during (5 seconds), // 3 constantUsersPerSec(20) during (15 seconds), // 4 constantUsersPerSec(20) during (15 seconds) randomized, // 5 rampUsersPerSec(10) to 20 during (10 minutes), // 6 rampUsersPerSec(10) to 20 during (10 minutes) randomized, // 7 heavisideUsers(1000) during (20 seconds) // 8 ).protocols(httpProtocol) )
The building blocks for profile injection the way you want are:
nothingFor(duration): Pause for a given duration.
atOnceUsers(nbUsers): Injects a given number of users at once.
rampUsers(nbUsers) during(duration): Injects a given number of users with a linear ramp over a given duration.
constantUsersPerSec(rate) during(duration): Injects users at a constant rate, defined in users per second, during a given duration. Users will be injected at regular intervals.
constantUsersPerSec(rate) during(duration) randomized: Injects users at a constant rate, defined in users per second, during a given duration. Users will be injected at randomized intervals.
rampUsersPerSec(rate1) to (rate2) during(duration): Injects users from starting rate to target rate, defined in users per second, during a given duration. Users will be injected at regular intervals.
rampUsersPerSec(rate1) to(rate2) during(duration) randomized: Injects users from starting rate to target rate, defined in users per second, during a given duration. Users will be injected at randomized intervals.
heavisideUsers(nbUsers) during(duration): Injects a given number of users following a smooth approximation of the heaviside step function stretched to a given duration.
setUp( scn.inject( constantConcurrentUsers(10) during (10 seconds), // 1 rampConcurrentUsers(10) to (20) during (10 seconds) // 2 ) )
constantConcurrentUsers(nbUsers) during(duration): Inject so that number of concurrent users in the system is constant
rampConcurrentUsers(fromNbUsers) to(toNbUsers) during(duration): Inject so that number of concurrent users in the system ramps linearly from a number to another
Then, you have to understand that Gatling’s default behavior is mimic human users with browsers, so each virtual user has its own connections. If you have a high creation rate of users with a short lifespan, you’ll end up opening and closing tons of connections every seconds. As a consequence, you might run out of resources (such as ephemeral ports, because your OS can’t recycle them fast enough). This behavior makes perfect sense when the load you’re modeling is internet traffic. Then, you might consider scaling out, for example with FrontLine, our Enterprise product.
If you’re actually trying to model a small fleet of webservice clients with connection pools, you might want to tune Gatling’s behavior and share the connection pool amongst virtual users.
Setting a smaller number of concurrent users won’t force existing users to abort. The only way for users to terminate is to complete their scenario.
It is possible to use elements of Meta DSL to write tests in an easier way. If you want to chain levels and ramps to reach the limit of your application (a test sometimes called capacity load testing), you can do it manually using the regular DSL and looping using map and flatMap. But there is now an alternative using the meta DSL.
// generate an open workload injection profile // with levels of 10, 15, 20, 25 and 30 arriving users per second // each level lasting 10 seconds // separated by linear ramps lasting 10 seconds setUp( scn.inject( incrementUsersPerSec(5) // Double .times(5) .eachLevelLasting(10 seconds) .separatedByRampsLasting(10 seconds) .startingFrom(10) // Double ) )
// generate a closed workload injection profile // with levels of 10, 15, 20, 25 and 30 concurrent users // each level lasting 10 seconds // separated by linear ramps lasting 10 seconds setUp( scn.inject( incrementConcurrentUsers(5) // Int .times(5) .eachLevelLasting(10 seconds) .separatedByRampsLasting(10 seconds) .startingFrom(10) // Int ) )
incrementUsersPerSec is for open workload and
incrementConcurrentUsers is for closed workload (users/sec vs concurrent users)
startingFrom are both optionals.
If you don’t precise a ramp, the test will jump from one level to another as soon as it is finished.
If you don’t precise an amount of starting users the test will start at 0 concurrent user or 0 user per sec and will go to the next step right away.
Global Pause configuration¶
The pauses can be configured on
Simulation with a bunch of methods:
disablePauses: disable the pauses for the simulation
constantPauses: the duration of each pause is precisely that specified in the
exponentialPauses: the duration of each pause is on average that specified in the
pause(duration)element and follow an exponential distribution.
normalPausesWithStdDevDuration(stdDev: Duration): the duration of each pause is on average that specified in the
pause(duration)element and follow an normal distribution.
stdDevis a Duration.
normalPausesWithPercentageDuration(stdDev: Double): the duration of each pause is on average that specified in the
pause(duration)element and follow an normal distribution.
stdDevis a percentage of the pause value.
customPauses(custom: Expression[Long]): the pause duration is computed by the provided
Expression[Long]. In this case the filled duration is bypassed.
uniformPausesPlusOrMinusDuration(plusOrMinus: Duration): the duration of each pause is on average that specified in the
pause(duration)element and follow a uniform distribution.
Pause definition can also be configured at scenario level.
If you want to reason in terms of requests per second and not in terms of concurrent users, consider using constantUsersPerSec(…) to set the arrival rate of users, and therefore requests, without need for throttling as well as it will be redundant in most cases.
If this is not sufficient for some reason then Gatling supports throttling with the
Throttling is implemented per protocol with support for regular HTTP and JMS.
- You still have to inject users at the scenario level. Throttling tries to ensure a targeted throughput with the given scenarios and their injection profiles (number of users and duration). It’s a bottleneck, ie an upper limit. If you don’t provide enough users, you won’t reach the throttle. If your injection lasts less than the throttle, your simulation will simply stop when all the users are done. If your injection lasts longer than the throttle, the simulation will stop at the end of the throttle.
- Throttling can also be configured per scenario.
setUp(scn.inject(constantUsersPerSec(100) during (30 minutes))).throttle( reachRps(100) in (10 seconds), holdFor(1 minute), jumpToRps(50), holdFor(2 hours) )
This simulation will reach 100 req/s with a ramp of 10 seconds, then hold this throughput for 1 minute, jump to 50 req/s and finally hold this throughput for 2 hours.
The building block for the throttling are:
reachRps(target) in (duration): target a throughput with a ramp over a given duration.
jumpToRps(target): jump immediately to a given targeted throughput.
holdFor(duration): hold the current throughput for a given duration.
Finally, you can configure the maximum duration of your simulation with the method
It is useful if you need to bound the duration of your simulation when you can’t predict it.
setUp(scn.inject(rampUsers(1000) during (20 minutes))).maxDuration(10 minutes)