How to achieve 32000 instance started per second?

Hi everyone.

After reading the blog entry for great success in the speed of your new platform, I decided to implement a PoC that would allow us to switch our classic workflow system to microservice architecture in the future with Zeebe (thanks to the Camunda developers for their great work and hope).

However, after several attempts to achieve at least close to the same indicators as you, the maximum that I have succeeded in is the speed of launching processes - 300 per second.

What i doing wrong?

Server configuration:
3.2 i7-8700 6 core
64 RAM
2*480 SSD
Ubuntu 18.04


Hey @AKuznetsov

please post also the Zeebe Configuration.


1 Like

Did you use this setup?

Hi, @Zelldon

Ok. Here you can find config and Go application that start process

No, we just use basic setup


So it is one of three things.

The server, your client code, or the machine (hardware/OS).

The obvious things to do are:

  1. Run the client code from the benchmark against your docker server.
  2. Run your client code against the server in the benchmark (was that dockerised?)
  3. Run the exact benchmark as it was done on the same hardware. (Verify the results and that you can replicated them)
  4. Run the benchmark as it was run on your hardware. (Get a baseline of your hardware)

Performance profiling like this is just basic science: Experimental design and observation, hypothesis generation and invalidation.

My hypotheses about your experimental design are these:

  1. You have degradation due to dockerisation.
  2. Dockerised server is resource starved.
  3. Running server in docker is slower than native (if your OS is not the same as the benchmark / the benchmark was run on a non-dockerised server)
  4. Your client code cannot generate the same input as the benchmark code.
  5. Your resources do not match the benchmark resources.

The four experimental combinations that I recommended above are sufficient to differentially invalidate or strengthen these hypotheses, and provide input for generating new hypotheses - or direct your attention to the part of the system where you need to experiment further to refine a hypothesis.

For example: if the benchmark client is faster than your Go client against the same server, then you would develop a hypothesis about your client code. Is it blocking? What happens if you run two instances of it in parallel?


There is no silver bullet configuration. You have to experimentally (in)validate your hypotheses and do exhaustive testing.

That was the point I was trying to make in that performance profiling blog post.

You should run a bunch of combinations and come up with a hypothesis with proof, like: “I think that the Go client has an issue where it can’t generate requests as fast as the Java client because I have these three experiments that show these results…”

Like that.

And so, after several days of testing i was not able to achieve the declared speed. Just to understand, I focus on this indicator.


Next I will write the test parameters:

We started testing with docker-compose and broker-only, with default settings. Speed was about 100 instance per second. When increasing the speed of the client (add worker number on Go application), the broker began to return an error: “Reached maximum capacity of requests handled”
Further, partitions were gradually increased, until server resources ceased to be enough, we achieved a speed of 300+ per second.
After that we started the broker directly, from the distribution (without docker). As in the previous case, the partitions and the speed of sending requests by the client were increased, as long as there were enough resources. Speed reached a little less than 400 per second.

So we still didn’t understand what needs to be changed so that the difference between the declared speed and our indicators would be at least not 100 times.

Hello friends!
Please share your working benchmark.
Maybe I’m trying in vain to repeat the results of the test bench?