As reported by @stu203404, when setting --jobInstance work is not shared between instances. Instead every instances starts an own job and processes the data of all Kafka partitions. We still have to find a solution here, so I'm reoping this issue.
I think, we finally solved this issue. I resetted the use of --jobInstance in !187 (merged). As far as I observed, instances are crashing if there are more instances than Kafka partitions Samza is reading from. This is okay for us as these instances would remain idle anyway (although it's strange that this is documented nowhere). We experienced the reported issue when running Kafka and the load generator from our Docker-Compose files and the load generator was starting "too fast". In that case, the load generator created the topic with only partition. Hence, when running two instances, one is crashing. I opened #306 to tackle this issue.
@stu203404 Feel free to close this ticket after verifying that everything works as intended.
Using the Docker-compose files the same problem applies to the beam Flink implementation. There are just as many Flink TaskManager working as partitions available. Except from that the flink implementations seems to work.