Flink on kubernetes
We are running flink(version 1.5.2) on k8s with rocksdb backend.
Each time when the job is cancelled and restarted, we face OOMKilled problem from the container.
In our case, we only assign 15% of container memory to JVM and leave others to rocksdb.
To us, it looks like memory used by rocksdb is not released after job cancelling. Anyone can gives some suggestions?
Currently our tmp fix is to restart the TM pod for each job cancelling, but it has to be manually.
This communication may contain privileged or other confidential information of Red. If you have received it in error, please advise the sender by reply e-mail and immediately delete the message and any attachments without copying or disclosing the contents. Thank you.