Scaled Jobs always leave behind extraneous jobs after work is finished. (AWS SQS, jobs deleted after completed) #3486
Unanswered
ZTGallagher
asked this question in
General
Replies: 0 comments
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
We're not sure why. It's not reliably reproducible. Keda simply ends up creating more jobs than we need, and they hang around for hours until we delete them. They do not create pods anymore after the 6 retries, but the jobs linger.
The use case is we want a single container to spin up, handle a single message, and then die. Each node supports one container right now (gpu bound). So we've arbitrarily chosen 6 max nodes, thus 6 max containers.
Do our settings now support that? We thought they would.
So code looks like:
Beta Was this translation helpful? Give feedback.
All reactions