Skip to content
This repository has been archived by the owner on Jan 9, 2020. It is now read-only.

Additional config settings for the memory overhead factor #481

Open
wants to merge 1 commit into
base: branch-2.2-kubernetes
Choose a base branch
from

Conversation

ash211
Copy link

@ash211 ash211 commented Sep 6, 2017

I'm seeing the default value of 0.10 fail for even reasonably-sized shuffle
jobs so expect this value to require some tuning to reliably succeed.

We copied this default value from YARN but it appears that kubernetes is more
strict on enforcing memory limits on containers than YARN has been: I have
two identically configured clusters of five AWS r3.4xls, one running YARN and
the other running kubernetes, with identical driver/executor settings, running
identical jobs, and the YARN job succeeds whereas the k8s job fails due to the
pod exceeding its memory limit.

I'm seeing the default value of 0.10 fail for even reasonably-sized shuffle
jobs so expect this value to require some tuning to reliably succeed.

We copied this default value from YARN but it appears that kubernetes is more
strict on enforcing memory limits on containers than YARN has been:  I have
two identically configured clusters of five AWS r3.4xls, one running YARN and
the other running kubernetes, with identical driver/executor settings, running
identical jobs, and the YARN job succeeds whereas the k8s job fails due to the
pod exceeding its memory limit.
@foxish
Copy link
Member

foxish commented Sep 7, 2017

To be consistent with YARN, maybe we should do memoryOverhead instead. The memory factor would make the computed value depend on another argument, driverMemory, which isn't ideal IMO.

Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

3 participants