Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Implement data prefetch #194

Open
jpsamaroo opened this issue Jan 21, 2021 · 0 comments · May be fixed by #199
Open

Implement data prefetch #194

jpsamaroo opened this issue Jan 21, 2021 · 0 comments · May be fixed by #199

Comments

@jpsamaroo
Copy link
Member

If we send input data to a node (that we know is going to execute a thunk in the future) before its thunk is ready to execute, and cache it on that node, then we can execute that thunk much quicker (assuming network transfers are fully asynchronous and don't impede other thunk executions). We should allow the scheduler to do a small amount of this "prefetching" of data when it has a large amount of data associated with a thunk.

We'll need to be able to pre-allocate a processor for each thunk, implement a worker-local cache of received inputs, and make thunks check this cache for their inputs before moving them. We should also start modeling the memory availability for a given worker, and the memory costs of inputs and estimated max memory allocations of each thunk.

@jpsamaroo jpsamaroo linked a pull request Jan 28, 2021 that will close this issue
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

Successfully merging a pull request may close this issue.

1 participant