-
Notifications
You must be signed in to change notification settings - Fork 212
Checkpointing plugins #3535
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
base: master
Are you sure you want to change the base?
Checkpointing plugins #3535
Conversation
drewoldag
left a comment
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Overall, this looks great and seems to be exactly what we would want to be able to programmatically avoid using cached/memoized data. I left one small question spurred by a comment in test_memoize_plugin.py, but it's not critical. Thanks a bunch for putting this together.
| # TODO: this .result() needs to be here, not in the loop | ||
| # because otherwise we race to complete... and then | ||
| # we might sometimes get a memoization before the loop | ||
| # and sometimes not... |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Is this generally something that users should keep in mind when working with caching and memoization in Parsl? Namely that there could be race conditions that crop up for functions that are in a loop and return very quickly?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
This will happen if you launch two apps with the same parameters at "the same time", where the same time means without waiting for one of them to complete:
- invoke my_app(7)
there's no checkpoint/memo so we submit it to be executed, as task 1
- invoke my_app(7)
there's no checkpoint/memo so we submit it to be executed, as task 2
-
task 1 completes and its result is memoized as my_app 7
-
task 2 completes and its result is memozied as my_app 7, replacing the result from step 3.
Maybe possible to implement to avoid this (again low priority for me) is more like Haskell thunks:
- invoke my_app(7)
there's no checkpoint/memo so we submit it to be executed, as task 1. we memoise its unpopulated future as my_app 7.
- invoke my_app(7)
there's a memo future for it (with no result yet) - so use that for the result
- task 1 completes, populating its result future.
This population of result future causes the 2nd invocation memo future to be populated too, and that completes.
so my_app is only run once.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Oh, ok. I see what you mean. From a selfish perspective, I don't think the extra effort of memoizing a future is necessary for my work.
Thanks a bunch for the explanation.
|
@drewoldag I'm interested to see any code you write that makes use of the interface - link it here if anything goes online. it doesn't need to look beautiful - i'm more interested in the realistic practical use. |
12b2168 to
1846145
Compare
|
by chance a different checkpoint related question/use case come up and I've added a bit more to this demo to address that. the question was can checkpointing optionally also checkpoint exceptions (so that they will be treated as permanent failures even across runs). almost all of the machinery to do that is already there - including the file format having a space for exceptions (that was unused and should have been tidied away long ago). This PR instead now makes use of that file format and lets a user specify policies for which completed apps should be checkpointed to disk. |
|
As mentioned by @benclifford, I wanted to cache certain exceptions as "valid" results so that these tasks would not be rerun. I've made some preliminary tests with this PR and it seems to work as expected. I used this custom Memoizer to cache tasks throwing SampleProcessingException or SampleValidationException: class ExceptionSafeMemoizer(BasicMemoizer):
def filter_for_checkpoint(self, app_fu):
# task record is available from app_fu.task_record
assert app_fu.task_record is not None
# Checkpoint either if there's a result, or if the exception is a
# SampleProcessingException or SampleValidationException:
exp = app_fu.exception()
# If an exception has occurred:
if exp is not None:
# Check if the exception type is one of our internal pipeline exceptions. If so, do cache this
# result, because this is expected behavior and we do not need to re-run this task in the future.
if isinstance( exp, SampleProcessingException ) or isinstance( exp, SampleValidationException ):
return True
else:
# If an unexpected exception occurred, do _not_ cache anything
return False
# If no exception occurred, cache result.
return True(Not sure about the line As mentioned on Slack, the only thing that was a bit inconvenient was that exceptions within the filter_for_checkpoint call - even code errors - would be silently ignored (i.e. they were only written into parsl.log, but they didn't communicate back to the main process called by the user or abort the program as I'd expect them to). |
|
a good other example of checkpointing is an out of memory store using sqlite3 - that would make a nice alternate implementation |
23ff9ce to
4726c81
Compare
697ee3d to
fd32707
Compare
fd32707 to
57fcfc6
Compare
57fcfc6 to
dbba605
Compare
6b0ac2b to
779b001
Compare
|
I made a blog post about this work: https://parsl-project.org/2025/09/30/checkpointing.html |
Thanks, I hope to check it out soon! Both database-checkpointing and the pluggable filter sound like they'll make my use-case a lot easier. |
3f2464a to
3e74939
Compare
d1d157f to
e1178be
Compare
This should not change behaviour and only move code around and add in more wiring. This is working towards all of checkpointing/memoizer code being plugable, with the DFK having no knowledge of the specifics - for example, the DFK no longer imports `pickle`, because that is now the business of the memoizer. Because of that goal, the parameters for Memoizer are arranged in two ways: Constructor parameters for Memoizer are parameters that the future user will supply at configuration when Memoizer is exposed as a user plugin, like other plugins. DFK internals that should be injected into the (eventually pluggable) memoizer (small m) are set as startup-time attributes, as for other plugins. Right now these two things happen right next to each other, but the Memoizer constructor call will move into user configuration space in a subsequent PR. As before this PR, there is still a separation between "checkpointing" and "memoization" that is slightly artificial. Subsequent PRs will merge these actions together more in this Memoizer implementation. This PR preserves a user-facing dfk.checkpoint() call, that becomes a passthrough to Memoizer.checkpoint(). I think that is likely to disappear in a future PR: manual checkpoint triggering will become a feature (or non-feature) of a specific memoizer implementation and so a method directly on that memoizer (or not). To go alongside the existing update_memo call, called by the DFK when a task is ready for memoization, this PR adds update_checkpoint, which captures the slightly different notion of a task being ready for checkpointing -- the current implementation can memoize a task that is not yet complete because it memoizes the future, not the result, while a checkpoint needs the actual result to be ready and so must be called later. This latter call happens in the race-condition-prone handle_app_update. A later PR will remove this distinction and move everything to around the same place as update_memo. See PR #3979 for description of fixing a related race condition related to update_memo. See PR #3535 for broader context.
This PR makes an interface class that defines the calls the DFK can make to the memoizer, distinct from the implementation of the implementation of the memoizer which is renamed to BasicMemoizer. The DFK still knows about how to construct a BasicMemoizer, but after that, self.memoizer is type-checked as the new Memoizer interface class. A future PR will separate out that configuration so that the user constructs a BasicMemoizer or some other plugin, using Parsl plugin style. At the same time, this new interface will have its checkpoint_queue method removed as being specific to the BasicMemoizer, rather than memoizer-generic.
Prior to this PR, checkpointing happened in the inherently race-prone handle_app_update where the user has already been informed that the task has completed. This means the user can observe a situation where checkpointing has not happened even though the task appears to have completed. Because the user-facing app future is now not set by the time of checkpointing, this PR changes the checkpointing code to pass results and exceptions explicitly as a CheckpointCommand object. This aligns with the on-disk checkpoint format and with the explicit separation of results and exceptions in earlier work - eg PR #4017 This new CheckpointCommand object contains support for representing exceptions because coming up in ~6 PRs time is support for choosing to checkpoint exceptions as well as results, as a user-configurable policy.
This follows the Future API, which has two completion variants, one for success and one for exception, and #4006 which introduces that distinction in DFK task completion helpers. Right now, the two new update_memo variants do the same thing. However, an upcoming PR will move checkpoint code out of handle_app_update (which is race-prone - #3762) and into update_memo which runs before the result/exception is exposed to the user. That checkpoint code will then need to see the result/exception explicitly.
This PR is not trying to simplify the whole nested structure,
just this one layer, as an incremental simplification.
It relies on:
if a:
x
else:
if b:
y
else:
z
being the same as:
if a:
x
elif b:
y
else:
z
This moves code around a little bit to make error handling in the dependency failure case more consistent with other failures. This unlocks potential for refactoring of the error handling code. Rought principles for error handling that are true for other errors and that this PR makes true for dependency failures: * _launch_if_ready_async creates a Future for the task, and does not modify the TaskRecord state or send monitoring information * handle_exec_update decides if a task has failed by whether the Future contains an exception or not. It updates the TaskRecord state and sends monitoring information. Removing one call of _send_task_log_info (in _launch_if_ready_async) and instead relying on the existing call in handle_exec_update removes a duplicate dep_fail entry in the monitoring database, that was previously unnoticed. $ pytest parsl/tests/test_python_apps/test_depfail_propagation.py --config parsl/tests/configs/htex_local_alternate.py Before this PR: sqlite> select * from status where task_status_name = 'dep_fail'; 1|dep_fail|2025-10-16 09:29:20.171145|1d65bd94-8639-4ebf-90f9-2b61cca36763|0 1|dep_fail|2025-10-16 09:29:20.171472|1d65bd94-8639-4ebf-90f9-2b61cca36763|0 3|dep_fail|2025-10-16 09:29:20.255851|1d65bd94-8639-4ebf-90f9-2b61cca36763|0 3|dep_fail|2025-10-16 09:29:20.256067|1d65bd94-8639-4ebf-90f9-2b61cca36763|0 4|dep_fail|2025-10-16 09:29:20.256353|1d65bd94-8639-4ebf-90f9-2b61cca36763|0 4|dep_fail|2025-10-16 09:29:20.256452|1d65bd94-8639-4ebf-90f9-2b61cca36763|0 5|dep_fail|2025-10-16 09:29:20.256732|1d65bd94-8639-4ebf-90f9-2b61cca36763|0 5|dep_fail|2025-10-16 09:29:20.256856|1d65bd94-8639-4ebf-90f9-2b61cca36763|0 7|dep_fail|2025-10-16 09:29:20.335936|1d65bd94-8639-4ebf-90f9-2b61cca36763|0 7|dep_fail|2025-10-16 09:29:20.336110|1d65bd94-8639-4ebf-90f9-2b61cca36763|0 8|dep_fail|2025-10-16 09:29:20.336398|1d65bd94-8639-4ebf-90f9-2b61cca36763|0 8|dep_fail|2025-10-16 09:29:20.336522|1d65bd94-8639-4ebf-90f9-2b61cca36763|0 After this PR: sqlite> select * from status where task_status_name = 'dep_fail'; 1|dep_fail|2025-10-16 09:31:35.782225|be2cd195-b13b-4b03-ac2f-95b73cc9a200|0 3|dep_fail|2025-10-16 09:31:35.865129|be2cd195-b13b-4b03-ac2f-95b73cc9a200|0 4|dep_fail|2025-10-16 09:31:35.865509|be2cd195-b13b-4b03-ac2f-95b73cc9a200|0 5|dep_fail|2025-10-16 09:31:35.865873|be2cd195-b13b-4b03-ac2f-95b73cc9a200|0 7|dep_fail|2025-10-16 09:31:35.945858|be2cd195-b13b-4b03-ac2f-95b73cc9a200|0 8|dep_fail|2025-10-16 09:31:35.946355|be2cd195-b13b-4b03-ac2f-95b73cc9a200|0
(These counts are not generally tested, but this PR is not introducing a general tests for them; instead it is about testing dependency handling behaviour more - based on regressions encountered during development of adjacent PRs)
The "else" case of the erstwhile-if statement was not defined properly: What are the circumstances and behaviour if an exec_fu is not set by this point? All don't-launch-yet behaviour is covered by `return` statements earlier on and the second part of the method should always proceed with making an exec_fu of some kind.
This PR adds _send_task_log_info as a mandatory part of task completion, which is already manually enforced. In one place, around line 500, one _send_task_log_info is split into two: one which then disappears into complete_task_result, and the other which remains in place for the non-factored error handling path. This PR renames _complete_task to _complete_task_result to reflect that it is for the result path not the exception path (taking naming from the Future concepts of the same name), and to open up the way for factorisation of the exception path into _complete_task_exception.
See PR #3999 which removes another one
This is set again 2 lines later, and the only potential for use is a call to update_task_state, which does not use time_returned. The time is set for the purposes of the subsequent _send_task_log_info.
This should not change behaviour and only move code around and add in more wiring. This is working towards all of checkpointing/memoizer code being plugable, with the DFK having no knowledge of the specifics - for example, the DFK no longer imports `pickle`, because that is now the business of the memoizer. Because of that goal, the parameters for Memoizer are arranged in two ways: Constructor parameters for Memoizer are parameters that the future user will supply at configuration when Memoizer is exposed as a user plugin, like other plugins. DFK internals that should be injected into the (eventually pluggable) memoizer (small m) are set as startup-time attributes, as for other plugins. Right now these two things happen right next to each other, but the Memoizer constructor call will move into user configuration space in a subsequent PR. As before this PR, there is still a separation between "checkpointing" and "memoization" that is slightly artificial. Subsequent PRs will merge these actions together more in this Memoizer implementation. This PR preserves a user-facing dfk.checkpoint() call, that becomes a passthrough to Memoizer.checkpoint(). I think that is likely to disappear in a future PR: manual checkpoint triggering will become a feature (or non-feature) of a specific memoizer implementation and so a method directly on that memoizer (or not). To go alongside the existing update_memo call, called by the DFK when a task is ready for memoization, this PR adds update_checkpoint, which captures the slightly different notion of a task being ready for checkpointing -- the current implementation can memoize a task that is not yet complete because it memoizes the future, not the result, while a checkpoint needs the actual result to be ready and so must be called later. This latter call happens in the race-condition-prone handle_app_update. A later PR will remove this distinction and move everything to around the same place as update_memo. See PR #3979 for description of fixing a related race condition related to update_memo. See PR #3535 for broader context.
Prior to this PR, the memoizer was told about the result in the AppFuture result callback. By the time this happens, the user workflow code can also observe that the task is completed and also observe that the memoizer has not got a copy of the result: for example, by invoking the same app again immediately and seeing a duplicate execution. The AppFuture callback shouldn't have anything in it that is related to user-observable task completion, but this PR does not remove other stuff in there. This PR moves update_memo earlier, to before setting the AppFuture result, so that the memoizer is strictly before future changes state.
This was introduced in 2018 in commit fd8ce2e already commented-out.
Prior to PR #2514, tasks were removed at completion when checkpointed, rather than when completed, in the case that checkpointing was enabled. This was race-y behaviour and was removed in PR #2414. This test also looks like it is subject to a related race condition around task completion, described in issue #1279, for which there is a sleep statement - this PR does not modify that but does add a reference to the issue.
This was introduced in 52b6b3a and looks like it was not used at that time either.
This comment talks about checkpointing taking priority over tasks, but it is unclear what priority for what? This comment was introduced in f2659c8 which implemented different checkpointing times.
update_memo is given the whole task record, which contains the relevant Future to be memoizing. The memoizer should never be being updated with a future that is not coming from a coherent task record with aligned hash sum and future.
…times) just like the periodic checkpointing is tests need to record the memoizer and invoke on that
and is more reusable when it isn't this isn't the only way to make a hash though. and hashing isn't the only way to compare checkpoint entries for equality.
because different memoizers (and different BasicMemoizer configs) won't necessarily have the identity semantics tested here
goal: results should not (never? in weak small cache?) be stored in an in-memory memo table. so that memo table should be not present in this implementation. instead all memo questions go to the sqlite3 database. this drives some blurring between in-memory caching and disk-based checkpointing: the previous disk based checkpointed model relied on repopulating the in-memory memo table cache... i hit some thread problems when using one sqlite3 connection across threads and the docs are unclear about what I can/cannot do, so i made this open the sqlite3 database on every access. that's probably got quite a performance hit, but its probably enough for basically validating the idea. the three usecases: sqlite3, in-memory and traditional on-disk checkpoint, maybe needs some thoughts about the completion semantics and data structures to go with it. TODO: this breaks: pytest parsl/tests/test_python_apps/test_memoize_exception.py --config parsl/tests/configs/htex_local_alternate.py because the behaviour asserted there is only behaviour for the default memoizer. So it should probably become a --config local test with known configuration(s).
e5ca7a9 to
dfe546d
Compare
This is the main breaking change of PR #3535 checkpoint plugins: this PR reworks construction of the memoizer so that the user now supplies an instance, plugin style, instead of the DFK constructing an instance. Memoization and configuration options, which turn out to be specific to the BasicMemoizer, are now supplied by the user directly to the instance they construct, rather than being passed through the DFK. Any checkpoint users will need to alter their Parsl configuration, although in a fairly straightforward way. If no memoizer is specified, the DFK constructs a BasicMemozier that preserves the no-parameters-supplied default behaviour: Memoization will occur, checkpointing will not occur. TODO: rewrite documentation that refers to any example checkpointing or app caching or memoization configuration with the documentation ready to receive documentation for the SQL plugin
This is the main breaking change of PR #3535 checkpoint plugins: this PR reworks construction of the memoizer so that the user now supplies an instance, plugin style, instead of the DFK constructing an instance. Memoization and configuration options, which turn out to be specific to the BasicMemoizer, are now supplied by the user directly to the instance they construct, rather than being passed through the DFK. Any checkpoint users will need to alter their Parsl configuration, although in a fairly straightforward way. If no memoizer is specified, the DFK constructs a BasicMemozier that preserves the no-parameters-supplied default behaviour: Memoization will occur, checkpointing will not occur.
This is the main breaking change of PR #3535 checkpoint plugins: this PR reworks construction of the memoizer so that the user now supplies an instance, plugin style, instead of the DFK constructing an instance. Memoization and configuration options, which turn out to be specific to the BasicMemoizer, are now supplied by the user directly to the instance they construct, rather than being passed through the DFK. Any checkpoint users will need to alter their Parsl configuration, although in a fairly straightforward way. If no memoizer is specified, the DFK constructs a BasicMemozier that preserves the no-parameters-supplied default behaviour: Memoization will occur, checkpointing will not occur.
This is the main breaking change of PR #3535 checkpoint plugins: this PR reworks construction of the memoizer so that the user now supplies an instance, plugin style, instead of the DFK constructing an instance. Memoization and configuration options, which turn out to be specific to the BasicMemoizer, are now supplied by the user directly to the instance they construct, rather than being passed through the DFK. Any checkpoint users will need to alter their Parsl configuration, although in a fairly straightforward way. If no memoizer is specified, the DFK constructs a BasicMemozier that preserves the no-parameters-supplied default behaviour: Memoization will occur, checkpointing will not occur. # Changed Behaviour User facing checkpointing configuration breaks completely and users will need to rewrite their checkpointing configuration. ## Type of change - New feature
Description
this is most immediately in the context of issue #3534 - but @WardLT might also be interested
this very rough PR:
moves more checkpoint/memo out of the data flow kernel into the existing memoizer implementation class
makes the DFK use an abstract Memoizer class, with the existing implementation now in BasicMemoizer
adds a test to demo perhaps for isse Method for overriding cached/checkpointed results #3534 showing how checkpoint/memo lookup can look at the args of the function it has been passed to decide whether to ask an underlying BasicMemoizer to look up a result, or to not return a memoized result without even asking the basic memoizer
This PR is intended for experimentation with this kind of API, but a lot of it drives towards a cleaner codebase and so for the most part should find its way into
masterbranchType of change
Choose which options apply, and delete the ones which do not apply.