blob: 96b9902d92d692b050e249cf31ea09a6b9b0468d [file] [log] [blame]
Polling: ``PollExecutor``
=========================
PollExecutor allows producing futures which should be resolved via a
custom poll function.
.. _poll function:
Poll function
~~~~~~~~~~~~~
The poll function has the following semantics:
- It's called with a single argument, a list of zero or more
:class:`~more_executors.PollDescriptor` objects.
- If the poll function can determine that a particular future should be
completed, either successfully or in error, it should call the
:meth:`yield_result` or :meth:`yield_exception` methods on the appropriate
:class:`~more_executors.PollDescriptor`.
- If the poll function raises an exception, all futures depending on that poll
will fail with that exception. The poll function will be retried later.
- If the poll function returns an int or float, it is used as the delay
in seconds until the next poll.
.. warning::
The poll function should avoid holding a reference to the corresponding
:class:`~more_executors.PollExecutor`. If it holds a reference to the
executor, invoking :meth:`~concurrent.futures.Executor.shutdown` becomes
mandatory in order to stop the polling thread.
.. _cancel function:
Cancel function
~~~~~~~~~~~~~~~
The cancel function has the following semantics:
- It's called when cancel is requested on a future returned by this executor,
if and only if the future is currently in the list of futures being polled,
i.e. it will not be called if the delegate future has not yet completed.
- It's called with a single argument: the value returned by the delegate
future.
- It should return `True` if the future can be cancelled.
- If the cancel function raises an exception, the future's cancel method will
return `False` and a message will be logged.
Example
~~~~~~~
Consider a web service which executes tasks asynchronously, with an API like
this:
* To create a task: `POST https://myservice/object/:id/publish`
* Returns an identifier for a task, such as `{"task_id": 123}`
* To get status of a single task: `GET https://myservice/tasks/:id`
* Returns task status, such as `{"task_id": 123, "state": "finished"}`
* To cancel a single task: `DELETE https://myservice/tasks/:id`
* To search for status of multiple tasks:
`POST https://myservice/tasks/search {"task_id": [123, 456, ...]}`
* Returns array of task statuses, such as:
.. code-block:: JSON
[{"task_id": 123, "state": "finished"},
{"task_id": 456, "state": "error", "error": "..."}]
Imagine we want to run many tasks concurrently and monitor their status.
Two obvious approaches include:
* We could make a function which starts a task and polls that task until
it completes, and have several threads run that function for different tasks.
* This could be easily encapsulated as a future using only the executors
from :mod:`concurrent.futures`, but is very inefficient - each task in
progress will consume a thread for polling.
* We could maintain a list of all outstanding tasks and implement a
special-purpose function to wait for all of them to complete.
* This is efficient as only one request is needed per poll interval, but
since tasks aren't represented as futures, useful features from
:mod:`concurrent.futures` cannot be used and different types of
asynchronous work will be handled inconsistently.
:class:`~more_executors.PollExecutor` bridges the gap between the two
approaches. By providing a custom poll function, the tasks can be monitored
by one thread efficiently, while at the same time allowing full usage of the
API provided by `Future` objects.
In this example, the poll function would look similar to this:
.. code-block:: python
def poll_tasks(poll_descriptors):
# Collect tasks we'll need to search for
task_ids = [d.result.get('task_id')
for d in poll_descriptors]
# Search for status of these tasks
search = {"task_id": task_ids}
response = requests.post(
'https://myservice/tasks/search', json=search)
# If request failed then all outstanding futures fail
response.raise_for_status()
# Request passed, resolve futures accordingly
tasks = {task.get('task_id'): task.get('state')
for task in response.json()}
for d in poll_descriptors:
task = tasks.get(d.result, {})
state = task.get('state')
if state == 'finished':
d.yield_result()
elif state == 'error':
d.yield_exception(TaskFailed())
To ensure that calling `future.cancel()` also cancels tasks on the
remote service, we can also provide a cancel function:
.. code-block:: python
def cancel_task(task):
task_id = task['task_id']
# Attempt to DELETE this task
response = requests.delete(
'https://myservice/tasks/%s' % task_id)
# Succeeded only if response was OK.
# Otherwise, we may have been too late to cancel.
return response.ok
With the poll and cancel functions in place, futures could be created
like this:
.. code-block:: python
def publish(object_id):
response = requests.post(
'https://myservice/object/%s/publish' % object_id)
response.raise_for_status()
return response.json()
# Submit up to 4 HTTP requests at a time,
# and poll for the returned tasks using our poll function.
executor = Executors.\
threadpool(max_workers=4).\
with_poll(poll_tasks, cancel_task)
futures = [executor.submit(publish, x)
for x in (10, 20, 30)]
# Now we can use concurrent.futures API to get tasks as
# the poll function determines they've completed
for completed in as_completed(futures):
# ...
.. autoclass:: more_executors.PollExecutor
:members:
.. automethod:: more_executors.Executors.with_poll(executor, poll_fn, cancel_fn=None, default_interval=5.0, logger=None)
.. autoclass:: more_executors.PollDescriptor()
:members: