Asynchronous Workflows¶
Asynchronous workflows describe situations where the loop between measurement and recommendation is more complex and needs to incorporate various other aspects. These could for instance be:
Distributed workflows: When recommendations are distributed across several operators, e.g. at different locations or in several reactors, some experiments might have been started, but are not ready when the next batch of recommendations is requested. Without further consideration, the algorithm would be likely to recommend the pending experiments again (since they were and still are considered most promising), as it is unaware they were already started.
Partial targets: When dealing with multiple targets that require very different amounts of time to measure, the targets of previously recommended points might only be partially available when requesting the next batch of recommendations. Still, these partial experiments should ideally be considered when generating the recommendations.
With pending experiments we mean experiments whose measurement process has been started, but not yet completed by time of triggering the next set of recommendations – this is typically the case when at least one of the configured targets has not yet been measured.
There are two levels of dealing with such situations:
Marking experiments as pending: If an experiment is not completed (meaning at least one target is not yet measured), its data cannot be added as a regular measurement. However, it can be marked as pending via
pending_experiments
inrecommend
.Adding partial results: If an experiment is partially completed (meaning at least one target has been measured), we can already update the model with the available information by adding a partial measurement.
Marking Experiments as Pending¶
To avoid repeated recommendations in the above scenario, BayBE provides the
pending_experiments
keyword. It is available wherever recommendations can be
requested, i.e. Campaign.recommend
or
RecommenderProtocol.recommend
.
Supported Acquisition Functions
pending_experiments
is only supported by Monte Carlo (MC) acquisition functions, i.e. the
ones that start with a q
in their name. Attempting to use a non-MC acquisition
function with pending_experiments
will result in an
IncompatibleAcquisitionFunctionError
.
Supported Recommenders
For technical reasons, not every recommender is able to make use of pending_experiments
.
For instance,
BotorchRecommender
takes all pending experiments into account, even if they do not match exactly with
points in the search space.
By contrast,
Non-predictive recommenders like
SKLearnClusteringRecommender
s,
RandomRecommender
or
FPSRecommender
do not consider pending_experiments
at all and raise an
UnusedObjectWarning
when such points are passed.
Akin to measurements
or recommendations
, pending_experiments
is a dataframe in
experimental representation.
In the following example, we get a set of recommendations, add results for half of them,
and start the next recommendation, marking the other half pending:
from baybe.utils.dataframe import add_fake_measurements
# Get a set of 10 recommendation
rec = campaign.recommend(batch_size=10)
# Split recommendations into two parts
rec_finished = rec.iloc[:5]
rec_pending = rec.iloc[5:]
# Add target measurements to the finished part. Here we add fake results
add_fake_measurements(rec_finished, campaign.targets)
campaign.add_measurements(rec_finished)
# Get the next set of recommendations, incorporating the still unfinished experiments.
# These will not include the experiments marked as pending again.
rec_next = campaign.recommend(10, pending_experiments=rec_pending)
Adding Partial Results¶
A partial result is possible if you have multiple targets, but only measured the outcome for some of those. This is a common occurrence, especially if the different target measurements correspond to experiments that differ in complexity or duration.
As a simple example, consider a campaign with medical background aimed at creating a drug formulation. Typically, there are quick initial analytics performed on the formulation, followed by in vitro experiments followed by mouse in vivo experiments. Without the ability to use partial measurements, you would have to wait until the slow mouse experiment for a given recommendation is measured until you could utilize any of the other (faster) experimental outcomes for that recommendation. Furthermore, if the fast measurements are already unpromising, the slower target measurements are possibly never performed at all.
In BayBE, you can leverage results even if they are only partial. This is indicated by setting the corresponding target measurement value to NaN. There are several ways to indicate this, e.g.:
Let us consider this 3-batch of recommendations, assuming we need to measure “Target_1”, “Target_2” and “Target_3”:
import numpy as np
import pandas as pd
rec = campaign.recommend(batch_size=3)
# Resetting the index to have easier access via .loc later
measurements = rec.reset_index(drop=True)
# Add measurement results
measurements.loc[0, "Target_1"] = 10.3
measurements.loc[0, "Target_2"] = 0.5
measurements.loc[0, "Target_3"] = 11.1
measurements.loc[1, "Target_1"] = 7.1
measurements.loc[1, "Target_2"] = np.nan # not measured yet
measurements.loc[1, "Target_3"] = 12.2
measurements.loc[2, "Target_1"] = 11.4
measurements.loc[2, "Target_2"] = pd.NA # not measured yet
measurements.loc[2, "Target_3"] = None # not measured yet
measurements
# Proceed with campaign.add_measurements ...
Param_1 |
Param_2 |
… |
Target_1 |
Target_2 |
Target_3 |
---|---|---|---|---|---|
on |
1.1 |
… |
10.3 |
0.5 |
11.1 |
on |
3.8 |
… |
7.1 |
NaN |
12.2 |
off |
2.9 |
… |
11.4 |
NaN |
NaN |
Internally, the incomplete rows are dropped when fitting a surrogate model for each target. If you use an unsupported surrogate model, an error will be thrown at runtime.
Limitations
The described method only works if the surrogate model uses a separate data basis
for each target. This is e.g. the case if you use the
CompositeSurrogate
to enable multi-output modeling required by the
ParetoObjective
. For details, see
multi-output modeling.
The DesirabilityObjective
does
not currently utilize multi-output models and hence does not support partial results.