Samples are grouped by Treatment and Collection time with the following group sizes:
Time | Treatment | n |
---|---|---|
4 | CTRL | 3 |
4 | SOC_TRT1 | 4 |
4 | SOC_TRT1_TRT2 | 3 |
4 | SOC_TRT1_TRT3 | 4 |
4 | SOC_TRT1_TRT3_TRT2 | 4 |
8 | SOC_TRT1 | 3 |
8 | SOC_TRT1_TRT2 | 3 |
8 | SOC_TRT1_TRT3 | 4 |
8 | SOC_TRT1_TRT3_TRT2 | 4 |
Total number of samples is: 32
Samples are to be blocked in batches for scRNA-seq.
This data set is also used in the nested dimensions example. Here, we focus on using different methods for the optimization.
We allocate surplus positions in the batch container and some excluded positions to check that all optimization methods support empty container positions.
# Setting up the batch container
bc <- BatchContainer$new(
dimensions = c(
batch = ceiling(nrow(multi_trt_day_samples) / 8),
run = 2, position = 5
),
exclude = tibble::tibble(batch = 4, run = c(1, 2), position = c(5, 5))
) |>
# Add samples to container
assign_in_order(samples = multi_trt_day_samples)
bc
#> Batch container with 38 locations and 32 samples (assigned).
#> Dimensions: batch, run, position
The samples are distributed to 4 batches (processing days). We use
the osat scoring on sample Treatment
and Time
,
using first a shuffling protocol with a fixed number of sample swaps on
each iteration.
Note that doing 32 swaps on 38 free container positions does not make sense, since each swapping operation affects two different positions anyway. The upper limit is reduced to the max number of meaningful swaps (19) on the fly.
Optimization finishes after the list of permutations is exhausted.
n_shuffle <- rep(c(32, 10, 5, 2, 1), c(20, 40, 40, 50, 50))
scoring_f <- osat_score_generator(c("batch"), c("Treatment", "Time"))
bc1 <- optimize_design(
bc,
scoring = scoring_f,
n_shuffle = n_shuffle # will implicitly generate a shuffling function according to the provided schedule
)
#> Re-defined number of swaps to 19 in swapping function.
#> Warning in osat_score(bc, batch_vars = batch_vars, feature_vars = feature_vars,
#> : NAs in features / batch columns; they will be excluded from scoring
#> Checking variances of 1-dim. score vector.
#> ... (27.944) - OK
#> Initial score: 73.03
#> Achieved score: 21.241 at iteration 1
#> Achieved score: 19.767 at iteration 3
#> Achieved score: 19.241 at iteration 7
#> Achieved score: 16.82 at iteration 8
#> Achieved score: 12.715 at iteration 10
#> Achieved score: 10.82 at iteration 11
#> Achieved score: 10.504 at iteration 46
#> Achieved score: 10.083 at iteration 67
#> Achieved score: 8.083 at iteration 103
#> Achieved score: 6.083 at iteration 116
#> Achieved score: 4.504 at iteration 120
#> Achieved score: 2.925 at iteration 121
bc1$trace$elapsed
#> Time difference of 0.80814 secs
Custom plot with some colours:
bc1$scores_table() |>
dplyr::mutate(
n_shuffle = c(NA, n_shuffle)
) |>
ggplot2::ggplot(
ggplot2::aes(step, value, color = factor(n_shuffle))
) +
ggplot2::geom_point() +
ggplot2::labs(
title = "Score 1 tracing",
subtitle = stringr::str_glue("Final score = {bc1$score(scoring_f)}"),
x = "Iteration",
y = "Score",
color = "n_shuffle"
)
Using the internal method…
We may safely apply the batch container methods get_samples() and score() also after using the new optimization code.
bc1$score(scoring_f)
#> score_1
#> 2.925208
bc1$get_samples(assignment = TRUE) |>
dplyr::filter(!is.na(Treatment)) |>
dplyr::mutate(anno = stringr::str_c(Time, " hr")) |>
ggplot2::ggplot(ggplot2::aes(x = batch, y = interaction(position, run), fill = Treatment)) +
ggplot2::geom_tile(color = "white") +
ggplot2::geom_hline(yintercept = 5.5, size = 1) +
ggplot2::geom_text(ggplot2::aes(label = anno)) +
ggplot2::labs(x = "Batch", y = "Position . Run")
#> Warning: Using `size` aesthetic for lines was deprecated in ggplot2 3.4.0.
#> ℹ Please use `linewidth` instead.
#> This warning is displayed once every 8 hours.
#> Call `lifecycle::last_lifecycle_warnings()` to see where this warning was
#> generated.
Further optimization (using a different shuffling protocol maybe) can be done immediately on the same batch container.
Starting optimization from scratch, we are passing now some stopping criteria that may terminate optimization before a shuffling protocol has been exhausted.
For demonstration, we use a shuffling function now that will do 3 sample (position) swaps per iteration and can be called an arbitrary number of times. Thus, iteration has to be stopped by either the max_iter criterion or by reaching a specific minimum delta threshold (score improvement from one selected solution to the next).
bc2 <- optimize_design(
bc,
scoring = scoring_f,
n_shuffle = 3, # will implicitly generate a shuffling function that will do 3 swaps at each iteration
max_iter = 2000,
min_delta = 0.1
)
#> Checking variances of 1-dim. score vector.
#> ... (32.089) - OK
#> Initial score: 73.03
#> Achieved score: 59.03 at iteration 1
#> Achieved score: 47.873 at iteration 2
#> Achieved score: 46.188 at iteration 3
#> Achieved score: 34.083 at iteration 4
#> Achieved score: 31.978 at iteration 5
#> Achieved score: 30.715 at iteration 6
#> Achieved score: 24.715 at iteration 7
#> Achieved score: 23.136 at iteration 8
#> Achieved score: 22.82 at iteration 15
#> Achieved score: 16.925 at iteration 16
#> Achieved score: 14.925 at iteration 17
#> Achieved score: 14.609 at iteration 24
#> Achieved score: 12.609 at iteration 31
#> Achieved score: 10.609 at iteration 44
#> Achieved score: 8.609 at iteration 45
#> Achieved score: 6.609 at iteration 53
#> Achieved score: 6.504 at iteration 120
#> Achieved score: 4.504 at iteration 152
#> Achieved score: 4.504 at iteration 175
#> Reached min delta in 175 iterations.
Instead of passing a single scoring function, a list of multiple scoring functions can be passed to the optimizer, each of which to return a scalar value on evaluation.
By default, a strict improvement rule is applied for classifying a potential solution as “better”: each of the individual scores has to be smaller than or equal to its previous value, and one of the scores has to be changed.
However, the user could specify other methods for aggregating the scores or defining the acceptance criterion. See later examples.
The second scoring function used here is by the way rather redundant and just serves for illustration.
multi_scoring_f <- list(
osat_score_generator(c("batch"), c("Treatment", "Time")),
osat_score_generator(c("batch"), c("Treatment"))
)
bc3 <- optimize_design(
bc,
scoring = multi_scoring_f,
n_shuffle = 3,
max_iter = 200,
min_delta = 0.1
)
#> Warning in osat_score(bc, batch_vars = batch_vars, feature_vars = feature_vars,
#> : NAs in features / batch columns; they will be excluded from scoring
#> Warning in osat_score(bc, batch_vars = batch_vars, feature_vars = feature_vars,
#> : NAs in features / batch columns; they will be excluded from scoring
#> Checking variances of 2-dim. score vector.
#> ... (24.179, 70.116) - OK
#> Initial score: c(73.03, 44.803)
#> Achieved score: c(47.03, 30.803) at iteration 1
#> Achieved score: c(37.452, 21.54) at iteration 2
#> Achieved score: c(31.452, 19.54) at iteration 7
#> Achieved score: c(21.662, 11.645) at iteration 9
#> Achieved score: c(15.662, 9.645) at iteration 14
#> Achieved score: c(14.083, 6.488) at iteration 15
#> Achieved score: c(10.083, 4.488) at iteration 34
#> Achieved score: c(7.978, 4.488) at iteration 40
#> Achieved score: c(7.978, 4.488) at iteration 180
#> Reached min delta in 180 iterations.
Note that the first score tends to yield higher values than the second one. This could be a problem when trying to select a solution based on an aggregated, overall score. We repeat the same optimization now by using the autoscaling functionality of the optimizer.
We’re just adding the autoscale_scores
option here to
estimate the distribution of individual scores on a number of completely
random sample assignments (200 in this case) and then apply a
transformation to rescale each score to a standard normal.
Note that by ‘normalizing’ the distribution of the scores we obtain values centered around zero, thus that the optimized scores are likely to be negative. We may also want to decrease the delta_min parameter to match the new numerical range.
bc3_as <- optimize_design(
bc,
scoring = multi_scoring_f,
n_shuffle = 3,
max_iter = 200,
min_delta = 0.01,
autoscale_scores = T,
autoscaling_permutations = 200
)
#> Checking variances of 2-dim. score vector.
#> ... (33.181, 48.223) - OK
#> Creating autoscaling function for 2-dim. score vector. (200 random permutations)
#> ... Performing boxcox lambda estimation.
#> Initial score: c(5.454, 2.669)
#> Achieved score: c(3.722, 1.125) at iteration 1
#> Achieved score: c(2.531, -0.305) at iteration 3
#> Achieved score: c(1.212, -1.342) at iteration 5
#> Achieved score: c(-0.928, -2.067) at iteration 12
#> Achieved score: c(-1.925, -2.067) at iteration 15
#> Achieved score: c(-2.536, -2.59) at iteration 25
#> Achieved score: c(-4.089, -2.59) at iteration 50
#> Achieved score: c(-4.902, -3.691) at iteration 81
Having directly comparable scores, it may be reasonable now to use a function that somehow aggregates the scores to decide on the best iteration (instead of looking at the scores individually).
An easy way to do this is to use the built-in worst_score function. This will simply set the aggregated score to whichever of the individual scores is larger (i.e. ‘worse’ in terms of the optimization).
bc4 <- optimize_design(
bc,
scoring = multi_scoring_f,
n_shuffle = 3,
aggregate_scores_func = worst_score,
max_iter = 200,
autoscale_scores = TRUE,
autoscaling_permutations = 200
)
#> Checking variances of 2-dim. score vector.
#> ... (30.658, 69.544) - OK
#> Creating autoscaling function for 2-dim. score vector. (200 random permutations)
#> ... Performing boxcox lambda estimation.
#> Initial score: c(7.755, 2.838)
#> Aggregated: 7.755
#> Achieved score: c(5.666, 1.362) at iteration 1
#> Aggregated: 5.666
#> Achieved score: c(3.297, 0.646) at iteration 2
#> Aggregated: 3.297
#> Achieved score: c(2.657, 0.646) at iteration 3
#> Aggregated: 2.657
#> Achieved score: c(0.939, 1.135) at iteration 5
#> Aggregated: 1.135
#> Achieved score: c(-0.117, 0.753) at iteration 6
#> Aggregated: 0.753
#> Achieved score: c(0.707, 0.592) at iteration 8
#> Aggregated: 0.707
#> Achieved score: c(0.252, -0.395) at iteration 9
#> Aggregated: 0.252
#> Achieved score: c(-0.978, -0.768) at iteration 11
#> Aggregated: -0.768
#> Achieved score: c(-1.412, -2.558) at iteration 16
#> Aggregated: -1.412
#> Achieved score: c(-1.794, -2.465) at iteration 23
#> Aggregated: -1.794
#> Achieved score: c(-2.772, -2.465) at iteration 34
#> Aggregated: -2.465
#> Achieved score: c(-2.772, -3.109) at iteration 98
#> Aggregated: -2.772
#> Achieved score: c(-3.195, -3.569) at iteration 107
#> Aggregated: -3.195
#> Achieved score: c(-3.648, -4.144) at iteration 147
#> Aggregated: -3.648
#> Achieved score: c(-4.28, -4.144) at iteration 178
#> Aggregated: -4.144
Another - more interesting - option would be to aggregate the two scores by taking their sum. This way both scores will influence the optimization at every step.
For illustration, we omit the n_shuffle
parameter here,
which will lead by default to pairwise sample swaps being done on each
iteration.
bc5 <- optimize_design(
bc,
scoring = multi_scoring_f,
aggregate_scores_func = sum_scores,
max_iter = 200,
autoscale_scores = TRUE,
autoscaling_permutations = 200
)
As a final example, we calculate the (squared) L2 norm to actually aggregate the two scores. Not that this choice is not really motivated in this case, but it could be used if optimization was carried on meaningful distance vectors or normalized n-tuples.
Note that we don’t use the auto-scaling in this case as the L2-norm based optimization would force both normalized scores towards zero, not the minimal (negative) value that would be desired in that case.
It is recommended to use the n_shuffle
parameter to
steer the optimization protocol. However, you may also provide a
dedicated shuffling function that on each call has to return a shuffling
order (as integer vector) or a list with the source and destination
positions (src and dst) of the sample positions to be swapped.
The following example uses a template for creating complete random shuffles across all available positions in the batch container. Note that this is usually not a good strategy for converging to a solution.
bc6 <- optimize_design(
bc,
scoring = scoring_f,
shuffle_proposal_func = complete_random_shuffling,
max_iter = 200
)
#> Checking variances of 1-dim. score vector.
#> ... (28.918) - OK
#> Initial score: 73.03
#> Achieved score: 18.504 at iteration 1
#> Achieved score: 14.504 at iteration 2
#> Achieved score: 12.399 at iteration 35
#> Achieved score: 12.083 at iteration 87
#> Achieved score: 11.136 at iteration 150
#> Achieved score: 10.82 at iteration 154
Esp. for very large search spaces, better solutions can be quite successfully obtained by a SA protocol which allows the optimizer to jump over ‘energy barriers’ to more likely converge at lower local minima.
The optimizer usually remembers the permutation with the best overall score to start with, but this behavior can be changed by supplying a simulated annealing protocol, most simply by generating a ready-made function template.
It is generally recommended for SA to make small changes at each step, like allowing just 1 sample swap per iteration.
Currently the simulated annealing protocol requires a single double value score to be optimized. Choose an appropriate aggregation function if you happen to have multiple scores initially.
bc7 <- optimize_design(
bc,
scoring = scoring_f,
n_shuffle = 1,
acceptance_func = mk_simanneal_acceptance_func(),
max_iter = 200
)
#> Checking variances of 1-dim. score vector.
#> ... (31.241) - OK
#> Initial score: 73.03
#> Achieved score: 65.03 at iteration 1
#> Achieved score: 57.03 at iteration 2
#> Achieved score: 47.03 at iteration 3
#> Achieved score: 41.03 at iteration 4
#> Achieved score: 41.03 at iteration 5
#> Achieved score: 34.925 at iteration 6
#> Achieved score: 26.925 at iteration 7
#> Achieved score: 28.925 at iteration 8
#> Achieved score: 28.925 at iteration 9
#> Achieved score: 24.925 at iteration 10
#> Achieved score: 22.925 at iteration 11
#> Achieved score: 22.925 at iteration 12
#> Achieved score: 18.925 at iteration 13
#> Achieved score: 17.241 at iteration 15
#> Achieved score: 15.241 at iteration 16
#> Achieved score: 17.241 at iteration 17
#> Achieved score: 19.241 at iteration 18
#> Achieved score: 19.241 at iteration 19
#> Achieved score: 19.241 at iteration 20
#> Achieved score: 19.241 at iteration 21
#> Achieved score: 19.241 at iteration 22
#> Achieved score: 19.241 at iteration 23
#> Achieved score: 19.241 at iteration 25
#> Achieved score: 19.241 at iteration 26
#> Achieved score: 19.241 at iteration 27
#> Achieved score: 17.241 at iteration 28
#> Achieved score: 17.241 at iteration 29
#> Achieved score: 17.241 at iteration 30
#> Achieved score: 15.241 at iteration 32
#> Achieved score: 11.241 at iteration 35
#> Achieved score: 11.241 at iteration 37
#> Achieved score: 9.346 at iteration 38
#> Achieved score: 7.767 at iteration 39
#> Achieved score: 7.767 at iteration 40
#> Achieved score: 7.767 at iteration 42
#> Achieved score: 6.083 at iteration 43
#> Achieved score: 4.504 at iteration 45
#> Achieved score: 4.504 at iteration 49
#> Achieved score: 4.504 at iteration 54
#> Achieved score: 4.609 at iteration 55
#> Achieved score: 4.609 at iteration 60
#> Achieved score: 4.609 at iteration 73
#> Achieved score: 4.609 at iteration 74
#> Achieved score: 4.609 at iteration 88
#> Achieved score: 4.609 at iteration 91
#> Achieved score: 4.609 at iteration 101
#> Achieved score: 4.609 at iteration 104
#> Achieved score: 4.609 at iteration 118
The trace may show a non strictly monotonic behavior now, reflecting the SA protocol at work.
Better results and quicker convergence may be achieved by playing with the starting temperature (T0) and cooling speed (alpha) in a specific case.
bc8 <- optimize_design(
bc,
scoring = scoring_f,
n_shuffle = 1,
acceptance_func = mk_simanneal_acceptance_func(mk_simanneal_temp_func(T0 = 100, alpha = 2)),
max_iter = 150
)
#> Checking variances of 1-dim. score vector.
#> ... (29.462) - OK
#> Initial score: 73.03
#> Achieved score: 65.03 at iteration 1
#> Achieved score: 59.03 at iteration 2
#> Achieved score: 51.03 at iteration 3
#> Achieved score: 51.03 at iteration 4
#> Achieved score: 47.03 at iteration 6
#> Achieved score: 47.03 at iteration 7
#> Achieved score: 43.346 at iteration 8
#> Achieved score: 33.346 at iteration 9
#> Achieved score: 31.346 at iteration 13
#> Achieved score: 27.346 at iteration 14
#> Achieved score: 25.346 at iteration 15
#> Achieved score: 25.346 at iteration 17
#> Achieved score: 25.346 at iteration 18
#> Achieved score: 25.346 at iteration 19
#> Achieved score: 25.241 at iteration 20
#> Achieved score: 25.241 at iteration 23
#> Achieved score: 23.241 at iteration 24
#> Achieved score: 17.241 at iteration 25
#> Achieved score: 17.241 at iteration 32
#> Achieved score: 11.241 at iteration 33
#> Achieved score: 9.662 at iteration 49
#> Achieved score: 7.662 at iteration 57
#> Achieved score: 6.083 at iteration 88
#> Achieved score: 4.188 at iteration 146
bc8$plot_trace()
The following example puts together all possible options to illustrate the flexibility of the optimization.
n_shuffle <- rep(c(3, 2, 1), c(20, 20, 200))
bc9 <- optimize_design(
bc,
scoring = list(
osat_score_generator(c("batch"), c("Treatment", "Time")),
osat_score_generator(c("batch"), c("Treatment")),
osat_score_generator(c("batch"), c("Time"))
),
n_shuffle = n_shuffle,
aggregate_scores_func = sum_scores,
acceptance_func = mk_simanneal_acceptance_func(mk_simanneal_temp_func(T0 = 500, alpha = 1)),
max_iter = 200,
min_delta = 1e-8,
autoscale_scores = T
)
#> Warning in osat_score(bc, batch_vars = batch_vars, feature_vars = feature_vars,
#> : NAs in features / batch columns; they will be excluded from scoring
#> Warning in osat_score(bc, batch_vars = batch_vars, feature_vars = feature_vars,
#> : NAs in features / batch columns; they will be excluded from scoring
#> Warning in osat_score(bc, batch_vars = batch_vars, feature_vars = feature_vars,
#> : NAs in features / batch columns; they will be excluded from scoring
#> Checking variances of 3-dim. score vector.
#> ... (37.344, 67.436, 112.389) - OK
#> Creating autoscaling function for 3-dim. score vector. (100 random permutations)
#> ... Performing boxcox lambda estimation.
#> Initial score: c(7.322, 2.702, 3.971)
#> Aggregated: 13.994
#> Achieved score: c(5.557, 2.137, 3.139) at iteration 1
#> Aggregated: 10.834
#> Achieved score: c(3.697, 0.657, 2.676) at iteration 2
#> Aggregated: 7.03
#> Achieved score: c(2.269, 1.656, 1.337) at iteration 3
#> Aggregated: 5.262
#> Achieved score: c(1.644, 0.748, 0.709) at iteration 4
#> Aggregated: 3.101
#> Achieved score: c(2.269, 2.033, 0.219) at iteration 5
#> Aggregated: 4.522
#> Achieved score: c(2.269, 2.033, -0.321) at iteration 6
#> Aggregated: 3.982
#> Achieved score: c(0.366, 1.025, -0.67) at iteration 7
#> Aggregated: 0.721
#> Achieved score: c(0.088, -0.103, -0.485) at iteration 8
#> Aggregated: -0.5
#> Achieved score: c(-0.93, -1.245, -0.434) at iteration 9
#> Aggregated: -2.609
#> Achieved score: c(-0.178, -1.245, 0.832) at iteration 10
#> Aggregated: -0.591
#> Achieved score: c(0.2, 0.333, 0.344) at iteration 11
#> Aggregated: 0.876
#> Achieved score: c(0.2, 0.591, 0.967) at iteration 12
#> Aggregated: 1.758
#> Achieved score: c(0.2, 0.591, 0.303) at iteration 13
#> Aggregated: 1.094
#> Achieved score: c(-0.547, 0.291, 0.303) at iteration 14
#> Aggregated: 0.047
#> Achieved score: c(0.181, 0.291, 0.303) at iteration 15
#> Aggregated: 0.775
#> Achieved score: c(0.181, 0.291, 0.303) at iteration 16
#> Aggregated: 0.775
#> Reached min delta in 16 iterations.
bc9$plot_trace()