Warning in options(stringsAsFactors = config$as_factors):
'options(stringsAsFactors = TRUE)' is deprecated and will be disabled
Here’s where I’ll keep a running tab of the incremental code updates and fixes
Oct 3
Fixed empty runway bug
I’ve fixed the error thrown when trying to evalueate the furthest occupied pot in an extinct replicate. This uses a new function, last_occupied_pot()
, in helpers.R
, which is called from within the model function but can also be called from the analysis code (see example usage in make_Ler_sims.R
). The optional argument zero
allows the return value for an empty replicate to be something other than zero (to avoid the risk of zero-index values).
I ran make_Ler_sims
successfully, and the warnings indicated that it ran into 26 instances of an extinct replicate.
Code profiling
I’ve profiled the model using make_Ler_sims
. The results are in logs/make_Ler_sims-2022-10-03.Rprofvis
(not on git as it’s large!)
- Total time in
iterage_genotype()
: 139300 ms - The bulk of that is in
seed_sampling()
: 134360 - Within that, the bulk is in the
aaply
call torgengamma
(25k), and theaaply
s todisp_table
(50k each) - Within
dist_table()
, 75k is in the lineraw_table <- table(dists)
. This is more than half the total time!
So the first thing to do is to find a faster replacement for table
. This StackOverflow page suggests that tabulate()
(in base R) is about 14 times faster than table()
, which would bring this operation down to 5-6k ms. According to the help page, “tabulate is the workhorse for the table function.” So the time is spent in all the robustness of table. It may take some work to figure out exactly how to call tabulate.
Note that this is not one of the issues I flagged in my manual code review!
Oct 4
Code profiling
I’ve changed disp_table()
to use tabulate
rather than table
, and confirmed that they return the same results. Microbenchmarking (see test_disp_table()
) suggests that there is a 100-fold speedup in that function!
Profiling make_Ler_sims.R
with this change now speeds things up a lot (make_Ler_sims-2022-10-04.Rprofvis). About half of the total time is in calculating the gengamma random numbers, which doesn’t make use of the internal vectorizing, instead using aaply
to go across parameters. Table building now takes about 40% of total time, but less than 10% of that seems to be actually inside disp_table
; apparently aaply
has a huge performance overhead! It seems that the base apply
may be much faster, but I wonder whether the bugs I was trapping by using .drop = FALSE
will reappear. It might also be that those commented-out aperm
statements were applied to an earlier attempt at using apply
.
Oct 5
General
I updated my home mac to R 4.4.1, and updated all libraries.
I found the code to revert blogdown posts to their old structure, and added it to the project’s .Rprofile
. I also added the option not to automatically rebuild the website every time a file is saved, as that is problematic if there are code errors and also takes nontrivial time with the (uncached) ProjectTemplate loading.
Bugs
I removed the empirical = TRUE
arguments to mvrnorm()
.
Efficiency
Let’s look at replacing aaply
with apply
. Things to check include the shape of the returned function and what happens to dimensions with value 1. The output uses up lots of space so I don’t show it; the comments below summarize the results of running the code.
xx <- array(1:24, dim = c(3,4,2))
# Scalar function return
y1 <- aaply(xx, c(1,2), sum)
y2 <- apply(xx, c(1,2), sum)
y1
y2
class(y1)
class(y2)
# Vector function return
y1 <- aaply(xx, c(1,2), function(x) c(sum(x), prod(x)))
y2 <- apply(xx, c(1,2), function(x) c(sum(x), prod(x)))
y1
y2
aperm(y2, c(2, 3, 1))
class(y1)
class(y2)
xx <- array(1:8, dim = c(1,4,2))
xx
aaply(xx, c(1,2), function(x) c(sum(x), prod(x)), .drop = FALSE)
apply(xx, c(1,2), function(x) c(sum(x), prod(x))) %>% aperm(c(2, 3, 1))
aaply(xx, c(1,2), function(x) c(sum(x), prod(x)))
xx <- array(1:8, dim = c(4,1,2))
xx
aaply(xx, c(1,2), function(x) c(sum(x), prod(x)), .drop = FALSE)
apply(xx, c(1,2), function(x) c(sum(x), prod(x))) %>% aperm(c(2, 3, 1))
aaply(xx, c(1,2), function(x) c(sum(x), prod(x)))
xx <- array(1:8, dim = c(2,4,1))
xx
aaply(xx, c(1,2), function(x) c(sum(x), prod(x)), .drop = FALSE)
apply(xx, c(1,2), function(x) c(sum(x), prod(x))) %>% aperm(c(2, 3, 1))
aaply(xx, c(1,2), function(x) c(sum(x), prod(x)))
- When the function returns a scalar, the outcomes are the same, except that rows and columns from
apply
are unnamed. The latter should not be an issue for me. - When the function returns a vector, that vector is put into the leading dimension of the returned array by
apply
, whereas it remains in its original position underaaply
. This is clearly why I had thoseaperm
statements. - When one of the dimensions being parsed has value 1, then
aaply
with the defaults drops a dimension. Butapply
by itself seems fine.
So it seems that apply %>% aperm
should be a drop-in replacement for aaply(.drop = FALSE)
. So why did I make the switch in the first place? Was I just trying to be modern, or does aaply
have some other advantage? I made the change on 5/29/19, in a commit that has massive numbers of code updates, with no comments on individual changes. It may just be that I liked the elegance of skipping the aperm
step.
I’ve tried adding the apply version back in (I had to fix the order of indices in aperm); compared with aaply, it’s about 2.5 times faster. It’s still the case that most of that time is in the overhead. apply
has a lot of cases, but if I’ve parsed it correctly, in my use case it’s doing a for
loop over the nreps x npots cases. Further, rather than calling the function directly in that loop, it’s calling something called forceAndCall
which is trapping for closures, which may add more time. And it’s hard to know how much is being added by aperm. This post suggests that apply is as fast as a for loop, but the system times were very crude (0.02 sec vs. 0.00 sec). I run a microbenchmark based on their example:
obs <- array(rnorm(10*13*12), c(10,13,12))
floop <- function(obs) {
l2 = dim(obs)[2]
l3 = dim(obs)[3]
minObs<-matrix(nrow=dim(obs)[2],ncol=dim(obs)[3])
for (i in 1:l2)
for (j in 1:l3)
minObs[i,j]<-min(obs[,i,j],na.rm = TRUE)
return(minObs)
}
mbm <- microbenchmark(aaply(obs, c(2,3), min), floop(obs), apply(obs, c(2,3), min))
Unit: microseconds
expr min lq mean median uq max neval
aaply(obs, c(2, 3), min) 26991.818 27617.330 29557.6689 29839.2335 30930.627 35212.618 100
floop(obs) 219.243 245.001 285.7647 264.8650 303.561 439.593 100
apply(obs, c(2, 3), min) 281.425 317.083 353.9656 341.7015 372.688 573.770 100
In this case, there’s only a 20% speedup of the for loop vs apply. So maybe it’s in the aperm call?
mbm2 <- microbenchmark(apply(obs, c(2,3), min), apply(obs, c(2,3), min) %>% aperm(c(2,1)))
Unit: microseconds
expr min lq mean median uq max neval
apply(obs, c(2, 3), min) 273.826 288.7125 324.2954 299.926 339.1565 482.853 100
apply(obs, c(2, 3), min) %>% aperm(c(2, 1)) 293.200 310.4655 345.1462 324.005 343.2050 554.648 100
Here there’s basically no difference. So the overhead is all in the for loop, and we probably can’t get away from it without working out a direct vectorization. But it is curious that the speedup from aaply to apply in the microbenchmark is so much greater than it is in the profile.
At any rate, I finalize the code with apply and aperm.
Oct 10
Efficiency
I just profiled apply
vs aaply
for the random number generation, and the former takes about half as long. So I’ll go with that.
Next, I compared a single call to rgengamma with a long vector of parameters to the apply call. Even with for loops before and after to simulate preparing and parsing the data, there was more than a 10-fold speedup (see test_rgengamma()
). So it seems worthwhile to pursue this.
Haven’t quite gotten there, but here are some code snippets that point where I’m heading:
disp_seeds <- matrix(c(5, 4, 0, 2, 10, 3), nrow = 2, byrow = TRUE)
QQ <- matrix(1:6, nrow = 2, byrow = TRUE)
x <- array(c(disp_seeds, QQ), dim = c(2,3,2))
y <- unlist(apply(x, c(2,1), function(x) rep(x[2], x[1]), simplify = FALSE))
max_ds <- max(disp_seeds) # to set the array dimension to pad to
yy <- array(0, dim = c(2,3,max_ds))
Then a double for loop to fill in appropriate parts of yy with appropriate parts of y. Note that the apply has the dimensions reversed to ensure that y comes out row-wise.
Oct 11
Efficiency
OK, I’ve implemented the revised code. The savings aren’t as dramatic as microbenchmark had suggested (perhaps because of memory issues?); maybe a 2-fold speedup. But it’s something!
Oct 13
Code inconsistencies
I’ve rationalized the way dispersal tails are handled by making det_kernel
simply calculate out to the value of controls$max_pots
and getting rid of the reflection in seed_sampling
. We still need a rationalized value of maximum dispersal distance.
Conceptual issues
I’ve added a value to pass into iterate_genotype
to use for the temporal component of environmental stochasticity in seed production, with a sensible default. This then gets passed through to ES_seeds
. This runs with both test_iterate_genotype
and with an appropriately updated make_Ler_sims
Oct 14
Documentation
- Cleaned up comments
- Fixed equation derivation in
gompertz_seeds()
- Added documentation for missing input list values
Code structure
I have separated the model inputs into plant_params
, expt_params
, and sim_settings
. Everything is appropriately propagated through model.R
, and I have updated test_iterate_genotype()
to work with the new structure. I’ll need to update some of the other test functions, as well as the simulation drivers.
I did realize on doing this that I’m often passing the entire list through simply to get a single value; in those cases I should simplify things.
- I’ve pulled
n_pots
out of the control arrays, as it is only used internally and is updated dynamically. - I’ve simplified the call to
gapify()
- Simplified call to
seed_sampling()
- Simplified call to
det_kernel()
- Simplified call to
kernel_stoch()
- Simplified call to
DS_seeds()
- Simplified call to
ES_seeds()
, and fixed mistake in application of ES_seed_time - Simplified call to
Gompertz_seeds()
- That’s everything for eliminating unessential list-passing!
New bugs
- There seems to be an issue with
apply()
dropping an array dimension if there is only one replicate that hasn’t gone extinct.
Oct 17
I updated make_Ler_sims to work with the updated version of the model. I also add a recording of the run number, so that we can get means and variances for each run.
I have started a new library script, viz.R
, with functions to provide visualizations. I have made a quick modification of the Ler plots to show violin plots of the multiple runs.
New bugs
In trying to run the visualization code, I was getting errors about on out-of-date grouping scheme for LerC_spread. I tried re-munging, but that threw an error in 04-Ler_dispersal_stats.R
. So apparently the package updated have left something non-functional in the munge code that I’ll need to update!