summaryrefslogtreecommitdiff
diff options
context:
space:
mode:
authorAndré Nusser <andre.nusser@googlemail.com>2020-02-02 15:53:20 +0100
committerAndré Nusser <andre.nusser@googlemail.com>2020-02-02 15:53:20 +0100
commitf11258049b9f3676e4d11c498e590d983f488e3e (patch)
treec0aa3ff56034c87ed5fcd9d8c63a49561f75c33f
parentc40280699582d85be898c6feba44e8d0674e3219 (diff)
Add a new section.
-rw-r--r--sampling_alg_lac2020/LAC-20.tex6
1 files changed, 6 insertions, 0 deletions
diff --git a/sampling_alg_lac2020/LAC-20.tex b/sampling_alg_lac2020/LAC-20.tex
index fc58394..c8d0230 100644
--- a/sampling_alg_lac2020/LAC-20.tex
+++ b/sampling_alg_lac2020/LAC-20.tex
@@ -369,10 +369,16 @@ We already explained the core part of the sample selection algorithm. The remain
Note that the worst-case complexity of evaluating the objective function is linear in the number of samples for the instrument that we are considering. However, in practice we can avoid evaluation for most samples by simply starting with the \enquote{most promising} sample and recursively evaluate the neighbors until the future possible evaluations cannot beat the currently best value.
+\section{Emulation Capabilities}
+\todo{Talk about which other algorithms we are a general case of, i.e., which algorithms can we emulate using the right power values and parameter settings.}
+
\section{Implementation}
\todobent{Give a short introduction to DrumGizmo, including a link to the git repository.}
\todo{Talk about the timeline, i.e., when were the releases and what is still unreleased?}
+
\todoandre{Talk about how the sampling algorithm was implemented. What do we need to store?}
+
+
\todoandre{Add some of the source code to the paper?}
\todoandre{Give less important implementation details, e.g., like adaptive search starting from the most promising value}