-
Notifications
You must be signed in to change notification settings - Fork 91
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Adding independent restarts as additional trials #1117
Comments
Not sure that "restarts" and their triggers should be in focus. On Tue, Jul 26, 2016 at 3:48 PM, nikohansen [email protected]
|
The instance-wise normalization could even be considered as a bug-fix, as this is how it should work if not all instances have the same number of repetitions. |
FTR, when we use simulated restarts, the number of instances should not cause any troubles, because in this case the number of samples is fixed independently of the number of instances. EDIT: see however #2042, and uniform sampling is still an issue. |
A simple way to address the problem of different number of repetitions of instances would be to copy the data of underrepresented instances within the data set, such that all instances have the same number of repetitions. EDIT: the option |
I wonder about the current status? I believe we can read in any data but we do not normalize by per-instance repetition? See #1972. EDIT: balancing by data duplication can now be toggled via |
The new experimental script adds trials instead of within-trial restarts. The number of trials for each instance is the same. The postprocessing balances instance if their count is unbalanced. |
We should consider to allow, or even encourage, to do additional trials of the very same instance instead of within-trial restarts. This would also have the advantage that restarts do not need to be logged additionally (see e.g. #1073).
First step: the post-processing must normalize the ECDF, the average RT, etc. by the number of same-instance repetitions. (EDIT: done and can be toggled via
genericsettings.balance_instances
) The normalization must be done such that first each (distinct) instance gets the same weight and then each problem gets the same weight in the result. In other words: we need to make sure that the post-processing can deal with data which have a different number of instances and/or same-instance repetitions appropriately.Remark: this might also be a good step in the direction where instances are not anymore conceptually considered as the very same function / problem. This might indeed already be the better approach for (some functions in) the
bbob-biobj
test suite (see also #1018).Open questions regarding the rule on the experimental setup: what are the eligible triggers for such additional trials, apart from the current budget? Could the trigger depend on the function or even the instance number without generating potential problems?
EDIT: for simulated restarts, there are two ways of instance data selection which lead potentially to different results:
Which one is better justifiable?
EDIT: changing the experimental recommendation for conditional instance repetitions can (and is somewhat likely to) "break" any other code that displays results, because an unweighted aggregation over all instances (e.g. an average) is now biased.
[DONE] EDIT: to avoid various issues, it seems advisable to repeat all instances of the same problem the same number of times. In this case, the used trigger seems rather uncritical.
The text was updated successfully, but these errors were encountered: