Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Measure one project 3x times #44

Open
martong opened this issue Nov 3, 2020 · 3 comments
Open

Measure one project 3x times #44

martong opened this issue Nov 3, 2020 · 3 comments
Labels
enhancement New feature or request

Comments

@martong
Copy link
Collaborator

martong commented Nov 3, 2020

It is an accepted measurement method to repeat the experiments at least 3x and then take the averages of the values. Would it be possible to implement such a config option? This could be really informative with respect to memory consumption and runtime cost evaluation.

@martong martong added the enhancement New feature or request label Nov 3, 2020
@Xazax-hun
Copy link
Owner

Yeah, I think this makes sense. A question is, how to handle such a measurement? We do not want to take the average of every run. Some statistics, like the number of bugs found, or the number of nodes in the exploded graph supposed to be the same across runs. Maybe we can also validate how deterministic the analyzer is.

@Xazax-hun
Copy link
Owner

I know this is not the best, but as a workaround it is possible to have 3 identical configurations and do the average by hand (or by a script).

@martong
Copy link
Collaborator Author

martong commented Dec 17, 2020

Yeah, I think this makes sense. A question is, how to handle such a measurement? ... We do not want to take the average of every run.

I think we should be able to specify how many times we want to measure the projects (or maybe a fine-grained per project option would be useful too).
IMO, for most statistic we need at least 3 values (min, average, max). But for some other values (e.g. run-time, peak resident memory usage) it could be useful to see the percentiles as well. Maybe it would be a nice option to be able to display these values with a candlestick in the charts.

Some statistics, like the number of bugs found, or the number of nodes in the exploded graph supposed to be the same across runs.

Yes, that's right, we should report any discrepancies.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
enhancement New feature or request
Projects
None yet
Development

No branches or pull requests

2 participants