Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Benchmark flexibility design discussion #85

Open
mmilenkoski opened this issue Apr 27, 2020 · 1 comment
Open

Benchmark flexibility design discussion #85

mmilenkoski opened this issue Apr 27, 2020 · 1 comment

Comments

@mmilenkoski
Copy link
Contributor

As the number of things we want to modify in the benchmarks increase, it is necessary that we discuss how do we handle this situation. On the one hand, it is important that we have a consistent benchmark that is exactly the same every time. On the other hand, we would like to allow users to try out and compare different backends, optimizers, etc. Introducing this flexibility raises design questions in the design of the benchmarks, but also in the design of the dashboard, CLI, and other components. I am opening this issue in order for it to serve as a place for discussing these design choices.

Some of the questions we need to address are:

  1. Will we create separate images for the official benchmark and one intended for experimentation?
  2. For the case of the experimentation image, how do we handle the parameter selection in the dashboard and the CLI?
  3. How do we handle the passing of the selected parameters to the experimentation image?

Feel free to propose solutions and add new questions as they come up.

@martinjaggi
Copy link
Member

**kwargs ?
none needed for the official benchmark (i.e. the reference implementation we provide for each task). we can still offer the official benchmark params with hardcoded params.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants