-
Notifications
You must be signed in to change notification settings - Fork 19
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Benchmarking results post-processing #70
Comments
@t-young31 has done more work on this in the DiRAC project than I had realised. As far as I can understand, he has already written a lot of the code needed for parsing the perflogs and plotting data for different use cases. His feedback was
|
Need to save spack spec as well. There's a hash for every spec, save it in the perflog as a field, as well as the spec in a separate file (name same as hash for simplicity) which will be updated only when something changes. |
Different ways to print results in the perflog:
After some tests, to see what's printed to the perflog for parameters and various |
For posterity, some of Tom's DiRAC work is in Lokesh's repo - the reading in the perflog part. |
When a benchmark is run, it generates a number of outputs: perflogs and environment logs. Those files have to be moved to an appropriate location (see #16 ), and then we'd like to provide some generic tools/scripts for the users to process them and produce tables and plots. The intention is for the scripts provided in this repository to be quite basic, and the user would clone/fork it and add their own. Scripts needed:
sysname/partition/environment/testname
- seeexcalibur-tests/modules/utils.py
Line 90 in 953bc6a
testname
includes the name of the app and the parameters used to run the specific benchmark.prefix
, just aboveformat
in thehandlers_perflog
block - see below for details and links.excalibur-tests/reframe_config.py
Line 393 in 87dde00
The content of this file is most of what we need to present in a performance report, things like date the benchmark was run (for regression testing purposes), the number of nodes and threads used, the benchmark parameters (possibly mangled into some info string, but it's there), and the FOM values for each test.
The text was updated successfully, but these errors were encountered: