-
Notifications
You must be signed in to change notification settings - Fork 1
Update visualization and data processing modules #181
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
for more information, see https://pre-commit.ci
While a time filter is great to have in general, specifically solving for that issue is better done through filtering of the |
Cool - looking forward to a follow-up that alleviates the hard-coded LBL machine ID and uses the entire database (of a particular database version, of course) |
|
More questions: for which seems to be what |
for more information, see https://pre-commit.ci
|
Thanks for the feedback @CodyCBakerPhD ! Based on your feedback and some further discussion with @oruebel, I have updated several plots and added captions to help with interpretability.
Good point, I will open up an issue or separate PR fix to add the database_version filtering parameter to the main CLI. I think we may have not updated the database version when we updated the LINDI files, so updating now to filter by the latest database version may exclude some of our results. But I would have to double check.
Yes, planning to address the hardcoded machine ID this week to include the dandihub runs. At the very least I will open up a related issue to use the entire database.
These outliers seem to be a systemic issue that we see with the S3-based fsspec cases and is also present in the network tracking results. Now that we have results from older package versions, it could also be a package version issue. I like the idea of varying the shade based on either time of run or version info, I will open up as an issue to address in the future.
These colors represented the different streaming methods. However after further consideration it doesn't make as much sense to separate out the individual packages into their own subplots, since we did not run tests that specifically controlled for one package version, but rather we ran environments at different timepoints across which multiple package versions could have changed. The new
The new "slicing_with_extrapolation" plot should match more with the initial conceptualization we had.
I've updated the captions and the plot names to help with interpretability. Let me know how that works or if I can add any additional clarifications. |
for more information, see https://pre-commit.ci
|
Latest results figures can be found in the google drive here |



Add a couple updates to the visualization and data processing modules:
These updates should now generate figures that cover all the main questions outlined in #89, though some refinement is needed before publication.
To generate the figures locally, run
nwb_benchmarks generate_figures. This will use the latest results stored in the~/.cache/nwb-benchmarks/nwb-benchmarks-results/folder, or will clone if it does not exist, or you can optionally specify a different results folder to use. Ideally we could add additional functionality to provide user-specific recommendations and figure reports, but I would address that in a separate PR and will leave this with the default LBL Mac results filtering for now.