asv_bench_memray0.1.2
asv_bench_memray0.1.2
Published
A memray benchmark plugin for asv
pip install asv-bench-memray
Package Downloads
Authors
Project URLs
Requires Python
>=3.8
asv_bench_memray
This a proof-of-concept externally defined memray
benchmark plugin for asv
.
Usage
In your asv.conf.json
add the following to the requirements matrix:
"matrix": {
"req": {
"numpy": ["1.21"],
"asv_plugin_memray": [""]
},
},
This should be enough for an example like the following to run:
import numpy as np
class MyBenchmark:
params = [10, int(2e4)]
def ray_sum(self, n):
self.data = np.random.rand(n)
np.sum(self.data)
Which should result in something roughly like:
ASV_RUNNER_PATH="../asv_runner" asv run --bench sum
· Creating environments.......
· Discovering benchmarks
·· Uninstalling from virtualenv-py3.9-asv_bench_memray.._asv_bench_memray-numpy1.21
·· Building 30a650cb <main> for virtualenv-py3.9-asv_bench_memray.._asv_bench_memray-numpy1.21.
·· Installing 30a650cb <main> into virtualenv-py3.9-asv_bench_memray.._asv_bench_memray-numpy1.21
· Running 1 total benchmarks (1 commits * 1 environments * 1 benchmarks)
[ 0.00%] · For project commit 30a650cb <main>:
[ 0.00%] ·· Benchmarking virtualenv-py3.9-asv_bench_memray.._asv_bench_memray-numpy1.21
[100.00%] ··· benchmarks.MyBenchmark.ray_sum ok
[100.00%] ··· ======== =======
param1
-------- -------
10 1.72k
20000 162k
======== =======
ASV_RUNNER_PATH="../asv_runner" asv run --bench sum 8.72s user 3.04s system 99% cpu 11.877 total
Development
During development, one needs to pass a relative path:
"matrix": {
"req": {
"asv_bench_memray": ["../asv_bench_memray"]
},
},
Also it may be necessary to cleanup older environments and explicitly point to an instance of asv_runner
:
rm -rf .asv && ASV_RUNNER_PATH="../asv_runner" asv run --bench sum
Personally I keep benchmarks for testing at an experiments repo, but any local test will do.
Conventions
Like all externally defined benchmark plugins for asv
, this has a strict hierarchy.
- The package name begins with
asv_bench
- Benchmarks are defined in a
benchmarks
folder under the package module - Each exported new benchmark type has the
export_as_benchmark = [NAMEBenchmark]
attribute
For more conventions follow the documention of asv_runner
.
License
MIT.