Re: [Speed] Should we change what benchmarks we have?

We on the Pyston team have created some new benchmarks which I can recommend using; I wouldn't call them "macrobenchmarks" since they don't test entire applications, but we've found them to be better than the existing benchmarks, which tend to be quite microbenchmarky. For example, our django-templating benchmark actually exercises the django templating system, as opposed to bm_django.py which just tests unicode concatenation. You can find them here https://github.com/dropbox/pyston-perf/tree/master/benchmarking/benchmark_su... The current ones we look at are django_template3_10x, sqlalchemy_imperative2_10x, and pyxl_bench_10x.
On Thu, Feb 11, 2016 at 10:36 AM, Brett Cannon <brett@python.org> wrote:
Are we happy with the current benchmarks? Are there some we want to drop? How about add? Do we want to have explanations as to why each benchmark is included? A better balance of micro vs. macro benchmarks (and probably matching groups)?
Speed mailing list Speed@python.org https://mail.python.org/mailman/listinfo/speed
participants (1)
-
Kevin Modzelewski