[Python-Dev] [Python-checkins] BAD Benchmark Results for Python Default 2016-01-26

Stefan Krah stefan at bytereef.org
Thu Jan 28 08:49:50 EST 2016


IMO the timings of the benchmark suite are a bit unstable -- this is not the
fault of Intel's setup, I noticed it also when running the suite myself.


On Tue, Jan 26, 2016 at 06:48:54PM +0000, Stewart, David C wrote:
> Wow, what happened to Python default to cause such a regression?
> 
> 
> 
> 
> On 1/26/16, 7:31 AM, "lp_benchmark_robot" <lp_benchmark_robot at intel.com> wrote:
> 
> >Results for project Python default, build date 2016-01-26 03:07:40 +0000
> >commit:		cbd4a6a2657e
> >previous commit:	f700bc0412bc
> >revision date:	2016-01-26 02:54:37 +0000
> >environment:	Haswell-EP
> >	cpu:		Intel(R) Xeon(R) CPU E5-2699 v3 @ 2.30GHz 2x18 cores, stepping 2, LLC 45 MB
> >	mem:		128 GB
> >	os:		CentOS 7.1
> >	kernel:	Linux 3.10.0-229.4.2.el7.x86_64
> >
> >Baseline results were generated using release v3.4.3, with hash b4cbecbc0781
> >from 2015-02-25 12:15:33+00:00
> >
> >----------------------------------------------------------------------------------
> >              benchmark   relative   change since   change since   current rev run
> >                          std_dev*       last run       baseline          with PGO
> >----------------------------------------------------------------------------------
> >:-)           django_v2      0.21%         -2.93%          8.95%            16.19%
> >:-|             pybench      0.10%          0.05%         -1.87%             5.40%
> >:-(            regex_v8      2.72%         -0.02%         -4.67%             4.57%
> >:-|               nbody      0.13%         -0.92%         -1.33%             7.40%
> >:-|        json_dump_v2      0.20%          0.87%         -1.59%            11.48%
> >:-|      normal_startup      0.90%         -0.57%          0.10%             5.35%
> >----------------------------------------------------------------------------------
> >* Relative Standard Deviation (Standard Deviation/Average)
> >
> >If this is not displayed properly please visit our results page here: http://languagesperformance.intel.com/bad-benchmark-results-for-python-default-2016-01-26/
> >
> >Note: Benchmark results are measured in seconds.
> >
> >Subject Label Legend:
> >Attributes are determined based on the performance evolution of the workloads
> >compared to the previous measurement iteration.
> >NEUTRAL: performance did not change by more than 1% for any workload
> >GOOD: performance improved by more than 1% for at least one workload and there
> >is no regression greater than 1%
> >BAD: performance dropped by more than 1% for at least one workload and there is
> >no improvement greater than 1%
> >UGLY: performance improved by more than 1% for at least one workload and also
> >dropped by more than 1% for at least one workload
> >
> >
> >Our lab does a nightly source pull and build of the Python project and measures
> >performance changes against the previous stable version and the previous nightly
> >measurement. This is provided as a service to the community so that quality
> >issues with current hardware can be identified quickly.
> >
> >Intel technologies' features and benefits depend on system configuration and may
> >require enabled hardware, software or service activation. Performance varies
> >depending on system configuration.
> _______________________________________________
> Python-checkins mailing list
> Python-checkins at python.org
> https://mail.python.org/mailman/listinfo/python-checkins



More information about the Python-Dev mailing list