
On 27 Mar 2022, at 18:16, Jonathan Fine <jfine2358@gmail.com> wrote:
Hi
Thank you Inada for your prompt and helpful reply. Here's a link for cached hash in bytes object: https://bugs.python.org/issue46864 <https://bugs.python.org/issue46864>
What I have in mind is making selected objects smaller, for example by using smaller pointers. But how to know the performance benefit this will give?
That will limit python to 2GiB or maybe 4GiB of memory - I routinely run beyond that size in production systems. There is a memory model that GCC supports that is 32bit pointers and 64bit ints. I do not recall the performance comparisons, buts it is not used very much.
I think it would be helpful to know how much SLOWER things are when we make Python objects say 8 or 16 bytes LARGER. This would give an estimate of the improvement from making all Python objects smaller.
I've not done much performance testing before. Anyone here interested in doing it, or helping me do it? (Warning - I've never built Python before.)
Performance tests is hard to get right. I do this as my day job for a big Cloud app written mostly in Python. There is an excellent book on performance measurement by Brendan Greg "Systems Performance; Enterprise and the Cloud". Barry
with best regards
Jonathan _______________________________________________ Python-ideas mailing list -- python-ideas@python.org To unsubscribe send an email to python-ideas-leave@python.org https://mail.python.org/mailman3/lists/python-ideas.python.org/ Message archived at https://mail.python.org/archives/list/python-ideas@python.org/message/PMDNJF... Code of Conduct: http://python.org/psf/codeofconduct/