[Distutils] Provisionally accepting PEP 517's declarative build system interface
Donald Stufft
donald at stufft.io
Mon May 29 16:48:30 EDT 2017
> On May 29, 2017, at 4:04 PM, Nathaniel Smith <njs at pobox.com> wrote:
>
> Ugh, sorry, fat-fingered that. Actual reply below...
>
> On Mon, May 29, 2017 at 12:56 PM, Nathaniel Smith <njs at pobox.com> wrote:
>> On Mon, May 29, 2017 at 12:50 PM, Donald Stufft <donald at stufft.io> wrote:
>>>
>>> To be honest, I’m not hardly going to feel particularly bad if one of the
>>> most compilation heavy packages that exist takes a whole 10 seconds to
>>> install from a VCS checkout.
>
> Rebuild latency is *really* important. People get really cranky at me
> when I argue that we should get rid of "editable installs", which
> create much greater problems for maintaining consistent environments,
> and that's only saving like 1 second of latency. I think I'm entitled
> to be cranky if your response is "well suck it up and maybe rewrite
> all your build tools”.
Well, distutils literally already has support for storing the “cache" someplace other than the current directory, the current directory is just the default. So “rewrite all your build tools” is fairly hyperbolic, it’s really just “change the default of your build tools”. See for example: https://gist.github.com/dstufft/a577c3c9d54a3bb3b88e9b20ba86c625 <https://gist.github.com/dstufft/a577c3c9d54a3bb3b88e9b20ba86c625> which shows that Numpy &tc are already capable of this.
Hell, the build backend could create an unpacked sdist in the target directory instead of an actual sdist that is already packed into a tarball, tools like twine could add a ``twine sdist`` command that just called the “create unpacked sdist” API and then just tar’d up the directory into the sdist. A quick rudimentary test on my machine (using ``python setup.py sdist —formats=`` in a numpy checkout [1]) suggests that this entire process takes ~0.7s which the copy operation on that same check out (shutil.copytree) also takes ~0.7s. That also eliminates the need to untar so unless someone is doing something in their sdist creation step that takes a significant amount of time, generating an unpacked sdist is really not any more time consuming than copying the files.
>
> NumPy really isn't that compilation heavy either... it's all C, which
> is pretty quick. SciPy is *much* slower, for example, as is pretty
> much any project using C++.
>
>>> Particularly when I assume that the build tool
>>> can be even smarter here than ccache is able to be to reduce the setup.py
>>> build step back down to the no-op incremental build case.
>>>
>>> I mean, unless numpy is doing something different, the default distutils
>>> incremental build stuff is incredibly dumb, it just stores the build output
>>> in a directory (by default it’s located in ./build/) and compares the mtime
>>> of a list of source files with the mtime of the target file, and if the
>>> sources files are newer, it recompiles it. If you replace mtime with blake2
>>> (or similar) then you can trivially support the exact same thing just
>>> storing the built target files in some user directory cache instead.
>
> Cache management is not a trivial problem.
>
> And it actually doesn't matter, because we definitely can't silently
> dump stuff into some user directory. An important feature of storing
> temporary artifacts in the source tree is that it means that if
> someone downloads the source, plays around with it a bit, and deletes
> it, then it's actually gone. We can't squirrel away a few hundred
> megabytes of data in some hidden directory that will hang around for
> years after the user stops using numpy.
I mean, you absolutely can do that. We store temporary wheels and HTTP responses silently in pip and have for years. I don’t think *anyone* has *ever* complained about it. I think macOS even explicitly will clean up stuff from ~/Library/Caches when it hasn’t been used in awhile. If you use the standard cache locations for Linux then IIRC similar systems exist for Linux too. In exchange for “I can delete the directory and it’s just all gone”, you get “faster builds in more scenarios, including straight from PyPI’s sdists”. If I were a user I’d care a lot more about the second then the first.
But even if I grant you that you can’t just do that silently, then go ahead and make it opt in. For people who need it, a simple boolean in a config file seems to be pretty low cost to me.
Combine this user cache with generating an unpacked sdist instead of copying the directory tree, and you get:
1) Safety from weirdness that comes from ``pip install`` from a sdist versus a VCS.
2) Not crapping up ./ with random debris from the installation process.
3) Fast incremental builds that even help speed up installs from PyPI etc (assuming we use something like blake2 to compute hashes for the files).
And you lose:
1) Deleting a clone doesn’t delete the cache directory, but your OS might already be managing this directory anyways.
Seems like an obvious trade off to me.
>
>>> Hell,
>>> we *might* even be able to preserve mtime (if we’re not already… we might
>>> be! But I’d need to dig into it) so literally the only thing that would need
>>> to change is instead of storing the built artifacts in ./build/ you store
>>> them in ~/.cache/my-cool-build-tool/{project-name}. Bonus points: this means
>>> you get incremental speeds even when building from a sdist from PyPI that
>>> doesn’t have wheels and hasn’t changed those files either.
>>>
>>> I’m of the opinion that first you need to make it *correct*, then you can
>>> try to make it *fast*. It is my opinion that a installer that shits random
>>> debris into your current directory is not correct. It’s kind of silly that
>>> we have to have a “random pip/distutils/setuptools” crap chunk of stuff to
>>> add to .gitignore to basically every Python package in existence. Nevermind
>>> the random stuff that doesn’t currently get written there, but will if we
>>> stop copying files out of the path and into a temporary location (I’m sure
>>> everyone wants a pip-egg-info directory in their current directory).
>>>
>>> I’m also of the opinion that avoiding foot guns is more important than
>>> shooting for the fastest operation possible. I regularly (sometimes multiple
>>> times a week!, but often every week or two) see people tripping up on the
>>> fact that ``git clone … && pip install .`` does something different than
>>> ``git clone … && python setup.py sdist && pip install dist/*``. Files
>>> suddenly go missing and they have no idea why. If they’re lucky, they’ll
>>> figure out they need to modify some combination of package_data, data_files,
>>> and MANIFEST.in to make it work, if they’re not lucky they just sit there
>>> dumbfounded at it.
>
> Yeah, setuptools is kinda sucky this way. But this is fixable with
> better build systems. And before we can get better build systems, we
> need buy-in from devs. And saying "sorry, we're unilaterally screwing
> up your recompile times because we don't care" is not a good way to
> get there :-(
I don’t think it has anything to do with setuptools TBH other than the fact that it’s interface for declaring what does and doesn’t go into a sdist is kind of crummy. This problem is going to exist as long as you have any mechanism for having some files not be included inside of a sdist.
>
>>>
>>>
>>> Also also, notice elsewhere in the thread where Thomas notes that flit
>>> can't build an sdist from an unpacked sdist. It seems like 'pip
>>> install unpacked-sdist/' is an important use case to support…
>>>
>>>
>>> If the build tool gives us a mechanism to determine if something is an
>>> unpacked sdist or not so we can fallback to just copying in that case, that
>>> is fine with me. The bad case is generally only going to be hit on VCS
>>> checkouts or other not sdist kinds of source trees.
>
> I guess numpy could just claim that all VCS checkouts are actually
> unpacked sdists…?
I mean, ``pip install .`` is still going to ``cp -r`` that VCS checkout into a temporary location if you do that, and making sure that the invariant of ``python setup.py build && pip install .`` doesn’t trigger a recompile isn’t going to be something that I would want pip to start doing. So would it _work_ for this use case? Possibly? Is it supported? Nope, if it breaks you get to keep both pieces.
—
Donald Stufft
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://mail.python.org/pipermail/distutils-sig/attachments/20170529/5bf423c7/attachment.html>
More information about the Distutils-SIG
mailing list