Awesome effort! ;) Is it maybe worth, during the refactor, to evaluate how NDimage could be supportive for auto-parallelization? I’m thinking that it somehow would be able to split off into embarrassingly parallel n tasks, if thrown to a n-core machine? I’m sorry if I talk non-sense, I’m just fighting to get GLCM faster, when running it via “generic_filter” over large-ish images, so far not successful, even trying to use numba, but as many things are already running in Cython, so I guess I can’t expect much auto-improvements. ;) So, I’m anticipating to have to go to a cluster for my hundreds of images I want to run multiple/many GLCM analyses on. Thanks again for the developer docs, they are great! Regards, Michael
On Jun 27, 2017, at 09:06, Stefan van der Walt <stefanv@berkeley.edu> wrote:
On Tue, Jun 27, 2017, at 07:49, Egor Panfilov wrote:
- Not sure that I'll find enough time to contribute to the code by myself, but it won't harm if you could provide to the public some directions on where to start with `scipy.ndimage` C-code.
Making ndimage developer notes along the way would be extremely helpful!
Stéfan
_______________________________________________ scikit-image mailing list scikit-image@python.org https://mail.python.org/mailman/listinfo/scikit-image