[Mailman-Users] Problem with archrunner using large %'s of cpu (read faq & archives)

Brad Knowles brad.knowles at skynet.be
Sat Nov 1 23:45:08 CET 2003

At 9:29 PM -0500 2003/10/31, Scott Lambert wrote:

>  If we were talking about more than 10,000 files, I might buy it.  But we
>  are talking about 1300 files.

	Many filesystems start significantly slowing down around 1,000 
files, not 10,000.  Moreover, are you sure that this is the largest 
number of files you've ever had in that directory?

>                                 Also the processing goes something like
>  O(n), in reverse, slower as it processes the files in the directory.

	That is a bit strange, but might be explained by holes in the 
directory structure that need to be skipped.

>                                                                        I
>  might buy it staying slow if it started slow but it doesn't.

	I've seen mail servers at large freemail providers that had 
previously grown to very large sizes, and worked reasonably well for 
numbers of files in the low thousands, but seriously flaked out when 
pushed much beyond that.

	Move the directory aside, move the files to a new directory, and 
restart -- suddenly everything works like magic again.

	Unless you know the filesystem code intimately, as well as the 
code that is using the filesystem, it can be difficult to predict how 
or when things will break or how badly they will break.

Brad Knowles, <brad.knowles at skynet.be>

"They that can give up essential liberty to obtain a little temporary
safety deserve neither liberty nor safety."
     -Benjamin Franklin, Historical Review of Pennsylvania.

GCS/IT d+(-) s:+(++)>: a C++(+++)$ UMBSHI++++$ P+>++ L+ !E-(---) W+++(--) N+
!w--- O- M++ V PS++(+++) PE- Y+(++) PGP>+++ t+(+++) 5++(+++) X++(+++) R+(+++)
tv+(+++) b+(++++) DI+(++++) D+(++) G+(++++) e++>++++ h--- r---(+++)* z(+++)

More information about the Mailman-Users mailing list