Best way to run multiple Python processes without overloading CPU or disk i/o

Philip Semanchuk philip at semanchuk.com
Wed Dec 3 05:29:32 CET 2008


On Dec 2, 2008, at 11:21 PM, python at bdurham.com wrote:

> Is there a cross-platform way to launch multiple Python processes
> and monitor CPU usage

os.getloadavg() might be useful. It certainly works on *nix, don't  
know about Windows. The documentation doesn't mention any platform  
limitations.


HTH
Philip



> and disk i/o so that the maximum number of
> independent processes can be running at all times without
> overloading their environment? By process I mean independent
> application sessions vs. multiple threads of a single
> application. I'm looking for suggestions on what Python modules
> and/or techniques to use to maximize the number of processes I
> can run on my system over a multi-day period.
> Background: I have a large collection of data analysis scripts
> that can run independently of one another. These scripts read
> very large log files in a sequential manner and calculate
> specific statistics. These scripts are hosted on a 8 core 64-bit
> server with 48G of memory that is dedicated to running these
> scripts. I am looking for a way to write a master script that
> monitors system CPU and disk i/o and when there is capacity
> launch another script from a pool of scripts so that I'm getting
> max utilization from my hardware, eg. so that I'm running the max
> number of scripts I can at any one point in time without
> overwhelming the system.
> The server in question is currently running Windows 2008
> Enterprise, but I have the option to switch to a 64-bit version
> of Linux if there are benefits to doing so.
> Thank you,
> Malcolmm
> --
> http://mail.python.org/mailman/listinfo/python-list




More information about the Python-list mailing list