在python中控制用于调用外部命令的子进程数

问题描述

我了解使用 subprocess 是调用外部命令的首选方式.

I understand using subprocess is the preferred way of calling external command.

但是,如果我想并行运行多个命令,但要限制生成的进程数怎么办?困扰我的是我无法阻止子进程.例如,如果我调用

But what if I want to run several commands in parall, but limit the number of processes being spawned? What bothers me is that I can't block the subprocesses. For example, if I call

subprocess.Popen(cmd, stderr=outputfile, stdout=outputfile)

然后该过程将继续,无需等待 cmd 完成.因此,我无法将其包装在 multiprocessing 库的工作人员中.

Then the process will continue, without waiting for cmd to finish. Therefore, I can't wrap it up in a worker of multiprocessing library.

例如,如果我这样做:

def worker(cmd): 
    subprocess.Popen(cmd, stderr=outputfile, stdout=outputfile);

pool = Pool( processes = 10 );
results =[pool.apply_async(worker, [cmd]) for cmd in cmd_list];
ans = [res.get() for res in results];

然后每个工作人员将在生成子进程后完成并返回.所以我不能通过使用Pool 来真正限制subprocess 生成的进程数.

then each worker will finish and return after spawning a subprocess. So I can't really limit the number of processes generated by subprocess by using Pool.

限制子进程数量的正确方法是什么?

What's the proper way of limiting the number of subprocesses?


解决方案

如果要等待命令完成,可以使用subprocess.call.有关详细信息,请参阅 pydoc 子进程.

You can use subprocess.call if you want to wait for the command to complete. See pydoc subprocess for more information.

您也可以调用 Popen.wait 你的工人中的方法:

You could also call the Popen.wait method in your worker:

def worker(cmd): 
    p = subprocess.Popen(cmd, stderr=outputfile, stdout=outputfile);
    p.wait()

相关文章