Friday, September 15, 2017

Multiprocessing with Python

I have been training a simple neural network on my desktop, and I realized that GPU wasn't running at its full capacity, i.e., there must be some bottleneck from, most likely, CPU side. My guess is image preprocessing from CPU is taking longer than GPU computation for each batch. In order to reduce the time for CPU to preprocess the images, I started investigating multiprocessing option in Python.

Below is a simple code for running OpenCV's Canny function across multiple processes using Python's built-in multiprocess module:

import cv2
import time
import glob
import multiprocessing as mp
def process_img(image_file):
if '_canny.jpg' not in image_file:
print 'processing', image_file
img = cv2.imread(image_file)
img = cv2.cvtColor(img, cv2.COLOR_BGR2GRAY)
img = cv2.blur(img, (3,3))
img = cv2.Canny(img, 10, 25)
cv2.imwrite(image_file[:-4] + '_canny.jpg', img)
def single_process(image_files):
start = time.time()
for image_file in image_files:
process_img(image_file)
print 'single process elapsed time:', int(time.time() - start)
def multi_processes(image_files, nproc=2):
start = time.time()
pool = mp.Pool(nproc)
for image_file in image_files:
pool.apply_async(process_img, (image_file,))
pool.close()
pool.join()
print nproc, 'processes elapsed time:', int(time.time() - start)
if __name__ == '__main__':
files = glob.glob('images/*.jpg')
single_process(files)
multi_processes(files, nproc=2)
multi_processes(files, nproc=4)
view raw process_pool.py hosted with ❤ by GitHub

Running the script yields approximately linear time reduction for 2 processes and sub-linear for 4 processes, due to other bottleneck, such as disk IO.

single process elapsed time: 364
2 processes elapsed time: 181
4 processes elapsed time: 108

No comments:

Post a Comment