spark-instrumented-optimizer/python/pyspark/daemon.py

165 lines
5 KiB
Python
Raw Normal View History

2013-05-06 19:34:30 -04:00
import os
import signal
import socket
2013-05-06 19:34:30 -04:00
import sys
import traceback
2013-05-06 19:34:30 -04:00
import multiprocessing
from ctypes import c_bool
2013-05-06 19:34:30 -04:00
from errno import EINTR, ECHILD
from socket import AF_INET, SOCK_STREAM, SOMAXCONN
from signal import SIGHUP, SIGTERM, SIGCHLD, SIG_DFL, SIG_IGN
2013-05-06 19:34:30 -04:00
from pyspark.worker import main as worker_main
from pyspark.serializers import write_int
try:
POOLSIZE = multiprocessing.cpu_count()
except NotImplementedError:
POOLSIZE = 4
exit_flag = multiprocessing.Value(c_bool, False)
def should_exit():
global exit_flag
return exit_flag.value
2013-05-06 19:34:30 -04:00
2013-06-21 12:13:48 -04:00
def compute_real_exit_code(exit_code):
# SystemExit's code can be integer or string, but os._exit only accepts integers
import numbers
if isinstance(exit_code, numbers.Integral):
return exit_code
else:
return 1
2013-05-06 19:34:30 -04:00
def worker(listen_sock):
# Redirect stdout to stderr
os.dup2(2, 1)
sys.stdout = sys.stderr # The sys.stdout object is different from file descriptor 1
2013-05-06 19:34:30 -04:00
# Manager sends SIGHUP to request termination of workers in the pool
def handle_sighup(*args):
assert should_exit()
signal.signal(SIGHUP, handle_sighup)
2013-05-06 19:34:30 -04:00
# Cleanup zombie children
def handle_sigchld(*args):
pid = status = None
try:
while (pid, status) != (0, 0):
pid, status = os.waitpid(0, os.WNOHANG)
except EnvironmentError as err:
if err.errno == EINTR:
# retry
handle_sigchld()
elif err.errno != ECHILD:
raise
signal.signal(SIGCHLD, handle_sigchld)
# Handle clients
while not should_exit():
2013-05-06 19:34:30 -04:00
# Wait until a client arrives or we have to exit
sock = None
while not should_exit() and sock is None:
2013-05-06 19:34:30 -04:00
try:
sock, addr = listen_sock.accept()
except EnvironmentError as err:
if err.errno != EINTR:
raise
if sock is not None:
# Fork a child to handle the client.
# The client is handled in the child so that the manager
# never receives SIGCHLD unless a worker crashes.
if os.fork() == 0:
2013-05-06 19:34:30 -04:00
# Leave the worker pool
signal.signal(SIGHUP, SIG_DFL)
2013-05-06 19:34:30 -04:00
listen_sock.close()
# Read the socket using fdopen instead of socket.makefile() because the latter
# seems to be very slow; note that we need to dup() the file descriptor because
# otherwise writes also cause a seek that makes us miss data on the read side.
infile = os.fdopen(os.dup(sock.fileno()), "a+", 65536)
outfile = os.fdopen(os.dup(sock.fileno()), "a+", 65536)
2013-06-21 12:13:48 -04:00
exit_code = 0
try:
worker_main(infile, outfile)
2013-06-21 12:13:48 -04:00
except SystemExit as exc:
exit_code = exc.code
2013-06-21 12:13:48 -04:00
finally:
outfile.flush()
sock.close()
os._exit(compute_real_exit_code(exit_code))
2013-05-06 19:34:30 -04:00
else:
sock.close()
def launch_worker(listen_sock):
if os.fork() == 0:
try:
worker(listen_sock)
except Exception as err:
traceback.print_exc()
os._exit(1)
else:
assert should_exit()
os._exit(0)
2013-05-06 19:34:30 -04:00
def manager():
# Create a new process group to corral our children
os.setpgid(0, 0)
# Create a listening socket on the AF_INET loopback interface
listen_sock = socket.socket(AF_INET, SOCK_STREAM)
2013-05-06 19:34:30 -04:00
listen_sock.bind(('127.0.0.1', 0))
listen_sock.listen(max(1024, 2 * POOLSIZE, SOMAXCONN))
listen_host, listen_port = listen_sock.getsockname()
write_int(listen_port, sys.stdout)
# Launch initial worker pool
for idx in range(POOLSIZE):
launch_worker(listen_sock)
2013-05-06 19:34:30 -04:00
listen_sock.close()
def shutdown():
global exit_flag
exit_flag.value = True
2013-05-06 19:34:30 -04:00
# Gracefully exit on SIGTERM, don't die on SIGHUP
signal.signal(SIGTERM, lambda signum, frame: shutdown())
signal.signal(SIGHUP, SIG_IGN)
2013-05-06 19:34:30 -04:00
# Cleanup zombie children
def handle_sigchld(*args):
2013-05-06 19:34:30 -04:00
try:
pid, status = os.waitpid(0, os.WNOHANG)
if status != 0 and not should_exit():
raise RuntimeError("worker crashed: %s, %s" % (pid, status))
2013-05-06 19:34:30 -04:00
except EnvironmentError as err:
if err.errno not in (ECHILD, EINTR):
raise
signal.signal(SIGCHLD, handle_sigchld)
2013-05-06 19:34:30 -04:00
# Initialization complete
sys.stdout.close()
try:
while not should_exit():
try:
# Spark tells us to exit by closing stdin
if os.read(0, 512) == '':
shutdown()
except EnvironmentError as err:
if err.errno != EINTR:
shutdown()
raise
finally:
signal.signal(SIGTERM, SIG_DFL)
exit_flag.value = True
# Send SIGHUP to notify workers of shutdown
os.kill(0, SIGHUP)
2013-05-06 19:34:30 -04:00
if __name__ == '__main__':
manager()