Source code
Revision control
Copy as Markdown
Other Tools
# This Source Code Form is subject to the terms of the Mozilla Public
# License, v. 2.0. If a copy of the MPL was not distributed with this file,
# The Marionette ProcessHandler and ProcessHandlerMixin classes are only
# utilized by Marionette as an alternative to the mozprocess package.
#
# This necessity arises because Marionette supports the application to
# restart itself and, under such conditions, fork its process. To maintain
# the ability to track the process, including permissions to terminate
# the process and receive log entries via stdout and stderr, the psutil
# package is utilized. To prevent any side effects for consumers of
# mozprocess, all necessary helper classes have been duplicated for now.
import codecs
import os
import signal
import subprocess
import sys
import threading
import time
import traceback
from queue import Empty, Queue
import psutil
import six
# Set the MOZPROCESS_DEBUG environment variable to 1 to see some debugging output
MOZPROCESS_DEBUG = os.getenv("MOZPROCESS_DEBUG")
INTERVAL_PROCESS_ALIVE_CHECK = 0.02
# For not self-managed processes the returncode seems to not be available.
# Use `8` to indicate this specific situation for now.
UNKNOWN_RETURNCODE = 8
isPosix = os.name == "posix" # includes MacOS X
class ProcessHandlerMixin(object):
"""
A class for launching and manipulating local processes.
:param cmd: command to run. May be a string or a list. If specified as a list, the first
element will be interpreted as the command, and all additional elements will be interpreted
as arguments to that command.
:param args: list of arguments to pass to the command (defaults to None). Must not be set when
`cmd` is specified as a list.
:param cwd: working directory for command (defaults to None).
:param env: is the environment to use for the process (defaults to os.environ).
:param kill_on_timeout: when True, the process will be killed when a timeout is reached.
When False, the caller is responsible for killing the process.
Failure to do so could cause a call to wait() to hang indefinitely. (Defaults to True.)
:param processOutputLine: function or list of functions to be called for
each line of output produced by the process (defaults to an empty
list).
:param processStderrLine: function or list of functions to be called
for each line of error output - stderr - produced by the process
(defaults to an empty list). If this is not specified, stderr lines
will be sent to the *processOutputLine* callbacks.
:param onTimeout: function or list of functions to be called when the process times out.
:param onFinish: function or list of functions to be called when the process terminates
normally without timing out.
:param kwargs: additional keyword args to pass directly into Popen.
NOTE: Child processes will be tracked by default.
"""
def __init__(
self,
cmd,
args=None,
cwd=None,
env=None,
kill_on_timeout=True,
processOutputLine=(),
processStderrLine=(),
onTimeout=(),
onFinish=(),
**kwargs,
):
self.args = args
self.cmd = cmd
self.cwd = cwd
self.keywordargs = kwargs
self.didTimeout = False
self.didOutputTimeout = False
self.proc = None
if env is None:
env = os.environ.copy()
self.env = env
# handlers
def to_callable_list(arg):
if callable(arg):
arg = [arg]
return CallableList(arg)
processOutputLine = to_callable_list(processOutputLine)
processStderrLine = to_callable_list(processStderrLine)
onTimeout = to_callable_list(onTimeout)
onFinish = to_callable_list(onFinish)
def on_timeout():
self.didTimeout = True
self.didOutputTimeout = self.reader.didOutputTimeout
if kill_on_timeout:
self.kill()
onTimeout.insert(0, on_timeout)
self._stderr = subprocess.STDOUT
if processStderrLine:
self._stderr = subprocess.PIPE
self.reader = ProcessReader(
stdout_callback=processOutputLine,
stderr_callback=processStderrLine,
finished_callback=onFinish,
timeout_callback=onTimeout,
)
# It is common for people to pass in the entire array with the cmd and
# the args together since this is how Popen uses it. Allow for that.
if isinstance(self.cmd, list):
if self.args is not None:
raise TypeError("cmd and args must not both be lists")
(self.cmd, self.args) = (self.cmd[0], self.cmd[1:])
elif self.args is None:
self.args = []
def _has_valid_proc(func):
def wrapper(self, *args, **kwargs):
if self.proc is None:
raise RuntimeError("Process hasn't been started yet")
return func(self, *args, **kwargs)
return wrapper
@property
@_has_valid_proc
def pid(self):
return self.proc.pid
@staticmethod
def pid_exists(pid):
return psutil.pid_exists(pid)
@property
def timedOut(self):
"""True if the process has timed out for any reason."""
return self.didTimeout
@property
def outputTimedOut(self):
"""True if the process has timed out for no output."""
return self.didOutputTimeout
@property
def commandline(self):
"""the string value of the command line (command + args)"""
return subprocess.list2cmdline([self.cmd] + self.args)
def _debug(self, msg):
if not MOZPROCESS_DEBUG:
return
print(f"DBG::MARIONETTE ProcessHandler {self.pid} | {msg}", file=sys.stdout)
@_has_valid_proc
def kill(self, sig=None, timeout=None):
"""Kills the managed process and all its child processes.
:param sig: Signal to use to kill the process. (Defaults to SIGKILL)
:param timeout: If not None, wait this number of seconds for the
process to exit.
Note that this does not manage any state, save any output etc,
it immediately kills the process.
"""
if hasattr(self, "returncode"):
return self.returncode
if self.proc.is_running():
processes = [self.proc] + self.proc.children(recursive=True)
if sig is None:
# TODO: try SIGTERM first to sanely shutdown the application
# and to not break later when Windows support gets added.
sig = signal.SIGKILL
# Do we need that?
for process in processes:
try:
self._debug(f"Killing process: {process}")
process.send_signal(sig)
except psutil.NoSuchProcess:
pass
psutil.wait_procs(processes, timeout=timeout)
# When we kill the the managed process we also have to wait for the
# reader thread to be finished. Otherwise consumers would have to assume
# that it still has not completely shutdown.
self.returncode = self.wait(0)
if self.returncode is None:
self._debug("kill: wait failed -- process is still alive")
return self.returncode
@_has_valid_proc
def poll(self):
"""Check if child process has terminated
Returns the current returncode value:
- None if the process hasn't terminated yet
- A negative number if the process was killed by signal N (Unix only)
- '0' if the process ended without failures
"""
if hasattr(self, "returncode"):
return self.returncode
# If the process that is observed wasn't started with Popen there is
# no `poll()` method available. Use `wait()` instead and do not wait
# for the reader thread because it would cause extra delays.
return self.wait(0, wait_reader=False)
def processOutput(self, timeout=None, outputTimeout=None):
"""
Handle process output until the process terminates or times out.
:param timeout: If not None, the process will be allowed to continue
for that number of seconds before being killed.
:outputTimeout: If not None, the process will be allowed to continue
for that number of seconds without producing any output before
being killed.
"""
# this method is kept for backward compatibility
if not hasattr(self, "proc"):
self.run(timeout=timeout, outputTimeout=outputTimeout)
# self.run will call this again
return
if not self.reader.is_alive():
self.reader.timeout = timeout
self.reader.output_timeout = outputTimeout
self.reader.start(self.proc)
def run(self, timeout=None, outputTimeout=None):
"""
Starts the process.
:param timeout: If not None, the process will be allowed to continue for
that number of seconds before being killed. If the process is killed
due to a timeout, the onTimeout handler will be called.
:outputTimeout: If not None, the process will be allowed to continue
for that number of seconds without producing any output before
being killed.
"""
self.didTimeout = False
self.didOutputTimeout = False
# default arguments
args = dict(
stdout=subprocess.PIPE,
stderr=self._stderr,
cwd=self.cwd,
env=self.env,
)
# build process arguments
args.update(self.keywordargs)
# launch the process
self.proc = psutil.Popen([self.cmd] + self.args, **args)
self.processOutput(timeout=timeout, outputTimeout=outputTimeout)
@_has_valid_proc
def update_process(self, new_pid, timeout=None):
"""Update the internally managed process for the provided process ID.
When the application restarts itself, such as during an update, the new
process is essentially a fork of itself. To continue monitoring this
process, the process ID needs to be updated accordingly.
:param new_pid: The ID of the new (forked) process to track.
:timeout: If not None, the old process will be allowed to continue for
that number of seconds before being killed.
"""
if isPosix:
if new_pid == self.pid:
return
print(
'Child process with id "%s" has been marked as detached because it is no '
"longer in the managed process group. Keeping reference to the process id "
'"%s" which is the new child process.' % (self.pid, new_pid),
file=sys.stdout,
)
returncode = self.wait(timeout, wait_reader=False)
if returncode is None:
# If the process is still running force kill it.
returncode = self.kill()
if hasattr(self, "returncode"):
del self.returncode
self.proc = psutil.Process(new_pid)
self._debug(
f"New process status: {self.proc} (terminal={self.proc.terminal()})"
)
return returncode
@_has_valid_proc
def wait(self, timeout=None, wait_reader=True):
"""
Waits until the process is terminated.
:param timeout: If not None, will return after timeout seconds.
This timeout only causes the wait function to return and
does not kill the process.
:param wait_reader: If set to True, it waits not only for the process
to exit but also for all output to be fully read. (Defaults to True).
Returns the process exit code value:
- None if the process hasn't terminated yet
- A negative number if the process was killed by signal N (Unix only)
- '0' if the process ended without failures
"""
# Thread.join() blocks the main thread until the reader thread is finished
# wake up once a second in case a keyboard interrupt is sent
if (
wait_reader
and self.reader.thread
and self.reader.thread is not threading.current_thread()
):
count = 0
while self.reader.is_alive():
if timeout is not None and count > timeout:
self._debug("wait timeout for reader thread")
return None
self.reader.join(timeout=1)
count += 1
try:
self.proc.wait(timeout)
self._debug(f"Process status after wait: {self.proc}")
if not isinstance(self.proc, psutil.Popen):
self._debug(
"Not self-managed processes do not have a returncode. "
f"Setting its value to {UNKNOWN_RETURNCODE}."
)
self.returncode = UNKNOWN_RETURNCODE
else:
self.returncode = self.proc.returncode
return self.returncode
except psutil.TimeoutExpired:
return None
class CallableList(list):
def __call__(self, *args, **kwargs):
for e in self:
e(*args, **kwargs)
def __add__(self, lst):
return CallableList(list.__add__(self, lst))
class ProcessReader(object):
def __init__(
self,
stdout_callback=None,
stderr_callback=None,
finished_callback=None,
timeout_callback=None,
timeout=None,
output_timeout=None,
):
self.stdout_callback = stdout_callback or (lambda line: True)
self.stderr_callback = stderr_callback or (lambda line: True)
self.finished_callback = finished_callback or (lambda: True)
self.timeout_callback = timeout_callback or (lambda: True)
self.timeout = timeout
self.output_timeout = output_timeout
self.thread = None
self.didOutputTimeout = False
def debug(self, msg):
if not MOZPROCESS_DEBUG:
return
print("DBG::MARIONETTE ProcessReader | {}".format(msg), file=sys.stdout)
def _create_stream_reader(self, name, stream, queue, callback):
thread = threading.Thread(
name=name, target=self._read_stream, args=(stream, queue, callback)
)
thread.daemon = True
thread.start()
return thread
def _read_stream(self, stream, queue, callback):
while True:
line = stream.readline()
if not line:
break
queue.put((line, callback))
stream.close()
def start(self, proc):
queue = Queue()
stdout_reader = None
if proc.stdout:
stdout_reader = self._create_stream_reader(
"ProcessReaderStdout", proc.stdout, queue, self.stdout_callback
)
stderr_reader = None
if proc.stderr and proc.stderr != proc.stdout:
stderr_reader = self._create_stream_reader(
"ProcessReaderStderr", proc.stderr, queue, self.stderr_callback
)
self.thread = threading.Thread(
name="ProcessReader",
target=self._read,
args=(stdout_reader, stderr_reader, queue),
)
self.thread.daemon = True
self.thread.start()
self.debug("ProcessReader started")
def _read(self, stdout_reader, stderr_reader, queue):
start_time = time.time()
timed_out = False
timeout = self.timeout
if timeout is not None:
timeout += start_time
output_timeout = self.output_timeout
if output_timeout is not None:
output_timeout += start_time
while (stdout_reader and stdout_reader.is_alive()) or (
stderr_reader and stderr_reader.is_alive()
):
has_line = True
try:
line, callback = queue.get(True, INTERVAL_PROCESS_ALIVE_CHECK)
except Empty:
has_line = False
now = time.time()
if not has_line:
if output_timeout is not None and now > output_timeout:
timed_out = True
self.didOutputTimeout = True
break
else:
if output_timeout is not None:
output_timeout = now + self.output_timeout
callback(line.rstrip())
if timeout is not None and now > timeout:
timed_out = True
break
self.debug("_read loop exited")
# process remaining lines to read
while not queue.empty():
line, callback = queue.get(False)
try:
callback(line.rstrip())
except Exception:
traceback.print_exc()
if timed_out:
try:
self.timeout_callback()
except Exception:
traceback.print_exc()
if stdout_reader:
stdout_reader.join()
if stderr_reader:
stderr_reader.join()
if not timed_out:
try:
self.finished_callback()
except Exception:
traceback.print_exc()
self.debug("_read exited")
def is_alive(self):
if self.thread:
return self.thread.is_alive()
return False
def join(self, timeout=None):
if self.thread:
self.thread.join(timeout=timeout)
# default output handlers
# these should be callables that take the output line
class StoreOutput(object):
"""accumulate stdout"""
def __init__(self):
self.output = []
def __call__(self, line):
self.output.append(line)
class StreamOutput(object):
"""pass output to a stream and flush"""
def __init__(self, stream, text=True):
self.stream = stream
self.text = text
def __call__(self, line):
ensure = six.ensure_text if self.text else six.ensure_binary
try:
self.stream.write(ensure(line, errors="ignore") + ensure("\n"))
except TypeError:
print(
"HEY! If you're reading this, you're about to encounter a "
"type error, probably as a result of a conversion from "
"Python 2 to Python 3. This is almost definitely because "
"you're trying to write binary data to a text-encoded "
"stream, or text data to a binary-encoded stream. Check how "
"you're instantiating your ProcessHandler and if the output "
"should be text-encoded, make sure you pass "
"universal_newlines=True.",
file=sys.stderr,
)
raise
self.stream.flush()
class LogOutput(StreamOutput):
"""pass output to a file"""
def __init__(self, filename):
self.file_obj = open(filename, "a")
StreamOutput.__init__(self, self.file_obj, True)
def __del__(self):
if self.file_obj is not None:
self.file_obj.close()
# front end class with the default handlers
class ProcessHandler(ProcessHandlerMixin):
"""
Convenience class for handling processes with default output handlers.
By default, all output is sent to stdout. This can be disabled by setting
the *stream* argument to None.
If processOutputLine keyword argument is specified the function or the
list of functions specified by this argument will be called for each line
of output; the output will not be written to stdout automatically then
if stream is True (the default).
If storeOutput==True, the output produced by the process will be saved
as self.output.
If logfile is not None, the output produced by the process will be
appended to the given file.
"""
def __init__(self, cmd, logfile=None, stream=True, storeOutput=True, **kwargs):
kwargs.setdefault("processOutputLine", [])
if callable(kwargs["processOutputLine"]):
kwargs["processOutputLine"] = [kwargs["processOutputLine"]]
if logfile:
logoutput = LogOutput(logfile)
kwargs["processOutputLine"].append(logoutput)
text = kwargs.get("universal_newlines", False) or kwargs.get("text", False)
if stream is True:
if text:
# The encoding of stdout isn't guaranteed to be utf-8. Fix that.
stdout = codecs.getwriter("utf-8")(sys.stdout.buffer)
else:
stdout = sys.stdout.buffer
if not kwargs["processOutputLine"]:
kwargs["processOutputLine"].append(StreamOutput(stdout, text))
elif stream:
streamoutput = StreamOutput(stream, text)
kwargs["processOutputLine"].append(streamoutput)
self.output = None
if storeOutput:
storeoutput = StoreOutput()
self.output = storeoutput.output
kwargs["processOutputLine"].append(storeoutput)
ProcessHandlerMixin.__init__(self, cmd, **kwargs)