Source code for gear

# Copyright 2013-2014 OpenStack Foundation
#
# Licensed under the Apache License, Version 2.0 (the "License"); you may
# not use this file except in compliance with the License. You may obtain
# a copy of the License at
#
#      http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
# License for the specific language governing permissions and limitations
# under the License.

import errno
import logging
import os
import random
import select
import six
import socket
import ssl
import struct
import threading
import time
import uuid as uuid_module

from gear import constants
from gear.acl import ACLError, ACLEntry, ACL  # noqa

try:
    import Queue as queue_mod
except ImportError:
    import queue as queue_mod

try:
    import statsd
except ImportError:
    statsd = None

PRECEDENCE_NORMAL = 0
PRECEDENCE_LOW = 1
PRECEDENCE_HIGH = 2


[docs]class ConnectionError(Exception): pass
[docs]class InvalidDataError(Exception): pass
[docs]class ConfigurationError(Exception): pass
[docs]class NoConnectedServersError(Exception): pass
[docs]class UnknownJobError(Exception): pass
[docs]class InterruptedError(Exception): pass
class TimeoutError(Exception): pass class GearmanError(Exception): pass class DisconnectError(Exception): pass class RetryIOError(Exception): pass def convert_to_bytes(data): try: data = data.encode('utf8') except AttributeError: pass return data def best_tls_version(): if hasattr(ssl, 'PROTOCOL_TLS'): return ssl.PROTOCOL_TLS # Note there is some risk in selecting tls 1.2 if available # as both the client and server may not support it and need 1.1 # or 1.0. However, a xenial installation with python 3.5 does # support 1.2 which is probably as old a setup as we need to worry # about. elif hasattr(ssl, 'PROTOCOL_TLSv1_2'): return ssl.PROTOCOL_TLSv1_2 elif hasattr(ssl, 'PROTOCOL_TLSv1_1'): return ssl.PROTOCOL_TLSv1_1 elif hasattr(ssl, 'PROTOCOL_TLSv1'): return ssl.PROTOCOL_TLSv1 else: raise ConnectionError('No supported TLS version available.') def create_ssl_context(): tls_version = best_tls_version() context = ssl.SSLContext(tls_version) # Disable TLSv1.3 # According to https://bugs.python.org/issue43622#msg389497, an event on # ssl socket can happen without data being available at application level. # As gear is using a polling loop with multiple file descriptors and ssl # socket used as a blocking one, a blocked state could happen. # This is highlighted by Zuul SSL test: TestSchedulerSSL, where such # blocked state appears consistently. # note: gear tests and zuul tests are ok for TLSv1.2 but this behavior # could also happen if (hasattr(ssl, 'PROTOCOL_TLS') and tls_version == ssl.PROTOCOL_TLS): context.options |= ssl.OP_NO_TLSv1_3 return context class Task(object): def __init__(self): self._wait_event = threading.Event() def setComplete(self): self._wait_event.set() def wait(self, timeout=None): """Wait for a response from Gearman. :arg int timeout: If not None, return after this many seconds if no response has been received (default: None). """ self._wait_event.wait(timeout) return self._wait_event.is_set() class SubmitJobTask(Task): def __init__(self, job): super(SubmitJobTask, self).__init__() self.job = job class OptionReqTask(Task): pass
[docs]class Connection(object): """A Connection to a Gearman Server. :arg str client_id: The client ID associated with this connection. It will be appending to the name of the logger (e.g., gear.Connection.client_id). Defaults to 'unknown'. :arg bool keepalive: Whether to use TCP keepalives :arg int tcp_keepidle: Idle time after which to start keepalives sending :arg int tcp_keepintvl: Interval in seconds between TCP keepalives :arg int tcp_keepcnt: Count of TCP keepalives to send before disconnect """ def __init__(self, host, port, ssl_key=None, ssl_cert=None, ssl_ca=None, client_id='unknown', keepalive=False, tcp_keepidle=7200, tcp_keepintvl=75, tcp_keepcnt=9): self.log = logging.getLogger("gear.Connection.%s" % (client_id,)) self.host = host self.port = port self.ssl_key = ssl_key self.ssl_cert = ssl_cert self.ssl_ca = ssl_ca self.keepalive = keepalive self.tcp_keepcnt = tcp_keepcnt self.tcp_keepintvl = tcp_keepintvl self.tcp_keepidle = tcp_keepidle self.use_ssl = False if all([self.ssl_key, self.ssl_cert, self.ssl_ca]): self.use_ssl = True self.input_buffer = b'' self.need_bytes = False self.echo_lock = threading.Lock() self.send_lock = threading.Lock() self._init() def _init(self): self.conn = None self.connected = False self.connect_time = None self.related_jobs = {} self.pending_tasks = [] self.admin_requests = [] self.echo_conditions = {} self.options = set() self.changeState("INIT") def changeState(self, state): # The state variables are provided as a convenience (and used by # the Worker implementation). They aren't used or modified within # the connection object itself except to reset to "INIT" immediately # after reconnection. self.log.debug("Setting state to: %s" % state) self.state = state self.state_time = time.time() def __repr__(self): return '<gear.Connection 0x%x host: %s port: %s>' % ( id(self), self.host, self.port)
[docs] def connect(self): """Open a connection to the server. :raises ConnectionError: If unable to open the socket. """ self.log.debug("Connecting to %s port %s" % (self.host, self.port)) s = None for res in socket.getaddrinfo(self.host, self.port, socket.AF_UNSPEC, socket.SOCK_STREAM): af, socktype, proto, canonname, sa = res try: s = socket.socket(af, socktype, proto) if self.keepalive and hasattr(socket, 'TCP_KEEPIDLE'): s.setsockopt(socket.SOL_SOCKET, socket.SO_KEEPALIVE, 1) s.setsockopt(socket.IPPROTO_TCP, socket.TCP_KEEPIDLE, self.tcp_keepidle) s.setsockopt(socket.IPPROTO_TCP, socket.TCP_KEEPINTVL, self.tcp_keepintvl) s.setsockopt(socket.IPPROTO_TCP, socket.TCP_KEEPCNT, self.tcp_keepcnt) elif self.keepalive: self.log.warning('Keepalive requested but not available ' 'on this platform') except socket.error: s = None continue if self.use_ssl: self.log.debug("Using SSL") context = create_ssl_context() context.verify_mode = ssl.CERT_REQUIRED context.check_hostname = False context.load_cert_chain(self.ssl_cert, self.ssl_key) context.load_verify_locations(self.ssl_ca) s = context.wrap_socket(s, server_hostname=self.host) try: s.connect(sa) except socket.error: s.close() s = None continue break if s is None: self.log.debug("Error connecting to %s port %s" % ( self.host, self.port)) raise ConnectionError("Unable to open socket") self.log.info("Connected to %s port %s" % (self.host, self.port)) self.conn = s self.connected = True self.connect_time = time.time() self.input_buffer = b'' self.need_bytes = False
[docs] def disconnect(self): """Disconnect from the server and remove all associated state data. """ if self.conn: try: self.conn.close() except Exception: pass self.log.info("Disconnected from %s port %s" % (self.host, self.port)) self._init()
[docs] def reconnect(self): """Disconnect from and reconnect to the server, removing all associated state data. """ self.disconnect() self.connect()
[docs] def sendRaw(self, data): """Send raw data over the socket. :arg bytes data The raw data to send """ with self.send_lock: sent = 0 while sent < len(data): try: sent += self.conn.send(data) except ssl.SSLError as e: if e.errno == ssl.SSL_ERROR_WANT_READ: continue elif e.errno == ssl.SSL_ERROR_WANT_WRITE: continue else: raise
[docs] def sendPacket(self, packet): """Send a packet to the server. :arg Packet packet: The :py:class:`Packet` to send. """ self.log.info("Sending packet to %s: %s" % (self, packet)) self.sendRaw(packet.toBinary())
def _getAdminRequest(self): return self.admin_requests.pop(0) def _readRawBytes(self, bytes_to_read): while True: try: buff = self.conn.recv(bytes_to_read) except ssl.SSLError as e: if e.errno == ssl.SSL_ERROR_WANT_READ: continue elif e.errno == ssl.SSL_ERROR_WANT_WRITE: continue else: raise break return buff def _putAdminRequest(self, req): self.admin_requests.insert(0, req)
[docs] def readPacket(self): """Read one packet or administrative response from the server. :returns: The :py:class:`Packet` or :py:class:`AdminRequest` read. :rtype: :py:class:`Packet` or :py:class:`AdminRequest` """ # This handles non-blocking or blocking IO. datalen = 0 code = None ptype = None admin = None admin_request = None need_bytes = self.need_bytes raw_bytes = self.input_buffer try: while True: try: if not raw_bytes or need_bytes: segment = self._readRawBytes(4096) if not segment: # This occurs when the connection is closed. The # the connect method will reset input_buffer and # need_bytes for us. return None raw_bytes += segment need_bytes = False except RetryIOError: if admin_request: self._putAdminRequest(admin_request) raise if admin is None: if raw_bytes[0:1] == b'\x00': admin = False else: admin = True admin_request = self._getAdminRequest() if admin: complete, remainder = admin_request.isComplete(raw_bytes) if remainder is not None: raw_bytes = remainder if complete: return admin_request else: length = len(raw_bytes) if code is None and length >= 12: code, ptype, datalen = struct.unpack('!4sii', raw_bytes[:12]) if length >= datalen + 12: end = 12 + datalen p = Packet(code, ptype, raw_bytes[12:end], connection=self) raw_bytes = raw_bytes[end:] return p # If we don't return a packet above then we need more data need_bytes = True finally: self.input_buffer = raw_bytes self.need_bytes = need_bytes
def hasPendingData(self): return self.input_buffer != b''
[docs] def sendAdminRequest(self, request, timeout=90): """Send an administrative request to the server. :arg AdminRequest request: The :py:class:`AdminRequest` to send. :arg numeric timeout: Number of seconds to wait until the response is received. If None, wait forever (default: 90 seconds). :raises TimeoutError: If the timeout is reached before the response is received. """ self.admin_requests.append(request) self.sendRaw(request.getCommand()) complete = request.waitForResponse(timeout) if not complete: raise TimeoutError()
[docs] def echo(self, data=None, timeout=30): """Perform an echo test on the server. This method waits until the echo response has been received or the timeout has been reached. :arg bytes data: The data to request be echoed. If None, a random unique byte string will be generated. :arg numeric timeout: Number of seconds to wait until the response is received. If None, wait forever (default: 30 seconds). :raises TimeoutError: If the timeout is reached before the response is received. """ if data is None: data = uuid_module.uuid4().hex.encode('utf8') self.echo_lock.acquire() try: if data in self.echo_conditions: raise InvalidDataError("This client is already waiting on an " "echo response of: %s" % data) condition = threading.Condition() self.echo_conditions[data] = condition finally: self.echo_lock.release() self.sendEchoReq(data) condition.acquire() condition.wait(timeout) condition.release() if data in self.echo_conditions: return data raise TimeoutError()
def sendEchoReq(self, data): p = Packet(constants.REQ, constants.ECHO_REQ, data) self.sendPacket(p) def handleEchoRes(self, data): condition = None self.echo_lock.acquire() try: condition = self.echo_conditions.get(data) if condition: del self.echo_conditions[data] finally: self.echo_lock.release() if not condition: return False condition.notifyAll() return True def handleOptionRes(self, option): self.options.add(option)
[docs]class AdminRequest(object): """Encapsulates a request (and response) sent over the administrative protocol. This is a base class that may not be instantiated dircectly; a subclass implementing a specific command must be used instead. :arg list arguments: A list of byte string arguments for the command. The following instance attributes are available: **response** (bytes) The response from the server. **arguments** (bytes) The argument supplied with the constructor. **command** (bytes) The administrative command. """ command = None arguments = [] response = None _complete_position = 0 def __init__(self, *arguments): self.wait_event = threading.Event() self.arguments = arguments if type(self) == AdminRequest: raise NotImplementedError("AdminRequest must be subclassed") def __repr__(self): return '<gear.AdminRequest 0x%x command: %s>' % ( id(self), self.command) def getCommand(self): cmd = self.command if self.arguments: cmd += b' ' + b' '.join(self.arguments) cmd += b'\n' return cmd def isComplete(self, data): x = -1 start = self._complete_position start = max(self._complete_position - 4, 0) end_index_newline = data.find(b'\n.\n', start) end_index_return = data.find(b'\r\n.\r\n', start) if end_index_newline != -1: x = end_index_newline + 3 elif end_index_return != -1: x = end_index_return + 5 elif data.startswith(b'.\n'): x = 2 elif data.startswith(b'.\r\n'): x = 3 self._complete_position = len(data) if x != -1: self.response = data[:x] return (True, data[x:]) else: return (False, None) def setComplete(self): self.wait_event.set() def waitForResponse(self, timeout=None): self.wait_event.wait(timeout) return self.wait_event.is_set()
[docs]class StatusAdminRequest(AdminRequest): """A "status" administrative request. The response from gearman may be found in the **response** attribute. """ command = b'status' def __init__(self): super(StatusAdminRequest, self).__init__()
[docs]class ShowJobsAdminRequest(AdminRequest): """A "show jobs" administrative request. The response from gearman may be found in the **response** attribute. """ command = b'show jobs' def __init__(self): super(ShowJobsAdminRequest, self).__init__()
[docs]class ShowUniqueJobsAdminRequest(AdminRequest): """A "show unique jobs" administrative request. The response from gearman may be found in the **response** attribute. """ command = b'show unique jobs' def __init__(self): super(ShowUniqueJobsAdminRequest, self).__init__()
[docs]class CancelJobAdminRequest(AdminRequest): """A "cancel job" administrative request. :arg str handle: The job handle to be canceled. The response from gearman may be found in the **response** attribute. """ command = b'cancel job' def __init__(self, handle): handle = convert_to_bytes(handle) super(CancelJobAdminRequest, self).__init__(handle) def isComplete(self, data): end_index_newline = data.find(b'\n') if end_index_newline != -1: x = end_index_newline + 1 self.response = data[:x] return (True, data[x:]) else: return (False, None)
[docs]class VersionAdminRequest(AdminRequest): """A "version" administrative request. The response from gearman may be found in the **response** attribute. """ command = b'version' def __init__(self): super(VersionAdminRequest, self).__init__() def isComplete(self, data): end_index_newline = data.find(b'\n') if end_index_newline != -1: x = end_index_newline + 1 self.response = data[:x] return (True, data[x:]) else: return (False, None)
class WorkersAdminRequest(AdminRequest): """A "workers" administrative request. The response from gearman may be found in the **response** attribute. """ command = b'workers' def __init__(self): super(WorkersAdminRequest, self).__init__()
[docs]class Packet(object): """A data packet received from or to be sent over a :py:class:`Connection`. :arg bytes code: The Gearman magic code (:py:data:`constants.REQ` or :py:data:`constants.RES`) :arg bytes ptype: The packet type (one of the packet types in constants). :arg bytes data: The data portion of the packet. :arg Connection connection: The connection on which the packet was received (optional). :raises InvalidDataError: If the magic code is unknown. """ def __init__(self, code, ptype, data, connection=None): if not isinstance(code, bytes) and not isinstance(code, bytearray): raise TypeError("code must be of type bytes or bytearray") if code[0:1] != b'\x00': raise InvalidDataError("First byte of packet must be 0") self.code = code self.ptype = ptype if not isinstance(data, bytes) and not isinstance(data, bytearray): raise TypeError("data must be of type bytes or bytearray") self.data = data self.connection = connection def __repr__(self): ptype = constants.types.get(self.ptype, 'UNKNOWN') try: extra = self._formatExtraData() except Exception: extra = '' return '<gear.Packet 0x%x type: %s%s>' % (id(self), ptype, extra) def __eq__(self, other): if not isinstance(other, Packet): return False if (self.code == other.code and self.ptype == other.ptype and self.data == other.data): return True return False def __ne__(self, other): return not self.__eq__(other) def _formatExtraData(self): if self.ptype in [constants.JOB_CREATED, constants.JOB_ASSIGN, constants.GET_STATUS, constants.STATUS_RES, constants.WORK_STATUS, constants.WORK_COMPLETE, constants.WORK_FAIL, constants.WORK_EXCEPTION, constants.WORK_DATA, constants.WORK_WARNING]: return ' handle: %s' % self.getArgument(0) if self.ptype == constants.JOB_ASSIGN_UNIQ: return (' handle: %s function: %s unique: %s' % (self.getArgument(0), self.getArgument(1), self.getArgument(2))) if self.ptype in [constants.SUBMIT_JOB, constants.SUBMIT_JOB_BG, constants.SUBMIT_JOB_HIGH, constants.SUBMIT_JOB_HIGH_BG, constants.SUBMIT_JOB_LOW, constants.SUBMIT_JOB_LOW_BG, constants.SUBMIT_JOB_SCHED, constants.SUBMIT_JOB_EPOCH]: return ' function: %s unique: %s' % (self.getArgument(0), self.getArgument(1)) if self.ptype in [constants.CAN_DO, constants.CANT_DO, constants.CAN_DO_TIMEOUT]: return ' function: %s' % (self.getArgument(0),) if self.ptype == constants.SET_CLIENT_ID: return ' id: %s' % (self.getArgument(0),) if self.ptype in [constants.OPTION_REQ, constants.OPTION_RES]: return ' option: %s' % (self.getArgument(0),) if self.ptype == constants.ERROR: return ' code: %s message: %s' % (self.getArgument(0), self.getArgument(1)) return ''
[docs] def toBinary(self): """Return a Gearman wire protocol binary representation of the packet. :returns: The packet in binary form. :rtype: bytes """ b = struct.pack('!4sii', self.code, self.ptype, len(self.data)) b = bytearray(b) b += self.data return b
[docs] def getArgument(self, index, last=False): """Get the nth argument from the packet data. :arg int index: The argument index to look up. :arg bool last: Whether this is the last argument (and thus nulls should be ignored) :returns: The argument value. :rtype: bytes """ parts = self.data.split(b'\x00') if not last: return parts[index] return b'\x00'.join(parts[index:])
[docs] def getJob(self): """Get the :py:class:`Job` associated with the job handle in this packet. :returns: The :py:class:`Job` for this packet. :rtype: Job :raises UnknownJobError: If the job is not known. """ handle = self.getArgument(0) job = self.connection.related_jobs.get(handle) if not job: raise UnknownJobError() return job
class BaseClientServer(object): def __init__(self, client_id=None): if client_id: self.client_id = convert_to_bytes(client_id) self.log = logging.getLogger("gear.BaseClientServer.%s" % (self.client_id,)) else: self.client_id = None self.log = logging.getLogger("gear.BaseClientServer") self.running = True self.active_connections = [] self.inactive_connections = [] self.connection_index = -1 # A lock and notification mechanism to handle not having any # current connections self.connections_condition = threading.Condition() # A pipe to wake up the poll loop in case it needs to restart self.wake_read, self.wake_write = os.pipe() self.poll_thread = threading.Thread(name="Gearman client poll", target=self._doPollLoop) self.poll_thread.daemon = True self.poll_thread.start() self.connect_thread = threading.Thread(name="Gearman client connect", target=self._doConnectLoop) self.connect_thread.daemon = True self.connect_thread.start() def _doConnectLoop(self): # Outer run method of the reconnection thread while self.running: self.connections_condition.acquire() while self.running and not self.inactive_connections: self.log.debug("Waiting for change in available servers " "to reconnect") self.connections_condition.wait() self.connections_condition.release() self.log.debug("Checking if servers need to be reconnected") try: if self.running and not self._connectLoop(): # Nothing happened time.sleep(2) except Exception: self.log.exception("Exception in connect loop:") def _connectLoop(self): # Inner method of the reconnection loop, triggered by # a connection change success = False for conn in self.inactive_connections[:]: self.log.debug("Trying to reconnect %s" % conn) try: conn.reconnect() except ConnectionError: self.log.debug("Unable to connect to %s" % conn) continue except Exception: self.log.exception("Exception while connecting to %s" % conn) continue try: self._onConnect(conn) except Exception: self.log.exception("Exception while performing on-connect " "tasks for %s" % conn) continue self.connections_condition.acquire() self.inactive_connections.remove(conn) self.active_connections.append(conn) self.connections_condition.notifyAll() os.write(self.wake_write, b'1\n') self.connections_condition.release() try: self._onActiveConnection(conn) except Exception: self.log.exception("Exception while performing active conn " "tasks for %s" % conn) success = True return success def _onConnect(self, conn): # Called immediately after a successful (re-)connection pass def _onActiveConnection(self, conn): # Called immediately after a connection is activated pass def _lostConnection(self, conn): # Called as soon as a connection is detected as faulty. Remove # it and return ASAP and let the connection thread deal with it. self.log.debug("Marking %s as disconnected" % conn) self.connections_condition.acquire() try: # NOTE(notmorgan): In the loop below it is possible to change the # jobs list on the connection. In python 3 .values() is an iter not # a static list, meaning that a change will break the for loop # as the object being iterated on will have changed in size. jobs = list(conn.related_jobs.values()) if conn in self.active_connections: self.active_connections.remove(conn) if conn not in self.inactive_connections: self.inactive_connections.append(conn) finally: self.connections_condition.notifyAll() self.connections_condition.release() for job in jobs: self.handleDisconnect(job) def _doPollLoop(self): # Outer run method of poll thread. while self.running: self.connections_condition.acquire() while self.running and not self.active_connections: self.log.debug("Waiting for change in available connections " "to poll") self.connections_condition.wait() self.connections_condition.release() try: self._pollLoop() except socket.error as e: if e.errno == errno.ECONNRESET: self.log.debug("Connection reset by peer") # This will get logged later at info level as # "Marking ... as disconnected" except Exception: self.log.exception("Exception in poll loop:") def _pollLoop(self): # Inner method of poll loop self.log.debug("Preparing to poll") poll = select.poll() bitmask = (select.POLLIN | select.POLLERR | select.POLLHUP | select.POLLNVAL) # Reverse mapping of fd -> connection conn_dict = {} for conn in self.active_connections: poll.register(conn.conn.fileno(), bitmask) conn_dict[conn.conn.fileno()] = conn # Register the wake pipe so that we can break if we need to # reconfigure connections poll.register(self.wake_read, bitmask) while self.running: self.log.debug("Polling %s connections" % len(self.active_connections)) ret = poll.poll() for fd, event in ret: if fd == self.wake_read: self.log.debug("Woken by pipe") while True: if os.read(self.wake_read, 1) == b'\n': break return conn = conn_dict[fd] if event & select.POLLIN: # Process all packets that may have been read in this # round of recv's by readPacket. while True: self.log.debug("Processing input on %s" % conn) p = conn.readPacket() if p: if isinstance(p, Packet): self.handlePacket(p) else: self.handleAdminRequest(p) else: self.log.debug("Received no data on %s" % conn) self._lostConnection(conn) return if not conn.hasPendingData(): break else: self.log.debug("Received error event on %s" % conn) self._lostConnection(conn) return def handlePacket(self, packet): """Handle a received packet. This method is called whenever a packet is received from any connection. It normally calls the handle method appropriate for the specific packet. :arg Packet packet: The :py:class:`Packet` that was received. """ self.log.info("Received packet from %s: %s" % (packet.connection, packet)) start = time.time() if packet.ptype == constants.JOB_CREATED: self.handleJobCreated(packet) elif packet.ptype == constants.WORK_COMPLETE: self.handleWorkComplete(packet) elif packet.ptype == constants.WORK_FAIL: self.handleWorkFail(packet) elif packet.ptype == constants.WORK_EXCEPTION: self.handleWorkException(packet) elif packet.ptype == constants.WORK_DATA: self.handleWorkData(packet) elif packet.ptype == constants.WORK_WARNING: self.handleWorkWarning(packet) elif packet.ptype == constants.WORK_STATUS: self.handleWorkStatus(packet) elif packet.ptype == constants.STATUS_RES: self.handleStatusRes(packet) elif packet.ptype == constants.GET_STATUS: self.handleGetStatus(packet) elif packet.ptype == constants.JOB_ASSIGN_UNIQ: self.handleJobAssignUnique(packet) elif packet.ptype == constants.JOB_ASSIGN: self.handleJobAssign(packet) elif packet.ptype == constants.NO_JOB: self.handleNoJob(packet) elif packet.ptype == constants.NOOP: self.handleNoop(packet) elif packet.ptype == constants.SUBMIT_JOB: self.handleSubmitJob(packet) elif packet.ptype == constants.SUBMIT_JOB_BG: self.handleSubmitJobBg(packet) elif packet.ptype == constants.SUBMIT_JOB_HIGH: self.handleSubmitJobHigh(packet) elif packet.ptype == constants.SUBMIT_JOB_HIGH_BG: self.handleSubmitJobHighBg(packet) elif packet.ptype == constants.SUBMIT_JOB_LOW: self.handleSubmitJobLow(packet) elif packet.ptype == constants.SUBMIT_JOB_LOW_BG: self.handleSubmitJobLowBg(packet) elif packet.ptype == constants.SUBMIT_JOB_SCHED: self.handleSubmitJobSched(packet) elif packet.ptype == constants.SUBMIT_JOB_EPOCH: self.handleSubmitJobEpoch(packet) elif packet.ptype == constants.GRAB_JOB_UNIQ: self.handleGrabJobUniq(packet) elif packet.ptype == constants.GRAB_JOB: self.handleGrabJob(packet) elif packet.ptype == constants.PRE_SLEEP: self.handlePreSleep(packet) elif packet.ptype == constants.SET_CLIENT_ID: self.handleSetClientID(packet) elif packet.ptype == constants.CAN_DO: self.handleCanDo(packet) elif packet.ptype == constants.CAN_DO_TIMEOUT: self.handleCanDoTimeout(packet) elif packet.ptype == constants.CANT_DO: self.handleCantDo(packet) elif packet.ptype == constants.RESET_ABILITIES: self.handleResetAbilities(packet) elif packet.ptype == constants.ECHO_REQ: self.handleEchoReq(packet) elif packet.ptype == constants.ECHO_RES: self.handleEchoRes(packet) elif packet.ptype == constants.ERROR: self.handleError(packet) elif packet.ptype == constants.ALL_YOURS: self.handleAllYours(packet) elif packet.ptype == constants.OPTION_REQ: self.handleOptionReq(packet) elif packet.ptype == constants.OPTION_RES: self.handleOptionRes(packet) else: self.log.error("Received unknown packet: %s" % packet) end = time.time() self.reportTimingStats(packet.ptype, end - start) def handleDisconnect(self, job): """Handle a Gearman server disconnection. If the Gearman server is disconnected, this will be called for any jobs currently associated with the server. :arg Job packet: The :py:class:`Job` that was running when the server disconnected. """ return job def reportTimingStats(self, ptype, duration): """Report processing times by packet type This method is called by handlePacket to report how long processing took for each packet. The default implementation does nothing. :arg bytes ptype: The packet type (one of the packet types in constants). :arg float duration: The time (in seconds) it took to process the packet. """ pass def _defaultPacketHandler(self, packet): self.log.error("Received unhandled packet: %s" % packet) def handleJobCreated(self, packet): return self._defaultPacketHandler(packet) def handleWorkComplete(self, packet): return self._defaultPacketHandler(packet) def handleWorkFail(self, packet): return self._defaultPacketHandler(packet) def handleWorkException(self, packet): return self._defaultPacketHandler(packet) def handleWorkData(self, packet): return self._defaultPacketHandler(packet) def handleWorkWarning(self, packet): return self._defaultPacketHandler(packet) def handleWorkStatus(self, packet): return self._defaultPacketHandler(packet) def handleStatusRes(self, packet): return self._defaultPacketHandler(packet) def handleGetStatus(self, packet): return self._defaultPacketHandler(packet) def handleJobAssignUnique(self, packet): return self._defaultPacketHandler(packet) def handleJobAssign(self, packet): return self._defaultPacketHandler(packet) def handleNoJob(self, packet): return self._defaultPacketHandler(packet) def handleNoop(self, packet): return self._defaultPacketHandler(packet) def handleSubmitJob(self, packet): return self._defaultPacketHandler(packet) def handleSubmitJobBg(self, packet): return self._defaultPacketHandler(packet) def handleSubmitJobHigh(self, packet): return self._defaultPacketHandler(packet) def handleSubmitJobHighBg(self, packet): return self._defaultPacketHandler(packet) def handleSubmitJobLow(self, packet): return self._defaultPacketHandler(packet) def handleSubmitJobLowBg(self, packet): return self._defaultPacketHandler(packet) def handleSubmitJobSched(self, packet): return self._defaultPacketHandler(packet) def handleSubmitJobEpoch(self, packet): return self._defaultPacketHandler(packet) def handleGrabJobUniq(self, packet): return self._defaultPacketHandler(packet) def handleGrabJob(self, packet): return self._defaultPacketHandler(packet) def handlePreSleep(self, packet): return self._defaultPacketHandler(packet) def handleSetClientID(self, packet): return self._defaultPacketHandler(packet) def handleCanDo(self, packet): return self._defaultPacketHandler(packet) def handleCanDoTimeout(self, packet): return self._defaultPacketHandler(packet) def handleCantDo(self, packet): return self._defaultPacketHandler(packet) def handleResetAbilities(self, packet): return self._defaultPacketHandler(packet) def handleEchoReq(self, packet): return self._defaultPacketHandler(packet) def handleEchoRes(self, packet): return self._defaultPacketHandler(packet) def handleError(self, packet): return self._defaultPacketHandler(packet) def handleAllYours(self, packet): return self._defaultPacketHandler(packet) def handleOptionReq(self, packet): return self._defaultPacketHandler(packet) def handleOptionRes(self, packet): return self._defaultPacketHandler(packet) def handleAdminRequest(self, request): """Handle an administrative command response from Gearman. This method is called whenever a response to a previously issued administrative command is received from one of this client's connections. It normally releases the wait lock on the initiating AdminRequest object. :arg AdminRequest request: The :py:class:`AdminRequest` that initiated the received response. """ self.log.info("Received admin data %s" % request) request.setComplete() def shutdown(self): """Close all connections and stop all running threads. The object may no longer be used after shutdown is called. """ if self.running: self.log.debug("Beginning shutdown") self._shutdown() self.log.debug("Beginning cleanup") self._cleanup() self.log.debug("Finished shutdown") else: self.log.warning("Shutdown called when not currently running. " "Ignoring.") def _shutdown(self): # The first part of the shutdown process where all threads # are told to exit. self.running = False self.connections_condition.acquire() try: self.connections_condition.notifyAll() os.write(self.wake_write, b'1\n') finally: self.connections_condition.release() def _cleanup(self): # The second part of the shutdown process where we wait for all # threads to exit and then clean up. self.poll_thread.join() self.connect_thread.join() for connection in self.active_connections: connection.disconnect() self.active_connections = [] self.inactive_connections = [] os.close(self.wake_read) os.close(self.wake_write) class BaseClient(BaseClientServer): def __init__(self, client_id='unknown'): super(BaseClient, self).__init__(client_id) self.log = logging.getLogger("gear.BaseClient.%s" % (self.client_id,)) # A lock to use when sending packets that set the state across # all known connections. Note that it doesn't necessarily need # to be used for all broadcasts, only those that affect multi- # connection state, such as setting options or functions. self.broadcast_lock = threading.RLock() def addServer(self, host, port=4730, ssl_key=None, ssl_cert=None, ssl_ca=None, keepalive=False, tcp_keepidle=7200, tcp_keepintvl=75, tcp_keepcnt=9): """Add a server to the client's connection pool. Any number of Gearman servers may be added to a client. The client will connect to all of them and send jobs to them in a round-robin fashion. When servers are disconnected, the client will automatically remove them from the pool, continuously try to reconnect to them, and return them to the pool when reconnected. New servers may be added at any time. This is a non-blocking call that will return regardless of whether the initial connection succeeded. If you need to ensure that a connection is ready before proceeding, see :py:meth:`waitForServer`. When using SSL connections, all SSL files must be specified. :arg str host: The hostname or IP address of the server. :arg int port: The port on which the gearman server is listening. :arg str ssl_key: Path to the SSL private key. :arg str ssl_cert: Path to the SSL certificate. :arg str ssl_ca: Path to the CA certificate. :arg bool keepalive: Whether to use TCP keepalives :arg int tcp_keepidle: Idle time after which to start keepalives sending :arg int tcp_keepintvl: Interval in seconds between TCP keepalives :arg int tcp_keepcnt: Count of TCP keepalives to send before disconnect :raises ConfigurationError: If the host/port combination has already been added to the client. """ self.log.debug("Adding server %s port %s" % (host, port)) self.connections_condition.acquire() try: for conn in self.active_connections + self.inactive_connections: if conn.host == host and conn.port == port: raise ConfigurationError("Host/port already specified") conn = Connection(host, port, ssl_key, ssl_cert, ssl_ca, self.client_id, keepalive, tcp_keepidle, tcp_keepintvl, tcp_keepcnt) self.inactive_connections.append(conn) self.connections_condition.notifyAll() finally: self.connections_condition.release() def _checkTimeout(self, start_time, timeout): if time.time() - start_time > timeout: raise TimeoutError() def waitForServer(self, timeout=None): """Wait for at least one server to be connected. Block until at least one gearman server is connected. :arg numeric timeout: Number of seconds to wait for a connection. If None, wait forever (default: no timeout). :raises TimeoutError: If the timeout is reached before any server connects. """ connected = False start_time = time.time() while self.running: self.connections_condition.acquire() try: while self.running and not self.active_connections: if timeout is not None: self._checkTimeout(start_time, timeout) self.log.debug("Waiting for at least one active " "connection") self.connections_condition.wait(timeout=1) if self.active_connections: self.log.debug("Active connection found") connected = True finally: self.connections_condition.release() if connected: return def getConnection(self): """Return a connected server. Finds the next scheduled connected server in the round-robin rotation and returns it. It is not usually necessary to use this method external to the library, as more consumer-oriented methods such as submitJob already use it internally, but is available nonetheless if necessary. :returns: The next scheduled :py:class:`Connection` object. :rtype: :py:class:`Connection` :raises NoConnectedServersError: If there are not currently connected servers. """ conn = None try: self.connections_condition.acquire() if not self.active_connections: raise NoConnectedServersError("No connected Gearman servers") self.connection_index += 1 if self.connection_index >= len(self.active_connections): self.connection_index = 0 conn = self.active_connections[self.connection_index] finally: self.connections_condition.release() return conn def broadcast(self, packet): """Send a packet to all currently connected servers. :arg Packet packet: The :py:class:`Packet` to send. """ connections = self.active_connections[:] for connection in connections: try: self.sendPacket(packet, connection) except Exception: # Error handling is all done by sendPacket pass def sendPacket(self, packet, connection): """Send a packet to a single connection, removing it from the list of active connections if that fails. :arg Packet packet: The :py:class:`Packet` to send. :arg Connection connection: The :py:class:`Connection` on which to send the packet. """ try: connection.sendPacket(packet) return except Exception: self.log.exception("Exception while sending packet %s to %s" % (packet, connection)) # If we can't send the packet, discard the connection self._lostConnection(connection) raise def handleEchoRes(self, packet): """Handle an ECHO_RES packet. Causes the blocking :py:meth:`Connection.echo` invocation to return. :arg Packet packet: The :py:class:`Packet` that was received. :returns: None """ packet.connection.handleEchoRes(packet.getArgument(0, True)) def handleError(self, packet): """Handle an ERROR packet. Logs the error. :arg Packet packet: The :py:class:`Packet` that was received. :returns: None """ self.log.error("Received ERROR packet: %s: %s" % (packet.getArgument(0), packet.getArgument(1))) try: task = packet.connection.pending_tasks.pop(0) task.setComplete() except Exception: self.log.exception("Exception while handling error packet:") self._lostConnection(packet.connection)
[docs]class Client(BaseClient): """A Gearman client. You may wish to subclass this class in order to override the default event handlers to react to Gearman events. Be sure to call the superclass event handlers so that they may perform job-related housekeeping. :arg str client_id: The client ID to provide to Gearman. It will appear in administrative output and be appended to the name of the logger (e.g., gear.Client.client_id). Defaults to 'unknown'. """ def __init__(self, client_id='unknown'): super(Client, self).__init__(client_id) self.log = logging.getLogger("gear.Client.%s" % (self.client_id,)) self.options = set() def __repr__(self): return '<gear.Client 0x%x>' % id(self) def _onConnect(self, conn): # Called immediately after a successful (re-)connection self.broadcast_lock.acquire() try: super(Client, self)._onConnect(conn) for name in self.options: self._setOptionConnection(name, conn) finally: self.broadcast_lock.release() def _setOptionConnection(self, name, conn): # Set an option on a connection packet = Packet(constants.REQ, constants.OPTION_REQ, name) task = OptionReqTask() try: conn.pending_tasks.append(task) self.sendPacket(packet, conn) except Exception: # Error handling is all done by sendPacket task = None return task
[docs] def setOption(self, name, timeout=30): """Set an option for all connections. :arg str name: The option name to set. :arg int timeout: How long to wait (in seconds) for a response from the server before giving up (default: 30 seconds). :returns: True if the option was set on all connections, otherwise False :rtype: bool """ tasks = {} name = convert_to_bytes(name) self.broadcast_lock.acquire() try: self.options.add(name) connections = self.active_connections[:] for connection in connections: task = self._setOptionConnection(name, connection) if task: tasks[task] = connection finally: self.broadcast_lock.release() success = True for task in tasks.keys(): complete = task.wait(timeout) conn = tasks[task] if not complete: self.log.error("Connection %s timed out waiting for a " "response to an option request: %s" % (conn, name)) self._lostConnection(conn) continue if name not in conn.options: success = False return success
[docs] def submitJob(self, job, background=False, precedence=PRECEDENCE_NORMAL, timeout=30): """Submit a job to a Gearman server. Submits the provided job to the next server in this client's round-robin connection pool. If the job is a foreground job, updates will be made to the supplied :py:class:`Job` object as they are received. :arg Job job: The :py:class:`Job` to submit. :arg bool background: Whether the job should be backgrounded. :arg int precedence: Whether the job should have normal, low, or high precedence. One of :py:data:`PRECEDENCE_NORMAL`, :py:data:`PRECEDENCE_LOW`, or :py:data:`PRECEDENCE_HIGH` :arg int timeout: How long to wait (in seconds) for a response from the server before giving up (default: 30 seconds). :raises ConfigurationError: If an invalid precendence value is supplied. """ if job.unique is None: unique = b'' else: unique = job.binary_unique data = b'\x00'.join((job.binary_name, unique, job.binary_arguments)) if background: if precedence == PRECEDENCE_NORMAL: cmd = constants.SUBMIT_JOB_BG elif precedence == PRECEDENCE_LOW: cmd = constants.SUBMIT_JOB_LOW_BG elif precedence == PRECEDENCE_HIGH: cmd = constants.SUBMIT_JOB_HIGH_BG else: raise ConfigurationError("Invalid precedence value") else: if precedence == PRECEDENCE_NORMAL: cmd = constants.SUBMIT_JOB elif precedence == PRECEDENCE_LOW: cmd = constants.SUBMIT_JOB_LOW elif precedence == PRECEDENCE_HIGH: cmd = constants.SUBMIT_JOB_HIGH else: raise ConfigurationError("Invalid precedence value") packet = Packet(constants.REQ, cmd, data) attempted_connections = set() while True: if attempted_connections == set(self.active_connections): break conn = self.getConnection() task = SubmitJobTask(job) conn.pending_tasks.append(task) attempted_connections.add(conn) try: self.sendPacket(packet, conn) except Exception: # Error handling is all done by sendPacket continue complete = task.wait(timeout) if not complete: self.log.error("Connection %s timed out waiting for a " "response to a submit job request: %s" % (conn, job)) self._lostConnection(conn) continue if not job.handle: self.log.error("Connection %s sent an error in " "response to a submit job request: %s" % (conn, job)) continue job.connection = conn return raise GearmanError("Unable to submit job to any connected servers")
[docs] def handleJobCreated(self, packet): """Handle a JOB_CREATED packet. Updates the appropriate :py:class:`Job` with the newly returned job handle. :arg Packet packet: The :py:class:`Packet` that was received. :returns: The :py:class:`Job` object associated with the job request. :rtype: :py:class:`Job` """ task = packet.connection.pending_tasks.pop(0) if not isinstance(task, SubmitJobTask): msg = ("Unexpected response received to submit job " "request: %s" % packet) self.log.error(msg) self._lostConnection(packet.connection) raise GearmanError(msg) job = task.job job.handle = packet.data packet.connection.related_jobs[job.handle] = job task.setComplete() self.log.debug("Job created; %s" % job) return job
[docs] def handleWorkComplete(self, packet): """Handle a WORK_COMPLETE packet. Updates the referenced :py:class:`Job` with the returned data and removes it from the list of jobs associated with the connection. :arg Packet packet: The :py:class:`Packet` that was received. :returns: The :py:class:`Job` object associated with the job request. :rtype: :py:class:`Job` """ job = packet.getJob() data = packet.getArgument(1, True) if data: job.data.append(data) job.complete = True job.failure = False del packet.connection.related_jobs[job.handle] self.log.debug("Job complete; %s data: %s" % (job, job.data)) return job
[docs] def handleWorkFail(self, packet): """Handle a WORK_FAIL packet. Updates the referenced :py:class:`Job` with the returned data and removes it from the list of jobs associated with the connection. :arg Packet packet: The :py:class:`Packet` that was received. :returns: The :py:class:`Job` object associated with the job request. :rtype: :py:class:`Job` """ job = packet.getJob() job.complete = True job.failure = True del packet.connection.related_jobs[job.handle] self.log.debug("Job failed; %s" % job) return job
[docs] def handleWorkException(self, packet): """Handle a WORK_Exception packet. Updates the referenced :py:class:`Job` with the returned data and removes it from the list of jobs associated with the connection. :arg Packet packet: The :py:class:`Packet` that was received. :returns: The :py:class:`Job` object associated with the job request. :rtype: :py:class:`Job` """ job = packet.getJob() job.exception = packet.getArgument(1, True) job.complete = True job.failure = True del packet.connection.related_jobs[job.handle] self.log.debug("Job exception; %s exception: %s" % (job, job.exception)) return job
[docs] def handleWorkData(self, packet): """Handle a WORK_DATA packet. Updates the referenced :py:class:`Job` with the returned data. :arg Packet packet: The :py:class:`Packet` that was received. :returns: The :py:class:`Job` object associated with the job request. :rtype: :py:class:`Job` """ job = packet.getJob() data = packet.getArgument(1, True) if data: job.data.append(data) self.log.debug("Job data; job: %s data: %s" % (job, job.data)) return job
[docs] def handleWorkWarning(self, packet): """Handle a WORK_WARNING packet. Updates the referenced :py:class:`Job` with the returned data. :arg Packet packet: The :py:class:`Packet` that was received. :returns: The :py:class:`Job` object associated with the job request. :rtype: :py:class:`Job` """ job = packet.getJob() data = packet.getArgument(1, True) if data: job.data.append(data) job.warning = True self.log.debug("Job warning; %s data: %s" % (job, job.data)) return job
[docs] def handleWorkStatus(self, packet): """Handle a WORK_STATUS packet. Updates the referenced :py:class:`Job` with the returned data. :arg Packet packet: The :py:class:`Packet` that was received. :returns: The :py:class:`Job` object associated with the job request. :rtype: :py:class:`Job` """ job = packet.getJob() job.numerator = packet.getArgument(1) job.denominator = packet.getArgument(2) try: job.fraction_complete = (float(job.numerator) / float(job.denominator)) except Exception: job.fraction_complete = None self.log.debug("Job status; %s complete: %s/%s" % (job, job.numerator, job.denominator)) return job
[docs] def handleStatusRes(self, packet): """Handle a STATUS_RES packet. Updates the referenced :py:class:`Job` with the returned data. :arg Packet packet: The :py:class:`Packet` that was received. :returns: The :py:class:`Job` object associated with the job request. :rtype: :py:class:`Job` """ job = packet.getJob() job.known = (packet.getArgument(1) == b'1') job.running = (packet.getArgument(2) == b'1') job.numerator = packet.getArgument(3) job.denominator = packet.getArgument(4) try: job.fraction_complete = (float(job.numerator) / float(job.denominator)) except Exception: job.fraction_complete = None return job
[docs] def handleOptionRes(self, packet): """Handle an OPTION_RES packet. Updates the set of options for the connection. :arg Packet packet: The :py:class:`Packet` that was received. :returns: None. """ task = packet.connection.pending_tasks.pop(0) if not isinstance(task, OptionReqTask): msg = ("Unexpected response received to option " "request: %s" % packet) self.log.error(msg) self._lostConnection(packet.connection) raise GearmanError(msg) packet.connection.handleOptionRes(packet.getArgument(0)) task.setComplete()
[docs]class FunctionRecord(object): """Represents a function that should be registered with Gearman. This class only directly needs to be instatiated for use with :py:meth:`Worker.setFunctions`. If a timeout value is supplied, the function will be registered with CAN_DO_TIMEOUT. :arg str name: The name of the function to register. :arg numeric timeout: The timeout value (optional). """ def __init__(self, name, timeout=None): self.name = name self.timeout = timeout def __repr__(self): return '<gear.FunctionRecord 0x%x name: %s timeout: %s>' % ( id(self), self.name, self.timeout)
class BaseJob(object): def __init__(self, name, arguments, unique=None, handle=None): self._name = convert_to_bytes(name) self._validate_arguments(arguments) self._arguments = convert_to_bytes(arguments) self._unique = convert_to_bytes(unique) self.handle = handle self.connection = None def _validate_arguments(self, arguments): if (not isinstance(arguments, bytes) and not isinstance(arguments, bytearray)): raise TypeError("arguments must be of type bytes or bytearray") @property def arguments(self): return self._arguments @arguments.setter def arguments(self, value): self._arguments = value @property def unique(self): return self._unique @unique.setter def unique(self, value): self._unique = value @property def name(self): if isinstance(self._name, six.binary_type): return self._name.decode('utf-8') return self._name @name.setter def name(self, value): if isinstance(value, six.text_type): value = value.encode('utf-8') self._name = value @property def binary_name(self): return self._name @property def binary_arguments(self): return self._arguments @property def binary_unique(self): return self._unique def __repr__(self): return '<gear.Job 0x%x handle: %s name: %s unique: %s>' % ( id(self), self.handle, self.name, self.unique)
[docs]class WorkerJob(BaseJob): """A job that Gearman has assigned to a Worker. Not intended to be instantiated directly, but rather returned by :py:meth:`Worker.getJob`. :arg str handle: The job handle assigned by gearman. :arg str name: The name of the job. :arg bytes arguments: The opaque data blob passed to the worker as arguments. :arg str unique: A byte string to uniquely identify the job to Gearman (optional). The following instance attributes are available: **name** (str) The name of the job. Assumed to be utf-8. **arguments** (bytes) The opaque data blob passed to the worker as arguments. **unique** (str or None) The unique ID of the job (if supplied). **handle** (bytes) The Gearman job handle. **connection** (:py:class:`Connection` or None) The connection associated with the job. Only set after the job has been submitted to a Gearman server. """ def __init__(self, handle, name, arguments, unique=None): super(WorkerJob, self).__init__(name, arguments, unique, handle)
[docs] def sendWorkData(self, data=b''): """Send a WORK_DATA packet to the client. :arg bytes data: The data to be sent to the client (optional). """ data = self.handle + b'\x00' + data p = Packet(constants.REQ, constants.WORK_DATA, data) self.connection.sendPacket(p)
[docs] def sendWorkWarning(self, data=b''): """Send a WORK_WARNING packet to the client. :arg bytes data: The data to be sent to the client (optional). """ data = self.handle + b'\x00' + data p = Packet(constants.REQ, constants.WORK_WARNING, data) self.connection.sendPacket(p)
[docs] def sendWorkStatus(self, numerator, denominator): """Send a WORK_STATUS packet to the client. Sends a numerator and denominator that together represent the fraction complete of the job. :arg numeric numerator: The numerator of the fraction complete. :arg numeric denominator: The denominator of the fraction complete. """ data = (self.handle + b'\x00' + str(numerator).encode('utf8') + b'\x00' + str(denominator).encode('utf8')) p = Packet(constants.REQ, constants.WORK_STATUS, data) self.connection.sendPacket(p)
[docs] def sendWorkComplete(self, data=b''): """Send a WORK_COMPLETE packet to the client. :arg bytes data: The data to be sent to the client (optional). """ data = self.handle + b'\x00' + data p = Packet(constants.REQ, constants.WORK_COMPLETE, data) self.connection.sendPacket(p)
[docs] def sendWorkFail(self): "Send a WORK_FAIL packet to the client." p = Packet(constants.REQ, constants.WORK_FAIL, self.handle) self.connection.sendPacket(p)
[docs] def sendWorkException(self, data=b''): """Send a WORK_EXCEPTION packet to the client. :arg bytes data: The exception data to be sent to the client (optional). """ data = self.handle + b'\x00' + data p = Packet(constants.REQ, constants.WORK_EXCEPTION, data) self.connection.sendPacket(p)
[docs]class Worker(BaseClient): """A Gearman worker. :arg str client_id: The client ID to provide to Gearman. It will appear in administrative output and be appended to the name of the logger (e.g., gear.Worker.client_id). :arg str worker_id: The client ID to provide to Gearman. It will appear in administrative output and be appended to the name of the logger (e.g., gear.Worker.client_id). This parameter name is deprecated, use client_id instead. """ job_class = WorkerJob def __init__(self, client_id=None, worker_id=None): if not client_id or worker_id: raise Exception("A client_id must be provided") if worker_id: client_id = worker_id super(Worker, self).__init__(client_id) self.log = logging.getLogger("gear.Worker.%s" % (self.client_id,)) self.worker_id = client_id self.functions = {} self.job_lock = threading.Lock() self.waiting_for_jobs = 0 self.job_queue = queue_mod.Queue() def __repr__(self): return '<gear.Worker 0x%x>' % id(self)
[docs] def registerFunction(self, name, timeout=None): """Register a function with Gearman. If a timeout value is supplied, the function will be registered with CAN_DO_TIMEOUT. :arg str name: The name of the function to register. :arg numeric timeout: The timeout value (optional). """ name = convert_to_bytes(name) self.functions[name] = FunctionRecord(name, timeout) if timeout: self._sendCanDoTimeout(name, timeout) else: self._sendCanDo(name) connections = self.active_connections[:] for connection in connections: if connection.state == "SLEEP": connection.changeState("IDLE") self._updateStateMachines()
[docs] def unRegisterFunction(self, name): """Remove a function from Gearman's registry. :arg str name: The name of the function to remove. """ name = convert_to_bytes(name) del self.functions[name] self._sendCantDo(name)
[docs] def setFunctions(self, functions): """Replace the set of functions registered with Gearman. Accepts a list of :py:class:`FunctionRecord` objects which represents the complete set of functions that should be registered with Gearman. Any existing functions will be unregistered and these registered in their place. If the empty list is supplied, then the Gearman registered function set will be cleared. :arg list functions: A list of :py:class:`FunctionRecord` objects. """ self._sendResetAbilities() self.functions = {} for f in functions: if not isinstance(f, FunctionRecord): raise InvalidDataError( "An iterable of FunctionRecords is required.") self.functions[f.name] = f for f in self.functions.values(): if f.timeout: self._sendCanDoTimeout(f.name, f.timeout) else: self._sendCanDo(f.name)
def _sendCanDo(self, name): self.broadcast_lock.acquire() try: p = Packet(constants.REQ, constants.CAN_DO, name) self.broadcast(p) finally: self.broadcast_lock.release() def _sendCanDoTimeout(self, name, timeout): self.broadcast_lock.acquire() try: data = name + b'\x00' + timeout p = Packet(constants.REQ, constants.CAN_DO_TIMEOUT, data) self.broadcast(p) finally: self.broadcast_lock.release() def _sendCantDo(self, name): self.broadcast_lock.acquire() try: p = Packet(constants.REQ, constants.CANT_DO, name) self.broadcast(p) finally: self.broadcast_lock.release() def _sendResetAbilities(self): self.broadcast_lock.acquire() try: p = Packet(constants.REQ, constants.RESET_ABILITIES, b'') self.broadcast(p) finally: self.broadcast_lock.release() def _sendPreSleep(self, connection): p = Packet(constants.REQ, constants.PRE_SLEEP, b'') self.sendPacket(p, connection) def _sendGrabJobUniq(self, connection=None): p = Packet(constants.REQ, constants.GRAB_JOB_UNIQ, b'') if connection: self.sendPacket(p, connection) else: self.broadcast(p) def _onConnect(self, conn): self.broadcast_lock.acquire() try: # Called immediately after a successful (re-)connection p = Packet(constants.REQ, constants.SET_CLIENT_ID, self.client_id) conn.sendPacket(p) super(Worker, self)._onConnect(conn) for f in self.functions.values(): if f.timeout: data = f.name + b'\x00' + f.timeout p = Packet(constants.REQ, constants.CAN_DO_TIMEOUT, data) else: p = Packet(constants.REQ, constants.CAN_DO, f.name) conn.sendPacket(p) conn.changeState("IDLE") finally: self.broadcast_lock.release() # Any exceptions will be handled by the calling function, and the # connection will not be put into the pool. def _onActiveConnection(self, conn): self.job_lock.acquire() try: if self.waiting_for_jobs > 0: self._updateStateMachines() finally: self.job_lock.release() def _updateStateMachines(self): connections = self.active_connections[:] for connection in connections: if (connection.state == "IDLE" and self.waiting_for_jobs > 0): self._sendGrabJobUniq(connection) connection.changeState("GRAB_WAIT") if (connection.state != "IDLE" and self.waiting_for_jobs < 1): connection.changeState("IDLE")
[docs] def getJob(self): """Get a job from Gearman. Blocks until a job is received. This method is re-entrant, so it is safe to call this method on a single worker from multiple threads. In that case, one of them at random will receive the job assignment. :returns: The :py:class:`WorkerJob` assigned. :rtype: :py:class:`WorkerJob`. :raises InterruptedError: If interrupted (by :py:meth:`stopWaitingForJobs`) before a job is received. """ self.job_lock.acquire() try: # self.running gets cleared during _shutdown(), before the # stopWaitingForJobs() is called. This check has to # happen with the job_lock held, otherwise there would be # a window for race conditions between manipulation of # "running" and "waiting_for_jobs". if not self.running: raise InterruptedError() self.waiting_for_jobs += 1 self.log.debug("Get job; number of threads waiting for jobs: %s" % self.waiting_for_jobs) try: job = self.job_queue.get(False) except queue_mod.Empty: job = None if not job: self._updateStateMachines() finally: self.job_lock.release() if not job: job = self.job_queue.get() self.log.debug("Received job: %s" % job) if job is None: raise InterruptedError() return job
[docs] def stopWaitingForJobs(self): """Interrupts all running :py:meth:`getJob` calls, which will raise an exception. """ self.job_lock.acquire() try: while True: connections = self.active_connections[:] now = time.time() ok = True for connection in connections: if connection.state == "GRAB_WAIT": # Replies to GRAB_JOB should be fast, give up if we've # been waiting for more than 5 seconds. if now - connection.state_time > 5: self._lostConnection(connection) else: ok = False if ok: break else: self.job_lock.release() time.sleep(0.1) self.job_lock.acquire() while self.waiting_for_jobs > 0: self.waiting_for_jobs -= 1 self.job_queue.put(None) self._updateStateMachines() finally: self.job_lock.release()
def _shutdown(self): self.job_lock.acquire() try: # The upstream _shutdown() will clear the "running" bool. Because # that is a variable which is used for proper synchronization of # the exit within getJob() which might be about to be called from a # separate thread, it's important to call it with a proper lock # being held. super(Worker, self)._shutdown() finally: self.job_lock.release() self.stopWaitingForJobs()
[docs] def handleNoop(self, packet): """Handle a NOOP packet. Sends a GRAB_JOB_UNIQ packet on the same connection. GRAB_JOB_UNIQ will return jobs regardless of whether they have been specified with a unique identifier when submitted. If they were not, then :py:attr:`WorkerJob.unique` attribute will be None. :arg Packet packet: The :py:class:`Packet` that was received. """ self.job_lock.acquire() try: if packet.connection.state == "SLEEP": self.log.debug("Sending GRAB_JOB_UNIQ") self._sendGrabJobUniq(packet.connection) packet.connection.changeState("GRAB_WAIT") else: self.log.debug("Received unexpecetd NOOP packet on %s" % packet.connection) finally: self.job_lock.release()
[docs] def handleNoJob(self, packet): """Handle a NO_JOB packet. Sends a PRE_SLEEP packet on the same connection. :arg Packet packet: The :py:class:`Packet` that was received. """ self.job_lock.acquire() try: if packet.connection.state == "GRAB_WAIT": self.log.debug("Sending PRE_SLEEP") self._sendPreSleep(packet.connection) packet.connection.changeState("SLEEP") else: self.log.debug("Received unexpected NO_JOB packet on %s" % packet.connection) finally: self.job_lock.release()
[docs] def handleJobAssign(self, packet): """Handle a JOB_ASSIGN packet. Adds a WorkerJob to the internal queue to be picked up by any threads waiting in :py:meth:`getJob`. :arg Packet packet: The :py:class:`Packet` that was received. """ handle = packet.getArgument(0) name = packet.getArgument(1) arguments = packet.getArgument(2, True) return self._handleJobAssignment(packet, handle, name, arguments, None)
[docs] def handleJobAssignUnique(self, packet): """Handle a JOB_ASSIGN_UNIQ packet. Adds a WorkerJob to the internal queue to be picked up by any threads waiting in :py:meth:`getJob`. :arg Packet packet: The :py:class:`Packet` that was received. """ handle = packet.getArgument(0) name = packet.getArgument(1) unique = packet.getArgument(2) if unique == b'': unique = None arguments = packet.getArgument(3, True) return self._handleJobAssignment(packet, handle, name, arguments, unique)
def _handleJobAssignment(self, packet, handle, name, arguments, unique): job = self.job_class(handle, name, arguments, unique) job.connection = packet.connection self.job_lock.acquire() try: packet.connection.changeState("IDLE") self.waiting_for_jobs -= 1 self.log.debug("Job assigned; number of threads waiting for " "jobs: %s" % self.waiting_for_jobs) self.job_queue.put(job) self._updateStateMachines() finally: self.job_lock.release()
[docs]class Job(BaseJob): """A job to run or being run by Gearman. :arg str name: The name of the job. :arg bytes arguments: The opaque data blob to be passed to the worker as arguments. :arg str unique: A byte string to uniquely identify the job to Gearman (optional). The following instance attributes are available: **name** (str) The name of the job. Assumed to be utf-8. **arguments** (bytes) The opaque data blob passed to the worker as arguments. **unique** (str or None) The unique ID of the job (if supplied). **handle** (bytes or None) The Gearman job handle. None if no job handle has been received yet. **data** (list of byte-arrays) The result data returned from Gearman. Each packet appends an element to the list. Depending on the nature of the data, the elements may need to be concatenated before use. This is returned as a snapshot copy of the data to prevent accidental attempts at modification which will be lost. **exception** (bytes or None) Exception information returned from Gearman. None if no exception has been received. **warning** (bool) Whether the worker has reported a warning. **complete** (bool) Whether the job is complete. **failure** (bool) Whether the job has failed. Only set when complete is True. **numerator** (bytes or None) The numerator of the completion ratio reported by the worker. Only set when a status update is sent by the worker. **denominator** (bytes or None) The denominator of the completion ratio reported by the worker. Only set when a status update is sent by the worker. **fraction_complete** (float or None) The fractional complete ratio reported by the worker. Only set when a status update is sent by the worker. **known** (bool or None) Whether the job is known to Gearman. Only set by handleStatusRes() in response to a getStatus() query. **running** (bool or None) Whether the job is running. Only set by handleStatusRes() in response to a getStatus() query. **connection** (:py:class:`Connection` or None) The connection associated with the job. Only set after the job has been submitted to a Gearman server. """ data_type = list def __init__(self, name, arguments, unique=None): super(Job, self).__init__(name, arguments, unique) self._data = self.data_type() self._exception = None self.warning = False self.complete = False self.failure = False self.numerator = None self.denominator = None self.fraction_complete = None self.known = None self.running = None @property def binary_data(self): for value in self._data: if isinstance(value, six.text_type): value = value.encode('utf-8') yield value @property def data(self): return self._data @data.setter def data(self, value): if not isinstance(value, self.data_type): raise ValueError( "data attribute must be {}".format(self.data_type)) self._data = value @property def exception(self): return self._exception @exception.setter def exception(self, value): self._exception = value
class TextJobArguments(object): """Assumes utf-8 arguments in addition to name If one is always dealing in valid utf-8, using this job class relieves one of the need to encode/decode constantly.""" def _validate_arguments(self, arguments): pass @property def arguments(self): args = self._arguments if isinstance(args, six.binary_type): return args.decode('utf-8') return args @arguments.setter def arguments(self, value): if not isinstance(value, six.binary_type): value = value.encode('utf-8') self._arguments = value class TextJobUnique(object): """Assumes utf-8 unique If one is always dealing in valid utf-8, using this job class relieves one of the need to encode/decode constantly.""" @property def unique(self): unique = self._unique if isinstance(unique, six.binary_type): return unique.decode('utf-8') return unique @unique.setter def unique(self, value): if not isinstance(value, six.binary_type): value = value.encode('utf-8') self._unique = value class TextList(list): def append(self, x): if isinstance(x, six.binary_type): x = x.decode('utf-8') super(TextList, self).append(x) def extend(self, iterable): def _iter(): for value in iterable: if isinstance(value, six.binary_type): yield value.decode('utf-8') else: yield value super(TextList, self).extend(_iter) def insert(self, i, x): if isinstance(x, six.binary_type): x = x.decode('utf-8') super(TextList, self).insert(i, x) class TextJob(TextJobArguments, TextJobUnique, Job): """ Sends and receives UTF-8 arguments and data. Use this instead of Job when you only expect to send valid UTF-8 through gearman. It will automatically encode arguments and work data as UTF-8, and any jobs fetched from this worker will have their arguments and data decoded assuming they are valid UTF-8, and thus return strings. Attributes and method signatures are thes ame as Job except as noted here: ** arguments ** (str) This will be returned as a string. ** data ** (tuple of str) This will be returned as a tuble of strings. """ data_type = TextList @property def exception(self): exception = self._exception if isinstance(exception, six.binary_type): return exception.decode('utf-8') return exception @exception.setter def exception(self, value): if not isinstance(value, six.binary_type): value = value.encode('utf-8') self._exception = value class TextWorkerJob(TextJobArguments, TextJobUnique, WorkerJob): """ Sends and receives UTF-8 arguments and data. See TextJob. sendWorkData and sendWorkWarning accept strings and will encode them as UTF-8. """ def sendWorkData(self, data=''): """Send a WORK_DATA packet to the client. :arg str data: The data to be sent to the client (optional). """ if isinstance(data, six.text_type): data = data.encode('utf8') return super(TextWorkerJob, self).sendWorkData(data) def sendWorkWarning(self, data=''): """Send a WORK_WARNING packet to the client. :arg str data: The data to be sent to the client (optional). """ if isinstance(data, six.text_type): data = data.encode('utf8') return super(TextWorkerJob, self).sendWorkWarning(data) def sendWorkComplete(self, data=''): """Send a WORK_COMPLETE packet to the client. :arg str data: The data to be sent to the client (optional). """ if isinstance(data, six.text_type): data = data.encode('utf8') return super(TextWorkerJob, self).sendWorkComplete(data) def sendWorkException(self, data=''): """Send a WORK_EXCEPTION packet to the client. :arg str data: The data to be sent to the client (optional). """ if isinstance(data, six.text_type): data = data.encode('utf8') return super(TextWorkerJob, self).sendWorkException(data) class TextWorker(Worker): """ Sends and receives UTF-8 only. See TextJob. """ job_class = TextWorkerJob class BaseBinaryJob(object): """ For the case where non-utf-8 job names are needed. It will function exactly like Job, except that the job name will not be decoded.""" @property def name(self): return self._name class BinaryWorkerJob(BaseBinaryJob, WorkerJob): pass class BinaryJob(BaseBinaryJob, Job): pass # Below are classes for use in the server implementation: class ServerJob(BinaryJob): """A job record for use in a server. :arg str name: The name of the job. :arg bytes arguments: The opaque data blob to be passed to the worker as arguments. :arg str unique: A byte string to uniquely identify the job to Gearman (optional). The following instance attributes are available: **name** (str) The name of the job. **arguments** (bytes) The opaque data blob passed to the worker as arguments. **unique** (str or None) The unique ID of the job (if supplied). **handle** (bytes or None) The Gearman job handle. None if no job handle has been received yet. **data** (list of byte-arrays) The result data returned from Gearman. Each packet appends an element to the list. Depending on the nature of the data, the elements may need to be concatenated before use. **exception** (bytes or None) Exception information returned from Gearman. None if no exception has been received. **warning** (bool) Whether the worker has reported a warning. **complete** (bool) Whether the job is complete. **failure** (bool) Whether the job has failed. Only set when complete is True. **numerator** (bytes or None) The numerator of the completion ratio reported by the worker. Only set when a status update is sent by the worker. **denominator** (bytes or None) The denominator of the completion ratio reported by the worker. Only set when a status update is sent by the worker. **fraction_complete** (float or None) The fractional complete ratio reported by the worker. Only set when a status update is sent by the worker. **known** (bool or None) Whether the job is known to Gearman. Only set by handleStatusRes() in response to a getStatus() query. **running** (bool or None) Whether the job is running. Only set by handleStatusRes() in response to a getStatus() query. **client_connection** :py:class:`Connection` The client connection associated with the job. **worker_connection** (:py:class:`Connection` or None) The worker connection associated with the job. Only set after the job has been assigned to a worker. """ def __init__(self, handle, name, arguments, client_connection, unique=None): super(ServerJob, self).__init__(name, arguments, unique) self.handle = handle self.client_connection = client_connection self.worker_connection = None del self.connection class ServerAdminRequest(AdminRequest): """An administrative request sent to a server.""" def __init__(self, connection): super(ServerAdminRequest, self).__init__() self.connection = connection def isComplete(self, data): end_index_newline = data.find(b'\n') if end_index_newline != -1: self.command = data[:end_index_newline] # Remove newline from data x = end_index_newline + 1 return (True, data[x:]) else: return (False, None) class NonBlockingConnection(Connection): """A Non-blocking connection to a Gearman Client.""" def __init__(self, host, port, ssl_key=None, ssl_cert=None, ssl_ca=None, client_id='unknown'): super(NonBlockingConnection, self).__init__( host, port, ssl_key, ssl_cert, ssl_ca, client_id) self.send_queue = [] def connect(self): super(NonBlockingConnection, self).connect() if self.connected and self.conn: self.conn.setblocking(0) def _readRawBytes(self, bytes_to_read): try: buff = self.conn.recv(bytes_to_read) except ssl.SSLError as e: if e.errno == ssl.SSL_ERROR_WANT_READ: raise RetryIOError() elif e.errno == ssl.SSL_ERROR_WANT_WRITE: raise RetryIOError() raise except socket.error as e: if e.errno == errno.EAGAIN: # Read operation would block, we're done until # epoll flags this connection again raise RetryIOError() raise return buff def sendPacket(self, packet): """Append a packet to this connection's send queue. The Client or Server must manage actually sending the data. :arg :py:class:`Packet` packet The packet to send """ self.log.debug("Queuing packet to %s: %s" % (self, packet)) self.send_queue.append(packet.toBinary()) self.sendQueuedData() def sendRaw(self, data): """Append raw data to this connection's send queue. The Client or Server must manage actually sending the data. :arg bytes data The raw data to send """ self.log.debug("Queuing data to %s: %s" % (self, data)) self.send_queue.append(data) self.sendQueuedData() def sendQueuedData(self): """Send previously queued data to the socket.""" try: while len(self.send_queue): data = self.send_queue.pop(0) r = 0 try: r = self.conn.send(data) except ssl.SSLError as e: if e.errno == ssl.SSL_ERROR_WANT_READ: raise RetryIOError() elif e.errno == ssl.SSL_ERROR_WANT_WRITE: raise RetryIOError() else: raise except socket.error as e: if e.errno == errno.EAGAIN: self.log.debug("Write operation on %s would block" % self) raise RetryIOError() else: raise finally: data = data[r:] if data: self.send_queue.insert(0, data) except RetryIOError: pass class ServerConnection(NonBlockingConnection): """A Connection to a Gearman Client.""" def __init__(self, addr, conn, use_ssl, client_id): if client_id: self.log = logging.getLogger("gear.ServerConnection.%s" % (client_id,)) else: self.log = logging.getLogger("gear.ServerConnection") self.send_queue = [] self.admin_requests = [] self.host = addr[0] self.port = addr[1] self.conn = conn self.conn.setblocking(0) self.input_buffer = b'' self.need_bytes = False self.use_ssl = use_ssl self.client_id = None self.functions = set() self.related_jobs = {} self.ssl_subject = None if self.use_ssl: for x in conn.getpeercert()['subject']: if x[0][0] == 'commonName': self.ssl_subject = x[0][1] self.log.debug("SSL subject: %s" % self.ssl_subject) self.changeState("INIT") def _getAdminRequest(self): return ServerAdminRequest(self) def _putAdminRequest(self, req): # The server does not need to keep track of admin requests # that have been partially received; it will simply create a # new instance the next time it tries to read. pass def __repr__(self): return '<gear.ServerConnection 0x%x name: %s host: %s port: %s>' % ( id(self), self.client_id, self.host, self.port)
[docs]class Server(BaseClientServer): """A simple gearman server implementation for testing (not for production use). :arg int port: The TCP port on which to listen. :arg str ssl_key: Path to the SSL private key. :arg str ssl_cert: Path to the SSL certificate. :arg str ssl_ca: Path to the CA certificate. :arg str statsd_host: statsd hostname. None means disabled (the default). :arg str statsd_port: statsd port (defaults to 8125). :arg str statsd_prefix: statsd key prefix. :arg str client_id: The ID associated with this server. It will be appending to the name of the logger (e.g., gear.Server.server_id). Defaults to None (unused). :arg ACL acl: An :py:class:`ACL` object if the server should apply access control rules to its connections. :arg str host: Host name or IPv4/IPv6 address to bind to. Defaults to "whatever getaddrinfo() returns", which might be IPv4-only. :arg bool keepalive: Whether to use TCP keepalives :arg int tcp_keepidle: Idle time after which to start keepalives sending :arg int tcp_keepintvl: Interval in seconds between TCP keepalives :arg int tcp_keepcnt: Count of TCP keepalives to send before disconnect """ edge_bitmask = select.EPOLLET error_bitmask = (select.EPOLLERR | select.EPOLLHUP | edge_bitmask) read_bitmask = (select.EPOLLIN | error_bitmask) readwrite_bitmask = (select.EPOLLOUT | read_bitmask) def __init__(self, port=4730, ssl_key=None, ssl_cert=None, ssl_ca=None, statsd_host=None, statsd_port=8125, statsd_prefix=None, server_id=None, acl=None, host=None, keepalive=False, tcp_keepidle=7200, tcp_keepintvl=75, tcp_keepcnt=9): self.port = port self.ssl_key = ssl_key self.ssl_cert = ssl_cert self.ssl_ca = ssl_ca self.high_queue = [] self.normal_queue = [] self.low_queue = [] self.jobs = {} self.running_jobs = 0 self.waiting_jobs = 0 self.total_jobs = 0 self.functions = set() self.max_handle = 0 self.acl = acl self.connect_wake_read, self.connect_wake_write = os.pipe() self.poll = select.epoll() # Reverse mapping of fd -> connection self.connection_map = {} self.use_ssl = False if all([self.ssl_key, self.ssl_cert, self.ssl_ca]): self.use_ssl = True # Get all valid passive listen addresses, then sort by family to prefer # ipv6 if available. addrs = socket.getaddrinfo(host, self.port, socket.AF_UNSPEC, socket.SOCK_STREAM, 0, socket.AI_PASSIVE | socket.AI_ADDRCONFIG) addrs.sort(key=lambda addr: addr[0], reverse=True) for res in addrs: af, socktype, proto, canonname, sa = res try: self.socket = socket.socket(af, socktype, proto) self.socket.setsockopt(socket.SOL_SOCKET, socket.SO_REUSEADDR, 1) if keepalive and hasattr(socket, 'TCP_KEEPIDLE'): self.socket.setsockopt(socket.SOL_SOCKET, socket.SO_KEEPALIVE, 1) self.socket.setsockopt(socket.IPPROTO_TCP, socket.TCP_KEEPIDLE, tcp_keepidle) self.socket.setsockopt(socket.IPPROTO_TCP, socket.TCP_KEEPINTVL, tcp_keepintvl) self.socket.setsockopt(socket.IPPROTO_TCP, socket.TCP_KEEPCNT, tcp_keepcnt) elif keepalive: self.log.warning('Keepalive requested but not available ' 'on this platform') except socket.error: self.socket = None continue try: self.socket.bind(sa) self.socket.listen(1) except socket.error: self.socket.close() self.socket = None continue break if self.socket is None: raise Exception("Could not open socket") if port == 0: self.port = self.socket.getsockname()[1] super(Server, self).__init__(server_id) # Register the wake pipe so that we can break if we need to # reconfigure connections self.poll.register(self.wake_read, self.read_bitmask) if server_id: self.log = logging.getLogger("gear.Server.%s" % (self.client_id,)) else: self.log = logging.getLogger("gear.Server") if statsd_host: if not statsd: self.log.error("Unable to import statsd module") self.statsd = None else: self.statsd = statsd.StatsClient(statsd_host, statsd_port, statsd_prefix) else: self.statsd = None def _doConnectLoop(self): while self.running: try: self.connectLoop() except Exception: self.log.exception("Exception in connect loop:") time.sleep(1) def connectLoop(self): poll = select.poll() bitmask = (select.POLLIN | select.POLLERR | select.POLLHUP | select.POLLNVAL) # Register the wake pipe so that we can break if we need to # shutdown. poll.register(self.connect_wake_read, bitmask) poll.register(self.socket.fileno(), bitmask) while self.running: ret = poll.poll() for fd, event in ret: if fd == self.connect_wake_read: self.log.debug("Accept woken by pipe") while True: if os.read(self.connect_wake_read, 1) == b'\n': break return if event & select.POLLIN: self.log.debug("Accepting new connection") c, addr = self.socket.accept() if self.use_ssl: context = create_ssl_context() context.verify_mode = ssl.CERT_REQUIRED context.load_cert_chain(self.ssl_cert, self.ssl_key) context.load_verify_locations(self.ssl_ca) c = context.wrap_socket(c, server_side=True) conn = ServerConnection(addr, c, self.use_ssl, self.client_id) self.log.info("Accepted connection %s" % (conn,)) self.connections_condition.acquire() try: self.active_connections.append(conn) self._registerConnection(conn) self.connections_condition.notifyAll() finally: self.connections_condition.release() def readFromConnection(self, conn): while True: self.log.debug("Processing input on %s" % conn) try: p = conn.readPacket() except RetryIOError: # Read operation would block, we're done until # epoll flags this connection again return if p: if isinstance(p, Packet): self.handlePacket(p) else: self.handleAdminRequest(p) else: self.log.debug("Received no data on %s" % conn) raise DisconnectError() def writeToConnection(self, conn): self.log.debug("Processing output on %s" % conn) conn.sendQueuedData() def _processPollEvent(self, conn, event): # This should do whatever is necessary to process a connection # that has triggered a poll event. It should generally not # raise exceptions so as to avoid restarting the poll loop. # The exception handlers here can raise exceptions and if they # do, it's okay, the poll loop will be restarted. try: if event & (select.EPOLLERR | select.EPOLLHUP): self.log.debug("Received error event on %s: %s" % ( conn, event)) raise DisconnectError() if event & (select.POLLIN | select.POLLOUT): self.readFromConnection(conn) self.writeToConnection(conn) except socket.error as e: if e.errno == errno.ECONNRESET: self.log.debug("Connection reset by peer: %s" % (conn,)) self._lostConnection(conn) return raise except DisconnectError: # Our inner method says we should quietly drop # this connection self._lostConnection(conn) return except Exception: self.log.exception("Exception reading or writing " "from %s:" % (conn,)) self._lostConnection(conn) return def _flushAllConnections(self): # If we need to restart the poll loop, we need to make sure # there are no pending data on any connection. Simulate poll # in+out events on every connection. # # If this method raises an exception, the poll loop wil # restart again. # # No need to get the lock since this is called within the poll # loop and therefore the list in guaranteed never to shrink. connections = self.active_connections[:] for conn in connections: self._processPollEvent(conn, select.POLLIN | select.POLLOUT) def _doPollLoop(self): # Outer run method of poll thread. while self.running: try: self._pollLoop() except Exception: self.log.exception("Exception in poll loop:") def _pollLoop(self): # Inner method of poll loop. self.log.debug("Preparing to poll") # Ensure there are no pending data. self._flushAllConnections() while self.running: self.log.debug("Polling %s connections" % len(self.active_connections)) ret = self.poll.poll() # Since we're using edge-triggering, we need to make sure # that every file descriptor in 'ret' is processed. for fd, event in ret: if fd == self.wake_read: # This means we're exiting, so we can ignore the # rest of 'ret'. self.log.debug("Woken by pipe") while True: if os.read(self.wake_read, 1) == b'\n': break return # In the unlikely event this raises an exception, the # loop will be restarted. conn = self.connection_map[fd] self._processPollEvent(conn, event) def _shutdown(self): super(Server, self)._shutdown() os.write(self.connect_wake_write, b'1\n') def _cleanup(self): super(Server, self)._cleanup() self.socket.close() os.close(self.connect_wake_read) os.close(self.connect_wake_write) def _registerConnection(self, conn): # Register the connection with the poll object # Call while holding the connection condition self.log.debug("Registering %s" % conn) self.connection_map[conn.conn.fileno()] = conn self.poll.register(conn.conn.fileno(), self.readwrite_bitmask) def _unregisterConnection(self, conn): # Unregister the connection with the poll object # Call while holding the connection condition self.log.debug("Unregistering %s" % conn) fd = conn.conn.fileno() if fd not in self.connection_map: return try: self.poll.unregister(fd) except KeyError: pass try: del self.connection_map[fd] except KeyError: pass def _lostConnection(self, conn): # Called as soon as a connection is detected as faulty. self.log.info("Marking %s as disconnected" % conn) self.connections_condition.acquire() self._unregisterConnection(conn) try: # NOTE(notmorgan): In the loop below it is possible to change the # jobs list on the connection. In python 3 .values() is an iter not # a static list, meaning that a change will break the for loop # as the object being iterated on will have changed in size. jobs = list(conn.related_jobs.values()) if conn in self.active_connections: self.active_connections.remove(conn) finally: self.connections_condition.notifyAll() self.connections_condition.release() for job in jobs: if job.worker_connection == conn: # the worker disconnected, alert the client try: p = Packet(constants.REQ, constants.WORK_FAIL, job.handle) if job.client_connection: job.client_connection.sendPacket(p) except Exception: self.log.exception("Sending WORK_FAIL to client after " "worker disconnect failed:") self._removeJob(job) try: conn.conn.shutdown(socket.SHUT_RDWR) except socket.error as e: if e.errno != errno.ENOTCONN: self.log.exception("Unable to shutdown socket " "for connection %s" % (conn,)) except Exception: self.log.exception("Unable to shutdown socket " "for connection %s" % (conn,)) try: conn.conn.close() except Exception: self.log.exception("Unable to close socket " "for connection %s" % (conn,)) self._updateStats() def _removeJob(self, job, dequeue=True): # dequeue is tri-state: True, False, or a specific queue if job.client_connection: try: del job.client_connection.related_jobs[job.handle] except KeyError: pass if job.worker_connection: try: del job.worker_connection.related_jobs[job.handle] except KeyError: pass try: del self.jobs[job.handle] except KeyError: pass if dequeue is True: # Search all queues for the job try: self.high_queue.remove(job) except ValueError: pass try: self.normal_queue.remove(job) except ValueError: pass try: self.low_queue.remove(job) except ValueError: pass elif dequeue is not False: # A specific queue was supplied dequeue.remove(job) # If dequeue is false, no need to remove from any queue self.total_jobs -= 1 if job.running: self.running_jobs -= 1 else: self.waiting_jobs -= 1
[docs] def getQueue(self): """Returns a copy of all internal queues in a flattened form. :returns: The Gearman queue. :rtype: list of :py:class:`WorkerJob`. """ ret = [] for queue in [self.high_queue, self.normal_queue, self.low_queue]: ret += queue return ret
[docs] def handleAdminRequest(self, request): self.log.info("Received admin request %s" % (request,)) if request.command.startswith(b'cancel job'): self.handleCancelJob(request) elif request.command.startswith(b'status'): self.handleStatus(request) elif request.command.startswith(b'workers'): self.handleWorkers(request) elif request.command.startswith(b'acl list'): self.handleACLList(request) elif request.command.startswith(b'acl grant'): self.handleACLGrant(request) elif request.command.startswith(b'acl revoke'): self.handleACLRevoke(request) elif request.command.startswith(b'acl self-revoke'): self.handleACLSelfRevoke(request) self.log.debug("Finished handling admin request %s" % (request,))
def _cancelJob(self, request, job, queue): if self.acl: if not self.acl.canInvoke(request.connection.ssl_subject, job.name): self.log.info("Rejecting cancel job from %s for %s " "due to ACL" % (request.connection.ssl_subject, job.name)) request.connection.sendRaw(b'ERR PERMISSION_DENIED\n') return self._removeJob(job, dequeue=queue) self._updateStats() request.connection.sendRaw(b'OK\n') return def handleCancelJob(self, request): words = request.command.split() handle = words[2] if handle in self.jobs: for queue in [self.high_queue, self.normal_queue, self.low_queue]: for job in queue: if handle == job.handle: return self._cancelJob(request, job, queue) request.connection.sendRaw(b'ERR UNKNOWN_JOB\n') def handleACLList(self, request): if self.acl is None: request.connection.sendRaw(b'ERR ACL_DISABLED\n') return for entry in self.acl.getEntries(): l = "%s\tregister=%s\tinvoke=%s\tgrant=%s\n" % ( entry.subject, entry.register, entry.invoke, entry.grant) request.connection.sendRaw(l.encode('utf8')) request.connection.sendRaw(b'.\n') def handleACLGrant(self, request): # acl grant register worker .* words = request.command.split(None, 4) verb = words[2] subject = words[3] if self.acl is None: request.connection.sendRaw(b'ERR ACL_DISABLED\n') return if not self.acl.canGrant(request.connection.ssl_subject): request.connection.sendRaw(b'ERR PERMISSION_DENIED\n') return try: if verb == 'invoke': self.acl.grantInvoke(subject, words[4]) elif verb == 'register': self.acl.grantRegister(subject, words[4]) elif verb == 'grant': self.acl.grantGrant(subject) else: request.connection.sendRaw(b'ERR UNKNOWN_ACL_VERB\n') return except ACLError as e: self.log.info("Error in grant command: %s" % (e.message,)) request.connection.sendRaw(b'ERR UNABLE %s\n' % (e.message,)) return request.connection.sendRaw(b'OK\n') def handleACLRevoke(self, request): # acl revoke register worker words = request.command.split() verb = words[2] subject = words[3] if self.acl is None: request.connection.sendRaw(b'ERR ACL_DISABLED\n') return if subject != request.connection.ssl_subject: if not self.acl.canGrant(request.connection.ssl_subject): request.connection.sendRaw(b'ERR PERMISSION_DENIED\n') return try: if verb == 'invoke': self.acl.revokeInvoke(subject) elif verb == 'register': self.acl.revokeRegister(subject) elif verb == 'grant': self.acl.revokeGrant(subject) elif verb == 'all': try: self.acl.remove(subject) except ACLError: pass else: request.connection.sendRaw(b'ERR UNKNOWN_ACL_VERB\n') return except ACLError as e: self.log.info("Error in revoke command: %s" % (e.message,)) request.connection.sendRaw(b'ERR UNABLE %s\n' % (e.message,)) return request.connection.sendRaw(b'OK\n') def handleACLSelfRevoke(self, request): # acl self-revoke register words = request.command.split() verb = words[2] if self.acl is None: request.connection.sendRaw(b'ERR ACL_DISABLED\n') return subject = request.connection.ssl_subject try: if verb == 'invoke': self.acl.revokeInvoke(subject) elif verb == 'register': self.acl.revokeRegister(subject) elif verb == 'grant': self.acl.revokeGrant(subject) elif verb == 'all': try: self.acl.remove(subject) except ACLError: pass else: request.connection.sendRaw(b'ERR UNKNOWN_ACL_VERB\n') return except ACLError as e: self.log.info("Error in self-revoke command: %s" % (e.message,)) request.connection.sendRaw(b'ERR UNABLE %s\n' % (e.message,)) return request.connection.sendRaw(b'OK\n') def _getFunctionStats(self): functions = {} for function in self.functions: # Total, running, workers functions[function] = [0, 0, 0] for job in self.jobs.values(): if job.name not in functions: functions[job.name] = [0, 0, 0] functions[job.name][0] += 1 if job.running: functions[job.name][1] += 1 for connection in self.active_connections: for function in connection.functions: if function not in functions: functions[function] = [0, 0, 0] functions[function][2] += 1 return functions def handleStatus(self, request): functions = self._getFunctionStats() for name, values in functions.items(): request.connection.sendRaw( ("%s\t%s\t%s\t%s\n" % (name.decode('utf-8'), values[0], values[1], values[2])).encode('utf8')) request.connection.sendRaw(b'.\n') def handleWorkers(self, request): for connection in self.active_connections: fd = connection.conn.fileno() ip = connection.host client_id = connection.client_id or b'-' functions = b' '.join(connection.functions).decode('utf8') request.connection.sendRaw(("%s %s %s : %s\n" % (fd, ip, client_id.decode('utf8'), functions)) .encode('utf8')) request.connection.sendRaw(b'.\n') def wakeConnection(self, connection): p = Packet(constants.RES, constants.NOOP, b'') if connection.state == 'SLEEP': connection.changeState("AWAKE") connection.sendPacket(p) def wakeConnections(self, job=None): p = Packet(constants.RES, constants.NOOP, b'') # Use a randomized copy of active_connections to try # to spread workload across the machines that workers are on. conns = self.active_connections[:] random.shuffle(conns) # Modifies the list for connection in conns: if connection.state == 'SLEEP': if ((job and job.name in connection.functions) or (job is None)): connection.changeState("AWAKE") connection.sendPacket(p)
[docs] def reportTimingStats(self, ptype, duration): """Report processing times by packet type This method is called by handlePacket to report how long processing took for each packet. If statsd is configured, timing and counts are reported with the key "prefix.packet.NAME". :arg bytes ptype: The packet type (one of the packet types in constants). :arg float duration: The time (in seconds) it took to process the packet. """ if not self.statsd: return ptype = constants.types.get(ptype, 'UNKNOWN') key = 'packet.%s' % ptype self.statsd.timing(key, int(duration * 1000)) self.statsd.incr(key)
def _updateStats(self): if not self.statsd: return # prefix.queue.total # prefix.queue.running # prefix.queue.waiting self.statsd.gauge('queue.total', self.total_jobs) self.statsd.gauge('queue.running', self.running_jobs) self.statsd.gauge('queue.waiting', self.waiting_jobs) def _handleSubmitJob(self, packet, precedence, background=False): name = packet.getArgument(0) unique = packet.getArgument(1) if not unique: unique = None arguments = packet.getArgument(2, True) if self.acl: if not self.acl.canInvoke(packet.connection.ssl_subject, name): self.log.info("Rejecting SUBMIT_JOB from %s for %s " "due to ACL" % (packet.connection.ssl_subject, name)) self.sendError(packet.connection, 0, 'Permission denied by ACL') return self.max_handle += 1 handle = ('H:%s:%s' % (packet.connection.host, self.max_handle)).encode('utf8') if not background: conn = packet.connection else: conn = None job = ServerJob(handle, name, arguments, conn, unique) p = Packet(constants.RES, constants.JOB_CREATED, handle) packet.connection.sendPacket(p) self.jobs[handle] = job self.total_jobs += 1 self.waiting_jobs += 1 if not background: packet.connection.related_jobs[handle] = job if precedence == PRECEDENCE_HIGH: self.high_queue.append(job) elif precedence == PRECEDENCE_NORMAL: self.normal_queue.append(job) elif precedence == PRECEDENCE_LOW: self.low_queue.append(job) self._updateStats() self.wakeConnections(job) def handleSubmitJob(self, packet): return self._handleSubmitJob(packet, PRECEDENCE_NORMAL) def handleSubmitJobHigh(self, packet): return self._handleSubmitJob(packet, PRECEDENCE_HIGH) def handleSubmitJobLow(self, packet): return self._handleSubmitJob(packet, PRECEDENCE_LOW) def handleSubmitJobBg(self, packet): return self._handleSubmitJob(packet, PRECEDENCE_NORMAL, background=True) def handleSubmitJobHighBg(self, packet): return self._handleSubmitJob(packet, PRECEDENCE_HIGH, background=True) def handleSubmitJobLowBg(self, packet): return self._handleSubmitJob(packet, PRECEDENCE_LOW, background=True) def getJobForConnection(self, connection, peek=False): for queue in [self.high_queue, self.normal_queue, self.low_queue]: for job in queue: if job.name in connection.functions: if not peek: queue.remove(job) connection.related_jobs[job.handle] = job job.worker_connection = connection job.running = True self.waiting_jobs -= 1 self.running_jobs += 1 self._updateStats() return job return None def handleGrabJobUniq(self, packet): job = self.getJobForConnection(packet.connection) if job: self.sendJobAssignUniq(packet.connection, job) else: self.sendNoJob(packet.connection) def sendJobAssignUniq(self, connection, job): unique = job.binary_unique if not unique: unique = b'' data = b'\x00'.join((job.handle, job.name, unique, job.arguments)) p = Packet(constants.RES, constants.JOB_ASSIGN_UNIQ, data) connection.sendPacket(p) def sendNoJob(self, connection): p = Packet(constants.RES, constants.NO_JOB, b'') connection.sendPacket(p) def handlePreSleep(self, packet): packet.connection.changeState("SLEEP") if self.getJobForConnection(packet.connection, peek=True): self.wakeConnection(packet.connection) def handleWorkComplete(self, packet): self.handlePassthrough(packet, True) def handleWorkFail(self, packet): self.handlePassthrough(packet, True) def handleWorkException(self, packet): self.handlePassthrough(packet, True) def handleWorkData(self, packet): self.handlePassthrough(packet) def handleWorkWarning(self, packet): self.handlePassthrough(packet) def handleWorkStatus(self, packet): handle = packet.getArgument(0) job = self.jobs.get(handle) if not job: self.log.info("Received packet %s for unknown job" % (packet,)) return job.numerator = packet.getArgument(1) job.denominator = packet.getArgument(2) self.handlePassthrough(packet) def handlePassthrough(self, packet, finished=False): handle = packet.getArgument(0) job = self.jobs.get(handle) if not job: self.log.info("Received packet %s for unknown job" % (packet,)) return packet.code = constants.RES if job.client_connection: job.client_connection.sendPacket(packet) if finished: self._removeJob(job, dequeue=False) self._updateStats() def handleSetClientID(self, packet): name = packet.getArgument(0) packet.connection.client_id = name def sendError(self, connection, code, text): data = (str(code).encode('utf8') + b'\x00' + str(text).encode('utf8') + b'\x00') p = Packet(constants.RES, constants.ERROR, data) connection.sendPacket(p) def handleCanDo(self, packet): name = packet.getArgument(0) if self.acl: if not self.acl.canRegister(packet.connection.ssl_subject, name): self.log.info("Ignoring CAN_DO from %s for %s due to ACL" % (packet.connection.ssl_subject, name)) # CAN_DO normally does not merit a response so it is # not clear that it is appropriate to send an ERROR # response at this point. return self.log.debug("Adding function %s to %s" % (name, packet.connection)) packet.connection.functions.add(name) self.functions.add(name) def handleCantDo(self, packet): name = packet.getArgument(0) self.log.debug("Removing function %s from %s" % (name, packet.connection)) packet.connection.functions.remove(name) def handleResetAbilities(self, packet): self.log.debug("Resetting functions for %s" % packet.connection) packet.connection.functions = set() def handleGetStatus(self, packet): handle = packet.getArgument(0) self.log.debug("Getting status for %s" % handle) known = 0 running = 0 numerator = b'' denominator = b'' job = self.jobs.get(handle) if job: known = 1 if job.running: running = 1 numerator = job.numerator or b'' denominator = job.denominator or b'' data = (handle + b'\x00' + str(known).encode('utf8') + b'\x00' + str(running).encode('utf8') + b'\x00' + numerator + b'\x00' + denominator) p = Packet(constants.RES, constants.STATUS_RES, data) packet.connection.sendPacket(p)