def __init__(self, configuration, poller): self.low = configuration.redirector.minimum # minimum concurrent redirector workers self.high = configuration.redirector.maximum # maximum concurrent redirector workers self.poller = poller self.configuration = configuration self.queue = Queue( ) # store requests we do not immediately have the resources to process self.nextid = 1 # unique id to give to the next spawned worker self.worker = {} # worker tasks for each spawned child self.processes = { } # worker tasks indexed by file descriptors we can poll self.available = set( ) # workers that are currently available to handle new requests self.active = { } # workers that are currently busy waiting for a response from the spawned process self.stopping = set( ) # workers we want to stop as soon as they stop being active program = configuration.redirector.program protocol = configuration.redirector.protocol self.redirector_factory = RedirectorFactory(configuration, program, protocol) self.log = Logger('manager', configuration.log.manager)
class RedirectorDispatcher (object): dispatcher_factory = DispatcherThread def __init__ (self, messagebox): self.messagebox = messagebox self.queue = Queue() self.thread = self.dispatcher_factory(messagebox, self.queue) def start (self): self.thread.start() def stop (self): self.queue.put(('STOP', '')) self.thread.join() return self.messagebox.stop() def sendRequest (self, client_id, peer, request, subrequest, source): message = client_id, peer, request, subrequest, source return self.queue.put(('REQUEST', message)) def getDecision (self): return self.messagebox.getDecision() def respawn (self): return self.messagebox.respawn() def decreaseSpawnLimit (self, *args): return self.messagebox.decreaseSpawnLimit(*args) def increaseSpawnLimit (self, *args): return self.messagebox.increaseSpawnLimit(*args) def getStats (self): return self.messagebox.getStats()
class RedirectorDispatcher(object): dispatcher_factory = DispatcherThread def __init__(self, messagebox): self.messagebox = messagebox self.queue = Queue() self.thread = self.dispatcher_factory(messagebox, self.queue) def start(self): self.thread.start() def stop(self): self.queue.put(('STOP', '')) self.thread.join() return self.messagebox.stop() def sendRequest(self, client_id, peer, request, subrequest, source): message = client_id, peer, request, subrequest, source return self.queue.put(('REQUEST', message)) def getDecision(self): return self.messagebox.getDecision() def respawn(self): return self.messagebox.respawn() def decreaseSpawnLimit(self, *args): return self.messagebox.decreaseSpawnLimit(*args) def increaseSpawnLimit(self, *args): return self.messagebox.increaseSpawnLimit(*args) def getStats(self): return self.messagebox.getStats()
def __init__ (self,configuration,poller): self.configuration = configuration self.low = configuration.redirector.minimum # minimum number of workers at all time self.high = configuration.redirector.maximum # maximum numbe of workers at all time self.program = configuration.redirector.program # what program speaks the squid redirector API self.nextid = 1 # incremental number to make the name of the next worker self.queue = Queue() # queue with HTTP headers to process self.poller = poller # poller interface that checks for events on sockets self.worker = {} # our workers threads self.closing = set() # workers that are currently closing self.running = True # we are running self.log = Logger('manager', configuration.log.manager)
class RedirectorDispatcher(object): dispatcher_factory = DispatcherThread def __init__(self, messagebox): self.messagebox = messagebox self.queue = Queue() self.thread = self.dispatcher_factory(messagebox, self.queue) def start(self): self.thread.start() def stop(self): self.queue.put(('STOP', '')) self.thread.join() try: res = self.messagebox.stop() except IOError: res = None return res def sendRequest(self, client_id, accept_addr, accept_port, peer, request, subrequest, source): message = client_id, accept_addr, accept_port, peer, request, subrequest, source return self.queue.put(('REQUEST', message)) def getDecision(self): return self.messagebox.getDecision() def respawn(self): return self.messagebox.respawn() def decreaseSpawnLimit(self, *args): return self.messagebox.decreaseSpawnLimit(*args) def increaseSpawnLimit(self, *args): return self.messagebox.increaseSpawnLimit(*args) def requestStats(self): return self.messagebox.requestStats() def readResponse(self): command, [data] = self.messagebox.readResponse() return command, data
class RedirectorDispatcher (object): dispatcher_factory = DispatcherThread def __init__ (self, messagebox): self.messagebox = messagebox self.queue = Queue() self.thread = self.dispatcher_factory(messagebox, self.queue) def start (self): self.thread.start() def stop (self): self.queue.put(('STOP', '')) self.thread.join() try: res = self.messagebox.stop() except IOError: res = None return res def sendRequest (self, client_id, accept_addr, accept_port, peer, request, subrequest, source): message = client_id, accept_addr, accept_port, peer, request, subrequest, source return self.queue.put(('REQUEST', message)) def getDecision (self): return self.messagebox.getDecision() def respawn (self): return self.messagebox.respawn() def decreaseSpawnLimit (self, *args): return self.messagebox.decreaseSpawnLimit(*args) def increaseSpawnLimit (self, *args): return self.messagebox.increaseSpawnLimit(*args) def requestStats (self): return self.messagebox.requestStats() def readResponse (self): command, [data] = self.messagebox.readResponse() return command, data
class RedirectorDispatcher (object): dispatcher_factory = DispatcherThread def __init__ (self, messagebox): self.messagebox = messagebox self.queue = Queue() self.thread = self.dispatcher_factory(messagebox, self.queue) def start (self): self.thread.start() def stop (self): self.queue.put(('STOP', '')) self.thread.join() try: res = self.messagebox.stop() except IOError, e: res = None return res
def __init__ (self, configuration, poller): self.low = configuration.redirector.minimum # minimum concurrent redirector workers self.high = configuration.redirector.maximum # maximum concurrent redirector workers self.poller = poller self.configuration = configuration self.queue = Queue() # store requests we do not immediately have the resources to process self.nextid = 1 # unique id to give to the next spawned worker self.worker = {} # worker tasks for each spawned child self.processes = {} # worker tasks indexed by file descriptors we can poll self.available = set() # workers that are currently available to handle new requests self.active = {} # workers that are currently busy waiting for a response from the spawned process self.stopping = set() # workers we want to stop as soon as they stop being active program = configuration.redirector.program protocol = configuration.redirector.protocol self.redirector_factory = RedirectorFactory(configuration, program, protocol) self.log = Logger('manager', configuration.log.manager)
def __init__ (self, messagebox): self.messagebox = messagebox self.queue = Queue() self.thread = self.dispatcher_factory(messagebox, self.queue)
def __init__(self, messagebox): self.messagebox = messagebox self.queue = Queue() self.thread = self.dispatcher_factory(messagebox, self.queue)
class RedirectorManager (object): def __init__ (self,configuration,poller): self.configuration = configuration self.low = configuration.redirector.minimum # minimum number of workers at all time self.high = configuration.redirector.maximum # maximum numbe of workers at all time self.program = configuration.redirector.program # what program speaks the squid redirector API self.nextid = 1 # incremental number to make the name of the next worker self.queue = Queue() # queue with HTTP headers to process self.poller = poller # poller interface that checks for events on sockets self.worker = {} # our workers threads self.closing = set() # workers that are currently closing self.running = True # we are running self.log = Logger('manager', configuration.log.manager) def _getid(self): id = str(self.nextid) self.nextid +=1 return id def _spawn (self): """add one worker to the pool""" wid = self._getid() worker = Redirector(self.configuration,wid,self.queue,self.program) self.poller.addReadSocket('read_workers', worker.response_box_read) self.worker[wid] = worker self.log.info("added a worker") self.log.info("we have %d workers. defined range is ( %d / %d )" % (len(self.worker),self.low,self.high)) self.worker[wid].start() def spawn (self,number=1): """create the set number of worker""" self.log.info("spawning %d more worker" % number) for _ in range(number): self._spawn() def respawn (self): """make sure we reach the minimum number of workers""" number = max(min(len(self.worker),self.high),self.low) for wid in set(self.worker): self.reap(wid) self.spawn(number) def reap (self,wid): self.log.info('we are killing worker %s' % wid) worker = self.worker[wid] self.closing.add(wid) worker.stop() # will cause the worker to stop when it can def decrease (self): if self.low < len(self.worker): worker = self._oldest() if worker: self.reap(worker.wid) def increase (self): if len(self.worker) < self.high: self.spawn() def start (self): """spawn our minimum number of workers""" self.log.info("starting workers.") self.spawn(max(0,self.low-len(self.worker))) def stop (self): """tell all our worker to stop reading the queue and stop""" self.running = False threads = self.worker.values() if len(self.worker): self.log.info("stopping %d workers." % len(self.worker)) for wid in set(self.worker): self.reap(wid) for thread in threads: self.request(None, None, None, 'nop') for thread in threads: thread.destroyProcess() thread.join() self.worker = {} def _oldest (self): """find the oldest worker""" oldest = None past = time.time() for wid in set(self.worker): creation = self.worker[wid].creation if creation < past and wid not in self.closing: past = creation oldest = self.worker[wid] return oldest def provision (self): """manage our workers to make sure we have enough to consume the queue""" if not self.running: return num_workers = len(self.worker) # bad we are bleeding workers ! if num_workers < self.low: self.log.info("we lost some workers, respawing %d new workers" % (self.low-num_workers)) self.spawn(self.low-num_workers) size = self.queue.qsize() # we need more workers if size >= num_workers: # nothing we can do we have reach our limit if num_workers >= self.high: self.log.warning("help ! we need more workers but we reached our ceiling ! %d request are queued for %d processes" % (size,num_workers)) return # try to figure a good number to add .. # no less than one, no more than to reach self.high, lower between self.low and a quarter of the allowed growth nb_to_add = int(min(max(1,min(self.low,(self.high-self.low)/4)),self.high-num_workers)) self.log.warning("we are low on workers adding a few (%d), the queue has %d unhandled url" % (nb_to_add,size)) self.spawn(nb_to_add) def deprovision (self): """manage our workers to make sure we have enough to consume the queue""" if not self.running: return size = self.queue.qsize() num_workers = len(self.worker) # we are now overprovisioned if size < 2 and num_workers > self.low: self.log.info("we have too many workers (%d), stopping the oldest" % num_workers) # if we have to kill one, at least stop the one who had the most chance to memory leak :) worker = self._oldest() if worker: self.reap(worker.wid) def request(self, client_id, peer, request, source): return self.queue.put((client_id,peer,request,source,False)) def getDecision(self, box): # NOTE: reads may block if we send badly formatted data try: r_buffer = box.read(3) while r_buffer.isdigit(): r_buffer += box.read(1) if ':' in r_buffer: size, response = r_buffer.split(':', 1) if size.isdigit(): size = int(size) else: size, response = None, None else: # not a netstring size, response = None, None if size is not None: required = size + 1 - len(response) response += box.read(required) if response is not None: if response.endswith(','): response = response[:-1] else: response = None except ValueError: # I/O operation on closed file worker = self.worker.get(box, None) if worker is not None: worker.destroyProcess() response = None except TypeError: response = None try: if response: client_id, command, decision = response.split('\0', 2) else: client_id = None command = None decision = None except (ValueError, TypeError): client_id = None command = None decision = None if command == 'requeue': _client_id, _peer, _source, _header = response.split('\0', 3) self.queue.put((_client_id,_peer,_header,_source,True)) client_id = None command = None decision = None elif command == 'hangup': wid = decision client_id = None command = None decision = None worker = self.worker.pop(wid, None) if worker: self.poller.removeReadSocket('read_workers', worker.response_box_read) if wid in self.closing: self.closing.remove(wid) worker.shutdown() worker.join() elif command == 'stats': wid, timestamp, stats = decision self.storeStats(timestamp, wid, stats) client_id = None command = None decision = None return client_id, command, decision def showInternalError(self): return 'file', '\0'.join(('200', 'internal_error.html')) def requestStats(self): for wid, worker in self.worker.iteritems(): worker.requestStats() def storeStats(self, timestamp, wid, stats): pairs = (d.split('=',1) for d in stats.split('?', 1).split('&')) d = self.cache.setdefault(timestamp, {}) for k, v in pairs: d.setdefault(k, []).append(v)
class RedirectorManager (object): def __init__ (self, configuration, poller): self.low = configuration.redirector.minimum # minimum concurrent redirector workers self.high = configuration.redirector.maximum # maximum concurrent redirector workers self.poller = poller self.configuration = configuration self.queue = Queue() # store requests we do not immediately have the resources to process self.nextid = 1 # unique id to give to the next spawned worker self.worker = {} # worker tasks for each spawned child self.processes = {} # worker tasks indexed by file descriptors we can poll self.available = set() # workers that are currently available to handle new requests self.active = {} # workers that are currently busy waiting for a response from the spawned process self.stopping = set() # workers we want to stop as soon as they stop being active program = configuration.redirector.program protocol = configuration.redirector.protocol self.redirector_factory = RedirectorFactory(configuration, program, protocol) self.log = Logger('manager', configuration.log.manager) def _getid(self): wid = str(self.nextid) self.nextid += 1 return wid def _spawn (self): """add one worker to the pool""" wid = self._getid() worker = self.redirector_factory.create(wid) self.worker[wid] = worker self.available.add(wid) if worker.process is not None: identifier = worker.process.stdout self.processes[identifier] = worker self.poller.addReadSocket('read_workers', identifier) self.log.info("added a worker") self.log.info("we have %d workers. defined range is ( %d / %d )" % (len(self.worker), self.low, self.high)) def spawn (self, number=1): """create the request number of worker processes""" self.log.info("spawning %d more workers" % number) for _ in range(number): self._spawn() def respawn (self): """make sure we reach the minimum number of workers""" number = max(min(len(self.worker), self.high), self.low) for wid in set(self.worker): self.stopWorker(wid) self.spawn(number) def stopWorker (self, wid): self.log.info('want worker %s to go away' % wid) if wid not in self.active: self.reap(wid) else: self.stopping.add(wid) def reap (self, wid): self.log.info('we are killing worker %s' % wid) worker = self.worker[wid] if wid in self.active: self.log.error('reaping worker %s even though it is still active' % wid) self.active.pop(wid) if wid in self.stopping: self.stopping.remove(wid) if wid in self.available: self.available.remove(wid) if worker.process is not None: self.poller.removeReadSocket('read_workers', worker.process.stdout) self.processes.pop(worker.process.stdout) worker.shutdown() self.worker.pop(wid) def _decrease (self): if self.low < len(self.worker): wid = self._oldest() if wid: self.stopWorker(wid) def _increase (self): if len(self.worker) < self.high: self.spawn() def decrease (self, count=1): for _ in xrange(count): self._decrease() def increase (self, count=1): for _ in xrange(count): self._increase() def start (self): """spawn our minimum number of workers""" self.log.info("starting workers.") self.spawn(max(0,self.low-len(self.worker))) def stop (self): """tell all our worker to stop reading the queue and stop""" for wid in self.worker: self.reap(wid) self.worker = {} def _oldest (self): """find the oldest worker""" oldest = None past = time.time() for wid in set(self.worker): creation = self.worker[wid].creation if creation < past and wid not in self.stopping: past = creation oldest = wid return oldest def provision (self): """manage our workers to make sure we have enough to consume the queue""" size = self.queue.qsize() num_workers = len(self.worker) # bad we are bleeding workers ! if num_workers < self.low: self.log.info("we lost some workers, respawing %d new workers" % (self.low - num_workers)) self.spawn(self.low - num_workers) # we need more workers if size >= num_workers: # nothing we can do we have reach our limit if num_workers >= self.high: self.log.warning("help ! we need more workers but we reached our ceiling ! %d request are queued for %d processes" % (size,num_workers)) return # try to figure a good number to add .. # no less than one, no more than to reach self.high, lower between self.low and a quarter of the allowed growth nb_to_add = int(min(max(1,min(self.low,(self.high-self.low)/4)),self.high-num_workers)) self.log.warning("we are low on workers adding a few (%d), the queue has %d unhandled url" % (nb_to_add,size)) self.spawn(nb_to_add) def deprovision (self): """manage our workers to make sure we have enough to consume the queue""" size = self.queue.qsize() num_workers = len(self.worker) # we are now overprovisioned if size < 2 and num_workers > self.low: self.log.info("we have too many workers (%d), stopping the oldest" % num_workers) # if we have to kill one, at least stop the one who had the most chance to memory leak :) wid = self._oldest() if wid: self.stopWorker(wid) def acquire (self): if self.available: identifier = self.available.pop() worker = self.worker[identifier] else: worker = None return worker def release (self, wid): if wid not in self.stopping: self.available.add(wid) else: self.reap(wid) def persist (self, wid, client_id, peer, data, header, subheader, source, tainted): self.active[wid] = client_id, peer, data, header, subheader, source, tainted def progress (self, wid): return self.active.pop(wid) def doqueue (self): if self.available and not self.queue.isempty(): client_id, peer, header, subheader, source, tainted = self.queue.get() _, command, decision = self.request(client_id, peer, header, subheader, source, tainted=tainted) else: client_id, command, decision = None, None, None return client_id, command, decision def request (self, client_id, peer, header, subheader, source, tainted=False): worker = self.acquire() if worker is not None: try: _, command, decision = worker.decide(client_id, peer, header, subheader, source) except: command, decision = None, None if command is None: self.reap(worker.wid) if tainted is False: _, command, decision = self.request(client_id, peer, header, subheader, source, tainted=True) else: _, command, decision = Respond.close(client_id) else: command, decision = None, None self.queue.put((client_id, peer, header, subheader, source, tainted)) if command == 'defer': self.persist(worker.wid, client_id, peer, decision, header, subheader, source, tainted) command, decision = None, None elif worker is not None: self.release(worker.wid) return client_id, command, decision def getDecision (self, pipe_in): worker = self.processes.get(pipe_in, None) if worker is not None and worker.wid in self.active: client_id, peer, request, header, subheader, source, tainted = self.progress(worker.wid) try: _, command, decision = worker.progress(client_id, peer, request, header, subheader, source) except Exception, e: command, decision = None, None self.release(worker.wid) if command is None: self.reap(worker.wid) if tainted is False: _, command, decision = self.request(client_id, peer, header, subheader, source, tainted=True) else: _, command, decision = Respond.close(client_id) else: