Python queue.put() Examples
The following are 30
code examples of queue.put().
You can vote up the ones you like or vote down the ones you don't like,
and go to the original project or source file by following the links above each example.
You may also want to check out all available functions/classes of the module
queue
, or try the search function
.
Example #1
Source File: mayhem_3.py From mayhem with MIT License | 6 votes |
def publish_sync(queue): """Simulates an external publisher of messages. Args: queue (queue.Queue): Queue to publish messages to. n (int): Number of messages to publish. """ choices = string.ascii_lowercase + string.digits while True: msg_id = str(uuid.uuid4()) host_id = "".join(random.choices(choices, k=4)) instance_name = f"cattle-{host_id}" msg = PubSubMessage(message_id=msg_id, instance_name=instance_name) # publish an item queue.put(msg) logging.info(f"Published {msg}") # simulate randomness of publishing messages time.sleep(random.random())
Example #2
Source File: test_dummy_thread.py From Fluid-Designer with GNU General Public License v3.0 | 6 votes |
def test_arg_passing(self): #Make sure that parameter passing works. def arg_tester(queue, arg1=False, arg2=False): """Use to test _thread.start_new_thread() passes args properly.""" queue.put((arg1, arg2)) testing_queue = queue.Queue(1) _thread.start_new_thread(arg_tester, (testing_queue, True, True)) result = testing_queue.get() self.assertTrue(result[0] and result[1], "Argument passing for thread creation using tuple failed") _thread.start_new_thread(arg_tester, tuple(), {'queue':testing_queue, 'arg1':True, 'arg2':True}) result = testing_queue.get() self.assertTrue(result[0] and result[1], "Argument passing for thread creation using kwargs failed") _thread.start_new_thread(arg_tester, (testing_queue, True), {'arg2':True}) result = testing_queue.get() self.assertTrue(result[0] and result[1], "Argument passing for thread creation using both tuple" " and kwargs failed")
Example #3
Source File: test_dummy_thread.py From ironpython3 with Apache License 2.0 | 6 votes |
def test_arg_passing(self): #Make sure that parameter passing works. def arg_tester(queue, arg1=False, arg2=False): """Use to test _thread.start_new_thread() passes args properly.""" queue.put((arg1, arg2)) testing_queue = queue.Queue(1) _thread.start_new_thread(arg_tester, (testing_queue, True, True)) result = testing_queue.get() self.assertTrue(result[0] and result[1], "Argument passing for thread creation using tuple failed") _thread.start_new_thread(arg_tester, tuple(), {'queue':testing_queue, 'arg1':True, 'arg2':True}) result = testing_queue.get() self.assertTrue(result[0] and result[1], "Argument passing for thread creation using kwargs failed") _thread.start_new_thread(arg_tester, (testing_queue, True), {'arg2':True}) result = testing_queue.get() self.assertTrue(result[0] and result[1], "Argument passing for thread creation using both tuple" " and kwargs failed")
Example #4
Source File: test_subscription_transport.py From graphql-python-subscriptions with MIT License | 6 votes |
def on_sub_mock(mocker): mgr = multiprocess.Manager() q = mgr.Queue() def on_subscribe(self, msg, params, websocket): new_params = copy.deepcopy(params) new_params.update({'context': msg.get('context', {})}) q.put(self) return new_params on_sub_mock = { 'on_subscribe': PickableMock(side_effect=promisify(on_subscribe), name='on_subscribe') } return on_sub_mock, q
Example #5
Source File: talon.py From TALON with MIT License | 6 votes |
def unpack_observed(annotation_info, queue, obs_file): """ Now that transcript has been annotated, unpack values and create an observed entry. Send the observed entry to the queue for output to obs_file.""" obs_ID = observed_counter.increment() observed = (obs_ID, annotation_info.gene_ID, annotation_info.transcript_ID, annotation_info.read_ID, annotation_info.dataset, annotation_info.start_vertex, annotation_info.end_vertex, annotation_info.start_exon, annotation_info.end_exon, annotation_info.start_delta, annotation_info.end_delta, annotation_info.read_length, annotation_info.fraction_As, annotation_info.custom_label, annotation_info.allelic_label, annotation_info.start_support, annotation_info.end_support) msg = (obs_file, "\t".join([str(x) for x in observed])) queue.put(msg) return
Example #6
Source File: test_messenger.py From networkzero with MIT License | 6 votes |
def test_wait_for_news(support): address = nw0.core.address() topic = uuid.uuid4().hex data = uuid.uuid4().hex sync_queue = queue.Queue() support.queue.put(("wait_for_news_from", [address, topic, data, sync_queue])) in_topic, in_data = nw0.wait_for_news_from(address, topic, wait_for_s=5) sync_queue.put(True) while in_data is None: in_topic, in_data = nw0.wait_for_news_from(address, topic, wait_for_s=5) assert (topic, data) == (in_topic, in_data) # # send to multiple addresses # For now, this is disallowed as the semantics aren't # intuitive. (It does a round-robin selection which is useful # for things like load-scheduling but not for broadcasting). #
Example #7
Source File: test_messenger.py From networkzero with MIT License | 6 votes |
def test_send_news(support): address = nw0.core.address() topic = uuid.uuid4().hex data = uuid.uuid4().hex reply_queue = queue.Queue() support.queue.put(("send_news_to", [address, topic, reply_queue])) while True: nw0.send_news_to(address, topic, None) try: in_topic, in_data = reply_queue.get_nowait() except queue.Empty: time.sleep(0.1) else: break nw0.send_news_to(address, topic, data) while in_data is None: in_topic, in_data = reply_queue.get() assert in_topic, in_data == (topic, data) # # wait_for_news_from #
Example #8
Source File: projection_subtraction.py From pyem with GNU General Public License v3.0 | 6 votes |
def producer(pool, queue, submap_ft, refmap_ft, fname, particles, sx, sy, s, a, apix, coefs_method, r, nr, fftthreads=1, crop=None, pfac=2): log = logging.getLogger('root') log.debug("Producing %s" % fname) zreader = mrc.ZSliceReader(particles[star.UCSF.IMAGE_ORIGINAL_PATH].iloc[0]) for i, ptcl in particles.iterrows(): log.debug("Produce %d@%s" % (ptcl[star.UCSF.IMAGE_ORIGINAL_INDEX], ptcl[star.UCSF.IMAGE_ORIGINAL_PATH])) # p1r = mrc.read_imgs(stack[i], idx[i] - 1, compat="relion") p1r = zreader.read(ptcl[star.UCSF.IMAGE_ORIGINAL_INDEX]) log.debug("Apply") ri = pool.apply_async( subtract_outer, (p1r, ptcl, submap_ft, refmap_ft, sx, sy, s, a, apix, coefs_method, r, nr), {"fftthreads": fftthreads, "crop": crop, "pfac": pfac}) log.debug("Put") queue.put((ptcl[star.UCSF.IMAGE_INDEX], ri), block=True) log.debug("Queue for %s is size %d" % (ptcl[star.UCSF.IMAGE_ORIGINAL_PATH], queue.qsize())) zreader.close() log.debug("Put poison pill") queue.put((-1, None), block=True)
Example #9
Source File: test_utils.py From bigchaindb with Apache License 2.0 | 6 votes |
def mock_queue(monkeypatch): class MockQueue: items = [] def get(self, timeout=None): try: return self.items.pop() except IndexError: if timeout: raise queue.Empty() raise def put(self, item): self.items.append(item) mockqueue = MockQueue() monkeypatch.setattr('queue.Queue', lambda: mockqueue) return mockqueue
Example #10
Source File: process.py From peach with Mozilla Public License 2.0 | 6 votes |
def _StartProcess(self): print("Command: {}".format(self.arguments)) self.process = Popen(self.arguments, stderr=STDOUT, stdout=PIPE, env=os.environ, bufsize=1, close_fds=isPosix()) self.pid = self.process.pid def enqueue_output(out, queue): for line in iter(out.readline, ""): queue.put(line) out.close() self.terminal_queue = Queue.Queue() self.terminal_producer = threading.Thread(target=enqueue_output, args=(self.process.stdout, self.terminal_queue)) self.terminal_consumer = threading.Thread(target=self._grab_sanitizer_trace) self.terminal_producer.setDaemon(True) self.terminal_consumer.setDaemon(True) self.terminal_producer.start() self.terminal_consumer.start()
Example #11
Source File: test_utils.py From Decentralized-Internet with MIT License | 6 votes |
def mock_queue(monkeypatch): class MockQueue: items = [] def get(self, timeout=None): try: return self.items.pop() except IndexError: if timeout: raise queue.Empty() raise def put(self, item): self.items.append(item) mockqueue = MockQueue() monkeypatch.setattr('queue.Queue', lambda: mockqueue) return mockqueue
Example #12
Source File: mayhem_1.py From mayhem with MIT License | 6 votes |
def publish(queue, n): """Simulates an external publisher of messages. Args: queue (queue.Queue): Queue to publish messages to. n (int): Number of messages to publish. """ choices = string.ascii_lowercase + string.digits for x in range(1, n + 1): host_id = "".join(random.choices(choices, k=4)) instance_name = f"cattle-{host_id}" msg = PubSubMessage(message_id=x, instance_name=instance_name) # publish an item queue.put(msg) logging.info(f"Published {x} of {n} messages") # indicate the publisher is done queue.put(None)
Example #13
Source File: main_2.8-12.py From motorized_zoom_lens with GNU General Public License v3.0 | 6 votes |
def grab(cam, queue, width, height, fps): global running capture = cv2.VideoCapture(cam) capture.set(cv2.CAP_PROP_FRAME_WIDTH, width) capture.set(cv2.CAP_PROP_FRAME_HEIGHT, height) capture.set(cv2.CAP_PROP_FPS, fps) while(running): frame = {} capture.grab() retval, img = capture.retrieve(0) frame["img"] = img frame["1"] = config["1"] frame["2"] = config["2"] blur = get_blur(img, 0.05) frame["blur"] = blur if queue.qsize() < 10: queue.put(frame) else: print(queue.qsize())
Example #14
Source File: mayhem_1.py From mayhem with MIT License | 6 votes |
def publish_sync(queue): """Simulates an external publisher of messages. Args: queue (queue.Queue): Queue to publish messages to. n (int): Number of messages to publish. """ choices = string.ascii_lowercase + string.digits while True: msg_id = str(uuid.uuid4()) host_id = "".join(random.choices(choices, k=4)) instance_name = f"cattle-{host_id}" msg = PubSubMessage(message_id=msg_id, instance_name=instance_name) # publish an item queue.put(msg) logging.info(f"Published {msg}") # simulate randomness of publishing messages time.sleep(random.random())
Example #15
Source File: mayhem_5.py From mayhem with MIT License | 6 votes |
def publish_sync(queue): """Simulates an external publisher of messages. Args: queue (queue.Queue): Queue to publish messages to. n (int): Number of messages to publish. """ choices = string.ascii_lowercase + string.digits while True: msg_id = str(uuid.uuid4()) host_id = "".join(random.choices(choices, k=4)) instance_name = f"cattle-{host_id}" msg = PubSubMessage(message_id=msg_id, instance_name=instance_name) # publish an item queue.put(msg) logging.info(f"Published {msg}") # simulate randomness of publishing messages time.sleep(random.random())
Example #16
Source File: mayhem_4.py From mayhem with MIT License | 6 votes |
def publish_sync(queue): """Simulates an external publisher of messages. Args: queue (queue.Queue): Queue to publish messages to. n (int): Number of messages to publish. """ choices = string.ascii_lowercase + string.digits while True: msg_id = str(uuid.uuid4()) host_id = "".join(random.choices(choices, k=4)) instance_name = f"cattle-{host_id}" msg = PubSubMessage(message_id=msg_id, instance_name=instance_name) # publish an item queue.put(msg) logging.info(f"Published {msg}") # simulate randomness of publishing messages time.sleep(random.random())
Example #17
Source File: source_queue.py From capture_reid with Apache License 2.0 | 6 votes |
def source_queue(self, procs, queues): while True: for index, proc in enumerate(procs): try: queue = queues[index] h, w, c = self.shape[index] frames_size = w * h * c nv12_data = proc.stdout.read(frames_size) if len(nv12_data) != frames_size: # logger.error("source_queue read error index %s len %s", index, len(nv12_data)) continue frame = np.frombuffer(nv12_data, dtype=np.uint8) img = np.array(frame, dtype=np.uint8).reshape((h, w, c)) if DEBUG: queue.put(img) else: queue.put_nowait(img) except Exception as e: # logger.error("source_queue queue full index %s", index) pass
Example #18
Source File: mayhem_2.py From mayhem with MIT License | 6 votes |
def publish_sync(queue): """Simulates an external publisher of messages. Args: queue (queue.Queue): Queue to publish messages to. n (int): Number of messages to publish. """ choices = string.ascii_lowercase + string.digits while True: msg_id = str(uuid.uuid4()) host_id = "".join(random.choices(choices, k=4)) instance_name = f"cattle-{host_id}" msg = PubSubMessage(message_id=msg_id, instance_name=instance_name) # publish an item queue.put(msg) logging.info(f"Published {msg}") # simulate randomness of publishing messages time.sleep(random.random())
Example #19
Source File: mayhem_6.py From mayhem with MIT License | 6 votes |
def publish_sync(queue): """Simulates an external publisher of messages. Args: queue (queue.Queue): Queue to publish messages to. n (int): Number of messages to publish. """ choices = string.ascii_lowercase + string.digits curr_thread = threading.current_thread() THREADS.add(curr_thread.ident) while True: msg_id = str(uuid.uuid4()) host_id = "".join(random.choices(choices, k=4)) instance_name = f"cattle-{host_id}" msg = PubSubMessage(message_id=msg_id, instance_name=instance_name) # publish an item queue.put(msg) logging.info(f"Published {msg}") # simulate randomness of publishing messages time.sleep(random.random())
Example #20
Source File: main_5-50.py From motorized_zoom_lens with GNU General Public License v3.0 | 6 votes |
def grab(cam, queue, width, height, fps): global running capture = cv2.VideoCapture(cam) capture.set(cv2.CAP_PROP_FRAME_WIDTH, width) capture.set(cv2.CAP_PROP_FRAME_HEIGHT, height) capture.set(cv2.CAP_PROP_FPS, fps) while(running): frame = {} capture.grab() retval, img = capture.retrieve(0) frame["img"] = img frame["1"] = config["1"] frame["2"] = config["2"] blur = get_blur(img, 0.05) frame["blur"] = blur if queue.qsize() < 10: queue.put(frame) else: print(queue.qsize())
Example #21
Source File: mayhem_7.py From mayhem with MIT License | 6 votes |
def publish_sync(queue): """Simulates an external publisher of messages. Args: queue (queue.Queue): Queue to publish messages to. n (int): Number of messages to publish. """ choices = string.ascii_lowercase + string.digits curr_thread = threading.current_thread() THREADS.add(curr_thread.ident) while True: msg_id = str(uuid.uuid4()) host_id = "".join(random.choices(choices, k=4)) instance_name = f"cattle-{host_id}" msg = PubSubMessage(message_id=msg_id, instance_name=instance_name) # publish an item queue.put(msg) logging.info(f"Published {msg}") # simulate randomness of publishing messages time.sleep(random.random())
Example #22
Source File: mayhem_9.py From mayhem with MIT License | 6 votes |
def publish_sync(queue): """Simulates an external publisher of messages. Args: queue (queue.Queue): Queue to publish messages to. n (int): Number of messages to publish. """ choices = string.ascii_lowercase + string.digits curr_thread = threading.current_thread() THREADS.add(curr_thread.ident) while True: msg_id = str(uuid.uuid4()) host_id = "".join(random.choices(choices, k=4)) instance_name = f"cattle-{host_id}" msg = PubSubMessage(message_id=msg_id, instance_name=instance_name) # publish an item queue.put(msg) logging.info(f"Published {msg}") # simulate randomness of publishing messages time.sleep(random.random())
Example #23
Source File: gen.py From SRNet-Datagen with Apache License 2.0 | 5 votes |
def enqueue_data(queue, capacity): np.random.seed() gen = datagen() while True: try: data = gen.gen_srnet_data_with_background() except Exception as e: pass if queue.qsize() < capacity: queue.put(data)
Example #24
Source File: iterators.py From dpu-utils with MIT License | 5 votes |
def __worker(queue: queue.Queue, original_iterator: Iterator[T])-> None: try: for element in original_iterator: assert element is not None, 'By convention, Iterables wrapped in ThreadedIterator may not contain None.' queue.put(element, block=True) queue.put(None, block=True) except Exception as e: _, __, tb = sys.exc_info() queue.put((e, tb), block=True)
Example #25
Source File: iterators.py From dpu-utils with MIT License | 5 votes |
def __next__(self) -> T: next_element = self.__queue.get(block=True) if next_element is None: self.__thread.join() self.__queue.put(None) # Make sure that we remember that we are done if we are called once more... raise StopIteration if isinstance(next_element, tuple) and isinstance(next_element[0], Exception): raise next_element[0].with_traceback(next_element[1]) return next_element
Example #26
Source File: test_subscription_transport.py From graphql-python-subscriptions with MIT License | 5 votes |
def enqueue_output(out, queue): with out: for line in iter(out.readline, b''): queue.put(line)
Example #27
Source File: iterators.py From dpu-utils with MIT License | 5 votes |
def __init__(self, argument_iterator: Iterator[Iterable], worker_callable: Callable, max_queue_size: int=1, num_workers: int = 5, use_threads: bool=True): self.__in_queue = queue.Queue() if use_threads else multiprocessing.Queue() # type: Union[queue.Queue, multiprocessing.Queue] self.__num_elements = 0 for callable_args in argument_iterator: self.__in_queue.put(callable_args) self.__num_elements += 1 self.__out_queue = queue.Queue(maxsize=max_queue_size) if use_threads else multiprocessing.Queue( maxsize=max_queue_size ) # type: Union[queue.Queue, multiprocessing.Queue] self.__threads = [ threading.Thread(target=lambda: self.__worker(worker_callable)) if use_threads else multiprocessing.Process(target=lambda: self.__worker(worker_callable)) for _ in range(num_workers) ] # type: List[Union[threading.Thread, multiprocessing.Process]] for worker in self.__threads: worker.start()
Example #28
Source File: train.py From CornerNet with BSD 3-Clause "New" or "Revised" License | 5 votes |
def prefetch_data(db, queue, sample_data, data_aug): ind = 0 print("start prefetching data...") np.random.seed(os.getpid()) while True: try: data, ind = sample_data(db, ind, data_aug=data_aug) queue.put(data) except Exception as e: traceback.print_exc() raise e
Example #29
Source File: api.py From KOBE with MIT License | 5 votes |
def enqueue_output(out, queue): for line in iter(out.readline, b""): queue.put(line) out.close()
Example #30
Source File: train.py From CornerNet with BSD 3-Clause "New" or "Revised" License | 5 votes |
def pin_memory(data_queue, pinned_data_queue, sema): while True: data = data_queue.get() data["xs"] = [x.pin_memory() for x in data["xs"]] data["ys"] = [y.pin_memory() for y in data["ys"]] pinned_data_queue.put(data) if sema.acquire(blocking=False): return