Python tensorflow.contrib.slim.python.slim.data.parallel_reader.parallel_read() Examples
The following are 6
code examples of tensorflow.contrib.slim.python.slim.data.parallel_reader.parallel_read().
You can vote up the ones you like or vote down the ones you don't like,
and go to the original project or source file by following the links above each example.
You may also want to check out all available functions/classes of the module
tensorflow.contrib.slim.python.slim.data.parallel_reader
, or try the search function
.
Example #1
Source File: parallel_reader_test.py From auto-alt-text-lambda-api with MIT License | 6 votes |
def testTFRecordReader(self): with self.test_session(): self._tfrecord_paths = test_utils.create_tfrecord_files( self.get_temp_dir(), num_files=3) key, value = parallel_reader.parallel_read( self._tfrecord_paths, reader_class=io_ops.TFRecordReader, num_readers=3) sv = supervisor.Supervisor(logdir=self.get_temp_dir()) with sv.prepare_or_wait_for_session() as sess: sv.start_queue_runners(sess) flowers = 0 num_reads = 100 for _ in range(num_reads): current_key, _ = sess.run([key, value]) if 'flowers' in str(current_key): flowers += 1 self.assertGreater(flowers, 0) self.assertEquals(flowers, num_reads)
Example #2
Source File: parallel_reader_test.py From keras-lambda with MIT License | 6 votes |
def testTFRecordReader(self): with self.test_session(): self._tfrecord_paths = test_utils.create_tfrecord_files( self.get_temp_dir(), num_files=3) key, value = parallel_reader.parallel_read( self._tfrecord_paths, reader_class=io_ops.TFRecordReader, num_readers=3) sv = supervisor.Supervisor(logdir=self.get_temp_dir()) with sv.prepare_or_wait_for_session() as sess: sv.start_queue_runners(sess) flowers = 0 num_reads = 100 for _ in range(num_reads): current_key, _ = sess.run([key, value]) if 'flowers' in str(current_key): flowers += 1 self.assertGreater(flowers, 0) self.assertEquals(flowers, num_reads)
Example #3
Source File: dataset_data_provider.py From auto-alt-text-lambda-api with MIT License | 5 votes |
def __init__(self, dataset, num_readers=1, shuffle=True, num_epochs=None, common_queue_capacity=256, common_queue_min=128, seed=None): """Creates a DatasetDataProvider. Args: dataset: An instance of the Dataset class. num_readers: The number of parallel readers to use. shuffle: Whether to shuffle the data sources and common queue when reading. num_epochs: The number of times each data source is read. If left as None, the data will be cycled through indefinitely. common_queue_capacity: The capacity of the common queue. common_queue_min: The minimum number of elements in the common queue after a dequeue. seed: The seed to use if shuffling. """ _, data = parallel_reader.parallel_read( dataset.data_sources, reader_class=dataset.reader, num_epochs=num_epochs, num_readers=num_readers, shuffle=shuffle, capacity=common_queue_capacity, min_after_dequeue=common_queue_min, seed=seed) items = dataset.decoder.list_items() tensors = dataset.decoder.decode(data, items) super(DatasetDataProvider, self).__init__( items_to_tensors=dict(zip(items, tensors)), num_samples=dataset.num_samples)
Example #4
Source File: dataset_data_provider.py From deep_image_model with Apache License 2.0 | 5 votes |
def __init__(self, dataset, num_readers=1, shuffle=True, num_epochs=None, common_queue_capacity=256, common_queue_min=128, seed=None): """Creates a DatasetDataProvider. Args: dataset: An instance of the Dataset class. num_readers: The number of parallel readers to use. shuffle: Whether to shuffle the data sources and common queue when reading. num_epochs: The number of times each data source is read. If left as None, the data will be cycled through indefinitely. common_queue_capacity: The capacity of the common queue. common_queue_min: The minimum number of elements in the common queue after a dequeue. seed: The seed to use if shuffling. """ _, data = parallel_reader.parallel_read( dataset.data_sources, reader_class=dataset.reader, num_epochs=num_epochs, num_readers=num_readers, shuffle=shuffle, capacity=common_queue_capacity, min_after_dequeue=common_queue_min, seed=seed) items = dataset.decoder.list_items() tensors = dataset.decoder.decode(data, items) super(DatasetDataProvider, self).__init__( items_to_tensors=dict(zip(items, tensors)), num_samples=dataset.num_samples)
Example #5
Source File: dataset_data_provider.py From keras-lambda with MIT License | 5 votes |
def __init__(self, dataset, num_readers=1, shuffle=True, num_epochs=None, common_queue_capacity=256, common_queue_min=128, seed=None): """Creates a DatasetDataProvider. Args: dataset: An instance of the Dataset class. num_readers: The number of parallel readers to use. shuffle: Whether to shuffle the data sources and common queue when reading. num_epochs: The number of times each data source is read. If left as None, the data will be cycled through indefinitely. common_queue_capacity: The capacity of the common queue. common_queue_min: The minimum number of elements in the common queue after a dequeue. seed: The seed to use if shuffling. """ _, data = parallel_reader.parallel_read( dataset.data_sources, reader_class=dataset.reader, num_epochs=num_epochs, num_readers=num_readers, shuffle=shuffle, capacity=common_queue_capacity, min_after_dequeue=common_queue_min, seed=seed) items = dataset.decoder.list_items() tensors = dataset.decoder.decode(data, items) super(DatasetDataProvider, self).__init__( items_to_tensors=dict(zip(items, tensors)), num_samples=dataset.num_samples)
Example #6
Source File: dataset_data_provider.py From lambda-packs with MIT License | 4 votes |
def __init__(self, dataset, num_readers=1, reader_kwargs=None, shuffle=True, num_epochs=None, common_queue_capacity=256, common_queue_min=128, record_key='record_key', seed=None, scope=None): """Creates a DatasetDataProvider. Args: dataset: An instance of the Dataset class. num_readers: The number of parallel readers to use. reader_kwargs: An optional dict of kwargs for the reader. shuffle: Whether to shuffle the data sources and common queue when reading. num_epochs: The number of times each data source is read. If left as None, the data will be cycled through indefinitely. common_queue_capacity: The capacity of the common queue. common_queue_min: The minimum number of elements in the common queue after a dequeue. record_key: The item name to use for the dataset record keys in the provided tensors. seed: The seed to use if shuffling. scope: Optional name scope for the ops. Raises: ValueError: If `record_key` matches one of the items in the dataset. """ key, data = parallel_reader.parallel_read( dataset.data_sources, reader_class=dataset.reader, num_epochs=num_epochs, num_readers=num_readers, reader_kwargs=reader_kwargs, shuffle=shuffle, capacity=common_queue_capacity, min_after_dequeue=common_queue_min, seed=seed, scope=scope) items = dataset.decoder.list_items() tensors = dataset.decoder.decode(data, items) if record_key in items: raise ValueError('The item name used for `record_key` cannot also be ' 'used for a dataset item: %s', record_key) items.append(record_key) tensors.append(key) super(DatasetDataProvider, self).__init__( items_to_tensors=dict(zip(items, tensors)), num_samples=dataset.num_samples)