Python pickle.HIGHEST_PROTOCOL Examples
The following are 30
code examples of pickle.HIGHEST_PROTOCOL().
You can vote up the ones you like or vote down the ones you don't like,
and go to the original project or source file by following the links above each example.
You may also want to check out all available functions/classes of the module
pickle
, or try the search function
.
Example #1
Source File: test_case.py From jawfish with MIT License | 6 votes |
def testPickle(self): # Issue 10326 # Can't use TestCase classes defined in Test class as # pickle does not work with inner classes test = unittest.TestCase('run') for protocol in range(pickle.HIGHEST_PROTOCOL + 1): # blew up prior to fix pickled_test = pickle.dumps(test, protocol=protocol) unpickled_test = pickle.loads(pickled_test) self.assertEqual(test, unpickled_test) # exercise the TestCase instance in a way that will invoke # the type equality lookup mechanism unpickled_test.assertEqual(set(), set())
Example #2
Source File: pascal_voc_rbg.py From cascade-rcnn_Pytorch with MIT License | 6 votes |
def gt_roidb(self): """ Return the database of ground-truth regions of interest. This function loads/saves from/to a cache file to speed up future calls. """ cache_file = os.path.join(self.cache_path, self.name + '_gt_roidb.pkl') if os.path.exists(cache_file): with open(cache_file, 'rb') as fid: try: roidb = pickle.load(fid) except: roidb = pickle.load(fid, encoding='bytes') print('{} gt roidb loaded from {}'.format(self.name, cache_file)) return roidb gt_roidb = [self._load_pascal_annotation(index) for index in self.image_index] with open(cache_file, 'wb') as fid: pickle.dump(gt_roidb, fid, pickle.HIGHEST_PROTOCOL) print('wrote gt roidb to {}'.format(cache_file)) return gt_roidb
Example #3
Source File: pascal_voc.py From Collaborative-Learning-for-Weakly-Supervised-Object-Detection with MIT License | 6 votes |
def gt_roidb(self): """ Return the database of ground-truth regions of interest. This function loads/saves from/to a cache file to speed up future calls. """ cache_file = os.path.join(self.cache_path, self.name + '_gt_roidb.pkl') if os.path.exists(cache_file): with open(cache_file, 'rb') as fid: try: roidb = pickle.load(fid) except: roidb = pickle.load(fid, encoding='bytes') print('{} gt roidb loaded from {}'.format(self.name, cache_file)) return roidb gt_roidb = [self._load_pascal_labels(index) for index in self.image_index] with open(cache_file, 'wb') as fid: pickle.dump(gt_roidb, fid, pickle.HIGHEST_PROTOCOL) print('wrote gt roidb to {}'.format(cache_file)) return gt_roidb
Example #4
Source File: cloudpickle.py From pywren-ibm-cloud with Apache License 2.0 | 6 votes |
def dumps(obj, protocol=None): """Serialize obj as a string of bytes allocated in memory protocol defaults to cloudpickle.DEFAULT_PROTOCOL which is an alias to pickle.HIGHEST_PROTOCOL. This setting favors maximum communication speed between processes running the same Python version. Set protocol=pickle.DEFAULT_PROTOCOL instead if you need to ensure compatibility with older versions of Python. """ file = StringIO() try: cp = CloudPickler(file, protocol=protocol) cp.dump(obj) return file.getvalue() finally: file.close() # including pickles unloading functions in this namespace
Example #5
Source File: coco.py From Collaborative-Learning-for-Weakly-Supervised-Object-Detection with MIT License | 6 votes |
def gt_roidb(self): """ Return the database of ground-truth regions of interest. This function loads/saves from/to a cache file to speed up future calls. """ cache_file = osp.join(self.cache_path, self.name + '_gt_roidb.pkl') if osp.exists(cache_file): with open(cache_file, 'rb') as fid: roidb = pickle.load(fid) print('{} gt roidb loaded from {}'.format(self.name, cache_file)) return roidb gt_roidb = [self._load_coco_annotation(index) for index in self._image_index] with open(cache_file, 'wb') as fid: pickle.dump(gt_roidb, fid, pickle.HIGHEST_PROTOCOL) print('wrote gt roidb to {}'.format(cache_file)) return gt_roidb
Example #6
Source File: filecache.py From cutout with MIT License | 6 votes |
def set(self, key, value, timeout=None): if timeout is None: timeout = self.default_timeout filename = self._get_filename(key) self._prune() try: fd, tmp = tempfile.mkstemp(suffix=self._fs_transaction_suffix, dir=self._path) f = os.fdopen(fd, 'wb') try: pickle.dump(int(time() + timeout), f, 1) pickle.dump(value, f, pickle.HIGHEST_PROTOCOL) finally: f.close() rename(tmp, filename) os.chmod(filename, self._mode) except (IOError, OSError): pass
Example #7
Source File: update_cache_compatibility.py From gated-graph-transformer-network with MIT License | 6 votes |
def main(cache_dir): files_list = list(os.listdir(cache_dir)) for file in files_list: full_filename = os.path.join(cache_dir, file) if os.path.isfile(full_filename): print("Processing {}".format(full_filename)) m, stored_kwargs = pickle.load(open(full_filename, 'rb')) updated_kwargs = util.get_compatible_kwargs(model.Model, stored_kwargs) model_hash = util.object_hash(updated_kwargs) print("New hash -> " + model_hash) model_filename = os.path.join(cache_dir, "model_{}.p".format(model_hash)) sys.setrecursionlimit(100000) pickle.dump((m,updated_kwargs), open(model_filename,'wb'), protocol=pickle.HIGHEST_PROTOCOL) os.remove(full_filename)
Example #8
Source File: pascal_voc.py From cascade-rcnn_Pytorch with MIT License | 6 votes |
def gt_roidb(self): """ Return the database of ground-truth regions of interest. This function loads/saves from/to a cache file to speed up future calls. """ cache_file = os.path.join(self.cache_path, self.name + '_gt_roidb.pkl') if os.path.exists(cache_file): with open(cache_file, 'rb') as fid: roidb = pickle.load(fid) print('{} gt roidb loaded from {}'.format(self.name, cache_file)) return roidb gt_roidb = [self._load_pascal_annotation(index) for index in self.image_index] with open(cache_file, 'wb') as fid: pickle.dump(gt_roidb, fid, pickle.HIGHEST_PROTOCOL) print('wrote gt roidb to {}'.format(cache_file)) return gt_roidb
Example #9
Source File: vg.py From cascade-rcnn_Pytorch with MIT License | 6 votes |
def gt_roidb(self): """ Return the database of ground-truth regions of interest. This function loads/saves from/to a cache file to speed up future calls. """ cache_file = os.path.join(self.cache_path, self.name + '_gt_roidb.pkl') if os.path.exists(cache_file): fid = gzip.open(cache_file,'rb') roidb = pickle.load(fid) fid.close() print('{} gt roidb loaded from {}'.format(self.name, cache_file)) return roidb gt_roidb = [self._load_vg_annotation(index) for index in self.image_index] fid = gzip.open(cache_file,'wb') pickle.dump(gt_roidb, fid, pickle.HIGHEST_PROTOCOL) fid.close() print('wrote gt roidb to {}'.format(cache_file)) return gt_roidb
Example #10
Source File: coco.py From cascade-rcnn_Pytorch with MIT License | 6 votes |
def gt_roidb(self): """ Return the database of ground-truth regions of interest. This function loads/saves from/to a cache file to speed up future calls. """ cache_file = osp.join(self.cache_path, self.name + '_gt_roidb.pkl') if osp.exists(cache_file): with open(cache_file, 'rb') as fid: roidb = pickle.load(fid) print('{} gt roidb loaded from {}'.format(self.name, cache_file)) return roidb gt_roidb = [self._load_coco_annotation(index) for index in self._image_index] with open(cache_file, 'wb') as fid: pickle.dump(gt_roidb, fid, pickle.HIGHEST_PROTOCOL) print('wrote gt roidb to {}'.format(cache_file)) return gt_roidb
Example #11
Source File: test_configparser.py From configparser with MIT License | 5 votes |
def test_interpolationdeptherror(self): import pickle e1 = configparser.InterpolationDepthError('option', 'section', 'rawval') for proto in range(pickle.HIGHEST_PROTOCOL + 1): pickled = pickle.dumps(e1, proto) e2 = pickle.loads(pickled) self.assertEqual(e1.message, e2.message) self.assertEqual(e1.args, e2.args) self.assertEqual(e1.section, e2.section) self.assertEqual(e1.option, e2.option) self.assertEqual(repr(e1), repr(e2))
Example #12
Source File: coco.py From ssds.pytorch with MIT License | 5 votes |
def _load_coco_img_path(self, coco_name, indexes): cache_file=os.path.join(self.cache_path,coco_name+'_img_path.pkl') if os.path.exists(cache_file): with open(cache_file, 'rb') as fid: img_path = pickle.load(fid) print('{} img path loaded from {}'.format(coco_name,cache_file)) return img_path print('parsing img path for {}'.format(coco_name)) img_path = [self.image_path_from_index(coco_name, index) for index in indexes] with open(cache_file, 'wb') as fid: pickle.dump(img_path,fid,pickle.HIGHEST_PROTOCOL) print('wrote img path to {}'.format(cache_file)) return img_path
Example #13
Source File: sessions.py From recruit with Apache License 2.0 | 5 votes |
def save(self, session): fn = self.get_session_filename(session.sid) fd, tmp = tempfile.mkstemp(suffix=_fs_transaction_suffix, dir=self.path) f = os.fdopen(fd, "wb") try: dump(dict(session), f, HIGHEST_PROTOCOL) finally: f.close() try: rename(tmp, fn) os.chmod(fn, self.mode) except (IOError, OSError): pass
Example #14
Source File: test_configparser.py From configparser with MIT License | 5 votes |
def test_interpolationsyntaxerror(self): import pickle e1 = configparser.InterpolationSyntaxError('option', 'section', 'msg') for proto in range(pickle.HIGHEST_PROTOCOL + 1): pickled = pickle.dumps(e1, proto) e2 = pickle.loads(pickled) self.assertEqual(e1.message, e2.message) self.assertEqual(e1.args, e2.args) self.assertEqual(e1.section, e2.section) self.assertEqual(e1.option, e2.option) self.assertEqual(repr(e1), repr(e2))
Example #15
Source File: cache.py From recruit with Apache License 2.0 | 5 votes |
def add(self, key, value, timeout=None): expires = self._normalize_timeout(timeout) self._prune() item = (expires, pickle.dumps(value, pickle.HIGHEST_PROTOCOL)) if key in self._cache: return False self._cache.setdefault(key, item) return True
Example #16
Source File: test_configparser.py From configparser with MIT License | 5 votes |
def test_parsingerror(self): import pickle e1 = configparser.ParsingError('source') e1.append(1, 'line1') e1.append(2, 'line2') e1.append(3, 'line3') for proto in range(pickle.HIGHEST_PROTOCOL + 1): pickled = pickle.dumps(e1, proto) e2 = pickle.loads(pickled) self.assertEqual(e1.message, e2.message) self.assertEqual(e1.args, e2.args) self.assertEqual(e1.source, e2.source) self.assertEqual(e1.errors, e2.errors) self.assertEqual(repr(e1), repr(e2)) e1 = configparser.ParsingError(filename='filename') e1.append(1, 'line1') e1.append(2, 'line2') e1.append(3, 'line3') for proto in range(pickle.HIGHEST_PROTOCOL + 1): pickled = pickle.dumps(e1, proto) e2 = pickle.loads(pickled) self.assertEqual(e1.message, e2.message) self.assertEqual(e1.args, e2.args) self.assertEqual(e1.source, e2.source) self.assertEqual(e1.errors, e2.errors) self.assertEqual(repr(e1), repr(e2))
Example #17
Source File: test_configparser.py From configparser with MIT License | 5 votes |
def test_missingsectionheadererror(self): import pickle e1 = configparser.MissingSectionHeaderError('filename', 123, 'line') for proto in range(pickle.HIGHEST_PROTOCOL + 1): pickled = pickle.dumps(e1, proto) e2 = pickle.loads(pickled) self.assertEqual(e1.message, e2.message) self.assertEqual(e1.args, e2.args) self.assertEqual(e1.line, e2.line) self.assertEqual(e1.source, e2.source) self.assertEqual(e1.lineno, e2.lineno) self.assertEqual(repr(e1), repr(e2))
Example #18
Source File: build.py From product-classifier with MIT License | 5 votes |
def build(self): """ Builds the model and writes to the outpath (which should be a directory). Two files are written: - the pickle of the model - a yaml file of associated data Note, if a file already exists at the outpath, this will raise an exception (don't want to overwrite a model by accident!) """ # Record the start time self.started = datetime.now() start = time.time() # Extract the features and train the model classifier, self.traintime = self.train() # Write the classifier to disk with open(self.model_path, 'w') as f: pickle.dump(classifier, f, pickle.HIGHEST_PROTOCOL) # Begin accuracy validation if self.validate: self.cross_validate() # Record the finish time self.finished = datetime.now() self.buildtime = time.time() - start # Write the information to disk self.write_details()
Example #19
Source File: coco.py From ssds.pytorch with MIT License | 5 votes |
def _load_coco_annotations(self, coco_name, indexes, _COCO): cache_file=os.path.join(self.cache_path,coco_name+'_gt_roidb.pkl') if os.path.exists(cache_file): with open(cache_file, 'rb') as fid: roidb = pickle.load(fid) print('{} gt roidb loaded from {}'.format(coco_name,cache_file)) return roidb print('parsing gt roidb for {}'.format(coco_name)) gt_roidb = [self._annotation_from_index(index, _COCO) for index in indexes] with open(cache_file, 'wb') as fid: pickle.dump(gt_roidb,fid,pickle.HIGHEST_PROTOCOL) print('wrote gt roidb to {}'.format(cache_file)) return gt_roidb
Example #20
Source File: data_handling.py From fancy-cnn with MIT License | 5 votes |
def create_data_sets(partition_list=range(1,100), pickle_base_name=DEFAULT_REVIEWS_PICKLE + '.'): """ Creates a 50% - 25% - 25% train/validation/test partition of the classification problem. Classes are balanced. It reads the list of partitions saved in pickles. Resulting data sets are saved as python pickles. """ load_partitions(partition_list, pickle_base_name) reviews, _, funny_votes, _, _ = get_reviews_data(partition_list, pickle_base_name) reviews, labels = give_balanced_classes(reviews, funny_votes) N = len(reviews) train_reviews = reviews[:N/2] train_labels = labels[:N/2] dev_reviews = reviews[N/2:3*N/4] dev_labels = labels[N/2:3*N/4] test_reviews = reviews[3*N/4:] test_labels = labels[3*N/4:] pickle.dump([train_reviews, train_labels], open("TrainSet_" + str(N), "wb"), pickle.HIGHEST_PROTOCOL) pickle.dump([dev_reviews, dev_labels], open("DevSet_" + str(N), "wb"), pickle.HIGHEST_PROTOCOL) pickle.dump([test_reviews, test_labels], open("TestSet_" + str(N), "wb"), pickle.HIGHEST_PROTOCOL)
Example #21
Source File: gipc.py From gipc with MIT License | 5 votes |
def _default_encoder(o): return pickle.dumps(o, pickle.HIGHEST_PROTOCOL)
Example #22
Source File: test.py From tandem with Apache License 2.0 | 5 votes |
def test_class_nested_enum_and_pickle_protocol_four(self): # would normally just have this directly in the class namespace class NestedEnum(Enum): twigs = 'common' shiny = 'rare' self.__class__.NestedEnum = NestedEnum self.NestedEnum.__qualname__ = '%s.NestedEnum' % self.__class__.__name__ test_pickle_dump_load(self.assertTrue, self.NestedEnum.twigs, protocol=(0, HIGHEST_PROTOCOL))
Example #23
Source File: test.py From tandem with Apache License 2.0 | 5 votes |
def test_class_nested_enum_and_pickle_protocol_four(self): # would normally just have this directly in the class namespace class NestedEnum(Enum): twigs = 'common' shiny = 'rare' self.__class__.NestedEnum = NestedEnum self.NestedEnum.__qualname__ = '%s.NestedEnum' % self.__class__.__name__ test_pickle_exception( self.assertRaises, PicklingError, self.NestedEnum.twigs, protocol=(0, 3)) test_pickle_dump_load(self.assertTrue, self.NestedEnum.twigs, protocol=(4, HIGHEST_PROTOCOL))
Example #24
Source File: test.py From tandem with Apache License 2.0 | 5 votes |
def test_pickle_exception(assertion, exception, obj, protocol=(0, HIGHEST_PROTOCOL)): start, stop = protocol failures = [] for protocol in range(start, stop+1): try: assertion(exception, dumps, obj, protocol=protocol) except Exception: exc = sys.exc_info()[1] failures.append('%d: %s %s' % (protocol, exc.__class__.__name__, exc)) if failures: raise ValueError('Failed with protocols: %s' % ', '.join(failures))
Example #25
Source File: test.py From tandem with Apache License 2.0 | 5 votes |
def test_pickle_dump_load(assertion, source, target=None, protocol=(0, HIGHEST_PROTOCOL)): start, stop = protocol failures = [] for protocol in range(start, stop+1): try: if target is None: assertion(loads(dumps(source, protocol=protocol)) is source) else: assertion(loads(dumps(source, protocol=protocol)), target) except Exception: exc, tb = sys.exc_info()[1:] failures.append('%2d: %s' %(protocol, exc)) if failures: raise ValueError('Failed with protocols: %s' % ', '.join(failures))
Example #26
Source File: utils.py From AerialDetection with Apache License 2.0 | 5 votes |
def getcategory( basepath, label, ): classedict = {} def initdic(): for clsname in classname_15: wordname = datamap_15[clsname] classedict[wordname] = [] initdic() picklepath = os.path.join(basepath, 'pickle') pickledir = os.path.join(picklepath, 'category-file.pickle') if not os.path.isfile(pickledir): labelpath = os.path.join(basepath, label) filelist = GetFileFromThisRootDir(labelpath) for fullname in filelist: name = mybasename(fullname) objects = parse_bod_poly(fullname) for obj in objects: #wordname = datamap[obj['name']] wordname = obj['name'] if name not in classedict[wordname]: classedict[wordname].append(name) with open(pickledir, 'wb') as f: pickle.dump(classedict, f, pickle.HIGHEST_PROTOCOL) else: with open(pickledir, 'rb') as f: classedict = pickle.load(f) return classedict
Example #27
Source File: reader.py From deepnlp with MIT License | 5 votes |
def save_instance(pickle_file, obj): fw = open(pickle_file, 'wb') pickle.dump(obj, fw, pickle.HIGHEST_PROTOCOL) fw.close() return
Example #28
Source File: dict_util.py From deepnlp with MIT License | 5 votes |
def gen_pickle_file(): dict_path = "./ner/dict/zh_o2o/entity_tags.dic" print ("NOTICE: Start Loading Default Entity Tag Dictionary...") prefix_dict = gen_prefix_dict(dict_path) print ("NOTICE: prefix_dict Size %d" % len(prefix_dict)) pickle_file = "./ner/dict/zh_o2o/entity_tags.dic.pkl" fw = open(pickle_file, 'wb') pickle.dump(prefix_dict, fw, pickle.HIGHEST_PROTOCOL)
Example #29
Source File: get_raw_skes_data.py From View-Adaptive-Neural-Networks-for-Skeleton-based-Human-Action-Recognition with MIT License | 5 votes |
def get_raw_skes_data(): # # save_path = './data' # # skes_path = '/data/pengfei/NTU/nturgb+d_skeletons/' # stat_path = osp.join(save_path, 'statistics') # # skes_name_file = osp.join(stat_path, 'skes_available_name.txt') # save_data_pkl = osp.join(save_path, 'raw_skes_data.pkl') # frames_drop_pkl = osp.join(save_path, 'frames_drop_skes.pkl') # # frames_drop_logger = logging.getLogger('frames_drop') # frames_drop_logger.setLevel(logging.INFO) # frames_drop_logger.addHandler(logging.FileHandler(osp.join(save_path, 'frames_drop.log'))) # frames_drop_skes = dict() skes_name = np.loadtxt(skes_name_file, dtype=str) num_files = skes_name.size print('Found %d available skeleton files.' % num_files) raw_skes_data = [] frames_cnt = np.zeros(num_files, dtype=np.int) for (idx, ske_name) in enumerate(skes_name): bodies_data = get_raw_bodies_data(skes_path, ske_name, frames_drop_skes, frames_drop_logger) raw_skes_data.append(bodies_data) frames_cnt[idx] = bodies_data['num_frames'] if (idx + 1) % 1000 == 0: print('Processed: %.2f%% (%d / %d)' % \ (100.0 * (idx + 1) / num_files, idx + 1, num_files)) with open(save_data_pkl, 'wb') as fw: pickle.dump(raw_skes_data, fw, pickle.HIGHEST_PROTOCOL) np.savetxt(osp.join(save_path, 'raw_data', 'frames_cnt.txt'), frames_cnt, fmt='%d') print('Saved raw bodies data into %s' % save_data_pkl) print('Total frames: %d' % np.sum(frames_cnt)) with open(frames_drop_pkl, 'wb') as fw: pickle.dump(frames_drop_skes, fw, pickle.HIGHEST_PROTOCOL)
Example #30
Source File: cache.py From recruit with Apache License 2.0 | 5 votes |
def set(self, key, value, timeout=None, mgmt_element=False): # Management elements have no timeout if mgmt_element: timeout = 0 # Don't prune on management element update, to avoid loop else: self._prune() timeout = self._normalize_timeout(timeout) filename = self._get_filename(key) try: fd, tmp = tempfile.mkstemp( suffix=self._fs_transaction_suffix, dir=self._path ) with os.fdopen(fd, "wb") as f: pickle.dump(timeout, f, 1) pickle.dump(value, f, pickle.HIGHEST_PROTOCOL) rename(tmp, filename) os.chmod(filename, self._mode) except (IOError, OSError): return False else: # Management elements should not count towards threshold if not mgmt_element: self._update_count(delta=1) return True