Python cv2.COLOR_RGB2YUV Examples
The following are 9
code examples of cv2.COLOR_RGB2YUV().
You can vote up the ones you like or vote down the ones you don't like,
and go to the original project or source file by following the links above each example.
You may also want to check out all available functions/classes of the module
cv2
, or try the search function
.
Example #1
Source File: utils.py From AdaIN-TF with MIT License | 6 votes |
def get_img_random_crop(src, resize=512, crop=256): '''Get & resize image and random crop''' img = get_img(src) img = resize_to(img, resize=resize) offset_h = random.randint(0, (img.shape[0]-crop)) offset_w = random.randint(0, (img.shape[1]-crop)) img = img[offset_h:offset_h+crop, offset_w:offset_w+crop, :] return img # def preserve_colors(content_rgb, styled_rgb): # """Extract luminance from styled image and apply colors from content""" # if content_rgb.shape != styled_rgb.shape: # new_shape = (content_rgb.shape[1], content_rgb.shape[0]) # styled_rgb = cv2.resize(styled_rgb, new_shape) # styled_yuv = cv2.cvtColor(styled_rgb, cv2.COLOR_RGB2YUV) # Y_s, U_s, V_s = cv2.split(styled_yuv) # image_YUV = cv2.cvtColor(content_rgb, cv2.COLOR_RGB2YUV) # Y_i, U_i, V_i = cv2.split(image_YUV) # styled_rgb = cv2.cvtColor(np.stack([Y_s, U_i, V_i], axis=-1), cv2.COLOR_YUV2RGB) # return styled_rgb
Example #2
Source File: pyfakewebcam.py From pyfakewebcam with GNU Lesser General Public License v3.0 | 5 votes |
def schedule_frame(self, frame): if frame.shape[0] != self._settings.fmt.pix.height: raise Exception('frame height does not match the height of webcam device: {}!={}\n'.format(self._settings.fmt.pix.height, frame.shape[0])) if frame.shape[1] != self._settings.fmt.pix.width: raise Exception('frame width does not match the width of webcam device: {}!={}\n'.format(self._settings.fmt.pix.width, frame.shape[1])) if frame.shape[2] != self._channels: raise Exception('num frame channels does not match the num channels of webcam device: {}!={}\n'.format(self._channels, frame.shape[2])) if cv2_imported: # t1 = timeit.default_timer() self._yuv = cv2.cvtColor(frame, cv2.COLOR_RGB2YUV) # t2 = timeit.default_timer() # sys.stderr.write('conversion time: {}\n'.format(t2-t1)) else: # t1 = timeit.default_timer() frame = np.concatenate((frame, self._ones), axis=2) frame = np.dot(frame, self._rgb2yuv.T) self._yuv[:,:,:] = np.clip(frame, 0, 255) # t2 = timeit.default_timer() # sys.stderr.write('conversion time: {}\n'.format(t2-t1)) # t1 = timeit.default_timer() for i in range(self._settings.fmt.pix.height): self._buffer[i,::2] = self._yuv[i,:,0] self._buffer[i,1::4] = self._yuv[i,::2,1] self._buffer[i,3::4] = self._yuv[i,::2,2] # t2 = timeit.default_timer() # sys.stderr.write('pack time: {}\n'.format(t2-t1)) os.write(self._video_device, self._buffer.tostring())
Example #3
Source File: utils.py From VerifAI with BSD 3-Clause "New" or "Revised" License | 5 votes |
def rgb2yuv(image): """ Convert the image from RGB to YUV (This is what the NVIDIA model does) """ return cv2.cvtColor(image, cv2.COLOR_RGB2YUV)
Example #4
Source File: utils.py From car-behavioral-cloning with MIT License | 5 votes |
def rgb2yuv(image): """ Convert the image from RGB to YUV (This is what the NVIDIA model does) """ return cv2.cvtColor(image, cv2.COLOR_RGB2YUV)
Example #5
Source File: hybrid_driver.py From sdc-live-trainer with MIT License | 5 votes |
def predict_steering(self, data): image_array = self.roi(cv2.cvtColor(data['image'], cv2.COLOR_RGB2YUV)) transformed_image_array = image_array[None, :, :, :] return float(model.predict(transformed_image_array, batch_size=1)) # Callback functions triggered by ControlServer
Example #6
Source File: drive.py From sdc-live-trainer with MIT License | 5 votes |
def preprocess_input(img): return roi(cv2.cvtColor(img, cv2.COLOR_RGB2YUV))
Example #7
Source File: live_trainer.py From sdc-live-trainer with MIT License | 5 votes |
def preprocess_input(self, img): return self.roi(cv2.cvtColor(img, cv2.COLOR_RGB2YUV))
Example #8
Source File: test_color.py From imgaug with MIT License | 4 votes |
def test_every_colorspace(self): def _image_to_channel(image, cspace): if cspace == iaa.CSPACE_YCrCb: image_cvt = cv2.cvtColor(image, cv2.COLOR_RGB2YCR_CB) return image_cvt[:, :, 0:0+1] elif cspace == iaa.CSPACE_HSV: image_cvt = cv2.cvtColor(image, cv2.COLOR_RGB2HSV) return image_cvt[:, :, 2:2+1] elif cspace == iaa.CSPACE_HLS: image_cvt = cv2.cvtColor(image, cv2.COLOR_RGB2HLS) return image_cvt[:, :, 1:1+1] elif cspace == iaa.CSPACE_Lab: if hasattr(cv2, "COLOR_RGB2Lab"): image_cvt = cv2.cvtColor(image, cv2.COLOR_RGB2Lab) else: image_cvt = cv2.cvtColor(image, cv2.COLOR_RGB2LAB) return image_cvt[:, :, 0:0+1] elif cspace == iaa.CSPACE_Luv: if hasattr(cv2, "COLOR_RGB2Luv"): image_cvt = cv2.cvtColor(image, cv2.COLOR_RGB2Luv) else: image_cvt = cv2.cvtColor(image, cv2.COLOR_RGB2LUV) return image_cvt[:, :, 0:0+1] else: assert cspace == iaa.CSPACE_YUV image_cvt = cv2.cvtColor(image, cv2.COLOR_RGB2YUV) return image_cvt[:, :, 0:0+1] # Max differences between input image and image after augmentation # when no child augmenter is used (for the given example image below). # For some colorspaces the conversion to input colorspace isn't # perfect. # Values were manually checked. max_diff_expected = { iaa.CSPACE_YCrCb: 1, iaa.CSPACE_HSV: 0, iaa.CSPACE_HLS: 0, iaa.CSPACE_Lab: 2, iaa.CSPACE_Luv: 4, iaa.CSPACE_YUV: 1 } image = np.arange(6*6*3).astype(np.uint8).reshape((6, 6, 3)) for cspace in self.valid_colorspaces: with self.subTest(colorspace=cspace): child = _BatchCapturingDummyAugmenter() aug = iaa.WithBrightnessChannels( children=child, to_colorspace=cspace) image_aug = aug(image=image) expected = _image_to_channel(image, cspace) diff = np.abs( image.astype(np.int32) - image_aug.astype(np.int32)) assert np.all(diff <= max_diff_expected[cspace]) assert np.array_equal(child.last_batch.images[0], expected)
Example #9
Source File: pyfakewebcam.py From snapchat-filters-opencv with MIT License | 4 votes |
def schedule_frame(self, frame): if frame.shape[0] != self._settings.fmt.pix.height: raise Exception( "frame height does not match the height of webcam device: {}!={}\n".format( self._settings.fmt.pix.height, frame.shape[0] ) ) if frame.shape[1] != self._settings.fmt.pix.width: raise Exception( "frame width does not match the width of webcam device: {}!={}\n".format( self._settings.fmt.pix.width, frame.shape[1] ) ) if frame.shape[2] != self._channels: raise Exception( "num frame channels does not match the num channels of webcam device: {}!={}\n".format( self._channels, frame.shape[2] ) ) if cv2_imported: # t1 = timeit.default_timer() self._yuv = cv2.cvtColor(frame, cv2.COLOR_RGB2YUV) # t2 = timeit.default_timer() # sys.stderr.write('conversion time: {}\n'.format(t2-t1)) else: # t1 = timeit.default_timer() frame = np.concatenate((frame, self._ones), axis=2) frame = np.dot(frame, self._rgb2yuv.T) self._yuv[:, :, :] = np.clip(frame, 0, 255) # t2 = timeit.default_timer() # sys.stderr.write('conversion time: {}\n'.format(t2-t1)) # t1 = timeit.default_timer() for i in range(self._settings.fmt.pix.height): self._buffer[i, ::2] = self._yuv[i, :, 0] self._buffer[i, 1::4] = self._yuv[i, ::2, 1] self._buffer[i, 3::4] = self._yuv[i, ::2, 2] # t2 = timeit.default_timer() # sys.stderr.write('pack time: {}\n'.format(t2-t1)) os.write(self._video_device, self._buffer.tostring())