Python sensor_msgs.msg.Image() Examples
The following are 30
code examples of sensor_msgs.msg.Image().
You can vote up the ones you like or vote down the ones you don't like,
and go to the original project or source file by following the links above each example.
You may also want to check out all available functions/classes of the module
sensor_msgs.msg
, or try the search function
.
![](https://www.programcreek.com/common/static/images/search.png)
Example #1
Source File: gCamera.py From Cherry-Autonomous-Racecar with MIT License | 7 votes |
def gCamera(): print "gstWebCam" bridge = CvBridge() video ="video4" pub = rospy.Publisher('stream', Image, queue_size=10) rospy.init_node('GstWebCam',anonymous=True) Gst.init(None) pipe = Gst.parse_launch("""v4l2src device=/dev/"""+video+""" ! video/x-raw, width=640, height=480,format=(string)BGR ! appsink sync=false max-buffers=2 drop=true name=sink emit-signals=true""") sink = pipe.get_by_name('sink') pipe.set_state(Gst.State.PLAYING) while not rospy.is_shutdown(): sample = sink.emit('pull-sample') img = gst_to_opencv(sample.get_buffer()) try: pub.publish(bridge.cv2_to_imgmsg(img, "bgr8")) except CvBridgeError as e: print(e)
Example #2
Source File: Dronet.py From rpg_public_dronet with MIT License | 7 votes |
def run(self): while not rospy.is_shutdown(): msg = CNN_out() msg.header.stamp = rospy.Time.now() data = None while data is None: try: data = rospy.wait_for_message("camera", Image, timeout=10) except: pass if self.use_network_out: print("Publishing commands!") else: print("NOT Publishing commands!") cv_image = utils.callback_img(data, self.target_size, self.crop_size, self.imgs_rootpath, self.use_network_out) outs = self.model.predict_on_batch(cv_image[None]) steer, coll = outs[0][0], outs[1][0] msg.steering_angle = steer msg.collision_prob = coll self.pub.publish(msg)
Example #3
Source File: dataset_to_rosbag.py From research with BSD 3-Clause "New" or "Revised" License | 7 votes |
def main(): if len(sys.argv) < 2: print("Usage: {} dataset_name".format(sys.argv[0])) exit(1) file_name = sys.argv[1] log_file = h5py.File('../dataset/log/{}.h5'.format(file_name)) camera_file = h5py.File('../dataset/camera/{}.h5'.format(file_name)) zipped_log = izip( log_file['times'], log_file['fiber_accel'], log_file['fiber_gyro']) with rosbag.Bag('{}.bag'.format(file_name), 'w') as bag: bar = Bar('Camera', max=len(camera_file['X'])) for i, img_data in enumerate(camera_file['X']): m_img = Image() m_img.header.stamp = rospy.Time.from_sec(0.01 * i) m_img.height = img_data.shape[1] m_img.width = img_data.shape[2] m_img.step = 3 * img_data.shape[2] m_img.encoding = 'rgb8' m_img.data = np.transpose(img_data, (1, 2, 0)).flatten().tolist() bag.write('/camera/image_raw', m_img, m_img.header.stamp) bar.next() bar.finish() bar = Bar('IMU', max=len(log_file['times'])) for time, v_accel, v_gyro in zipped_log: m_imu = Imu() m_imu.header.stamp = rospy.Time.from_sec(time) [setattr(m_imu.linear_acceleration, c, v_accel[i]) for i, c in enumerate('xyz')] [setattr(m_imu.angular_velocity, c, v_gyro[i]) for i, c in enumerate('xyz')] bag.write('/fiber_imu', m_imu, m_imu.header.stamp) bar.next() bar.finish()
Example #4
Source File: head_display.py From intera_sdk with Apache License 2.0 | 6 votes |
def _setup_image(self, image_path): """ Load the image located at the specified path @type image_path: str @param image_path: the relative or absolute file path to the image file @rtype: sensor_msgs/Image or None @param: Returns sensor_msgs/Image if image convertable and None otherwise """ if not os.access(image_path, os.R_OK): rospy.logerr("Cannot read file at '{0}'".format(image_path)) return None img = cv2.imread(image_path) # Return msg return cv_bridge.CvBridge().cv2_to_imgmsg(img, encoding="bgr8")
Example #5
Source File: omnirobot_server.py From robotics-rl-srl with MIT License | 6 votes |
def imageCallback(self, msg): try: # Convert your ROS Image message to OpenCV cv2_img = bridge.imgmsg_to_cv2(msg, "rgb8") if self.first_msg: shape = cv2_img.shape min_length = min(shape[0], shape[1]) up_margin = int((shape[0] - min_length) / 2) # row left_margin = int((shape[1] - min_length) / 2) # col self.valid_box = [up_margin, up_margin + min_length, left_margin, left_margin + min_length] print("origin size: {}x{}".format(shape[0],shape[1])) print("crop each image to a square image, cropped size: {}x{}".format(min_length, min_length)) self.first_msg = False undistort_image = cv2.undistort(cv2_img, self.camera_matrix, self.distortion_coefficients) self.valid_img = undistort_image[self.valid_box[0]:self.valid_box[1], self.valid_box[2]:self.valid_box[3]] except CvBridgeError as e: print("CvBridgeError:", e)
Example #6
Source File: test.py From image_recognition with MIT License | 6 votes |
def trigger_configuration(self): """ Callback when the configuration button is clicked """ topic_name, ok = QInputDialog.getItem(self._widget, "Select topic name", "Topic name", rostopic.find_by_type('sensor_msgs/Image')) if ok: self._create_subscriber(topic_name) available_rosservices = [] for s in rosservice.get_service_list(): try: if rosservice.get_service_type(s) in _SUPPORTED_SERVICES: available_rosservices.append(s) except: pass srv_name, ok = QInputDialog.getItem(self._widget, "Select service name", "Service name", available_rosservices) if ok: self._create_service_client(srv_name)
Example #7
Source File: lidarEvasion.py From Cherry-Autonomous-Racecar with MIT License | 6 votes |
def __init__(self): self.evadeSet = False self.controller = XBox360() self.bridge = CvBridge() self.throttle = 0 self.grid_img = None ##self.throttleLock = Lock() print "evasion" rospy.Subscriber("/lidargrid", Image, self.gridCB, queue_size=1) rospy.Subscriber("/cmd_vel", TwistStamped , self.twistCB , queue_size = 1) rospy.Subscriber("/joy", Joy, self.joyCB, queue_size=5) self.pub_img = rospy.Publisher("/steering_img", Image) self.pub_twist = rospy.Publisher("/lidar_twist", TwistStamped, queue_size=1) self.sound = rospy.Publisher("/sound_server/speech_synth", String, queue_size=1) rospy.init_node ('lidar_cmd',anonymous=True) rospy.spin()
Example #8
Source File: annotation.py From image_recognition with MIT License | 6 votes |
def trigger_configuration(self): """ Callback when the configuration button is clicked """ topic_name, ok = QInputDialog.getItem(self._widget, "Select topic name", "Topic name", rostopic.find_by_type('sensor_msgs/Image')) if ok: self._create_subscriber(topic_name) available_rosservices = [] for s in rosservice.get_service_list(): try: if rosservice.get_service_type(s) in _SUPPORTED_SERVICES: available_rosservices.append(s) except: pass srv_name, ok = QInputDialog.getItem(self._widget, "Select service name", "Service name", available_rosservices) if ok: self._create_service_client(srv_name)
Example #9
Source File: Merger.py From openag_cv with GNU General Public License v3.0 | 6 votes |
def __init__(self, args): self.node_name = "ImgMerger" rospy.init_node(self.node_name) # What we do during shutdown rospy.on_shutdown(self.cleanup) # Create the cv_bridge object self.bridge = CvBridge() self.frame_img1 = np.zeros((1280, 1024), np.uint8) self.frame_img2 = np.zeros((1280, 1024), np.uint8) # Subscribe to the camera image topics and set # the appropriate callbacks self.image_sub_first_image = rospy.Subscriber( args[1], Image, self.image_callback_img1) self.image_sub_second_image = rospy.Subscriber( args[2], Image, self.image_callback_img2) self.image_pub = rospy.Publisher(args[3], Image, queue_size=10) rospy.loginfo("Waiting for image topics...")
Example #10
Source File: MaskPlantTray.py From openag_cv with GNU General Public License v3.0 | 6 votes |
def __init__(self, args): self.node_name = "MaskPlantTray" rospy.init_node(self.node_name) # What we do during shutdown rospy.on_shutdown(self.cleanup) # Create the cv_bridge object self.bridge = CvBridge() # Subscribe to the camera image topics and set # the appropriate callbacks self.image_sub = rospy.Subscriber(args[1], Image, self.image_callback) self.image_pub = rospy.Publisher( "%s/MaskPlantTray" % (args[1]), Image, queue_size=10) rospy.loginfo("Waiting for image topics...")
Example #11
Source File: ObjectMeasurer.py From openag_cv with GNU General Public License v3.0 | 6 votes |
def __init__(self, argv): self.node_name = "ObjectMeasurer" rospy.init_node(self.node_name) # What we do during shutdown rospy.on_shutdown(self.cleanup) # Create the cv_bridge object self.bridge = CvBridge() # Subscribe to the camera image topics and set # the appropriate callbacks self.image_sub = rospy.Subscriber(argv[1], Image, self.image_callback) self.image_pub = rospy.Publisher( "%s/ObjectMeasurer" % (argv[1]), Image, queue_size=10) rospy.loginfo("Waiting for image topics...") # Initialization of the 'pixels per metric variable' self.pixelsPerMetric = None
Example #12
Source File: dataRecorder.py From Cherry-Autonomous-Racecar with MIT License | 6 votes |
def __init__(self): print "dataRecorder" self.record = False self.twist = None self.twistLock = Lock() self.bridge = CvBridge() self.globaltime = None self.controller = XBox360() ##rospy.Subscriber("/camera/depth/points" , PointCloud2, self.ptcloudCB) rospy.Subscriber("/camera/depth/image_rect_raw_drop", Image, self.depthCB) rospy.Subscriber("/camera/rgb/image_rect_color_drop", Image, self.streamCB) #rospy.Subscriber("/camera/rgb/image_rect_mono_drop", Image, self.streamCB) ##for black and white images see run.launch and look at the drop fps nodes near the bottom rospy.Subscriber("/lidargrid", Image, self.lidargridCB) rospy.Subscriber("/cmd_vel", TwistStamped, self.cmd_velCB) rospy.Subscriber("/joy", Joy ,self.joyCB) self.sound = rospy.Publisher("/sound_server/speech_synth", String, queue_size=1) rospy.init_node('dataRecorder',anonymous=True) rospy.spin()
Example #13
Source File: runModel.py From Cherry-Autonomous-Racecar with MIT License | 6 votes |
def __init__(self): rospy.loginfo("runner.__init__") # ---------------------------- self.sess = tf.InteractiveSession() self.model = cnn_cccccfffff() saver = tf.train.Saver() saver.restore(self.sess, "/home/ubuntu/catkin_ws/src/car/scripts/model.ckpt") rospy.loginfo("runner.__init__: model restored") # ---------------------------- self.bridge = CvBridge() self.netEnable = False self.controller = XBox360() rospy.Subscriber("/camera/rgb/image_rect_color", Image, self.runCB) rospy.Subscriber("/joy", Joy ,self.joyCB) self.pub_twist = rospy.Publisher("/neural_twist", TwistStamped, queue_size=1) self.sound = rospy.Publisher("/sound_server/speech_synth", String, queue_size=1) rospy.init_node('neural_cmd',anonymous=True) rospy.spin()
Example #14
Source File: cv_bridge_demo.py From Learning-Robotics-using-Python-Second-Edition with MIT License | 6 votes |
def __init__(self): self.node_name = "cv_bridge_demo" rospy.init_node(self.node_name) # What we do during shutdown rospy.on_shutdown(self.cleanup) # Create the OpenCV display window for the RGB image self.cv_window_name = self.node_name # Create the cv_bridge object self.bridge = CvBridge() # Subscribe to the camera image and depth topics and set # the appropriate callbacks rospy.Subscriber("/camera/rgb/image_raw", Image, self.image_callback) rospy.Subscriber("/camera/depth/image_raw", Image, self.depth_callback) rospy.Timer(rospy.Duration(0.03), self.show_img_cb) rospy.loginfo("Waiting for image topics...")
Example #15
Source File: ros_tensorflow_mnist.py From ros_tensorflow with Apache License 2.0 | 6 votes |
def __init__(self): self._cv_bridge = CvBridge() self.x = tf.placeholder(tf.float32, [None,28,28,1], name="x") self.keep_prob = tf.placeholder("float") self.y_conv = makeCNN(self.x,self.keep_prob) self._saver = tf.train.Saver() self._session = tf.InteractiveSession() init_op = tf.global_variables_initializer() self._session.run(init_op) ROOT_PATH = os.path.abspath(os.path.join(os.path.dirname(__file__), os.pardir)) PATH_TO_CKPT = ROOT_PATH + '/include/mnist/model.ckpt' self._saver.restore(self._session, PATH_TO_CKPT) self._sub = rospy.Subscriber('usb_cam/image_raw', Image, self.callback, queue_size=1) self._pub = rospy.Publisher('/result_ripe', Int16, queue_size=1)
Example #16
Source File: tensorflow_in_ros_mnist.py From Tensorflow_in_ROS with Apache License 2.0 | 6 votes |
def __init__(self): self._cv_bridge = CvBridge() self.x = tf.placeholder(tf.float32, [None,28,28,1], name="x") self.keep_prob = tf.placeholder("float") self.y_conv = makeCNN(self.x,self.keep_prob) self._saver = tf.train.Saver() self._session = tf.InteractiveSession() init_op = tf.global_variables_initializer() self._session.run(init_op) self._saver.restore(self._session, "model/model.ckpt") self._sub = rospy.Subscriber('image', Image, self.callback, queue_size=1) self._pub = rospy.Publisher('result', Int16, queue_size=1)
Example #17
Source File: naoqi_camera.py From naoqi_bridge with BSD 3-Clause "New" or "Revised" License | 6 votes |
def __init__(self, node_name='naoqi_camera'): NaoqiNode.__init__(self, node_name) self.camProxy = self.get_proxy("ALVideoDevice") if self.camProxy is None: exit(1) self.nameId = None self.camera_infos = {} def returnNone(): return None self.config = defaultdict(returnNone) # ROS publishers self.pub_img_ = rospy.Publisher('~image_raw', Image, queue_size=5) self.pub_info_ = rospy.Publisher('~camera_info', CameraInfo, queue_size=5) # initialize the parameter server self.srv = Server(NaoqiCameraConfig, self.reconfigure) # initial load from param server self.init_config() # initially load configurations self.reconfigure(self.config, 0)
Example #18
Source File: flock_driver.py From flock with BSD 3-Clause "New" or "Revised" License | 6 votes |
def video_worker(self): # Get video stream, open in PyAV container = av.open(self._drone.get_video_stream()) # Decode h264 rospy.loginfo('starting video pipeline') for frame in container.decode(video=0): # Convert PyAV frame => PIL image => OpenCV Mat color_mat = cv2.cvtColor(numpy.array(frame.to_image()), cv2.COLOR_RGB2BGR) # Convert OpenCV Mat => ROS Image message and publish self._image_pub.publish(self._cv_bridge.cv2_to_imgmsg(color_mat, 'bgr8')) # Check for normal shutdown if self._stop_request.isSet(): return
Example #19
Source File: cv_bridge_demo.py From Learning-Robotics-using-Python-Second-Edition with MIT License | 6 votes |
def __init__(self): self.node_name = "cv_bridge_demo" rospy.init_node(self.node_name) # What we do during shutdown rospy.on_shutdown(self.cleanup) # Create the OpenCV display window for the RGB image self.cv_window_name = self.node_name # Create the cv_bridge object self.bridge = CvBridge() # Subscribe to the camera image and depth topics and set # the appropriate callbacks rospy.Subscriber("/camera/rgb/image_raw", Image, self.image_callback) rospy.Subscriber("/camera/depth/image_raw", Image, self.depth_callback) rospy.Timer(rospy.Duration(0.03), self.show_img_cb) rospy.loginfo("Waiting for image topics...")
Example #20
Source File: rosbag2video.py From GtS with MIT License | 6 votes |
def filter_image_msgs(topic, datatype, md5sum, msg_def, header): if(datatype=="sensor_msgs/CompressedImage"): if (opt_topic != "" and opt_topic == topic) or opt_topic == "": print( "############# USING ######################") print( topic,' with datatype:', str(datatype)) return True; if(datatype=="theora_image_transport/Packet"): if (opt_topic != "" and opt_topic == topic) or opt_topic == "": print( topic,' with datatype:', str(datatype)) # print "############# USING ######################" print( '!!! theora not supportet, sorry !!!') return False; if(datatype=="sensor_msgs/Image"): if (opt_topic != "" and opt_topic == topic) or opt_topic == "": print( "############# USING ######################") print( topic,' with datatype:', str(datatype) ) return True; return False;
Example #21
Source File: rosbag2video.py From rosbag2video with GNU General Public License v2.0 | 6 votes |
def filter_image_msgs(topic, datatype, md5sum, msg_def, header): if(datatype=="sensor_msgs/CompressedImage"): if (opt_topic != "" and opt_topic == topic) or opt_topic == "": print "############# COMPRESSED IMAGE ######################" print topic,' with datatype:', str(datatype) return True; if(datatype=="theora_image_transport/Packet"): if (opt_topic != "" and opt_topic == topic) or opt_topic == "": print topic,' with datatype:', str(datatype) print '!!! theora is not supported, sorry !!!' return False; if(datatype=="sensor_msgs/Image"): if (opt_topic != "" and opt_topic == topic) or opt_topic == "": print "############# UNCOMPRESSED IMAGE ######################" print topic,' with datatype:', str(datatype) return True; return False;
Example #22
Source File: goggle.py From midlevel-reps with MIT License | 6 votes |
def __init__(self): #self.rgb = None rospy.init_node('gibson-goggle') self.depth = None self.image_pub = rospy.Publisher("/gibson_ros/camera_goggle/rgb/image", Image, queue_size=10) self.depth_pub = rospy.Publisher("/gibson_ros/camera_goggle/depth/image", Image, queue_size=10) self.bridge = CvBridge() self.model = self.load_model() self.imgv = Variable(torch.zeros(1, 3, 240, 320), volatile=True).cuda() self.maskv = Variable(torch.zeros(1, 2, 240, 320), volatile=True).cuda() self.mean = torch.from_numpy(np.array([0.57441127, 0.54226291, 0.50356019]).astype(np.float32)) self.mean = self.mean.view(3, 1, 1).repeat(1, 240, 320) self.rgb_sub = rospy.Subscriber("/camera/rgb/image_raw", Image, self.rgb_callback) self.depth_sub = rospy.Subscriber("/camera/depth/image_raw", Image, self.depth_callback)
Example #23
Source File: core.py From pyrobot with MIT License | 6 votes |
def __init__(self, configs): """ Constructor for Camera parent class. :param configs: configurations for camera :type configs: YACS CfgNode """ self.configs = configs self.cv_bridge = CvBridge() self.camera_info_lock = threading.RLock() self.camera_img_lock = threading.RLock() self.rgb_img = None self.depth_img = None self.camera_info = None self.camera_P = None rospy.Subscriber( self.configs.CAMERA.ROSTOPIC_CAMERA_INFO_STREAM, CameraInfo, self._camera_info_callback, ) rgb_topic = self.configs.CAMERA.ROSTOPIC_CAMERA_RGB_STREAM self.rgb_sub = message_filters.Subscriber(rgb_topic, Image) depth_topic = self.configs.CAMERA.ROSTOPIC_CAMERA_DEPTH_STREAM self.depth_sub = message_filters.Subscriber(depth_topic, Image) img_subs = [self.rgb_sub, self.depth_sub] self.sync = message_filters.ApproximateTimeSynchronizer( img_subs, queue_size=10, slop=0.2 ) self.sync.registerCallback(self._sync_callback)
Example #24
Source File: ggcnn_service.py From mvp_grasp with BSD 3-Clause "New" or "Revised" License | 6 votes |
def __init__(self): # Get the camera parameters cam_info_topic = rospy.get_param('~camera/info_topic') camera_info_msg = rospy.wait_for_message(cam_info_topic, CameraInfo) self.cam_K = np.array(camera_info_msg.K).reshape((3, 3)) self.img_pub = rospy.Publisher('~visualisation', Image, queue_size=1) rospy.Service('~predict', GraspPrediction, self.compute_service_handler) self.base_frame = rospy.get_param('~camera/robot_base_frame') self.camera_frame = rospy.get_param('~camera/camera_frame') self.img_crop_size = rospy.get_param('~camera/crop_size') self.img_crop_y_offset = rospy.get_param('~camera/crop_y_offset') self.cam_fov = rospy.get_param('~camera/fov') self.counter = 0 self.curr_depth_img = None self.curr_img_time = 0 self.last_image_pose = None rospy.Subscriber(rospy.get_param('~camera/depth_topic'), Image, self._depth_img_callback, queue_size=1) self.waiting = False self.received = False
Example #25
Source File: run_husky.py From costar_plan with Apache License 2.0 | 5 votes |
def __init__(self): self.index = 0 self.gazeboModels = None self.img_np = None self.pose = Odometry() self.action = Twist() rospy.Subscriber(overheadImageTopic, Image, self.imageCallback) rospy.Subscriber(huskyCmdVelTopic, Twist, self.cmdVelCallback) self.listener = tf.TransformListener() self.writer = NpzDataset('husky_data') self.trans = None self.rot = None self.roll, self.pitch, self.yaw = 0., 0., 0. self.trans_threshold = 0.25 rospy.wait_for_service("/gazebo/pause_physics") self.pause = rospy.ServiceProxy("/gazebo/pause_physics", EmptySrv) self.unpause = rospy.ServiceProxy("/gazebo/unpause_physics", EmptySrv) self.set_model_state = rospy.ServiceProxy("/gazebo/set_model_state", SetModelState) self.set_link_state = rospy.ServiceProxy("/gazebo/set_link_state", SetLinkState) self.get_link_state = rospy.ServiceProxy("/gazebo/get_link_state", GetLinkState) self.pose = None
Example #26
Source File: annotation.py From image_recognition with MIT License | 5 votes |
def store_image(self, roi_image): """ Store the image :param roi_image: Image we would like to store """ if roi_image is not None and self.label is not None and self.output_directory is not None: image_writer.write_annotated(self.output_directory, roi_image, self.label, True)
Example #27
Source File: image_recognition.py From rostensorflow with Apache License 2.0 | 5 votes |
def __init__(self): classify_image.maybe_download_and_extract() self._session = tf.Session() classify_image.create_graph() self._cv_bridge = CvBridge() self._sub = rospy.Subscriber('image', Image, self.callback, queue_size=1) self._pub = rospy.Publisher('result', String, queue_size=1) self.score_threshold = rospy.get_param('~score_threshold', 0.1) self.use_top_k = rospy.get_param('~use_top_k', 5)
Example #28
Source File: gazebo_server.py From robotics-rl-srl with MIT License | 5 votes |
def imageCallback(self, msg): try: # Convert your ROS Image message to OpenCV cv2_img = bridge.imgmsg_to_cv2(msg, "rgb8") self.valid_img = cv2_img except CvBridgeError as e: print("CvBridgeError:", e)
Example #29
Source File: real_baxter_server.py From robotics-rl-srl with MIT License | 5 votes |
def imageCallback(self, msg): try: # Convert your ROS Image message to OpenCV cv2_img = bridge.imgmsg_to_cv2(msg, "rgb8") self.valid_img = cv2_img except CvBridgeError as e: print("CvBridgeError:", e)
Example #30
Source File: loop.py From turtlebot3_applications with Apache License 2.0 | 5 votes |
def imageCallback(msg): global in_progress rospy.loginfo("Final Image") in_progress = False