List of ROS Topics#
This page list all the topics exposed by the SDK 23.1.
Caution
Only the topics contributing to the public API of the pal-sdk-23.1 are listed here.
Additional ROS topics might be present on the robot, for internal purposes. They are however not part of the documented and supported robot API.
Alphabetic index#
/active_listening/arm_left_controller/command/arm_left_controller/safe_command/arm_right_controller/command/arm_right_controller/safe_command/audio/channel0/audio/channel1/audio/channel2/audio/channel3/audio/channel4/audio/channel5/audio/raw/audio/sound_direction/audio/sound_localization/audio/speech/audio/voice_detected/current_zone_of_interest/diagnostics/diagnostics_agg/eyes/expression/hand_left_controller/command/hand_right_controller/command/head_controller/command/head_front_camera/color/camera_info/head_front_camera/color/image_raw/*/head_front_camera/image_throttle/compressed/hri_face_detect/ready/hri_face_identification/ready/humans/bodies/*/cropped/humans/bodies/*/joint_states/humans/bodies/*/position/humans/bodies/*/roi/humans/bodies/*/skeleton2d/humans/bodies/tracked/humans/bodies/*/velocity/humans/candidate_matches/humans/faces/*/aligned/humans/faces/*/cropped/humans/faces/*/landmarks/humans/faces/*/roi/humans/faces/tracked/humans/persons/*/alias/humans/persons/*/anonymous/humans/persons/*/body_id/humans/persons/*/engagement_status/humans/persons/*/face_id/humans/persons/known/humans/persons/*/location_confidence/humans/persons/tracked/humans/persons/*/voice_id/humans/voices/*/audio/humans/voices/*/is_speaking/humans/voices/*/speech/humans/voices/tracked/intents/interaction_logger/interaction_profile_manager/parameter_updates/joint_states/kb/add_fact/kb/events/*/kb/remove_fact/look_at/look_at_with_style/map/mobile_base_controller/cmd_vel/mobile_base_controller/odom/move_base/current_goal/move_base_simple/goal/pause_navigation/power/battery_level/power/is_charging/power/is_docked/power/is_emergency/power/is_plugged/power_status/scan/torso_back_camera/fisheye1/camera_info/torso_back_camera/fisheye1/image_raw/*/torso_back_camera/fisheye2/camera_info/torso_back_camera/fisheye2/image_raw/*/torso_front_camera/aligned_depth_to_color/camera_info/torso_front_camera/aligned_depth_to_color/image_raw/*/torso_front_camera/color/camera_info/torso_front_camera/color/image_raw/*/torso_front_camera/depth/camera_info/torso_front_camera/depth/color/points/torso_front_camera/depth/image_rect_raw/*/torso_front_camera/infra1/camera_info/torso_front_camera/infra1/image_rect_raw/*/torso_front_camera/infra2/image_rect_raw/compressed/touch_web_state/user_input/web/go_to/web_subtitles
By capability#
Developing applications#
/intents(documentation) An intent, encoding a desired activity to be scheduled by the robot (not to be confused by the chatbot intents). Read more about Intents.
Expressive interactions#
/eyes/expression(documentation) Set the expression of ARI eyes. See How-to: Control ARI’s eyes for details./look_at(documentation) Set a target for the robot to look at. Uses both the eyes and the head position./look_at_with_style(documentation) Set a target for the robot to look at, while letting you configure how the gaze will be performed.
ARI hardware#
/audio/raw(documentation) Merged audio channel of the ReSpeaker’s 4 microphones (alias for /audio/channel0)./audio/channel0(documentation) Merged audio channel of the ReSpeaker’s 4 microphones/audio/channel1(documentation) Audio stream from the ReSpeaker’s first microphone./audio/channel2(documentation) Audio stream from the ReSpeaker’s second microphone./audio/channel3(documentation) Audio stream from the ReSpeaker’s third microphone./audio/channel4(documentation) Audio stream from the ReSpeaker’s fourth microphone./audio/channel5(documentation) Monitor audio stream from the ReSpeaker’s audio input (used for self-echo cancellation)./audio/voice_detected(documentation) Publishes a boolean indicating if a voice is currently detected (ie, whether someone is currently speaking)/audio/sound_direction(documentation) The estimated Direction of Arrival of the detected sound./audio/sound_localization(documentation) The estimated sound source location./audio/speech(documentation) Raw audio data of detected speech (published once the person has finished speaking)./head_front_camera/color/camera_info(documentation) Camera calibration and metadata/head_front_camera/color/image_raw/*(documentation) Color rectified image. RGB format/head_front_camera/image_throttle/compressed(documentation) Compressed head image./joint_states(documentation) The current state of the robot’s joints (eg angular position of each joint)./mobile_base_controller/cmd_vel(documentation) Set the desired linear and angular velocity of the robot (in meters per second)./scan(documentation) Laser scan readings of ARI’s back LIDAR./torso_back_camera/fisheye1/camera_info(documentation) Camera calibration and metadata (fisheye2)/torso_back_camera/fisheye1/image_raw/*(documentation) Fisheye image/torso_back_camera/fisheye2/camera_info(documentation) Camera calibration and metadata (fisheye2)/torso_back_camera/fisheye2/image_raw/*(documentation) Fisheye image (/torso_front_camera/aligned_depth_to_color/camera_info(documentation) Intrinsics parameters of the aligned dept to color image/torso_front_camera/aligned_depth_to_color/image_raw/*(documentation) Aligned depth to color image/torso_front_camera/color/camera_info(documentation) Camera calibration and metadata/torso_front_camera/color/image_raw/*(documentation) Color rectified image. RGB format/torso_front_camera/depth/camera_info(documentation) Camera calibration and metadata/torso_front_camera/depth/color/points(documentation) Registered XYZRGB point cloud./torso_front_camera/depth/image_rect_raw/*(documentation) Rectified depth image/torso_front_camera/infra1/camera_info(documentation) Camera calibration and metadata (infra1 and infra2)/torso_front_camera/infra1/image_rect_raw/*(documentation) Raw uint16 IR image/torso_front_camera/infra2/image_rect_raw/compressed(documentation)
Robot management#
/diagnostics(documentation)/diagnostics_agg(documentation)/interaction_logger(documentation) Logs arbitrary strings to a CSV file. See How-to: Log and retrieve data from the robot for details./interaction_profile_manager/parameter_updates(documentation)/power_status(documentation)/power/battery_level(documentation)/power/is_charging(documentation)/power/is_docked(documentation)/power/is_emergency(documentation)/power/is_plugged(documentation)
Gestures and motions#
/arm_left_controller/command(documentation)/arm_left_controller/safe_command(documentation)/arm_right_controller/command(documentation)/arm_right_controller/safe_command(documentation)/hand_left_controller/command(documentation)/hand_right_controller/command(documentation)/head_controller/command(documentation)
Knowledge and reasoning#
/kb/add_fact(documentation) Statements published to this topic are added to the knowledge base. The string must represent a <s, p, o> triple, with terms separated by a space. See KnowledgeCore documentation for details./kb/events/*(documentation) Event notifications for previously subscribed events. See service_kb_events./kb/remove_fact(documentation) Statements published to this topic are removed from the knowledge base. The string must represent a <s, p, o> triple, with terms separated by a space. See KnowledgeCore documentation for details.
Speech and language processing#
/humans/voices/*/audio(documentation) The audio stream of the voice./humans/voices/*/is_speaking(documentation) Whether verbal content is currently recognised in this voice’s audio stream./humans/voices/*/speech(documentation) The recognised text, as spoken by this voice./humans/voices/tracked(documentation) The list of voices currently detected by the robot./web_subtitles(documentation)/active_listening(documentation) Whether or not recognized speech should be further processed (eg by the chatbot). See Dialogue management for details.
Touchscreen#
/touch_web_state(documentation)/user_input(documentation)/web/go_to(documentation) Sets the webpage to be displayed on the touchscreen.