The GazeSense ROS bridge exposes 3D eye tracking data from the GazeSense application into the Robot Operating System (ROS).
GazeSense is an application developed by Eyeware Tech SA (http://www.eyeware.tech) that provides real-time 3D eye tracking by relying on consumer 3D sensors. GazeSense allows to define virtual 3D objects of the environment, and measure the attention of people towards those objects.
The GazeSense ROS bridge publishes 3D eye tracking data on ROS topics. In addition, this tracking data can be visualized in Rviz:
Left: GazeSense eye tracking. Right: ROS bridge eye tracking.Both visualizations show the position of a camera (purple glyph) and a tracked person's estimated gaze (orange rays), as the person is looking at different areas of the environment (yellow screens).
The area that is being fixated at a given instant is highlighted in red. The black cube represents the point that the person is looking at.
The GazeSense ROS bridge was developed and tested on a system with:
- Ubuntu 18.04
- GazeSense (graphical application) 2.1.0
- ROS Melodic with Python 3 support
In the rest of this guide, we will assume that:
- GazeSense (graphical application) is uncompressed in
~/gazesense/
- the ROS workspace is located in
~/gazesense_ws/
However, the two paths can be adjusted as desired.
Uncompress GazeSense (graphical application) in ~/gazesense
. In particular, the directory ~/gazesense/API/python
should exist.
First, we install ROS Melodic as described in the ROS installation page:
sudo sh -c 'echo "deb http://packages.ros.org/ros/ubuntu $(lsb_release -sc) main" > /etc/apt/sources.list.d/ros-latest.list'
sudo apt-key adv --keyserver 'hkp://keyserver.ubuntu.com:80' --recv-key C1CF6E31E6BADE8868B172B4F42ED6FBAB17C654
sudo apt update
sudo apt install ros-melodic-desktop-full
The Python API of GazeSense requires Python 3.6, but ROS Melodic ships with Python 2 by default. Therefore, we need to install the Python 3 version of ROS Melodic. We follow the instructions from here:
sudo apt install python3-pip python3-all-dev python3-rospkg
# The previous command might uninstall parts of ROS Melodic. We now restore it:
sudo apt install ros-melodic-desktop-full --fix-missing
sudo apt install python3-numpy
sudo apt install ros-melodic-ddynamic-reconfigure
We create the ROS workspace in ~/gazesense_ws/
in this way:
source /opt/ros/melodic/setup.bash
mkdir -p ~/gazesense_ws/src/
catkin_init_workspace ~/gazesense_ws/src/
git -C ~/gazesense_ws/src/ clone https://github.com/eyeware/eyeware-ros
catkin_make -C ~/gazesense_ws/
Successful result:
...
[100%] Built target gazesense_msgs_generate_messages_cpp
[100%] Built target gazesense_msgs_generate_messages_py
[100%] Built target gazesense_msgs_generate_messages_eus
Scanning dependencies of target gazesense_msgs_generate_messages
[100%] Built target gazesense_msgs_generate_messages
Below, we first describe how to run a complete demo (including graphical visualization). Then, we will go over specific components (publisher, subscriber) and their available options.
Assuming that everything is installed and compiled as described in the Installation section, run GazeSense in one terminal:
cd ~/gazesense/
./run_GazeSense.sh
Within GazeSense, it is useful to export the 3D Setup description file on the disk as a JSON file. This file encodes the geometric description of the environment, such as the locations and sizes of possible attention targets. We can export the file with this button:
Let us assume the file is exported to the location /PATH/TO/SETUP.json
.
Finally, within GazeSense, tick the "External Communication" box with port 12000, and press "Connect". The application will start tracking the person's head and gaze.
Then, in another terminal, run:
source ~/gazesense_ws/devel/setup.bash
export PYTHONPATH=$HOME/gazesense/API/python:$PYTHONPATH
roslaunch gazesense_bridge demo.launch port:=12000 host:=localhost setup:=/PATH/TO/SETUP.json
where setup:=/PATH/TO/SETUP.json
points to the 3D Setup description file (geometric description of the environment) currently being used by GazeSense for its tracking, exported as described before.
In RViz, we can trigger the display of the following elements:
- TF trees
- screens (topic
/gazesense/screens
) - camera glyph (topic
/gazesense/cameras
) - gaze rays (topic
/gazesense/gaze
) - gaze intersection point on the screen (topic
/gazesense/gaze_intersection
)
Example:
Next, we will describe the individual parts that make up the complete demo.
We run the ROS bridge publisher of head and eye tracking information as follows:
source ~/gazesense_ws/devel/setup.bash
export PYTHONPATH=$HOME/gazesense/API/python:$PYTHONPATH
roslaunch gazesense_bridge gazesense_bridge.launch
Messages are published under topic
/gazesense_bridge/persons
.
Optionally, we can run the ROS bridge publisher by specifying the following options:
source ~/gazesense_ws/devel/setup.bash
export PYTHONPATH=$HOME/gazesense/API/python:$PYTHONPATH
roslaunch gazesense_bridge gazesense_bridge.launch port:=12000 host:=localhost root:=Camera setup:=/PATH/TO/SETUP.json
where:
host
andport
define the network configuration where the ROS bridge receives tracking data from the GazeSense application. The port number must match the one specified in the "External Communication" panel in GazeSense:
setup:=/PATH/TO/SETUP.json
loads the 3D Setup description file from GazeSense (geometric description of environment screens and camera, exported as described before), and publishes the TF tree of the screens and cameraroot:=Camera
indicates which screen or camera should be used as the TF tree root.
We provide a simple listener to show how to subscribe to the GazeSense ROS API and receive tracking data from it.
In a terminal we run the GazeSense bridge publisher:
source ~/gazesense_ws/devel/setup.bash
export PYTHONPATH=$HOME/gazesense/API/python:$PYTHONPATH
roslaunch gazesense_bridge gazesense_bridge.launch
In another terminal, we run the listener:
source ~/gazesense_ws/devel/setup.bash
export PYTHONPATH=$HOME/gazesense/API/python:$PYTHONPATH
rosrun gazesense_bridge gazesense_listener.py
This simple listener demo subscribes to /gazesense/persons
and prints the
ID of the screen that the user is looking at.
RViz markers permit to show elements such as screen rectangles, reference frame markers, intersection points, head and gaze lines of the user via
To display these markers, run:
source ~/gazesense_ws/devel/setup.bash
export PYTHONPATH=$HOME/gazesense/API/python:$PYTHONPATH
roslaunch gazesense_bridge gazesense_bridge.launch root:=Camera with_markers:=1 setup:=/PATH/TO/SETUP.json
where
with_markers:=1
activates markers (they are off by default)setup:=/PATH/TO/SETUP.json
loads the 3D Setup description file from GazeSense (geometric description of environment screens and camera, exported as described before). Make sure it is the same as the one currently used by GazeSense.
The following markers are published:
- screens (topic
/gazesense/screens
) - camera (topic
/gazesense/cameras
) - gaze rays (topic
/gazesense/gaze
) - gaze intersection point on the screen (topic
/gazesense/gaze_intersection
)
We can customize the way in which GazeSense ROS bridge and its options are run from file
launch/demo.launch
:
<include
file="$(find gazesense_bridge)/launch/gazesense_bridge.launch" >
<arg name="port" value="$(arg port)" />
<arg name="host" value="$(arg host)" />
<arg name="setup" value="$(arg setup)" />
<arg name="root" value="World" />
<arg name="with_markers" value="true" />
</include>
In addition, we can connect the GazeSense TF tree to any other TF tree (e.g., positioned on a robot part) as follows:
<node
pkg="tf2_ros"
type="static_transform_publisher"
name="gazesense_broadcaster"
args="0 0 0 0 0 0 1 YOUR_NODE GAZESENSE_ROOT" >
</node>
For instance, if the camera is placed on a ROS node called robot_hand
, we can use:
# Argument for gazesense_bridge
<arg name="root" value="Camera" />
# Argument for transform
args="0 0 0.05 0 0 0 1 robot_hand Camera"
Depending on which root is more convenient to use in a custom application, one can use either "Camera", or "World", or any of the other screens as GAZESENSE_ROOT
.
The GazeSense ROS bridge works by sending data in one direction, from GazeSense (graphical application) to ROS.
This means that GazeSense app is the part responsible for selecting the input camera hardware type, as well as managing the 3D Setup environment. The ROS bridge receives tracking data from GazeSense app over a network connection. If you wish to change the camera input type or the location of the camera and attention screens in the 3D Setup, change those properties in GazeSense app first, then restart the ROS bridge so that it receives the updated information.
It is currently not possible to use a single camera as input to 1) GazeSense ROS bridge and 2) another program simultaneously. GazeSense (the graphical application) takes exclusive control of the camera driver.
Currently, the RGB image or point cloud from the camera cannot be displayed in RViz.
The reason is that the camera driver is started and managed by the GazeSense application, not by the ROS bridge. The ROS bridge receives tracking data from GazeSense (this data does not include images for efficiency reasons), then re-publishes that tracking information on ROS topics.
The glyph for the camera was generated with this script. The .stl
file can be generated as follows:
cd gazesense_bridge/scripts/
python3 generate_camera_stl.py > camera.stl
If you are interested in using 3D eye tracking in an application, let us know at contact@eyeware.tech or https://eyeware.tech/