NavWareSet: A Dataset of Socially Compliant and Non-Compliant Robot Navigation

Top-left: Lidar point cloud of the robot's environment;
Top-right: Robot's own sensory data;
Bottom-left: Synced point cloud with pedestrian's positions and robot's pose;
Bottom-right: Extracted pedestrian and robot trajectories with environment map.

Abstract

NavWareSet is a novel dataset designed to advance socially aware robot navigation. It offers multi-modal recordings of socially compliant and non-compliant robot trajectories in realistic indoor scenarios. Using two different robots across seven diverse setups, NavWareSet captures rich human-robot interactions and navigation challenges. With lidar, RGB-D, video, odometry, and annotated human positions, it provides a valuable source for analyzing and training navigation algorithms that prioritize human comfort and safety.

Overview

  • Wide range of social navigation scenarios with both individual and group interactions.
  • Over 192 minutes of interaction data and over 172 minutes of annotated trajectories.
  • Over 1000 individually annotated human tracks and over 600 robot tracks.
  • Sensory data from the perspective of 2 different robots (Toyota HSR and Clearpath Jackal).
  • Stationary Ground-Truth Recording Station (GRS) with video camera and 3D LiDAR scans.
  • Occupancy grid of the obstacles in the environment.
  • Recording of both social and non-social behavior of the robots.

Experimental Setup

Environment

Experimental Setup
The dataset was recorded in a broad garage space measuring 3.9 m × 10.7 m with a lateral opening making it possible to elicit tracks in a straight line or around corners. The environment is equipped with a Ground-Truth Recording Station (GRS) that includes a video recorder and a 3D LiDAR scanner. For each scene goal positions were placed around the map according to different social navigation scenarios.


Ground-Truth Recording Station (GRS)

Experimental Setup

The Ground-Truth Recording Station (GRS) is equipped with an Intel® RealSense™ Depth Camera D455 and a Robosense RS-LiDAR-16 3D LiDAR. Both sensors are mounted on a custom 3D-printed frame, which is positioned on a tripod for stable data collection. You can interact with the frame in the 3D viewport above (click and drag to rotate) or access the 3D model here.

To minimize occlusions, the GRS was positioned at a height of 2.17 m above the ground. To enhance the resolution of the LiDAR data, the GRS was tilted 15° downward toward the area of interest.


Robots

Experimental Setup
The robots used in the NavWareSet dataset are the Clearpath Jackal and the Toyota HSR. The Jackal is a small, rugged robot designed for outdoor and indoor environments, equipped with a 3D LiDAR sensor and an RGB-D camera. The HSR is a versatile human-support robot capable of performing household tasks, featuring a manipulator arm, a mobile base, and various sensors for safe and intelligent interaction with people and objects. In most scenarios both robots were teleoperated by a human operator, first in a socially aware manner, and then in a non-aware manner.


Experiment description

Experimental Setup

The experiment was designed to capture common social navigation scenarios. These scenarios were proposed by Francis et al. in their paper, Principles and Guidelines for Evaluating Social Robot Navigation Algorithms. They serve not only to facilitate data collection but also to provide a standardized basis for evaluating the performance of social navigation algorithms. A detailed description of the scenarios used in NavWareSet is provided in the table below.


In total 17 participants were assigned numbers and divided in 2 groups of 5 and 5 pairs. Each group performed all scenarios but the object handover scenario. All scenarios were recorded for 4 minutes. Most scenarios were performed with both robots, except for the object handover scenario which was only performed with the Toyota HSR. Most scenarios were recorded twice, once with the robot navigating in a socially compliant manner and once in a non-compliant manner. Social behavior adhered to predefined rules (maintaining social distances, avoiding abrupt movements), while non-social behavior involved direct waypoint-to-waypoint navigation, disregarding human presence. A detailed description of all recorded scenes used in NavWareSet is provided in the table below.

Data Pipeline

Experimental Setup
  1. Experiment Design: The pipeline begins with careful planning of social navigation scenarios in a controlled indoor environment. At this step it is defined how robots and human participants will interact in both socially compliant and non-compliant modes. This phase ensures that a diverse set of realistic interactions is captured.

  2. Data Collection: The robots (HSR and Jackal) and a Ground-Truth Recording Station (GRS) simultaneously collect raw sensory data. The robots log LiDAR, RGB-D, stereo vision, IMU, odometry, and velocity commands. Meanwhile, the stationary GRS records the entire scene from an external vantage point, providing an independent reference for later annotation and validation.

  3. Raw Data Organization: All recorded sensor data is saved in ROS bag files. Each scene’s raw data is packaged into separate files for the robot and the GRS. This organized structure allows researchers to easily replay the scenarios and verify the consistency of recorded sensor streams.

  4. Data Processing: The raw data undergoes processing. Robot poses are extracted using SLAM. Human participant positions are manually annotated frame-by-frame using the CVAT tool applied to the GRS point cloud. This ensures accurate human trajectory labeling, crucial for studying robot social navigation behavior. All CVAT annotations were saved using the Supervisely format.

  5. Data Product Generation: After processing, clean and usable data products are created for each scene:
    • Annotated point clouds synchronized with robot poses.
    • Robot and participant positions in easily accessible CSV files.
    • Occupancy grids detailing static obstacles.
    • All metadata and calibration offsets are packaged to ensure reproducibility.
    These structured files make the dataset valuable for training, benchmarking, and testing navigation algorithms.

  6. Distribution: The complete, organized dataset — including raw bags, annotations, CSVs, and occupancy grids — is shared publicly. A series of tutorials and helper scripts are provided to assist researchers in using the dataset effectively. The dataset is designed to be user-friendly, with clear documentation on how to access and utilize the data for various research purposes. Researchers can access everything through the NavWareSet website and the NavWare GitHub repository. All data can be downloaded in bulk or individually using the dat@UBFC website.

Data Overview

Each scene is divided into four files, each containing its own data. All files start with the number of its respective scene name.

For generic scene "x" we have the following files:


Scenes 53 to 57 don't have person position annotation because the participants don't walk around the environment; only the robot moves between the participants.

BibTeX

BibTex Code Here Soon...

Authors

1Universidade Federal de Minas Gerais
2Université de technologie de Belfort Montbéliard
3Czech Technical University in Prague
4Inria
5ENSTA - Institut Polytechnique de Paris

Funding

Experimental Setup
This project has received funding from the French National Research Agency (ANR) under grant agreement No ANR-23-CE10-0016.