ROADWork dataset contains annotated images and sequences taken by driving through nearly 5000 different work zones in 18 US cities.
Perceiving and navigating through work zones is challenging and under-explored, even with major strides in self-driving research. An important reason is the lack of open datasets for developing new algorithms to address this long-tailed scenario. We propose the ROADWork dataset to learn how to recognize, observe and analyze and drive through work zones. We find that state-of-the-art foundation models perform poorly on work zones. With our dataset,
Why are work zones so hard for self-driving cars? No two work zones are truly alike and objects like barriers, and work vehicles widely vary by the type, status, location and geography of work zones. Navigational aids (like signs) are customized to the particular work zone and require fine-grained understanding. Spatial configurations of work zone objects do not conform to the lane, road and sidewalk layouts. Often, rules of traffic are suspended and new rules are enforced that may change over time. All these reasons make work zone understanding and navigation difficult.
Work zones are dynamic. For example, in this situation, a self-driving is expected to read the signs held by the workers saying STOP, wait till the car from the opposite direction passes, and observe that the workers have changed the sign to SLOW and then proceed. This requires the car to understand the global scene context, fine grained observations like signs, and the larger context of the work zone and workers which makes navigation challenging.
Work zone images and sequences from 18 U.S. cities. We have segmented 15 object instances like workers, vehicles and barriers. We provide scene level and object attributes (for signs and arrow boards) to enable fine-grained understanding. Work zone scene descriptions help analyze the scene globally and one passable trajectory automatically estimated from video to learn how to drive through work zones.
Using detectors trained on ROADWork dataset, we discovered work zones around the world in Mappilary and BDD datasets.
ROADWork dataset improves fine-grained understanding of work zone signs, arrow boards and other rare objects. These objects are poorly detected by pre-trained foundation models like Detic and OpenSEED.
Our sign attributes contain 62 types of graphics and 360 different text annotations.
We annotate rare and diverse object instances like police cars, tubular markers, barriers and work vehicles.
search
Please hover on the images to zoom in.
Pre-trained foundation models like LLaVA poorly understand work zones. Using ROADWork Dataset, we improved their performance via ground truth descriptions and work zone objects as context.
Using drivable paths from ROADWork dataset, we can learn drivable goals and pathways for navigating work zones.
This work was supported by a research contract from General Motors Research-Israel, NSF Grant CNS-2038612, a US DOT grant 69A3551747111 through the Mobility21 UTC and grants 69A3552344811 and 69A3552348316 through the Safety21 UTC. We thank N. Dinesh Reddy, Khiem Vuong, Shefali Srivastava, Neha Boloor, Tiffany Ma for insightful discussions.