Automatic Recognition and Understanding of the Driving Environment for Driver Feedback


People


Description


1. Introduction

Understanding the driving environment

In this project we investigate advanced concepts to develop techniques for building internal models of both the vehicle's static environment (objects, features, terrain) and of the vehicle's dynamic environment (people and vehicle moving in the vehicle's environment) from sensor data, which can operate online and can be used to assist the driver by providing the information necessary to make recommendations, to generate alarms, or to take emergency actions. Our overall approach is to combine recent progress in machine perception with the rapid advent of onboard sensors, and the availability of external data sources, such as maps.

2. Approach

The first component of our approach is to extend state-of-the-art machine perception techniques in three areas:
  1. Scene understanding from images in which objects, regions, and features are identified based on image input.
  2. Scene understanding from the type of 3D point clouds acquired from, for example, stereo or Lidar systems.
  3. Analysis of moving objects which includes the ability to predict likely future motions in addition to modeling the current trajectory.

The second component of our approach is to maximize the use of external sources of information. We have started by using current map data from navigation systems to generate priors on distribution of features and objects in the environments, and to generate priors of pedestrian and tracking activity. An example is shown in the following figure, where priors of roads and buildings, obtained from an external map, are overlaid on top of the image.

Priors for roads and buildings obtained from maps


Static scene understanding

One of the goals of our research is to investigate how to boost the performance of machine perception techniques through the use of priors obtained from external sources of information. Our initial objective is to test our algorithms using priors generated from maps by comparing the labeling performance with and without this information. Since maps do not contain information about dynamic elements of the scene (e.g. cars and pedestrians), we focus on the static elements, whose information is available in the maps.
To this end, we have initially concentrated on the use of video cameras as the primary sensor. We have leveraged work done recently in semantic labeling from images. In particular, we use the Stacked Hierarchical Labeling algorithm [1], which is based on the decomposition of an image to encode spatial and relational information.
A sample output from the semantic labeling from images is shown in the following figure. As shown, regions in the image are colorized according to which class they belong to.

Semantic labeling from images

Generation of priors

As mentioned above, we use maps to generate priors on locations where buildingas and roads are likely to occur. Given the vehicle's location, the information from the context in the vicinity is imported from Open Street Map, and then a synthetic image is generated. From this virtual image, a set of priors for roads and buildings is obtained. This process is illustrated in the following image:

Using maps to generate priors


To generate the synthetic image, it is important to acocunt for the vehicle's postion uncertainty. Otherwise, the synthetic image will not match the observation from the vehicle, and the resulting priors will be incrorrect. We process the vehicle's position measurement from GPS using a Kalman Filter, and use the estimates of position and heading uncertainty to calculate the prior. This has the effect of "diluting" the priors when the uncertainty is high, as shown in the following figure:



3. Results

We investigated approaches to generate priors based on a maximum entropy framework. This approach is suitable for our purposes because different sources of information can contribute with different elements for any given scenario, yet, it is not guaranteed that all the elements in the scene will be represented in every source.
Similarly, we have explored several approaches to incorporate the prior distribution into the scene parsing algorithm. These include a combination of probabilities using Product of Experts, re-weighting the label probability distribution using prior contexts using a ensamble of classifiers, and appending the prior probability onto the feature stacks of the Hierarchical Inference Machine[1]. Preliminary results obtained using a small set of urban street scenes from the KITTI Vision Benchmark dataset are shown in the following figure, where two approaches of incorporating priors are compared: re-weighting using logistic regression, and appending prior into hierarchical inference machine:

Sample comparison of performance

Reference

    [1] Stacked Hierarchical Labeling, European Conference on Computer Vision
    D. Munoz, J.A. Bagnell, and M. Hebert.
    ECCV, September, 2010.
    [Paper (PDF)], [BibTeX]
ECCV-2010
Paper
                       

Funding

  • U.S. DoT University Transportation Center Grant DTRT12GUTC11