Skip to content

autoware_traffic_light_multi_camera_fusion#

Overview#

This node fuses traffic light recognition results from multiple cameras to produce a single, reliable traffic light state. By integrating information from different viewpoints and ROIs, it ensures robust performance even in challenging scenarios, such as partial occlusions or recognition errors from an individual camera.

graph LR
    subgraph "Multi Camera Feeds"
        direction TB
        Cam1[" <br> <b>Camera 1</b> <br> State: GREEN <br> Confidence: 0.95"]
        Cam2[" <br> <b>Camera 2</b> <br> State: GREEN <br> Confidence: 0.94"]
        Cam3[" <br> <b>Camera 3</b> <br> State: RED <br> Confidence: 0.95"]
    end

    subgraph "Processing"
        direction TB
        Fusion["<b>Multi-Camera Fusion Node</b> <br><i>Fuses evidence using <br> Bayesian updating</i>"]
    end

    subgraph "Unified & Robust State"
        direction TB
        Result[" <br> <b>Final State: GREEN</b>"]
    end

    Cam1 --> Fusion
    Cam2 --> Fusion
    Cam3 --> Fusion
    Fusion --> Result

    style Fusion fill:#e0f7fa,stroke:#00796b,stroke-width:2px,color:#004d40
    style Result fill:#e8f5e9,stroke:#2e7d32,stroke-width:3px,color:#1b5e20

How It Works#

The fusion algorithm operates in two main stages.

graph TD
    subgraph "Input: Multiple Camera Results"
        A["Camera 1<br>Recognition Result"]
        B["Camera 2<br>Recognition Result"]
        C["..."]
    end

    subgraph "Stage 1: Per-Camera Fusion"
        D{"Best ROIs Selection<br><br>For each ROI,<br>select the single most<br>reliable detection result."}
    end

    E["Best Detection per ROIs"]

    subgraph "Stage 2: Group Fusion"
        F{"Group Consensus<br><br>Fuse all 'best detections'<br>into a single state for<br>the entire traffic light group<br>using Bayesian updating."}
    end

    subgraph "Final Output"
        G["Final Group State<br>(e.g., GREEN)"]
    end

    A --> D
    B --> D
    C --> D
    D --> E
    E --> F
    F --> G

    style D fill:#e0f7fa,stroke:#00796b,stroke-width:2px,color:black
    style F fill:#e0f7fa,stroke:#00796b,stroke-width:2px,color:black
    style E fill:#fff,stroke:#333,stroke-width:2px,stroke-dasharray: 5 5,color:black
    style G fill:#e8f5e9,stroke:#2e7d32,stroke-width:3px,color:black

Stage 1: Best View Selection (Per-Camera Fusion)#

First, for each individual ROIs, the node selects the single most reliable detection—the "best shot"—from all available camera views.

This selection is based on a strict priority queue:

  • Latest Timestamp: Detections with the most recent timestamp are prioritized for the same sensor.
  • Known State: Results with a known color (Red, Green, etc.) are prioritized over 'Unknown'.
  • Full Visibility: Detections from non-truncated ROIs (fully visible ROIs) are prioritized.
  • Highest Confidence: The result with the highest detection confidence score is prioritized.

This process yields the single most plausible recognition for every ROIs.

Stage 2: Group Consensus (Bayesian Fusion)#

Next, the "best shot" detections from Stage 1 are fused to determine a single, coherent state for the entire traffic light group. Instead of simple voting or averaging, this node employs a more principled method: Bayesian updating.

  • Belief Score: Each color (Red, Green, Yellow) maintains a "belief score" represented in log-odds for numerical stability and ease of updating.
  • Evidence Update: Each selected detection from Stage 1 is treated as a piece of "evidence." Its confidence score is converted into a log-odds value representing the strength of that evidence.
  • Score Accumulation: This evidence is added to the corresponding color's belief score.
  • Final Decision: After accumulating all evidence, the color with the highest final score is chosen as the definitive state for the group.

Input topics#

For every camera, the following three topics are subscribed:

Name Type Description
~/<camera_namespace>/camera_info sensor_msgs::msg::CameraInfo camera info from map_based_detector
~/<camera_namespace>/detection/rois tier4_perception_msgs::msg::TrafficLightRoiArray detection roi from fine_detector
~/<camera_namespace>/classification/traffic_signals tier4_perception_msgs::msg::TrafficLightArray classification result from classifier

You don't need to configure these topics manually. Just provide the camera_namespaces parameter and the node will automatically extract the <camera_namespace> and create the subscribers.

Output topics#

Name Type Description
~/output/traffic_signals autoware_perception_msgs::msg::TrafficLightGroupArray traffic light signal fusion result

Node parameters#

Name Type Description Default Range
message_lifespan float The maximum timestamp span to be fused. 0 N/A
approximate_sync boolean Whether to work in Approximate Synchronization Mode. 0 N/A
prior_log_odds float The uniform prior log-odds to initialize the belief for any new traffic light state. 0.0 (log(1)) means a neutral (50%) initial belief. Positive values make all states initially more likely, negative values make them less likely. 0 N/A