Maritime Computer Vision Workshop @ CVPR 2026

Challenges / Vision-to-Chart Data Association

Vision-to-Chart Data Association

Quick links: Dataset download Submit Leaderboard Ask for help

Quick Start

  1. Download the dataset from the dataset page .
  2. Train your model on the provided dataset (or use the 120-epoch checkpoint). Find starter code here .
  3. Evaluate it on the validation split using provided evaluation script (coming soon).
  4. Details for final submission will be announced soon.

Overview

This challenge is about augmenting navigation data for maritime vehicles (boats). Given an image captured by boat-mounted camera, and chart markers of nearby navigational aids (buoys), the task is to identify visible buoys and match them with their corresponding chart markers. The goal is to maximize detected buoys as well as correct matches while minimizing localization error of detected buoys and wrong detections. The challenge is based on the paper "Real-Time Fusion of Visual and Chart Data for Enhanced Maritime Vision".

Task

The task is to develop a deep learning model capable of detecting buoys in a monocular image, and matching them to associated chart positions. Models will be evaluated on both detection and matching accuracy as well as localization error.

Dataset

The dataset consists of 5189 entries, with each entry containing a single RGB-based image, a variable amount of quieries, and a variable amount of labels. Each query denotes chart data of a nearby buoy, and has the format: (query id, distance to ship [meters], bearing from ship centerline [degrees], longitude, latidude). Each label denotes a nearby buoy visible in the image, and has the format (query id, BB center x, BB center y, BB width, BB height), where BB location specification follows the YOLO format and denotes the corresponding queries position in the image. Note that this dataset contains entries where no buoy is visible.
The dataset is provided in a predefined format split into training and validation sets for use with the baseline training code. Participants are free to further manipulate or reorganize the data according to their own experimental needs. The leaderboard evaluation will be conducted on a separate, unseen dataset that is not shared with the participants. In addition, to enable participants to experiment with their own methods, raw IMU data corresponding to each sequence is also provided. The IMU data is not integrated into the baseline training code; however, participants may extract and incorporate these raw measurements into their own algorithms during training if desired.

Evaluation metrics

The submitted models are evaluated on a test set that is not publicly available.
The following metrics are employed to assess model capabilities:

Participate

To participate in the challenge follow these steps:

  1. Download the dataset from the dataset page .
  2. Train a deep learning model on the dataset.
  3. Use the provided evaluation script to get preliminary performance on the validation split
  4. Upload the results using upload page to be represented on the leaderboard .

Get Started

To help you get started and provide a brief introduction to the topic, we have developed a fusion transformer based on DETR. You can find the code here and the 120-epoch pre-trained checkpoint here.

Final Submission

The final submission will be evaluated on the private test data and hence determines the final performance score of the participants. This challenges aims to provide participants with as much freedom as possible. However, to ensure a fair and streamlined evaluation, the following restrictions have to be imposed for the final submission.

Terms and Conditions

If you have any questions, please join the MaCVi Support forum.