Skip to content

Latest commit

 

History

History
81 lines (55 loc) · 3.22 KB

README.md

File metadata and controls

81 lines (55 loc) · 3.22 KB

Multi Robot Scene Completion: Towards Task-agnostic Collaborative Perception

Yiming Li, Juexiao Zhang, Dekun Ma, Yue Wang, Chen Feng

See our paper on OpenReview.

main

News:

[2022-11] Our paper is camera-ready!

[2022-10] The project website is online.

[2022-09] Our work is accepted at the 6th Conference on Robot Learning (CoRL 2022).

Abstract:

Collaborative perception learns how to share information among multiple robots to perceive the environment better than individually done. Past research on this has been task-specific, such as detection or segmentation. Yet this leads to different information sharing for different tasks, hindering the large-scale deployment of collaborative perception. We propose the first task-agnostic collaborative perception paradigm that learns a single collaboration module in a self-supervised manner for different downstream tasks. This is done by a novel task termed multi-robot scene completion, where each robot learns to effectively share information for reconstructing a complete scene viewed by all robots. Moreover, we propose a spatiotemporal autoencoder (STAR) that amortizes over time the communication cost by spatial sub-sampling and temporal mixing. Extensive experiments validate our method's effectiveness on scene completion and collaborative perception in autonomous driving scenarios.

Installation

The work is tested with:

  • python 3.7
  • pytorch 1.8.1
  • torchvision 0.9.1
  • timm 0.3.2

Download the GitHub repository:

git clone https://github.com/coperception/star.git
cd star

Create a conda environment with the dependencies:

conda env create -f environment.yml
conda activate star

If conda installation failed, install the dependencies through pip:
(Make sure your Python version is 3.7)

pip install -r requirements.txt

Usage:

To train, run:

cd completion/
make train_completion

To test the trained model on scene completion:

cd completion/
make test_completion

More commands and experiment settings are included in the Makefile.

You can find the training and test scripts at: completion.

Dataset:

Our experiments are conducted on the V2X-Sim[1] simulated dataset. Find more about the dataset on the website.

[1] Li, Yiming, et al. "V2X-Sim: Multi-agent collaborative perception dataset and benchmark for autonomous driving." IEEE Robotics and Automation Letters 7.4 (2022): 10914-10921.

Citation:

@inproceedings{li2022multi,
  title={Multi-Robot Scene Completion: Towards Task-Agnostic Collaborative Perception},
  author={Li, Yiming and Zhang, Juexiao and Ma, Dekun and Wang, Yue and Feng, Chen},
  booktitle={6th Annual Conference on Robot Learning}
}