Skip to content

Latest commit

 

History

History
95 lines (69 loc) · 5.86 KB

README.md

File metadata and controls

95 lines (69 loc) · 5.86 KB

FOCAL(Ford - OLIVES Collaboration on Active Learning): A Cost-Aware Video Dataset for Active Learning


This work was done in the Omni Lab for Intelligent Visual Engineering and Science (OLIVES) @ Georgia Tech in collaboration with the Ford Motor Company. It has recently been accepted for publication in the IEEE International Conference on Big Data (Acceptance Rate 17.4%)!! Feel free to check our lab's Website and GitHub for other interesting work!!!

The paper can be read at this arxiv link.


Citation

K. Kokilepersaud*, Y. Logan*, R. Benkert, C. Zhou, M. Prabhushankar, G. AlRegib, E. Corona, K. Singh, A. Parchami, "FOCAL: A Cost-Aware, Video Dataset for Active Learning," in IEEE Conference on Big Data 2023, Sorento, Italy, Dec. 15-18, 2023.

@article{kokilepersaud2023focal,
  title={FOCAL: A Cost-Aware, Video Dataset for Active Learning},
  author={Kokilepersaud, Kiran and Logan, Yash-Yee and Benkert, Ryan and Zhou, Chen and Prabhushankar, Mohit and AlRegib, Ghassan and Corona, Enrique and Singh, Kunjan and Parchami, Armin},
  journal={IEEE International Conference on Big Data},
  year={2023},
  publisher={IEEE}
}

Abstract

In this paper, we introduce the FOCAL (Ford-OLIVES Collaboration on Active Learning) dataset which enables the study of the impact of annotation-cost within a video active learning setting. Annotation-cost refers to the time it takes an annotator to label and quality-assure a given video sequence. A practical motivation for active learning research is to minimize annotation-cost by selectively labeling informative samples that will maximize performance within a given budget constraint. However, previous work in video active learning lacks real-time annotation labels for accurately assessing cost minimization and instead operates under the assumption that annotation-cost scales linearly with the amount of data to annotate. This assumption does not take into account a variety of real-world confounding factors that contribute to a nonlinear cost such as the effect of an assistive labeling tool and the variety of interactions within a scene such as occluded objects, weather, and motion of objects. FOCAL addresses this discrepancy by providing real annotation-cost labels for 126 video sequences across 69 unique city scenes with a variety of weather, lighting, and seasonal conditions. These videos have a wide range of interactions that are at the intersection of infrastructure-assisted autonomy and autonomous vehicle communities. We show through a statistical analysis of the FOCAL dataset that cost is more correlated with a variety of factors beyond just the length of a video sequence. We also introduce a set of conformal active learning algorithms that take advantage of the sequential structure of video data in order to achieve a better trade-off between annotation-cost and performance while also reducing floating point operations (FLOPS) overhead by at least 77.67%. We show how these approaches better reflect how annotations on videos are done in practice through a sequence selection framework. We further demonstrate the advantage of these approaches by introducing two performance-cost metrics and show that the best conformal active learning method is cheaper than the best traditional active learning method by 113 hours.

Visual Abstract

This figure shows the differences between traditional active learning algorithms and the considerations that exist in real-world annotation workflows.

Discrepancy Between Traditional and Real-World Video Active Learning

This diagram shows the differences in cost of the queried sequences by a variety of Active Learning Algorithms.

Difference in Cost of Chosen Sequences by Active Learning Strategy

Data

The data for this work can be found at this zenodo location.

Code Usage

  1. Clone the repository with:
git clone https://github.com/olivesgatech/FOCAL_Dataset.git
  1. Set the python path in the starting directory of the repo with:
export PYTHONPATH=$PYTHONPATH:$PWD
  1. Configure the .toml file according to the desired settings of the project. An example is provided in the provided repository. Important parameters of interest are shown below:
# Important Parameters for Object Detection Network

yolocfg = './Models/detection/yolov5/models/yolov5n.yaml' 
hyp = './Models/detection/yolov5/data/hyps/hyp.scratch-low.yaml'
data = './Models/detection/yolov5/data/focal.yaml'
pretrained = 'yolov5n.pt'

# Important Parameters for Sequential Active Learning Experiments
# n_start is intial number of sequences to begin training
n_start = 2

# n_end is the number of sequences in the final round of training
n_end = 14

# n_query is the number of sequences added after each round
n_query = 1

# strategy is the query strategy used for the experiment
# Implemented strategies include: Entropy, Least Confidence, [GauSS](https://arxiv.org/abs/2302.12018), Margin, Random, and a variety of Conformal Methods
strategy = 'entropy'
init_seed = 0

# Convergence Map and Max Epochs Determine how long each round will train for
convergence_map = 0.15
max_epochs = 9

# Sampling Type sets the active learning to framewise or sequential modes
sampling_type = 'sequence'
  1. Set the desired .toml as the config file for the experiment and perform the active learning experiment with the following command:
python applications/activelearning/run.py --config example.toml

Acknowledgements

This work was done in collaboration with the Ford Motor Company. Code for object detection experiments can be found at the official Yolo v5 repository.