AutoFlow: Learning a Better Training Set for Optical Flow
CVPR 2021

Disentangling Architecture and Training for Optical Flow
ECCV 2022

Self-supervised AutoFlow
CVPR 2023


 

Self-supervised AutoFlow

Hsin-Ping Huang Charles Herrmann Junhwa Hur Erika Lu
Kyle Sargent Austin Stone Ming-Hsuan Yang Deqing Sun
Google Research
| Paper | Code |

Self-supervised AutoFlow learns to generate an optical flow training set through self-supervision on the target domain. It performs comparable to supervised AutoFlow on Sintel and KITTI without requiring ground truth (GT) and learns a better dataset for real-world DAVIS, where GT is not available. We report optical flow accuracy on Sintel and KITTI, and keypoint propagation accuracy on DAVIS.

Abstract

Recently, AutoFlow has shown promising results on learning a training set for optical flow, but requires ground truth labels in the target domain to compute its search metric. Observing a strong correlation between the ground truth search metric and self-supervised losses, we introduce self-supervised AutoFlow to handle real-world videos without ground truth labels. Using self-supervised loss as the search metric, our self-supervised AutoFlow performs on par with AutoFlow on Sintel and KITTI where ground truth is available, and performs better on the real-world DAVIS dataset. We further explore using self-supervised AutoFlow in the (semi-)supervised setting and obtain competitive results against the state of the art.

 

Papers

 

"Self-supervised AutoFlow"
Hsin-Ping Huang, Charles Herrmann, Junhwa Hur, Erika Lu, Kyle Sargent, Austin Stone, Ming-Hsuan Yang, and Deqing Sun
CVPR 2023.
[Arxiv][CVF]

Code

    Link to code for "Self-supervised AutoFlow".

Bibtex

@inproceedings{huang2023self,
  title={Self-supervised AutoFlow},
  author={Huang, Hsin-Ping and Herrmann, Charles and Hur, Junhwa and Lu, Erika
	and Sargent, Kyle and Stone, Austin and Yang, Ming-Hsuan and Sun, Deqing}, 
  booktitle={CVPR},
  year={2023}
}
[Back To Top]

 

Disentangling Architecture and Training for Optical Flow

Deqing Sun Charles Herrmann Fitsum Reda
Michael Rubinstein David Fleet William T. Freeman
Google Research
| Paper | Code |

Left: Large improvements with newly trained PWC-Net, IRR-PWC and RAFT. (left: originally published results in blue; results of our newly trained models in red). The newly trained RAFT is more accurate than all published methods on KITTI 2015 at the time of writing. Right: Visual comparison on a Davis sequence: between the original [43] and our newly trained PWC-Net and RAFT, shows improved flow details, e.g. the hole between the cart and the person at the back. The newly trained PWC-Net recovers the hole between the cart and the front person better than RAFT.

Abstract

How important are training details and datasets to recent optical flow models like RAFT? And do they generalize? To explore these questions, rather than develop a new model, we revisit three prominent models, PWC-Net, IRR-PWC and RAFT, with a common set of modern training techniques and datasets, and observe significant performance gains, demonstrating the importance and generality of these training details. Our newly trained PWC-Net and IRR-PWC models show surprisingly large improvements, up to 30% versus original published results on Sintel and KITTI 2015 benchmarks. They outperform the more recent Flow1D on KITTI 2015 while being 3× faster during inference. Our newly trained RAFT achieves an Fl-all score of 4.31% on KITTI 2015, more accurate than all published optical flow methods at the time of writing. Our results demonstrate the benefits of separating the contributions of models, training techniques and datasets when analyzing performance gains of optical flow methods.

 

Papers

 

"Disentangling Architecture and Training for Optical Flow"
Deqing SunT,*, Charles Herrmann*, Fitsum Reda, Michael Rubinstein, David J. Fleet, and William T. Freeman
ECCV 2022. T project lead, * equal technical contribution
[Arxiv][CVF]

Code

    Link to code for "Disentangling Architecture and Training for Optical Flow".

Bibtex

@inproceedings{sun2022disentangling,
  title={Disentangling Architecture and Training for Optical Flow},
  author={Sun, Deqing and Herrmann, Charles and Reda, Fitsum and Rubinstein, Michael
   and Fleet, David J. and Freeman, William T}, 
    booktitle={ECCV},
  year={2022}
}
[Back To Top]

 

AutoFlow: Learning a Better Training Set for Optical Flow

Deqing Sun Daniel Vlasic Charles Herrmann Varun Jampani Michael Krainin Huiwen Chang
Ramin Zabih William T. Freeman Ce Liu
Google Research
| Paper | Samples | Code (available now!) | Dataset (available now!) |

Left: Pipelines for optical flow. A typical pipeline pre-trains models on static datasets,e.g., FlyingChairs, and then evaluates the performance on a target dataset,e.g., Sintel. AutoFlow learns pre-training data which is optimized ona target dataset. Right: Accuracy w.r.t. number of pre-training examples on Sintel.final. Four AutoFlow pre-training examples with augmentation achieve lower errors than 22,872 FlyingChairs pre-training examples with augmentation. The gap between PWC-Net and RAFT becomes small when pre-trained on enough AutoFlow examples.

Abstract

Synthetic datasets play a critical role in pre-training CNN models for optical flow, but they are painstaking to generate and hard to adapt to new applications. To automate the process, we present AutoFlow, a simple and effective method to render training data for optical flow that optimizes the performance of a model on a target dataset. AutoFlow takes a layered approach to render synthetic data, where the motion, shape, and appearance of each layer are controlled by learnable hyperparameters. Experimental results show that AutoFlow achieves state-of-the-art accuracy in pre-training both PWC-Net and RAFT.

 

Papers

 

"AutoFlow: Learning a Better Training Set for Optical Flow"
Deqing Sun, Daniel Vlasic, Charles Herrmann, Varun Jampani, Michael Krainin, Huiwen Chang, Ramin Zabih, William T. Freeman, and Ce Liu
Oral presentation, CVPR 2021.
[arXiv][CVF]

Samples

Code

    Link to code for "AutoFlow: Learning a Better Training Set for Optical Flow".

Dataset

    Static dataset with 40,000 training examples [ part 1, part 2, part 3, part 4, ~96G in total]. License: CC-BY

Bibtex

@inproceedings{sun2021autoflow,
  title={AutoFlow: Learning a Better Training Set for Optical Flow},
  author={Sun, Deqing and Vlasic, Daniel and Herrmann, Charles and Jampani, Varun and Krainin, Michael
   and Chang, Huiwen and Zabih, Ramin and Freeman, William T and Liu, Ce}, 
    booktitle={CVPR},
  year={2021}
}
[Back To Top]