Spaces:
Running
Running
File size: 6,807 Bytes
10dcc2e |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 |
# COTR: Correspondence Transformer for Matching Across Images (ICCV 2021)
This repository is a reference implementation for COTR.
COTR establishes correspondence in a functional and end-to-end fashion. It solves dense and sparse correspondence problem in the same framework.
[[arXiv]](https://arxiv.org/abs/2103.14167), [[video]](https://jiangwei221.github.io/vids/cotr/README.html), [[presentation]](https://youtu.be/bOZ12kgfn3E), [[pretrained_weights]](https://www.cs.ubc.ca/research/kmyi_data/files/2021/cotr/default.zip), [[distance_matrix]](https://www.cs.ubc.ca/research/kmyi_data/files/2021/cotr/MegaDepth_v1.zip)
## Training
### 1. Prepare data
See `prepare_data.md`.
### 2. Setup configuration json
Add an entry inside `COTR/global_configs/dataset_config.json`, make sure it is correct on your system. In the provided `dataset_config.json`, we have different configurations for different clusters.
Explanations on some json parameters:
`valid_list_json`: The valid list json file, see `2. Valid list` in `Scripts to generate dataset`.
`train_json/val_json/test_json`: The splits json files, see `3. Train/val/test split` in `Scripts to generate dataset`.
`scene_dir`: Path to Megadepth SfM folder(rectified ones!). `{0}{1}` are scene and sequence id used by f-string.
`image_dir/depth_dir`: Path to images and depth maps of Megadepth.
### 3. Example command
```python train_cotr.py --scene_file sample_data/jsons/debug_megadepth.json --dataset_name=megadepth --info_level=rgbd --use_ram=no --batch_size=2 --lr_backbone=1e-4 --max_iter=200 --valid_iter=10 --workers=4 --confirm=no```
**Important arguments:**
`use_ram`: Set to "yes" to load data into main memory.
`crop_cam`: How to crop the image, it will change the camera intrinsic accordingly.
`scene_file`: The sequence control file.
`suffix`: Give the model a unique suffix.
`load_weights`: Load a pretrained weights, only need the model name, it will automatically find the folder with the same name under the output folder, and load the "checkpoint.pth.tar".
### 4. Our training commands
As stated in the paper, we have 3 training stages. The machine we used has 1 RTX 3090, i7-10700, and 128G RAM. We store the training data inside the main memory during the first two stages.
Stage 1: `python train_cotr.py --scene_file sample_data/jsons/200_megadepth.json --info_level=rgbd --use_ram=yes --use_cc=no --batch_size=24 --learning_rate=1e-4 --lr_backbone=0 --max_iter=300000 --workers=8 --cycle_consis=yes --bidirectional=yes --position_embedding=lin_sine --layer=layer3 --confirm=no --dataset_name=megadepth_sushi --suffix=stage_1 --valid_iter=1000 --enable_zoom=no --crop_cam=crop_center_and_resize --out_dir=./out/cotr`
Stage 2: `python train_cotr.py --scene_file sample_data/jsons/200_megadepth.json --info_level=rgbd --use_ram=yes --use_cc=no --batch_size=16 --learning_rate=1e-4 --lr_backbone=1e-5 --max_iter=2000000 --workers=8 --cycle_consis=yes --bidirectional=yes --position_embedding=lin_sine --layer=layer3 --confirm=no --dataset_name=megadepth_sushi --suffix=stage_2 --valid_iter=10000 --enable_zoom=no --crop_cam=crop_center_and_resize --out_dir=./out/cotr --load_weights=model:cotr_resnet50_layer3_1024_dset:megadepth_sushi_bs:24_pe:lin_sine_lrbackbone:0.0_suffix:stage_1`
Stage 3: `python train_cotr.py --scene_file sample_data/jsons/200_megadepth.json --info_level=rgbd --use_ram=no --use_cc=no --batch_size=16 --learning_rate=1e-4 --lr_backbone=1e-5 --max_iter=300000 --workers=8 --cycle_consis=yes --bidirectional=yes --position_embedding=lin_sine --layer=layer3 --confirm=no --dataset_name=megadepth_sushi --suffix=stage_3 --valid_iter=2000 --enable_zoom=yes --crop_cam=no_crop --out_dir=./out/cotr --load_weights=model:cotr_resnet50_layer3_1024_dset:megadepth_sushi_bs:16_pe:lin_sine_lrbackbone:1e-05_suffix:stage_2`
<p align="center">
<img src="./sample_data/imgs/loss_curves.png" height="200">
</p>
## Demos
Check out our demo video at [here](https://jiangwei221.github.io/vids/cotr/README.html).
### 1. Install environment
Our implementation is based on PyTorch. Install the conda environment by: `conda env create -f environment.yml`.
Activate the environment by: `conda activate cotr_env`.
### 2. Download the pretrained weights
Download the pretrained weights at [here](https://www.cs.ubc.ca/research/kmyi_data/files/2021/cotr/default.zip). Extract in to `./out`, such that the weights file is at `/out/default/checkpoint.pth.tar`.
### 3. Single image pair demo
```python demo_single_pair.py --load_weights="default"```
Example sparse output:
<p align="center">
<img src="./sample_data/imgs/sparse_output.png" height="400">
</p>
Example dense output with triangulation:
<p align="center">
<img src="./sample_data/imgs/dense_output.png" height="200">
</p>
**Note:** This example uses 10K valid sparse correspondences to densify.
### 4. Facial landmarks demo
`python demo_face.py --load_weights="default"`
Example:
<p align="center">
<img src="./sample_data/imgs/face_output.png" height="200">
</p>
### 5. Homography demo
`python demo_homography.py --load_weights="default"`
<p align="center">
<img src="./sample_data/imgs/paint_output.png" height="300">
</p>
### 6. Guided matching demo
`python demo_guided_matching.py --load_weights="default"`
<p align="center">
<img src="./sample_data/imgs/guided_matching_output.png" height="400">
</p>
### 7. Two view reconstruction demo
Note: this demo uses both known camera intrinsic and extrinsic.
`python demo_reconstruction.py --load_weights="default" --max_corrs=2048 --faster_infer=yes`
<p align="center">
<img src="./sample_data/imgs/recon_output.png" height="250">
</p>
### 8. Annotation suggestions
If the annotator knows the scale difference of two buildings, then COTR can skip the scale estimation step.
`python demo_wbs.py --load_weights="default"`
<p align="center">
<img src="./sample_data/imgs/annotation_output.png" height="250">
</p>
## Faster Inference
We added a faster inference engine.
The idea is that for each network invocation, we want to solve more queries. We search for nearby queries and group them on the fly.
*Note: Faster inference engine has slightly worse spatial accuracy.*
Guided matching demo now supports faster inference.
The time consumption for default inference engine is ~216s, and the time consumption for faster inference engine is ~79s, on 1080Ti.
Try `python demo_guided_matching.py --load_weights="default" --faster_infer=yes`.
## Citation
If you use this code in your research, please cite our paper:
```
@inproceedings{jiang2021cotr,
title={{COTR: Correspondence Transformer for Matching Across Images}},
author={Wei Jiang and Eduard Trulls and Jan Hosang and Andrea Tagliasacchi and Kwang Moo Yi},
booktitle=ICCV,
year={2021}
}
```
|