Vincentqyw
update: features and matchers
a80d6bb
# Rethinking Low-level Features for Interest Point Detection and Description
## Dependency
- pytorch
- torchvision
- cv2
- tqdm
We use cuda 11.4/python 3.8.13/torch 1.10.0/torchvision 0.11.0/opencv 3.4.8 for training and testing.
## Pre-trained models
We provide two versions of LANet with different structure in [network_v0](network_v0) and [network_v1](network_v1), the corresponding pre-trained models are in [checkpoints](checkpoints).
- v0: The original version used in our paper.
- v1: An improved version that has a better over all performance.
## Training
Download the COCO dataset:
```
cd datasets/COCO/
wget http://images.cocodataset.org/zips/train2017.zip
unzip train2017.zip
```
Prepare the training file:
```
python datasets/prepare_coco.py --raw_dir datasets/COCO/train2017/ --saved_dir datasets/COCO/
```
To train the model (v0) on COCO dataset, run:
```
python main.py --train_root datasets/COCO/train2017/ --train_txt datasets/COCO/train2017.txt
```
## Evaluation
### Evaluation on HPatches dataset
Download the HPatches dataset:
```
cd datasets/HPatches/
wget http://icvl.ee.ic.ac.uk/vbalnt/hpatches/hpatches-sequences-release.tar.gz
tar -xvf hpatches-sequences-release.tar.gz
```
To evaluate the pre-trained model, run:
```
python test.py --test_dir ./datasets/HPatches/hpatches-sequences-release
```
## License
The code is released under the [MIT license](LICENSE).
## Citation
Please use the following citation when referencing our work:
```
@InProceedings{Wang_2022_ACCV,
author = {Changhao Wang and Guanwen Zhang and Zhengyun Cheng and Wei Zhou},
title = {Rethinking Low-level Features for Interest Point Detection and Description},
booktitle = {Computer Vision - {ACCV} 2022 - 16th Asian Conference on Computer
Vision, Macao, China, December 4-8, 2022, Proceedings, Part {II}},
series = {Lecture Notes in Computer Science},
volume = {13842},
pages = {108--123},
year = {2022}
}
```
## Related Projects
https://github.com/TRI-ML/KP2D