Scale-Net: Learning to Reduce Scale Differences for Large-Scale Invariant Image Matching
tl;dr: same idea as with other ScaleNet: estimate scale difference in 2 images-> resize them. Eval on their version of Scale IMC 1/2 (check 2nd)
Yujie Fu, Yihong Wu arxiv.org/abs/2112.10485
In order to prove priority and originality of the idea w.r.t. @axelbarrosotw ScaleNet, authors included #ICCV2021 submission screenshot as a proof.
Weird, however that current paper has less authors. 2/2
After 16 hours working on a video for presentation, it crossed my mind that it would be a great idea before going to bed, to upgrade my desktop from Ubuntu 16.04 to 20.04 (yes, at 2:30 am).
What could possibly go wrong?
To summarize:
1st problem were some binaries from zoom installed.
Then it clashed with libgl -mesa-something. Nvidia drivers got involved
What helped, is running apt -o Dpkg::Options::="--force-overwrite"
2/
3)Oh, and everything with very weird "zoomed-in" thing and no windows manager working.
After update to 20.04 and cuda/nvidia driver installation, windows were up. But huuuge
3/
The situation is following, in thread: 1) relatively small group gets the ideas of AE+ViT=SiT, validates on CIFAR, STL,puts on arXiv in April. Also submits to PAMI
1/
3) Masked AE + ViT from Meta AI, comes out in November, gets nice ImageNet results and get all attention. Not citing SiT
So arXiv does not protect SiT from scooping.
Oh, and PAMI rejects SiT
2/
However, arXiv + social media allow us to at least have this conversation and maybe fix a credit assignment.
Without arXiv as a proof of work, and social media to point out the situation, complete scooping.
3/
Thanks to @david_picard , I can now formulate why I am for social media.
That is an instrument, where many things depends on you and how good you have done your job.
Writing a good paper is another such thing, but then you are dependent on (random) reviewer choice.
1/
ArXiv is better than the conference in a sense that everyone can see and judge. But it also quite random: your target audience may not check the feed today. However, it is googlable, so not full failure
2/
Now, if we add social media, then we can almost make sure (in long term at least) that target audience will see the paper.
Maybe they will not like it, but they will see it.
3/3
Let me introduce pixelstitch: simple correspondence annotator based on @matplotlib + jupyter notebook.
Just propvide list of image_fnames and
import CorrespondenceAnnotator.
You can add and erase points, zoom, move, and visualize epipolar geometry, induced by correspondences. 1/
You can install it by
pip install pixelstitch
and the documentation is here: ducha-aiki.github.io/pixelstitch/
Ofc, powered by nbdev.
All suggestions are welcomed
2/3
Why did I do this? Well, the built-in labelling tools in SfM apps like RealityCapture are nice, but one cannot quickly go through unrelated image pairs.
New artisan datasets for WxBS from me are coming soon,
specifically WxBS-relabeled and EVD 2.0.
3/3
For example, you can detect ORB features is OpenCV and convert them to kornia lafs with
lafs = laf_from_opencv_ORB_kpts(kps)
Or you can match descriptors on GPU with kornia.feature.match_snn() and transfer to OpenCV with cv2_matches_from_kornia 2/3
The repo is not @kornia_foss "official" in the terms that we don't promise to maintain everything in a timely manner. However, it should work for most of the times :) 3/3