Sharing ideas on how to disseminate your research.
"I am THRILLED to share that our paper is accepted to ..."
Congrats! So what's next? No one is going to browse through the list of thousands of accepted papers. Ain't nobody got time for that.
Check out 🧵below for examples.
*Website*
Use memorable domain names for your project website so that people can easily find/share the link. No university account? That's okay. Register a new name for GitHub pages.
Make it easy for people to remember and refer to your work. As David Patterson said, the vowel is important.
For example, NeRF sounds waaaaaay cooler than NRF..
*Result video*
Make a simple video showing the killer results from your work! Based on my back-of-the-envelope calculation, I would have to present this ECCV paper in the zoom poster session for 18 years straight to reach the same level of visibility.
Do not put your result videos on YouTube and then embed them on your website. Make it super easy for people to download (and share) your results. Help people help share your work.
No image/video size limits so share the highest quality possible.
*Additional results*
Very often you need to work hard constructing the baseline results across multiple datasets. Make them available so that people can easily follow up. For example, many citations of this work are not for specific technical contributions.
*Supplementary website*
Organize all the results across multiple datasets, methods, and ablation on a webpage. This allows EVERYONE (including myself) to interactively explore the results.
Not everyone has the knowledge/resources to set up the environment required for your code on GitHub. Preparing a colab notebook demo (or other platforms) allows everyone to play around with your method.
*arXiv*
Host all your papers on arXiv. It's very frustrating to papers behind the paywall. Posting your paper at a specific time further increases the visibility/readership/impact of your work.
On your publication page, show teaser images/videos so that people can quickly browse through all your work. Work hard to optimize the quality of your teaser! Trust me, it's definitely worth your time.
When sharing on social media or other sites (e.g., Twitter, YouTube, HackerNews, Reddit...), engage with people who comment on your work even tho sometimes you may encounter comments with bad intentions. Over time, they will be your best allies.
* Hyperlinks*
Make sure every page has hyperlinks to every other pages. For examples, add links to authors’ pages, related projects, GitHub/Colab, datasets, YouTube videos additional results. Make it easy to navigate the contents via multiple paths.
*BibTeX*
Everyone knows that bibtex from google scholar is erroneous. Do your readers a favor and help people cite your paper correctly.
*Paper title*
A title should capture what is SPECIAL about the paper. Check out the talk by Jitendra Malik about the paper title.
Whaaaaat?! How is music related to my research? YES, it helps make your video more engaging and fun to watch. If possible, match the transitions with music beats.
I recently found an easy setup to get into my slides. Compared to the standard zoom setup, it's fun, engaging, and allows me to interact with the slide contents directly.
Check out the thread below and set it up for your own presentation!
I mainly follow the excellent video tutorial by @cem_yuksel
but with a poor man's variant (i.e., without a white background or green screen).
Make sure to check out the videos for the best quality!
Step 1: Download Open Broadcaster Software (OBS) studio. obsproject.com
Why: We will use OBS to composite the slides and your camera video feed together and create a "virtual camera".
You can then use the virtual camera for your video conferencing presentation.
NeRF has shown incredible view synthesis results, but it requires multi-view captures for STATIC scenes.
How can we achieve view synthesis for DYNAMIC scenes from a single video? Here is what I learned from several recent efforts.
Instead of presenting Video-NeRF, Nerfie, NR-NeRF, D-NeRF, NeRFlow, NSFF (and many others!) as individual algorithms, here I try to view them from a unifying perspective and understand the pros/cons of various design choices.
Okay, here we go.
*Background*
NeRF represents the scene as a 5D continuous volumetric scene function that maps the spatial position and viewing direction to color and density. It then projects the colors/densities to form an image with volume rendering.
How do we get pseudo labels from unlabeled images?
Unlike classification, directly thresholding the network outputs for dense prediction doesn't work well.
Our idea: start with weakly sup. localization (Grad-CAM) and refine it with self-attention for propagating the scores.
Using two different prediction mechanisms is great bc they make errors in different ways. With our fusion strategy, we get WELL-CALIBRATED pseudo labels (see the expected calibration errors in E below) and IMPROVED accuracy under 1/4, 1/8, 1/16 of labeled examples.
I feel that 10% of my job is to replace \hline with \toprule, \midrule, and \bottomrule. Formatting your table well will help you convey your messages much more clearly.
Have you ever wondered why papers from top universities/research labs often appear in the top few positions in the daily email and web announcements from arXiv?
Why is that the case? Why should I care?
Wait a minute! Does the article position even matter?
The method achieves AWESOME results but requires precise camera poses as inputs.
Isn't SLAM/SfM a SOLVED problem? You might ask.
Yes, it works pretty well for static and controlled environments. For causal videos, existing methods usually fail to register all frames or produce outlier poses with large errors.