Sharing ideas on how to disseminate your research.
"I am THRILLED to share that our paper is accepted to ..."
Congrats! So what's next? No one is going to browse through the list of thousands of accepted papers. Ain't nobody got time for that.
Check out 🧵below for examples.
*Website*
Use memorable domain names for your project website so that people can easily find/share the link. No university account? That's okay. Register a new name for GitHub pages.
I recently found an easy setup to get into my slides. Compared to the standard zoom setup, it's fun, engaging, and allows me to interact with the slide contents directly.
Check out the thread below and set it up for your own presentation!
I mainly follow the excellent video tutorial by @cem_yuksel
but with a poor man's variant (i.e., without a white background or green screen).
Make sure to check out the videos for the best quality!
Step 1: Download Open Broadcaster Software (OBS) studio. obsproject.com
Why: We will use OBS to composite the slides and your camera video feed together and create a "virtual camera".
You can then use the virtual camera for your video conferencing presentation.
NeRF has shown incredible view synthesis results, but it requires multi-view captures for STATIC scenes.
How can we achieve view synthesis for DYNAMIC scenes from a single video? Here is what I learned from several recent efforts.
Instead of presenting Video-NeRF, Nerfie, NR-NeRF, D-NeRF, NeRFlow, NSFF (and many others!) as individual algorithms, here I try to view them from a unifying perspective and understand the pros/cons of various design choices.
Okay, here we go.
*Background*
NeRF represents the scene as a 5D continuous volumetric scene function that maps the spatial position and viewing direction to color and density. It then projects the colors/densities to form an image with volume rendering.
How do we get pseudo labels from unlabeled images?
Unlike classification, directly thresholding the network outputs for dense prediction doesn't work well.
Our idea: start with weakly sup. localization (Grad-CAM) and refine it with self-attention for propagating the scores.
Using two different prediction mechanisms is great bc they make errors in different ways. With our fusion strategy, we get WELL-CALIBRATED pseudo labels (see the expected calibration errors in E below) and IMPROVED accuracy under 1/4, 1/8, 1/16 of labeled examples.
I feel that 10% of my job is to replace \hline with \toprule, \midrule, and \bottomrule. Formatting your table well will help you convey your messages much more clearly.
Have you ever wondered why papers from top universities/research labs often appear in the top few positions in the daily email and web announcements from arXiv?
Why is that the case? Why should I care?
Wait a minute! Does the article position even matter?