π’ Introducing SynSE, a language-guided approach for generalized zero shot learning of pose-based action representations! Great effort by @bublaasaur and @divyanshu1709#actionrecognition
For enabling compositional generalization to novel action-object combinations, the action description is transformed into individual Part-of-Speech based embeddings.
The PoS-based embeddings are aligned with action sequence embedding via a VAE-based generative space. This alignment is optimized using within and cross modality constraints.
The default ZSL paradigm is biased towards seen classes. We use the elegant gating approach by Atzmon&co.
for Generalized ZSL. Essentially, we learn a binary classifier which distinguishes between seen and unseen class samples.
SynSE obtains state of the art ZSL and GZSL performance on the large-scale NTU-RGBD skeleton action dataset.
π£ JPoSE (mwray.github.io/FGAR/), CADA-VAE (github.com/edgarschnfld/Cβ¦) which inspired our work. JPoSE: alignment of per-PoS language embedding with visual counterpart but in non-generative setting. CADA-VAE: visuo-lingual alignment in VAE-based setting, but no PoS-awareness.
Presenting BoundaryNet - a resizing-free approach for high-precision weakly supervised document layout parsing. BoundaryNet will be an ORAL presentation (Oral Session 3) today at @icdar2021 . Project page: ihdia.iiit.ac.in/BoundaryNet/ . Details π
Precise boundary annotations can be crucial for downstream applications which rely on region-class semantics. Some document collections contain irregular and overlapping region instances. Fully automatic approaches require resizing and often produce suboptimal parsing results.
Our semi-automatic approach takes region bounding box as input and predicts boundary polygon as output. Importantly, BoundaryNet can handle variable sized images without any need for resizing.
π’ In our #ACMMM21 paper, we highlight issues with training and evaluation of π°πΏπΌππ± π°πΌππ»ππΆπ»π΄ deep networks. π§΅π
For far too long, π°πΏπΌππ± π°πΌππ»ππΆπ»π΄ works in #CVPR, #AAAI, #ICCV, #NeurIPS have reported only MAE, but not standard deviation.
Looking at MAE and standard deviation from MAE, a very grim picture emerges. E.g. Imagine a SOTA net with MAE 71.7 but deviation is a whopping 376.4 !