Syntactically Guided Generative Embeddings for Zero-Shot Skeleton Action Recognition
Abstract
We introduce SynSE, a novel syntactically guided generative approach for Zero-Shot Learning (ZSL). Our end-to-end approach learns progressively refined generative embedding spaces constrained within and across the involved modalities (visual, language). The inter-modal constraints are defined between action sequence embedding and embeddings of Parts of Speech (PoS) tagged words in the corresponding action description. We deploy SynSE for the task of skeleton-based action sequence recognition. It has been accepted for publication at the 2021 IEEE ICIP.
Architecture

Citation
@misc{gupta2021syntactically,
title={Syntactically Guided Generative Embeddings for Zero-Shot Skeleton Action Recognition},
author={Pranay Gupta and Divyanshu Sharma and Ravi Kiran Sarvadevabhatla},
year={2021},
eprint={2101.11530},
archivePrefix={arXiv},
primaryClass={cs.CV}
}