- Harris Partaourides17,
- Andreas Voskou17,
- Dimitrios Kosmopoulos18,
- Sotirios Chatzis17 &
- …
- Dimitris N. Metaxas19
Part of the book series:Lecture Notes in Computer Science ((LNIP,volume 12510))
Included in the following conference series:
1989Accesses
Abstract
Memory-efficient continuous Sign Language Translation is a significant challenge for the development of assisted technologies with real-time applicability for the deaf. In this work, we introduce a paradigm of designing recurrent deep networks whereby the output of the recurrent layer is derived from appropriate arguments from nonparametric statistics. A novel variational Bayesian sequence-to-sequence network architecture is proposed that consists of a) a full Gaussian posterior distribution for data-driven memory compression and b) a nonparametric Indian Buffet Process prior for regularization applied on the Gated Recurrent Unit non-gate weights. We dub our approach Stick-Breaking Recurrent network and show that it can achieve a substantial weight compression without diminishing modeling performance.
This is a preview of subscription content,log in via an institution to check access.
Access this chapter
Subscribe and save
- Get 10 units per month
- Download Article/Chapter or eBook
- 1 Unit = 1 Article or 1 Chapter
- Cancel anytime
Buy Now
- Chapter
- JPY 3498
- Price includes VAT (Japan)
- eBook
- JPY 11439
- Price includes VAT (Japan)
- Softcover Book
- JPY 14299
- Price includes VAT (Japan)
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
Abadi, M., et al.: TensorFlow: large-scale machine learning on heterogeneous systems (2015). Softwarehttp://tensorflow.org
Ba, J., Caruana, R.: Do deep nets really need to be deep? In: Advances in Neural Information Processing Systems, pp. 2654–2662 (2014)
Bahdanau, D., Cho, K., Bengio, Y.: Neural machine translation by jointly learning to align and translate. In: 3rd International Conference on Learning Representations, ICLR 2015 (2015)
Camgoz, N.C., Hadfield, S., Koller, O., Bowden, R.: SubUNets: end-to-end hand shape and continuous sign language recognition. In: 2017 IEEE International Conference on Computer Vision (ICCV), pp. 3075–3084. IEEE (2017)
Chatzis, S.: Indian buffet process deep generative models for semi-supervised classification. In: IEEE ICASSP (2018)
Chatzis, S.P., Korkinof, D., Demiris, Y.: A quantum-statistical approach toward robot learning by demonstration. IEEE Trans. Rob.28(6), 1371–1381 (2012)
Cho, K., van Merrienboer, B., Gulcehre, C., Bougares, F., Schwenk, H., Bengio, Y.: Learning phrase representations using RNN encoder-decoder for statistical machine translation. EMNLP (2014)
Cho, K., Van Merriënboer, B., Bahdanau, D., Bengio, Y.: On the properties of neural machine translation: encoder-decoder approaches. arXiv preprintarXiv:1409.1259 (2014)
Chung, J., Gulcehre, C., Cho, K., Bengio, Y.: Empirical evaluation of gated recurrent neural networks on sequence modeling. In: NIPS 2014 Workshop on Deep Learning, December 2014 (2014)
Cihan Camgoz, N., Hadfield, S., Koller, O., Ney, H., Bowden, R.: Neural sign language translation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 7784–7793 (2018)
Dreuw, P., et al.: The signspeak project-bridging the gap between signers and speakers. In: Proceedings of the Seventh International Conference on Language Resources and Evaluation (LREC 2010) (2010)
Gal, Y., Ghahramani, Z.: Dropout as a Bayesian approximation: representing model uncertainty in deep learning.arXiv:1506.02142 (2015)
Graves, A.: Practical variational inference for neural networks. In: Proceedings of the NIPS (2011)
Griffiths, T.L., Ghahramani, Z.: Infinite latent feature models and the Indian buffet process. In: Proceedings of the NIPS (2005)
Guo, D., Zhou, W., Li, H., Wang, M.: Hierarchical LSTM for sign language translation. In: Thirty-Second AAAI Conference on Artificial Intelligence (2018)
Hinton, G., Vinyals, O., Dean, J.: Distilling the knowledge in a neural network. In: NIPS Deep Learning and Representation Learning Workshop (2015)
Hochreiter, S., Schmidhuber, J.: Long short-term memory. Neural Comput.9(8), 1735–1780 (1997)
Ishwaran, H., James, L.F.: Gibbs sampling methods for stick-breaking priors. J. Am. Stat. Assoc.96, 161–173 (2001)
Jang, E., Gu, S., Poole, B.: Categorical reparameterization using Gumbel-Softmax. In: Proceedings of the ICLR (2017)
Kingma, D.P., Ba, J.: Adam: a method for stochastic optimization. arXiv preprintarXiv:1412.6980 (2014)
Kumaraswamy, P.: A generalized probability density function for double-bounded random processes. J. Hydrol.46(1–2), 79–88 (1980)
Louizos, C., Ullrich, K., Welling, M.: Bayesian compression for deep learning. In: Proceedings of the NIPS (2017)
Luong, M.T., Brevdo, E., Zhao, R.: Neural machine translation (seq2seq) tutorial (2017)
Maddison, C., Mnih, A., Teh, Y.: The concrete distribution: a continuous relaxation of discrete random variables. In: International Conference on Learning Representations (2017)
Murray, K.: Lack of British sign language interpreters putting deaf people at risk. Guardian8 (2013)
Nalisnick, E., Smyth, P.: Stick-breaking variational autoencoders. In: Proceedings of the ICLR (2016)
Neklyudov, K., Molchanov, D., Ashukha, A., Vetrov, D.P.: Structured Bayesian pruning via log-normal multiplicative noise. In: Advances in Neural Information Processing Systems, pp. 6775–6784 (2017)
Panousis, K., Chatzis, S., Theodoridis, S.: Nonparametric Bayesian deep networks with local competition. In: International Conference on Machine Learning, pp. 4980–4988 (2019)
Partaourides, H., Chatzis, S.P.: Asymmetric deep generative models. Neurocomputing241, 90–96 (2017)
Partaourides, H., Chatzis, S.P.: Deep network regularization via Bayesian inference of synaptic connectivity. In: Kim, J., Shim, K., Cao, L., Lee, J.-G., Lin, X., Moon, Y.-S. (eds.) PAKDD 2017. LNCS (LNAI), vol. 10234, pp. 30–41. Springer, Cham (2017).https://doi.org/10.1007/978-3-319-57454-7_3
Sutskever, I., Vinyals, O., Le, Q.V.: Sequence to sequence learning with neural networks. In: Advances in Neural Information Processing Systems, pp. 3104–3112 (2014)
Teh, Y.W., Görür, D., Ghahramani, Z.: Stick-breaking construction for the Indian buffet process. In: Proceedings of the AISTATS (2007)
Theodoridis, S.: Machine Learning: A Bayesian and Optimization Perspective. Academic Press, Cambridge (2015)
Vaswani, A., et al.: Attention is all you need. In: Advances in Neural Information Processing Systems, pp. 5998–6008 (2017)
Acknowledgments
This research was partially supported by the Research Promotion Foundation of Cyprus, through the grant: INTERNATIONAL/USA/0118/0037.
Author information
Authors and Affiliations
Cyprus University of Technology, Limassol, Cyprus
Harris Partaourides, Andreas Voskou & Sotirios Chatzis
University of Patras, Patras, Greece
Dimitrios Kosmopoulos
Rutgers University, New Brunswick, NJ, USA
Dimitris N. Metaxas
- Harris Partaourides
You can also search for this author inPubMed Google Scholar
- Andreas Voskou
You can also search for this author inPubMed Google Scholar
- Dimitrios Kosmopoulos
You can also search for this author inPubMed Google Scholar
- Sotirios Chatzis
You can also search for this author inPubMed Google Scholar
- Dimitris N. Metaxas
You can also search for this author inPubMed Google Scholar
Corresponding author
Correspondence toHarris Partaourides.
Editor information
Editors and Affiliations
University of Nevada Reno, Reno, NV, USA
George Bebis
Stony Brook University, Stony Brook, NY, USA
Zhaozheng Yin
Drexel University, Philadelphia, PA, USA
Edward Kim
RWTH Aachen University, Aachen, Germany
Jan Bender
University of Edinburgh, Edinburgh, UK
Kartic Subr
IBM Research – Cambridge, Cambridge, MA, USA
Bum Chul Kwon
University of Waterloo, Waterloo, ON, Canada
Jian Zhao
Graz University of Technology, Graz, Austria
Denis Kalkofen
The Hong Kong Polytechnic University, Hong Kong, Hong Kong
George Baciu
Rights and permissions
Copyright information
© 2020 Springer Nature Switzerland AG
About this paper
Cite this paper
Partaourides, H., Voskou, A., Kosmopoulos, D., Chatzis, S., N. Metaxas, D. (2020). Variational Bayesian Sequence-to-Sequence Networks for Memory-Efficient Sign Language Translation. In: Bebis, G.,et al. Advances in Visual Computing. ISVC 2020. Lecture Notes in Computer Science(), vol 12510. Springer, Cham. https://doi.org/10.1007/978-3-030-64559-5_19
Download citation
Published:
Publisher Name:Springer, Cham
Print ISBN:978-3-030-64558-8
Online ISBN:978-3-030-64559-5
eBook Packages:Computer ScienceComputer Science (R0)
Share this paper
Anyone you share the following link with will be able to read this content:
Sorry, a shareable link is not currently available for this article.
Provided by the Springer Nature SharedIt content-sharing initiative