Open Access   Article Go Back

Life Logging Using Egocentric Perception

A. Anandakrishnan1 , A. Walia2 , A. Jha3 , J. J. Pandya4 , C. V. Raj5

Section:Research Paper, Product Type: Journal Paper
Volume-07 , Issue-14 , Page no. 335-338, May-2019

CrossRef-DOI:   https://doi.org/10.26438/ijcse/v7si14.335338

Online published on May 15, 2019

Copyright © A. Anandakrishnan, A. Walia, A. Jha, J. J. Pandya, C. V. Raj . This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

View this paper at   Google Scholar | DPI Digital Library

How to Cite this Paper

  • IEEE Citation
  • MLA Citation
  • APA Citation
  • BibTex Citation
  • RIS Citation

IEEE Style Citation: A. Anandakrishnan, A. Walia, A. Jha, J. J. Pandya, C. V. Raj, “Life Logging Using Egocentric Perception,” International Journal of Computer Sciences and Engineering, Vol.07, Issue.14, pp.335-338, 2019.

MLA Style Citation: A. Anandakrishnan, A. Walia, A. Jha, J. J. Pandya, C. V. Raj "Life Logging Using Egocentric Perception." International Journal of Computer Sciences and Engineering 07.14 (2019): 335-338.

APA Style Citation: A. Anandakrishnan, A. Walia, A. Jha, J. J. Pandya, C. V. Raj, (2019). Life Logging Using Egocentric Perception. International Journal of Computer Sciences and Engineering, 07(14), 335-338.

BibTex Style Citation:
@article{Anandakrishnan_2019,
author = {A. Anandakrishnan, A. Walia, A. Jha, J. J. Pandya, C. V. Raj},
title = {Life Logging Using Egocentric Perception},
journal = {International Journal of Computer Sciences and Engineering},
issue_date = {5 2019},
volume = {07},
Issue = {14},
month = {5},
year = {2019},
issn = {2347-2693},
pages = {335-338},
url = {https://www.ijcseonline.org/full_spl_paper_view.php?paper_id=1148},
doi = {https://doi.org/10.26438/ijcse/v7i14.335338}
publisher = {IJCSE, Indore, INDIA},
}

RIS Style Citation:
TY - JOUR
DO = {https://doi.org/10.26438/ijcse/v7i14.335338}
UR - https://www.ijcseonline.org/full_spl_paper_view.php?paper_id=1148
TI - Life Logging Using Egocentric Perception
T2 - International Journal of Computer Sciences and Engineering
AU - A. Anandakrishnan, A. Walia, A. Jha, J. J. Pandya, C. V. Raj
PY - 2019
DA - 2019/05/15
PB - IJCSE, Indore, INDIA
SP - 335-338
IS - 14
VL - 07
SN - 2347-2693
ER -

           

Abstract

This paper aims to provide a solution that uses a multimodal approach to analyse large intake of audio and video data and use it to understand the emotions of a subject and to describe the current surroundings to the subject in question. The model is trained on the egocentric data, which contains audio and video signals. The model contains emotion recognition and a speech recognition which extract features of their own allowing to perform a classification on the emotions. The large inflow of data from up and coming technologies like Google Lens and onset of Internet of things are key application points for this solution.

Key-Words / Index Term

Emotion Recognition, Face Extraction, Speech Recognition,Scene Description,Life Logging

References

[1] A. Gutierrez and Z. Robert. Lip Reading Word Classification
[2] Ashish B. Ingale,.D.S.Chaudhari "Speech Emotion Recognition" in International Journal of Soft Computing and Engineering (IJSCE)
ISSN: 2231-2307, Volume-2, Issue-1, March 2012
[3] S. Tamura, H. Ninomiya, N. Kitaoka, S. Osuga, Y. Iribe, K. Takeda, and S. Hayamizu, “Audio-visual speech recognition using deep bottleneck features and high-performance lipreading,” in Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA), 2015, pp. 575–582
[4] Z. Ma, H. Yu, W. Chen, and J. Guo, “Short utterance based speech language identification in intelligent vehicles with time-scale modifications and deep bottleneck features,” IEEE Trans. Veh. Technol., vol. 68, no. 1, pp. 1–13, 2019.
[5] M. Xu, X. Deng, S. Li and Z. Wang, "Region-of-Interest Based Conversational HEVC Coding with Hierarchical Perception Model of Face," in IEEE Journal of Selected Topics in Signal Processing, vol. 8, no. 3, pp. 475-489, June 2014.
[6] H. Palangi, L. Deng, Y. Shen, J. Gao, X. He, J. Chen, X. Song, and R. Ward. Deep sentence embedding using long short-term memory networks: Analysis and application to information retrieval. IEEE/ACM Transactions on Audio, Speech, and Language Processing, 24(4):694–707, 2016.
[7] Wang X, Jiang W, Luo Z. Combination of convolutional and recurrent neural network for sentiment analysis of short texts. In Proceedings of the InternationalConference on Computational Linguistics (COLING 2016),2016
[8] P. Tzirakis, G. Trigeorgis, M. A. Nicolaou, B. W. Schuller and S. Zafeiriou, "End-to-End Multimodal Emotion Recognition Using Deep Neural Networks," in IEEE Journal of Selected Topics in Signal Processing, vol. 11, no. 8, pp. 1301-1309, Dec. 2017.