Open Access   Article Go Back

Asynchronous Multimodal Text Entry Using Speech and Gesture Keyboards

V.Maniraj 1 , E.Manohari 2

Section:Research Paper, Product Type: Journal Paper
Volume-3 , Issue-5 , Page no. 371-376, May-2015

Online published on May 30, 2015

Copyright © V.Maniraj , E.Manohari . This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

View this paper at   Google Scholar | DPI Digital Library

How to Cite this Paper

  • IEEE Citation
  • MLA Citation
  • APA Citation
  • BibTex Citation
  • RIS Citation

IEEE Style Citation: V.Maniraj , E.Manohari, “Asynchronous Multimodal Text Entry Using Speech and Gesture Keyboards,” International Journal of Computer Sciences and Engineering, Vol.3, Issue.5, pp.371-376, 2015.

MLA Style Citation: V.Maniraj , E.Manohari "Asynchronous Multimodal Text Entry Using Speech and Gesture Keyboards." International Journal of Computer Sciences and Engineering 3.5 (2015): 371-376.

APA Style Citation: V.Maniraj , E.Manohari, (2015). Asynchronous Multimodal Text Entry Using Speech and Gesture Keyboards. International Journal of Computer Sciences and Engineering, 3(5), 371-376.

BibTex Style Citation:
@article{_2015,
author = { V.Maniraj , E.Manohari},
title = {Asynchronous Multimodal Text Entry Using Speech and Gesture Keyboards},
journal = {International Journal of Computer Sciences and Engineering},
issue_date = {5 2015},
volume = {3},
Issue = {5},
month = {5},
year = {2015},
issn = {2347-2693},
pages = {371-376},
url = {https://www.ijcseonline.org/full_paper_view.php?paper_id=535},
publisher = {IJCSE, Indore, INDIA},
}

RIS Style Citation:
TY - JOUR
UR - https://www.ijcseonline.org/full_paper_view.php?paper_id=535
TI - Asynchronous Multimodal Text Entry Using Speech and Gesture Keyboards
T2 - International Journal of Computer Sciences and Engineering
AU - V.Maniraj , E.Manohari
PY - 2015
DA - 2015/05/30
PB - IJCSE, Indore, INDIA
SP - 371-376
IS - 5
VL - 3
SN - 2347-2693
ER -

VIEWS PDF XML
2506 2317 downloads 2370 downloads
  
  
           

Abstract

We propose plummeting errors in text entry by uniting speech and gesture keyboard input. We describe a combine model that combines gratitude conorders in an asynchronous and flexible manner. We composed speech and gesture data of manipulators entering both short email sentences and web search queries. By amalgamation gratitude conorders from both modalities, word error rate was abridged by 53% comparative for email sentences and 29% comparative for web searches. For email exclamations with speech errors, we examined providing gesture keyboard alterations of only the mistaken words. Deprived of the user openly indicating the improper words, our model was able to decrease the word error rate by 44% relative.

Key-Words / Index Term

Mobile text entry, Multimodal interfaces

References

[1] Arif, A.S. ; Dept. of Comput. Sci. & Eng., York Univ., Toronto, ON, Canada, “A survey on mobile text entry handedness: How do users input text on handheld devices while nomadic?”, Published in: Intelligent Human Computer Interaction (IHCI), 2012 4th International Conference on Date of Conference: 27-29 Dec. 2012 Page(s): 1 – 6.
[2] Sandnes, F.E. ; Dept. of Comput. Sci., Oslo Univ. Coll., Norway ; Thorkildssen, H.W. ; Arvei, A. ; Buverad, J.O., “Techniques for fast and easy mobile text-entry with three-keys”, Published in: System Sciences, 2004. Proceedings of the 37th Annual Hawaii International Conference on Date of Conference: 5-8 Jan. 2004.
[3] Poirier, F. ; Univ. de Bretagne-Sud, Vannes ; Sad, H.H., “A Platform for Mobile Text Entry Methods Evaluation”, Published in: Advances in Computer-Human Interaction, 2008 First International Conference on Date of Conference: 10-15 Feb. 2008 Page(s): 183 – 188.
[4] Starner, T., “Keyboards redux: fast mobile text entry”, Published in: Pervasive Computing, IEEE (Volume:3 , Issue: 3 ) Page(s): 97 – 101 ISSN : 1536-1268.
[5] Tantug, A.C. ; Istanbul Tech. Univ., Istanbul, Turkey, “A probabilistic mobile text entry system for agglutinative languages”, Published in: Consumer Electronics, IEEE Transactions on (Volume:56 , Issue: 2 ) Page(s): 1018 – 1024 ISSN : 0098-3063 DOI: 10.1109/TCE.2010.5506034 Date of Publication : May 2010.
[6] Clawson, J. ; Coll. of Comput., Georgia Inst. of Technol., Atlanta, GA, USA ; Lyons, K. ; Starner, T. ; Clarkson, E., “The impacts of limited visual feedback on mobile text entry for the Twiddler and mini-QWERTY keyboards”, Published in:Wearable Computers, 2005. Proceedings. Ninth IEEE International Symposium on Date of Conference: 18-21 Oct. 2005 Page(s): 170 – 177.
[7] Ma Yang ; Sch. of Comput. Eng., Nanyang Technol. Univ., Singapore, Singapore ; Leedham, G. ; Higgins, C. ; Htwe, S.M., “On-line recognition of Pitman shorthand for fast mobile text entry”, Published in: Information Technology and Applications, 2005. ICITA 2005. Third International Conference on (Volume:1 ) Date of Conference: 4-7 July 2005 Page(s): 686 - 691 vol.1.
[8] Ha, R.W.K. ; Dept. of Electr. & Comput. Eng., Waterloo Univ., Ont., Canada ; Pin-Han Ho ; Shen, X.S., “SIMKEYS: an efficient approach in text entry for mobile communications”, Published in: Consumer Communications and Networking Conference, 2004. CCNC 2004. First IEEE Date of Conference: 5-8 Jan. 2004 Page(s): 687 - 689.
[9] Misra, S. ; Covenant Univ., Ota, Nigeria ; Banubakode, S.M. ; Dhawale, C.A., “Novel user interface for text entry on touch screen mobile device for visually impaired users”, Published in: Computer & Information Technology (GSCIT), 2014 Global Summit on Date of Conference: 14-16 June 2014 Page(s): 1 – 5.
[10] Heng-Li Yang ; Deptartment of Manage. Inf. Syst., Nat. Chengchi Univ., Taipei, Taiwan ; Ren-Xiang Lin, “A Chinese predictive text entry method for mobile devices”, Published in: Machine Learning and Cybernetics (ICMLC), 2014 International Conference on (Volume:1 ) Date of Conference: 13-16 July 2014 Page(s): 309 – 314.
[11] Prince, M.E.H. ; Dept. of Comput. Sci. & Eng., Chittagong Univ. of Eng. & Technol. (CUET), Chittagong, Bangladesh ; Hossain, G. ; Dewan, A.A. ; Debnath, P., “An audible Bangla text-entry method in Mobile phones with intelligent keypad”, Published in: Computers and Information Technology, 2009. ICCIT '09. 12th International Conference on Date of Conference: 21-23 Dec. 2009 Page(s): 279 – 284.
[12] McDermott-Wells, P. ; Graduate Sch. of Comput. & Inf. Sci., Nova Southeastern Univ., Dania, FL, “Evaluation of Three Stylus-Based Text Entry Methods on a Pocket PCtm Mobile Device”, Published in: SoutheastCon, 2006. Proceedings of the IEEE Date of Conference: March 31 2005-April 2 2005 Page(s): 228 – 234.
[13] Sharma, R. ; Adv. Interface Technol. Inc., State College, PA, USA ; Yeasin, M. ; Krahnstoever, N. ; Rauschert, I., “Speech-gesture driven multimodal interfaces for crisis management”, Published in: Proceedings of the IEEE (Volume:91 , Issue: 9 ) Page(s): 1327 – 1354 Date of Publication : Sept. 2003.
[14] Oviatt, S. ; Dept. of Comput. Sci., Oregon Health & Sci. Univ., Beaverton, OR, USA, “User-centered modeling and evaluation of multimodal interfaces”, Published in: Proceedings of the IEEE (Volume:91 , Issue: 9 ) Page(s): 1457 - 1468 Date of Publication : Sept. 2003.
[15] Pacheco, B.A. ; Comput. & Inf. Fac., Mackenzie Presbyterian Univ., São Paulo, Brazil ; Souza-Concilio, I.A., “Multimodal interfaces: An enactive approach”, Published in: Cognitive Infocommunications (CogInfoCom), 2013 IEEE 4th International Conference on Date of Conference: 2-5 Dec. 2013 Page(s): 51 – 58.
[16] Czyzewski, A. ; Gdansk Univ. of Technol., Gdansk, Poland ; Dalka, P. ; Kosikowski, L. ; Kunka, B., “Multimodal human-computer interfaces based on advanced video and audio analysis”, Published in: Human System Interaction (HSI), 2013 The 6th International Conference on Date of Conference: 6-8 June 2013 Page(s): 18 – 25.
[17] Karpov, A. ; SPIIRAS, St. Petersburg Inst. for Inf. & Autom., St. Petersburg, Russia ; Carbini, S. ; Ronzhin, A. ; Viallet, J.E., “Comparison of two different similar speech and gestures multimodal interfaces”, Published in: Signal Processing Conference, 2008 16th European Date of Conference: 25-29 Aug. 2008 Page(s): 1 – 5.
[18] Czyzewski, A. ; Multimedia Syst. Dept., Gdansk Univ. of Technol., Gdansk, Poland, “New applications of multimodal human-computer interfaces”, Published in: New Trends in Audio & Video and Signal Processing: Algorithms, Architectures, Arrangements, and Applications (NTAV/SPA), 2012 Joint Conference Date of Conference: 27-29 Sept. 2012 Page(s): 19 – 24.
[19] Jing Liu ; Dept. of Comput., Macquarie Univ., North Ryde, VIC, Australia ; Kavakli, M., “A survey of speech-hand gesture recognition for the development of multimodal interfaces in computer games”, Published in: Multimedia and Expo (ICME), 2010 IEEE International Conference on Date of Conference: 19-23 July 2010 Page(s): 1564 – 1569.
[20] Toth, B. ; Budapest Univ. of Technol. & Econ., Budapest ; Nemeth, G., “Challenges of creating multimodal interfaces on mobile devices”, Published in: ELMAR, 2007 Date of Conference: 12-14 Sept. 2007 Page(s): 171 – 174.