Getting rid of OK Google : Individual Multimodal Input Adaption in Real World Applications

Size: px
Start display at page:

Download "Getting rid of OK Google : Individual Multimodal Input Adaption in Real World Applications"

Transcription

1 Getting rid of OK Google : Individual Multimodal Input Adaption in Real World Applications Felix Schüssel 1, Frank Honold 1, Nikola Bubalo 2, Anke Huckauf 2, and Michael Weber 1 1 Institute of Media Informatics, Ulm University, Germany 2 Department of General Psychology, Ulm University, Germany {felix.schuessel, frank.honold, nikola.bubalo, anke.huckauf, michael.weber}@uni-ulm.de Abstract. Multimodal Interaction has the potential to significantly increase the ease of use of human computer interaction (HCI). At the same time, due to error-prone recognition based inputs, it is merely used in real-world applications. While literature on multimodal input fusion describes modeling of different user behaviors as a key for increased robustness, it still failed to prove it s practical use. This article presents the design of a user study that applies previous theoretical work on individual user adaption in a smartwatch scenario. We describe the practical implementation of a process for error recognition and recovery based on the history of multimodal inputs and a concrete scenario suitable for evaluating its practical impact on the ease of use. This could prove the real-world use of individual multimodal input adaption and finally lead to multimodal systems less cumbersome than today. Keywords: multimodal interaction, input fusion, user behavior, interaction history, input error detection and recovery 1 Introduction Multimodal systems, that make use of parallel inputs from different modalities, are topic of extensive research for several years now, but still, their adoption in real-world applications is almost non existent. This could be due to the fact, that error rates of sensor inputs are still too high for practical use. Recent work has shown the theoretical possibility to detect and recover from sensor errors within the fusion of multimodal user inputs by making use of a so called interaction history (cf. [2]). While these results are promising, they are based on a rather abstract gaming task. The actual implementation and evaluation of such an approach in a realistic setting has not been conducted so far. To overcome this, we propose a setting that uses a popular interaction with smartwatches as realistic scenario: quickly displaying some kind of information like upcoming tasks, news headlines, or recent messages. The next section elucidates the idea behind using an individual interaction history within multimodal input fusion. It is a summary of the work in [2] that provides

2 a rather abstract process for error detection and recovery. This is followed by the proposed implementation of such a process in an existing input fusion component [1] and it s necessary enhancements. After that, the real-world scenario for an evaluation is described. Finally, these contributions are summarized and their relation to the state of the art in commercially available systems is discussed. 2 Individual User Adaption of Input Fusion Research from Oviatt et al. and others repeatedly showed the existence of different user behaviors in terms of temporal relations between multimodal constructions [12, 11, 13, 6, 7, 9, 3, 5]. These are called integration patterns and describe the dominant use of modalities in simultaneous or sequential temporal manner (cf. [10]). Two modalities are used sequentially, when their time intervals do not overlap, but instead show a temporal gap between each other. In the opposite case, i. e. when their time intervals overlap, they are used simultaneously. This is depicted in Fig. 1. An integration pattern is called dominant for a specific user, if more than 60% of inputs are uttered in the respective manner. Though this classification scheme has been applied by others [8, 4], it s potential to increase the robustness of a system, as suggested in [13], has not been proven yet. Inputs Time intervals Integra-on Simultaneous Sequen-al Fig. 1. Simultaneous and sequential integration of modalities. In order to investigate the occurrence and use of such integration patterns in relatively short lasting multimodal interactions, in [2] an experiment where users repeatedly had to perform a visual search task was conducted. Given a set of colored geometric objects, users had to spot the unique color/shape combination present in each trial (see Fig. 2). Once this combination was discovered, users had to specify the position and color of the object in a multimodal way. Results showed, that users did not exhibit such dominant integration patterns as found in the more lengthy tasks of Oviatt and others. It is concluded that for short lasting multimodal inputs, an individual user history of temporal distributions may be more appropriate than a simple classification into dominantly simultaneous or sequential users. Schüssel et al. then present a process for the use of such individual interaction history for error detection and prevention within the input fusion processing. This process can be used to detect and recover from three types of errors: false negatives, false positives, and conflicts. Applied to the recorded data set, a potential decrease of error rates from 4.9% to 1.2% is reported. But these values remain only theoretical and contain a large

3 number of unknowns, as parts of the suggested process could not be applied on the recorded data, as acknowledged by the authors. In addition, the task at hand is rather artificial and thus results may not necessarily be transferred to real world applications. To overcome these limitations, next we will describe the actual implementation of the theoretical process before we describe the scenario that we propose to finally evaluate the increase of robustness. matrix with number- labeled objects /me/money bar color- labeled bu,ons User: blue Fig. 2. A user indicating the individual item s location using touch, and its color by using speech. (taken from [2]) 3 Realisation As stated before, the realization of the error detection and recovery process extends the existing input fusion component of [1] and is build around it, as visualized in Fig. 3. Although the actual fusion methodology is of no importance for the interaction history module, still there are some requirements that must be met by the overall implementation. To begin with, the temporal onsets and offsets of all recognized events that lead to a fused input must be forwarded to the history data base, because they form the basis of the individual history for each user. In addition, the event times, i.e. the time an input event was raised by a sensor, must be tracked, too. This is needed in order to detect false negatives, i.e. when a sensor missed an input (cf. [2]). Of course, all those timestamps must be synchronized. The fusion itself must be able to respect those timestamps in the actual fusion process by fusing events based on their real-world onsets and offsets. This is of particular importance, when events are raised long after the input was performed by the user. This can happen, if a sensor takes some processing time to decide on an input (e.g. the speech recognition used in [2] takes about 642 milliseconds), or if a missed input is discovered by the error detection process and recovered much later than the user performed the input. The multimodal input fusion component from [1], that is based on evidential reasoning, has been extended to meet these requirements.

4 Input fusion extended by user individual interac5on history Input Sensors recogni'on false nega've? no Mul'modal Input Fusion result conflict? no false posi've? no: final result Applica'on yes yes yes determine 'me frame of missed event select most likely result discard event request reprocessing individual interac'on history of current user update history Fig. 3. Overview of the implemented process of error recognition and recovery extending an existing input fusion component. 4 A Real-World Scenario for Evaluation To actually proof the advantages of an individualized interaction history and to overcome the drawback of the rather abstract scenario chosen in [2], a realistic application is inevitable. We propose to use a smartwatch scenario for several reasons. First, smartwatches offer very limited interaction possibilities due to their small screen size that limits the possibilities and usefulness of touch interactions. Thus, they could benefit a lot from multimodal interactions. Second, as a very personal device, smartwatches seem the perfect choice when it comes to user individual adaption. Third, typical tasks performed on a smartwatch are relatively short and therefore transferability of the previous findings seems reasonable. The basic task chosen is that of quick information retrieval. In order to avoid boredom, nine different types of informations are offered: inbox, weather status, traffic information, news headlines, flight status, sports scores, current heart-rate, todays step counter and upcoming tasks. The most natural way of retrieving this information is to raise one s arm and tell the smartwatch which information should be displayed. So the multimodal interaction is modeled with two interactions: the lifting of the arm and a simple speech command specifying the kind of information requested. Both interactions can easily be gathered by common motion sensors and speech recognition engines found in all modern smartwatches. For the actual implementation in Google s Android Wear however, we chose to reuse an existing speech recognition module running on a desktop pc, as it already brings the necessary onset and offset information not accessible with the current speech API offered by Google. For the detection of the arm lifting movement, we implemented a simple finite state machine that is fed with live data from the smartwatch sensors. These data are then sent to the multimodal fusion system extended by the interaction history processing as explained in the previous section. Once a valid combination of arm lifting and voice command is detected, the display of the respective information is triggered on the smartwatch.

5 To keep participants interest on a high level for the necessary number of repetitions, a slight gamification will be applied. Participants will have to answer simple questions about each requested information on a separate touch screen. For a more realistic context of use, participants will have to move between three positions in the laboratory all equipped with a touch screen for answering the current question. This should also avoid a too monotonous, unrealistic behavior. In addition, we will establish a not completely silent environment by constantly playing some kind of radio station in the background. For the actual evaluation of the benefits of user individual interaction history, a between subject design will be applied. One group of participants will perform the experiment without any adaptions, while the other group will use the actual adaption system that should reduce error rates in the course of the experiment. 5 Summary and Discussion While previous work [2] has only theoretically shown the usefulness of a user individual interaction history to detect and recover from sensory errors that occur during multimodal input interaction, the actual implementation and final proof in a real world scenario is still missing. For the actual implementation of the detection and recovery process, we identified several requirements that must be met by an actual implementation not covered in previous work. With the implementation meeting these requirements, a follow up study can now be realized for final evaluation. For this, we propose to use a smartwatch scenario for several reasons, as described in the previous section. The proposed quick information retrieval task is one of the main use cases of a smartwatch. Applying a multimodal interaction consisting of an implicit arm lifting and a short explicit speech command makes information retrieval much more easier than using interactions offered by the state of the art in commercially available systems. For Android Wear, for example, the user must raise his arm, explicitly activate voice input by uttering OK Google, before he can finally state the wanted information. The same holds true for the Apple Watch, where one has to activate voice recognition with Hey Siri or via an explicit press on the crown of the watch. Both ways to retrieve information take much longer and seem less natural than the short interaction we propose. Of course, the rationale for an explicit command to activate voice recognition is the avoidance of input errors. With the use of a user individual interaction history, however, many errors may be avoided in the first place. That is what we feel confident to prove with the work presented here, independently of the scenario. However, if error rates are low enough, introductory commands found in todays smartwatches may become entirely needless. 6 Acknowledgments This work was supported by the Transregional Collaborative Research Center SFB/ TRR 62 Companion-Technology for Cognitive Technical Systems, which is funded by the German Research Foundation (DFG).

6 References 1. Anonymized authors: Anonymized title. In: Anonymized editors (ed.) Anonymized book, LNCS, vol. Anonymized volume, p. Anonymized pages. Springer Berlin Heidelberg (2013) 2. Anonymized authors: Anonymized title. In: Anonymized proceedings. p. Anonymized pages. Anonymized series, ACM, New York, NY, USA (2014) 3. Dey, P., Madhvanath, S., Ranjan, A., Das, S.: An exploration of gesture-speech multimodal patterns for touch interfaces. In: Proc. of the 3rd Int. Conference on Human Computer Interaction. pp IndiaHCI 11, ACM, New York (2011) 4. Dumas, B., Ingold, R., Lalanne, D.: Benchmarking fusion engines of multimodal interactive systems. In: ICMI-MLMI 09: Proceedings of the 2009 international conference on Multimodal interfaces. pp ACM, New York, NY, USA (2009) 5. Haas, E.C., Pillalamarri, K.S., Stachowiak, C.C., McCullough, G.: Temporal binding of multimodal controls for dynamic map displays: a systems approach. In: Proc. of the 13th international conference on multimodal interfaces. pp ICMI 11, ACM, New York (2011) 6. Huang, X., Oviatt, S.L.: Toward adaptive information fusion in multimodal systems. In: Renals, S., Bengio, S. (eds.) Machine Learning for Multimodal Interaction, MLMI 2005, Edinburgh, UK. Lecture Notes in Computer Science, vol. 3869/2006, pp Springer (2005) 7. Huang, X., Oviatt, S.L., Lunsford, R.: Combining user modeling and machine learning to predict users multimodal integration patterns. In: Renals, S., Bengio, S., Fiscus, J.G. (eds.) Machine Learning for Multimodal Interaction, MLMI 2006, Bethesda, USA. LNCS, vol. 4299/2006, pp Springer (2006) 8. Johnston, M., Cohen, P.R., McGee, D., Oviatt, S.L., Pittman, J.A., Smith, I.: Unification-based multimodal integration. In: Proceedings of the Eighth Conference on European Chapter of the Association for Computational Linguistics. pp EACL 97, Association for Computational Linguistics, Stroudsburg, PA, USA (1997) 9. Lee, M., Billinghurst, M.: A wizard of oz study for an ar multimodal interface. In: Proc. of the 10th international conference on Multimodal interfaces. pp ICMI 08, ACM, New York (2008) 10. Oviatt, S., Cohen, P., Wu, L., Vergo, J., Duncan, L., Suhm, B., Bers, J., Holzman, T., Winograd, T., Landay, J., Larson, J., Ferro, D.: Designing the user interface for multimodal speech and pen-based gesture applications: state-of-the-art systems and future research directions. Human-Computer Interaction 15(4), (2000) 11. Oviatt, S., Coulston, R., Tomko, S., Xiao, B., Lunsford, R., Wesson, M., Carmichael, L.: Toward a theory of organized multimodal integration patterns during human-computer interaction. In: ICMI 03: Proc. of the 5th international conference on Multimodal interfaces. pp ACM, New York (2003) 12. Oviatt, S., DeAngeli, A., Kuhn, K.: Integration and synchronization of input modes during multimodal human-computer interaction. In: Proceedings of the SIGCHI conference on Human factors in computing systems. pp CHI 97, ACM, New York, NY, USA (1997) 13. Oviatt, S., Lunsford, R., Coulston, R.: Individual differences in multimodal integration patterns: what are they and why do they exist? In: CHI 05: Proc. of the SIGCHI conference on Human factors in computing systems. pp ACM, New York (2005)

An Adaptive Approach to Collecting Multimodal Input

An Adaptive Approach to Collecting Multimodal Input An Adaptive Approach to Collecting Multimodal Input Anurag Gupta University of New South Wales School of Computer Science and Engineering Sydney, NSW 2052 Australia akgu380@cse.unsw.edu.au Abstract Multimodal

More information

SOFTWARE DESIGN AND DEVELOPMENT OF MUTIMODAL INTERACTION

SOFTWARE DESIGN AND DEVELOPMENT OF MUTIMODAL INTERACTION SOFTWARE DESIGN AND DEVELOPMENT OF MUTIMODAL INTERACTION Marie-Luce Bourguet Queen Mary University of London Abstract: Key words: The multimodal dimension of a user interface raises numerous problems that

More information

Modality Fusion in a Route Navigation System

Modality Fusion in a Route Navigation System Modality Fusion in a Route Navigation System Topi Hurtig Kristiina Jokinen topi.hurtig@helsinki.fi kristiina.jokinen@helsinki.fi University of Helsinki PL9 (Siltavuorenpenger 20A) FIN-00014 University

More information

Keywords: Multimodal voice/ink applications, speech recognition, online handwriting recognition, mutual disambiguation, VoiceXML, InkXML.

Keywords: Multimodal voice/ink applications, speech recognition, online handwriting recognition, mutual disambiguation, VoiceXML, InkXML. Multimodal Integration of Voice and Ink for Pervasive Computing Zouheir Trabelsi *, Sung-Hyuk Cha, Darshan Desai, Charles Tappert CSIS, Pace University, Goldstein Center, 861 Bedford Road, Pleasantville

More information

Design and Engineering of Software Applications for Touch Input Collection on Mobile Devices

Design and Engineering of Software Applications for Touch Input Collection on Mobile Devices Design and Engineering of Software Applications for Touch Input Collection on Mobile Devices Bogdan Gheran MintViz Lab MANSiD Research Center University Ștefan cel Mare of Suceava, ROMANIA E-mail: bogdan.gheran[at]gmail.com

More information

Trade-offs in the Design of Multimodal Interaction for Older Adults

Trade-offs in the Design of Multimodal Interaction for Older Adults Trade-offs in the Design of Multimodal Interaction for Older Adults Gianluca Schiavo, Michela Ferron, Ornella Mich and Nadia Mana FBK - Fondazione Bruno Kessler, Trento, Italy {gschiavo, ferron, mich,

More information

Identifying Interaction Problems on Web Applications due to the Change of Input Modality

Identifying Interaction Problems on Web Applications due to the Change of Input Modality Identifying Interaction Problems on Web Applications due to Change of Input Modality André Constantino da Silva, André Luis Correa Viana, Samuel Gomes de Lima Instituto Federal de São Paulo Campus Hortolândia

More information

Multimodal Interfaces. Remotroid

Multimodal Interfaces. Remotroid Multimodal Interfaces Remotroid Siavash Bigdeli / Christian Lutz University of Neuchatel and University of Fribourg 1. June 2012 Table of contents 1 Introduction...3 2 Idea of the application...3 3 Device

More information

Integration of Handwriting Recognition in Butterfly Net

Integration of Handwriting Recognition in Butterfly Net Integration of Handwriting Recognition in Butterfly Net Sye-Min Christina Chan Department of Computer Science Stanford University Stanford, CA 94305 USA sychan@stanford.edu Abstract ButterflyNet allows

More information

An Analysis of the Usage of Mobile Phones for Personalized Interactions with Ubiquitous Public Displays

An Analysis of the Usage of Mobile Phones for Personalized Interactions with Ubiquitous Public Displays An Analysis of the Usage of Mobile Phones for Personalized Interactions with Ubiquitous Public Displays Enrico Rukzio, Albrecht Schmidt, Heinrich Hussmann Media Informatics Institute for Computer Science

More information

Measuring the Capability of Smartphones for Executing Context Algorithms

Measuring the Capability of Smartphones for Executing Context Algorithms Maximilian Eibl, Martin Gaedke. (Hrsg.): INFORMATIK 2017, Lecture Lecture Notes in Notes Informatics in Informatics (LNI), Gesellschaft (LNI), Gesellschaft für Informatik, für Informatik, Bonn 2017 Bonn

More information

15 Things You Can Do With Cortana

15 Things You Can Do With Cortana 15 Things You Can Do With Cortana How-to Geek www.howtogeek.com SCV COMPUTER CLUB DECEMBER 14, 2016 What is Cortana Cortana is one of Windows 10 s most visible new features Made the leap from Windows Phone

More information

Ontology driven voice-based interaction in mobile environment

Ontology driven voice-based interaction in mobile environment Ontology driven voice-based interaction in mobile environment Jiri Kopsa 1, Zdenek Mikovec 1, Pavel Slavik 1 1 Czech Technical University in Prague Karlovo namesti 13, Prague 2, Czech Republic j.kopsa@fee.ctup.cz,

More information

COMPARATIVE ANALYSIS OF EYE DETECTION AND TRACKING ALGORITHMS FOR SURVEILLANCE

COMPARATIVE ANALYSIS OF EYE DETECTION AND TRACKING ALGORITHMS FOR SURVEILLANCE Volume 7 No. 22 207, 7-75 ISSN: 3-8080 (printed version); ISSN: 34-3395 (on-line version) url: http://www.ijpam.eu ijpam.eu COMPARATIVE ANALYSIS OF EYE DETECTION AND TRACKING ALGORITHMS FOR SURVEILLANCE

More information

SmartWeb Handheld Multimodal Interaction with Ontological Knowledge Bases and Semantic Web Services

SmartWeb Handheld Multimodal Interaction with Ontological Knowledge Bases and Semantic Web Services IJCAI Workshop AI4HC, Hyderabad, 6/1/2007 SmartWeb Handheld Multimodal Interaction with Ontological Knowledge Bases and Semantic Web Services Daniel Sonntag, Ralf Engel, Gerd Herzog, Alexander Pfalzgraf,

More information

A Language Independent Author Verifier Using Fuzzy C-Means Clustering

A Language Independent Author Verifier Using Fuzzy C-Means Clustering A Language Independent Author Verifier Using Fuzzy C-Means Clustering Notebook for PAN at CLEF 2014 Pashutan Modaresi 1,2 and Philipp Gross 1 1 pressrelations GmbH, Düsseldorf, Germany {pashutan.modaresi,

More information

HAND GESTURE RECOGNITION USING MEMS SENSORS

HAND GESTURE RECOGNITION USING MEMS SENSORS HAND GESTURE RECOGNITION USING MEMS SENSORS S.Kannadhasan 1, R.Suresh 2,M.Shanmuganatham 3 1,2 Lecturer, Department of EEE, Tamilnadu Polytechnic College, Madurai, Tamilnadu, (India) 3 Senior Lecturer,

More information

Towards Systematic Usability Verification

Towards Systematic Usability Verification Towards Systematic Usability Verification Max Möllers RWTH Aachen University 52056 Aachen, Germany max@cs.rwth-aachen.de Jonathan Diehl RWTH Aachen University 52056 Aachen, Germany diehl@cs.rwth-aachen.de

More information

A modified and fast Perceptron learning rule and its use for Tag Recommendations in Social Bookmarking Systems

A modified and fast Perceptron learning rule and its use for Tag Recommendations in Social Bookmarking Systems A modified and fast Perceptron learning rule and its use for Tag Recommendations in Social Bookmarking Systems Anestis Gkanogiannis and Theodore Kalamboukis Department of Informatics Athens University

More information

MMWA-ae: boosting knowledge from Multimodal Interface Design, Reuse and Usability Evaluation

MMWA-ae: boosting knowledge from Multimodal Interface Design, Reuse and Usability Evaluation MMWA-ae: boosting knowledge from Multimodal Interface Design, Reuse and Usability Evaluation Americo Talarico Neto, Renata Pontin M. Fortes, Rafael Rossi and Solange Rezende Av. Trabalhador Saocarlense,400.

More information

Rule Formats for Nominal Modal Transition Systems

Rule Formats for Nominal Modal Transition Systems Rule Formats for Nominal Modal Transition Systems Anke Stüber Universitet Uppsala, Uppsala, Sweden anke.stuber@it.uu.se Abstract. Modal transition systems are specification languages that allow the expression

More information

Adding Usability to Web Engineering Models and Tools

Adding Usability to Web Engineering Models and Tools Adding Usability to Web Engineering Models and Tools Richard Atterer 1 and Albrecht Schmidt 2 1 Media Informatics Group Ludwig-Maximilians-University Munich, Germany richard.atterer@ifi.lmu.de 2 Embedded

More information

Leveraging Mobile Interaction with Sensor-Driven and Multimodal User Interfaces"

Leveraging Mobile Interaction with Sensor-Driven and Multimodal User Interfaces Leveraging Mobile Interaction with Sensor-Driven and Multimodal User Interfaces Andreas Möller Betreuer: Prof. Dr. Matthias Kranz Doktorandenseminar an der LMU München Institute for Media Technology My

More information

Evaluating Algorithms for Shared File Pointer Operations in MPI I/O

Evaluating Algorithms for Shared File Pointer Operations in MPI I/O Evaluating Algorithms for Shared File Pointer Operations in MPI I/O Ketan Kulkarni and Edgar Gabriel Parallel Software Technologies Laboratory, Department of Computer Science, University of Houston {knkulkarni,gabriel}@cs.uh.edu

More information

An open source process engine framework for realtime pattern recognition and information fusion tasks

An open source process engine framework for realtime pattern recognition and information fusion tasks An open source process engine framework for realtime pattern recognition and information fusion tasks Volker Fritzsch, Stefan Scherer, Friedhelm Schwenker Institute of Neural Information Processing Ulm

More information

Multimodal Interfaces That Process What Comes Naturally. Sharon Oviatt and Philip Cohen

Multimodal Interfaces That Process What Comes Naturally. Sharon Oviatt and Philip Cohen Multimodal Interfaces That Process What Comes Naturally Sharon Oviatt and Philip Cohen Center for Human-Computer Communication, Department of Computer Science Oregon Graduate Institute of Science & Technology

More information

Temporal Aspects of CARE-based Multimodal Fusion: From a Fusion Mechanism to Composition Components and WoZ Components

Temporal Aspects of CARE-based Multimodal Fusion: From a Fusion Mechanism to Composition Components and WoZ Components Temporal Aspects of CARE-based Multimodal Fusion: From a Fusion Mechanism to Composition Components and WoZ Components Marcos Serrano and Laurence Nigay University of Grenoble, CNRS, LIG B.P. 53, 38041,

More information

International Journal of Computer Science Trends and Technology (IJCST) Volume 5 Issue 5, Sep Oct 2017

International Journal of Computer Science Trends and Technology (IJCST) Volume 5 Issue 5, Sep Oct 2017 RESEARCH ARTICLE OPEN ACCESS Iris and Palmprint Decision Fusion to Enhance Human Ali M Mayya [1], Mariam Saii [2] PhD student [1], Professor Assistance [2] Computer Engineering Tishreen University Syria

More information

Voronoi Diagrams, Vectors and the Visually Impaired

Voronoi Diagrams, Vectors and the Visually Impaired Voronoi Diagrams, Vectors and the Visually Impaired Christopher Power Computer Science University of Western Ontario London, ON N6A 5B7 Canada power1@csd.uwo.ca Dawn Gill Epidemiology and Biostatistics

More information

Tai Chi Motion Recognition Using Wearable Sensors and Hidden Markov Model Method

Tai Chi Motion Recognition Using Wearable Sensors and Hidden Markov Model Method Tai Chi Motion Recognition Using Wearable Sensors and Hidden Markov Model Method Dennis Majoe 1, Lars Widmer 1, Philip Tschiemer 1 and Jürg Gutknecht 1, 1 Computer Systems Institute, ETH Zurich, Switzerland

More information

Real Time Eye Movement Identification Protocol

Real Time Eye Movement Identification Protocol Real Time Identification Protocol Do Hyong Koh Department of Computer Science Texas State University-San Marcos San Marcos, TX 78666 USA dk1132@txstate.edu Sandeep Munikrishne Gowda Department of Computer

More information

Wearables CS 4720 Mobile Application Development

Wearables CS 4720 Mobile Application Development Wearables Mobile Application Development It was news to me In preparing this lecture, I did research into wearable computing in general Turns out, there s been a conference on wearable computing since

More information

Bringing Usability to Industrial Control Systems

Bringing Usability to Industrial Control Systems Bringing Usability to Industrial Control Systems Marcus Reul RWTH Aachen University 52056 Aachen, Germany marcus.reul@rwth-aachen.de Abstract Within my ongoing work at a manufacturer for industrial test

More information

POLYBIO Multibiometrics Database: Contents, description and interfacing platform

POLYBIO Multibiometrics Database: Contents, description and interfacing platform POLYBIO Multibiometrics Database: Contents, description and interfacing platform Anixi Antonakoudi 1, Anastasis Kounoudes 2 and Zenonas Theodosiou 3 1 Philips College, 4-6 Lamias str. 2001, P.O. Box 28008,

More information

UniGest: Text Entry Using Three Degrees of Motion

UniGest: Text Entry Using Three Degrees of Motion UniGest: Text Entry Using Three Degrees of Motion Steven J. Castellucci Department of Computer Science and Engineering York University 4700 Keele St. Toronto, Ontario M3J 1P3 Canada stevenc@cse.yorku.ca

More information

Command composition approach in crossing interfaces. with more than two continuous goals

Command composition approach in crossing interfaces. with more than two continuous goals Some Student November 30, 2010 CS 5317 Command composition approach in crossing interfaces with more than two continuous goals 1. INTRODUCTION Over the last decade, pen and touch-based interactions began

More information

Research Article A Two-Level Cache for Distributed Information Retrieval in Search Engines

Research Article A Two-Level Cache for Distributed Information Retrieval in Search Engines The Scientific World Journal Volume 2013, Article ID 596724, 6 pages http://dx.doi.org/10.1155/2013/596724 Research Article A Two-Level Cache for Distributed Information Retrieval in Search Engines Weizhe

More information

Interaction Design and Implementation for Multimodal Mobile Semantic Web Interfaces

Interaction Design and Implementation for Multimodal Mobile Semantic Web Interfaces HCI International, Beijing, China, 27th July 2007 Interaction Design and Implementation for Multimodal Mobile Semantic Web Interfaces Daniel Sonntag German Research Center for Artificial Intelligence 66123

More information

Pervasive Computing. OpenLab Jan 14 04pm L Institute of Networked and Embedded Systems

Pervasive Computing. OpenLab Jan 14 04pm L Institute of Networked and Embedded Systems Pervasive Computing Institute of Networked and Embedded Systems OpenLab 2010 Jan 14 04pm L4.1.01 MISSION STATEMENT Founded in 2007, the Pervasive Computing Group at Klagenfurt University is part of the

More information

A Mobile Scenario for the XX Olympic Winter Games Torino 2006

A Mobile Scenario for the XX Olympic Winter Games Torino 2006 A Mobile Scenario for the XX Olympic Winter Games Torino 2006 LUCIA TERRENGHI Fraunhofer Institute for Applied Information Technology, Information in Context Department, Schloss Brilinghoven, 53754 Sankt

More information

Learning to Match. Jun Xu, Zhengdong Lu, Tianqi Chen, Hang Li

Learning to Match. Jun Xu, Zhengdong Lu, Tianqi Chen, Hang Li Learning to Match Jun Xu, Zhengdong Lu, Tianqi Chen, Hang Li 1. Introduction The main tasks in many applications can be formalized as matching between heterogeneous objects, including search, recommendation,

More information

Security and Usability Challenges of Moving-Object CAPTCHAs

Security and Usability Challenges of Moving-Object CAPTCHAs Security and Usability Challenges of Moving-Object CAPTCHAs Decoding Codewords in Motion Yi Xu, Gerardo Reynaga, Sonia Chiasson, Jan-Michael Frahm, Fabian Monrose and Paul Van Oorschot 1 2 CAPTCHAs Application

More information

CIRGDISCO at RepLab2012 Filtering Task: A Two-Pass Approach for Company Name Disambiguation in Tweets

CIRGDISCO at RepLab2012 Filtering Task: A Two-Pass Approach for Company Name Disambiguation in Tweets CIRGDISCO at RepLab2012 Filtering Task: A Two-Pass Approach for Company Name Disambiguation in Tweets Arjumand Younus 1,2, Colm O Riordan 1, and Gabriella Pasi 2 1 Computational Intelligence Research Group,

More information

MIKE: a Multimodal Cinematographic Editor for Virtual Worlds

MIKE: a Multimodal Cinematographic Editor for Virtual Worlds MIKE: a Multimodal Cinematographic Editor for Virtual Worlds Bruno de Araújo, André Campos, Joaquim A. Jorge Department of Information Systems and Computer Science INESC-ID/IST/Technical University of

More information

Encyclopedia of Information Science and Technology

Encyclopedia of Information Science and Technology Encyclopedia of Information Science and Technology Second Edition Mehdi Khosrow-Pour Information Resources Management Association, USA Volume IV G-Internet INFORMATION SCIENCE REFERENCE Hershey New York

More information

MATRIX BASED INDEXING TECHNIQUE FOR VIDEO DATA

MATRIX BASED INDEXING TECHNIQUE FOR VIDEO DATA Journal of Computer Science, 9 (5): 534-542, 2013 ISSN 1549-3636 2013 doi:10.3844/jcssp.2013.534.542 Published Online 9 (5) 2013 (http://www.thescipub.com/jcs.toc) MATRIX BASED INDEXING TECHNIQUE FOR VIDEO

More information

Available online at ScienceDirect. Procedia Computer Science 59 (2015 )

Available online at  ScienceDirect. Procedia Computer Science 59 (2015 ) Available online at www.sciencedirect.com ScienceDirect Procedia Computer Science 59 (2015 ) 550 558 International Conference on Computer Science and Computational Intelligence (ICCSCI 2015) The Implementation

More information

A Constrained Spreading Activation Approach to Collaborative Filtering

A Constrained Spreading Activation Approach to Collaborative Filtering A Constrained Spreading Activation Approach to Collaborative Filtering Josephine Griffith 1, Colm O Riordan 1, and Humphrey Sorensen 2 1 Dept. of Information Technology, National University of Ireland,

More information

Human-Computer Interaction (HCI) Interaction modality and Multimodality Laurence Nigay

Human-Computer Interaction (HCI) Interaction modality and Multimodality Laurence Nigay Human-Computer Interaction (HCI) Interaction modality and Multimodality Laurence Nigay University of Grenoble CLIPS-IMAG Laboratory User Interface Engineering Team Introduction: the Domain Human-Computer

More information

A Prototype Robot Speech Interface with Multimodal Feedback

A Prototype Robot Speech Interface with Multimodal Feedback Proceedings of the 2002 IEEE Int. Workshop on Robot and Human Interactive Communication Berlin, Germany, Sept. 25-27, 2002 A Prototype Robot Speech Interface with Multimodal Feedback Mathias Haage +, Susanne

More information

Implementation of Kinetic Typography by Motion Recognition Sensor

Implementation of Kinetic Typography by Motion Recognition Sensor Implementation of Kinetic Typography by Motion Recognition Sensor Sooyeon Lim, Sangwook Kim Department of Digital Media Art, Kyungpook National University, Korea School of Computer Science and Engineering,

More information

Flexible Multimodal Human-Machine Interaction in Mobile Environments

Flexible Multimodal Human-Machine Interaction in Mobile Environments Flexible Multimodal Human-Machine Interaction in Mobile Environments Dirk Bühler, Wolfgang Minker and Jochen Häußler, Sven Krüger Abstract. This article describes requirements and a prototype system for

More information

Automatic Enhancement of Correspondence Detection in an Object Tracking System

Automatic Enhancement of Correspondence Detection in an Object Tracking System Automatic Enhancement of Correspondence Detection in an Object Tracking System Denis Schulze 1, Sven Wachsmuth 1 and Katharina J. Rohlfing 2 1- University of Bielefeld - Applied Informatics Universitätsstr.

More information

TOWARDS THE IMPLEMENTATION OF TEMPORAL-BASED SOFTWARE VERSION MANAGEMENT AT UNIVERSITI DARUL IMAN MALAYSIA

TOWARDS THE IMPLEMENTATION OF TEMPORAL-BASED SOFTWARE VERSION MANAGEMENT AT UNIVERSITI DARUL IMAN MALAYSIA TOWARDS THE IMPLEMENTATION OF TEMPORAL-BASED SOFTWARE VERSION MANAGEMENT AT UNIVERSITI DARUL IMAN MALAYSIA M Nordin A Rahman, Azrul Amri Jamal and W Dagang W Ali Faculty of Informatics Universiti Darul

More information

6 TOOLS FOR A COMPLETE MARKETING WORKFLOW

6 TOOLS FOR A COMPLETE MARKETING WORKFLOW 6 S FOR A COMPLETE MARKETING WORKFLOW 01 6 S FOR A COMPLETE MARKETING WORKFLOW FROM ALEXA DIFFICULTY DIFFICULTY MATRIX OVERLAP 6 S FOR A COMPLETE MARKETING WORKFLOW 02 INTRODUCTION Marketers use countless

More information

Process Model Consistency Measurement

Process Model Consistency Measurement IOSR Journal of Computer Engineering (IOSRJCE) ISSN: 2278-0661, ISBN: 2278-8727Volume 7, Issue 6 (Nov. - Dec. 2012), PP 40-44 Process Model Consistency Measurement Sukanth Sistla CSE Department, JNTUniversity,

More information

Investigating the Effects of User Age on Readability

Investigating the Effects of User Age on Readability Investigating the Effects of User Age on Readability Kyung Hoon Hyun, Ji-Hyun Lee, and Hwon Ihm Korea Advanced Institute of Science and Technology, Korea {hellohoon,jihyunl87,raccoon}@kaist.ac.kr Abstract.

More information

Adaptive Socio-Recommender System for Open Corpus E-Learning

Adaptive Socio-Recommender System for Open Corpus E-Learning Adaptive Socio-Recommender System for Open Corpus E-Learning Rosta Farzan Intelligent Systems Program University of Pittsburgh, Pittsburgh PA 15260, USA rosta@cs.pitt.edu Abstract. With the increase popularity

More information

Fully Automatic Methodology for Human Action Recognition Incorporating Dynamic Information

Fully Automatic Methodology for Human Action Recognition Incorporating Dynamic Information Fully Automatic Methodology for Human Action Recognition Incorporating Dynamic Information Ana González, Marcos Ortega Hortas, and Manuel G. Penedo University of A Coruña, VARPA group, A Coruña 15071,

More information

Reliability Analysis of Gamma Interconnection Network Systems

Reliability Analysis of Gamma Interconnection Network Systems International Journal of Performability Engineering, Vol. 5, No. 5, October 9, pp. 485-49. RAMS Consultants Printed in India Reliability Analysis of Gamma Interconnection Network Systems 1. Introduction

More information

Please write neatly. We cannot give credit for what we cannot read

Please write neatly. We cannot give credit for what we cannot read CS 160: USER INTERFACE DESIGN UC BERKELEY SPRING 2016 MIDTERM EXAM Professor Eric Paulos 16 March 2016 Your name (please print): This is a closed book, individual test. You are not allowed to use your

More information

CS Human Computer Interaction

CS Human Computer Interaction Part A 1. Define HCI CS6008 - Human Computer Interaction UNIT-I Question Bank FOUNDATIONS OF HCI 2. What are the basic requirements of an Successful Interactive System? 3. What is STM & LTM? 4. List out

More information

MOBILE COMPUTING 2/14/17. System Structure. Context as Implicit Input. explicit input. explicit output. explicit input.

MOBILE COMPUTING 2/14/17. System Structure. Context as Implicit Input. explicit input. explicit output. explicit input. MOBILE COMPUTING CSE 40814/60814 Spring 2017 System Structure explicit input explicit output Context as Implicit Input explicit input explicit output Context: state of the user state of the physical environment

More information

Writing a Research Paper

Writing a Research Paper Writing a Research Paper I. Scott MacKenzie 1 Research Paper The final step Research is not finished until the results are published! 2 1 Organization of a Research Paper Title Abstract Body Main sections

More information

Contactless Hand Gesture Recognition System

Contactless Hand Gesture Recognition System www.ijecs.in International Journal Of Engineering And Computer Science ISSN:2319-7242 Volume 3 Issue 11 November, 2014 Page No. 9238-9242 Contactless Hand Gesture Recognition System Puja D. Kumbhare Email-id:

More information

How to make CAD tools more useful to designers through re- representation

How to make CAD tools more useful to designers through re- representation How to make CAD tools more useful to designers through re- representation John S Gero and Nick Kelly Key Centre of Design Computing and Cognition, University of Sydney, Australia ABSTRACT: This paper describes

More information

Debugging Abstract State Machine Specifications: An Extension of CoreASM

Debugging Abstract State Machine Specifications: An Extension of CoreASM Debugging Abstract State Machine Specifications: An Extension of CoreASM Marcel Dausend, Michael Stegmaier and Alexander Raschke Institute of Software Engineering and Compiler Construction, University

More information

Coin Size Wireless Sensor Interface for Interaction with Remote Displays

Coin Size Wireless Sensor Interface for Interaction with Remote Displays Coin Size Wireless Sensor Interface for Interaction with Remote Displays Atia Ayman, Shin Takahashi, and Jiro Tanaka Department of Computer Science, Graduate school of systems and information engineering,

More information

Parallel Implementation of the NIST Statistical Test Suite

Parallel Implementation of the NIST Statistical Test Suite Parallel Implementation of the NIST Statistical Test Suite Alin Suciu, Iszabela Nagy, Kinga Marton, Ioana Pinca Computer Science Department Technical University of Cluj-Napoca Cluj-Napoca, Romania Alin.Suciu@cs.utcluj.ro,

More information

Evaluating Multi-modal Input Modes in a Wizard-of-Oz Study for the Domain of Web Search

Evaluating Multi-modal Input Modes in a Wizard-of-Oz Study for the Domain of Web Search Evaluating Multi-modal Input Modes in a Wizard-of-Oz Study for the Domain of Web Search Alexandra Klein, Ingrid Schwank, Michel Généreux & Harald Trost Austrian Research Institute for Artificial Intelligence,

More information

Oleksandr Kuzomin, Bohdan Tkachenko

Oleksandr Kuzomin, Bohdan Tkachenko International Journal "Information Technologies Knowledge" Volume 9, Number 2, 2015 131 INTELLECTUAL SEARCH ENGINE OF ADEQUATE INFORMATION IN INTERNET FOR CREATING DATABASES AND KNOWLEDGE BASES Oleksandr

More information

Toward Part-based Document Image Decoding

Toward Part-based Document Image Decoding 2012 10th IAPR International Workshop on Document Analysis Systems Toward Part-based Document Image Decoding Wang Song, Seiichi Uchida Kyushu University, Fukuoka, Japan wangsong@human.ait.kyushu-u.ac.jp,

More information

Future Trends in Search User Interfaces

Future Trends in Search User Interfaces Future Trends in Search User Interfaces Dr. Marti Hearst UC Berkeley i-know Conference Keynote Sept 1, 2010 Forecasting the Future First: What are the larger trends? In technology? In society? Next: Project

More information

Presenting Non-verbal Communication to Blind Users in Brainstorming Sessions

Presenting Non-verbal Communication to Blind Users in Brainstorming Sessions Presenting Non-verbal Communication to Blind Users in Brainstorming Sessions Stephan Pölzer and Klaus Miesenberger Johannes Kepler Universität Linz, Institute Integriert Studieren, Linz, Austria {stephan.poelzer,klaus.miesenberger}@jku.at

More information

Gesture-Based 3D Mesh Modeler

Gesture-Based 3D Mesh Modeler Gesture-Based 3D Mesh Modeler Roland Carlos 1, Clarence Dalan 1, Aaron Sanchez 1, Kevin Tolentino 1 and Florante R. Salvador 1, * 1De La Salle University, Manila *Corresponding Author: florante.salvador@dlsu.edu.ph

More information

Web Applications Usability Testing With Task Model Skeletons

Web Applications Usability Testing With Task Model Skeletons Web Applications Usability Testing With Task Model Skeletons Ivo Maly, Zdenek Mikovec, Czech Technical University in Prague, Faculty of Electrical Engineering, Karlovo namesti 13, 121 35 Prague, Czech

More information

Implicit Personalization of Public Environments using Bluetooth

Implicit Personalization of Public Environments using Bluetooth Implicit Personalization of Public Environments using Bluetooth Hema Mahato RWTH Aachen 52062 Aachen hema.mahato@rwth-aachen.de Dagmar Kern Pervasive Computing University of Duisburg-Essen 45117 Essen,

More information

MoViSign: A novel authentication mechanism using mobile virtual signatures

MoViSign: A novel authentication mechanism using mobile virtual signatures MoViSign: A novel authentication mechanism using mobile virtual signatures Nikhil Handigol, Te-Yuan Huang, Gene Chi Liu 1 Introduction Signatures are a popular means of authentication. They, ideally, are

More information

Query-Sensitive Similarity Measure for Content-Based Image Retrieval

Query-Sensitive Similarity Measure for Content-Based Image Retrieval Query-Sensitive Similarity Measure for Content-Based Image Retrieval Zhi-Hua Zhou Hong-Bin Dai National Laboratory for Novel Software Technology Nanjing University, Nanjing 2193, China {zhouzh, daihb}@lamda.nju.edu.cn

More information

Demonstrating Context-aware Process Injection with the CaPI Tool

Demonstrating Context-aware Process Injection with the CaPI Tool Demonstrating Context-aware Process Injection with the CaPI Tool Klaus Kammerer, Nicolas Mundbrod, and Manfred Reichert Institute of Databases and ation Systems Ulm University, Germany {klaus.kammerer,

More information

From Handwriting Recognition to Ontologie-Based Information Extraction of Handwritten Notes

From Handwriting Recognition to Ontologie-Based Information Extraction of Handwritten Notes From Handwriting Recognition to Ontologie-Based Information Extraction of Handwritten Notes Marcus Liwicki 1, Sebastian Ebert 1,2, and Andreas Dengel 1,2 1 DFKI, Trippstadter Str. 122, Kaiserslautern,

More information

Summarization of Egocentric Moving Videos for Generating Walking Route Guidance

Summarization of Egocentric Moving Videos for Generating Walking Route Guidance Summarization of Egocentric Moving Videos for Generating Walking Route Guidance Masaya Okamoto and Keiji Yanai Department of Informatics, The University of Electro-Communications 1-5-1 Chofugaoka, Chofu-shi,

More information

ENTITIES IN THE OBJECT-ORIENTED DESIGN PROCESS MODEL

ENTITIES IN THE OBJECT-ORIENTED DESIGN PROCESS MODEL INTERNATIONAL DESIGN CONFERENCE - DESIGN 2000 Dubrovnik, May 23-26, 2000. ENTITIES IN THE OBJECT-ORIENTED DESIGN PROCESS MODEL N. Pavković, D. Marjanović Keywords: object oriented methodology, design process

More information

A Prospect of Websites Evaluation Tools Based on Event Logs

A Prospect of Websites Evaluation Tools Based on Event Logs A Prospect of Websites Evaluation Tools Based on Event Logs Vagner Figuerêdo de Santana 1, and M. Cecilia C. Baranauskas 2 1 Institute of Computing, UNICAMP, Brazil, v069306@dac.unicamp.br 2 Institute

More information

International Journal of Advance Research in Computer Science and Management Studies

International Journal of Advance Research in Computer Science and Management Studies Volume 2, Issue 12, December 2014 ISSN: 2321 7782 (Online) International Journal of Advance Research in Computer Science and Management Studies Research Article / Survey Paper / Case Study Available online

More information

MIBA: Multitouch Image-Based Authentication on Smartphones

MIBA: Multitouch Image-Based Authentication on Smartphones MIBA: Multitouch Image-Based Authentication on Smartphones Daniel Ritter daniel.ritter@uni-ulm.de Florian Schaub florian.schaub@uni-ulm.de Marcel Walch marcel.walch@uni-ulm.de Michael Weber michael.weber@uni-ulm.de

More information

VHDL MODEL OF SMART SENSOR

VHDL MODEL OF SMART SENSOR VHDL MODEL OF SMART SENSOR 1 MS. VAISHALI M. BAGADE, 2 MR M.B.LIMKAR 1Electronics Department 1, TERNA College, Nerul 2 Electronics & Telecommunication Department, TERNA College, Navi Mumbai Email: 1 vaishali.bagade2@gmail.com,

More information

Towards a Search-based Interactive Configuration of Cyber Physical. System Product Lines 1

Towards a Search-based Interactive Configuration of Cyber Physical. System Product Lines 1 Towards a Search-based Interactive Configuration of Cyber Physical System Product Lines Kunming Nie, Tao Yue, Shaukat Ali Software Engineering Institute, Beihang University, Beijing, China niekunming@cse.buaa.edu.cn

More information

Image Classification Using Wavelet Coefficients in Low-pass Bands

Image Classification Using Wavelet Coefficients in Low-pass Bands Proceedings of International Joint Conference on Neural Networks, Orlando, Florida, USA, August -7, 007 Image Classification Using Wavelet Coefficients in Low-pass Bands Weibao Zou, Member, IEEE, and Yan

More information

Automatic visual recognition for metro surveillance

Automatic visual recognition for metro surveillance Automatic visual recognition for metro surveillance F. Cupillard, M. Thonnat, F. Brémond Orion Research Group, INRIA, Sophia Antipolis, France Abstract We propose in this paper an approach for recognizing

More information

Towards Automatic Recognition of Fonts using Genetic Approach

Towards Automatic Recognition of Fonts using Genetic Approach Towards Automatic Recognition of Fonts using Genetic Approach M. SARFRAZ Department of Information and Computer Science King Fahd University of Petroleum and Minerals KFUPM # 1510, Dhahran 31261, Saudi

More information

Guiding System Modelers in Multi View Environments: A Domain Engineering Approach

Guiding System Modelers in Multi View Environments: A Domain Engineering Approach Guiding System Modelers in Multi View Environments: A Domain Engineering Approach Arnon Sturm Department of Information Systems Engineering Ben-Gurion University of the Negev, Beer Sheva 84105, Israel

More information

Learning the Three Factors of a Non-overlapping Multi-camera Network Topology

Learning the Three Factors of a Non-overlapping Multi-camera Network Topology Learning the Three Factors of a Non-overlapping Multi-camera Network Topology Xiaotang Chen, Kaiqi Huang, and Tieniu Tan National Laboratory of Pattern Recognition, Institute of Automation, Chinese Academy

More information

Mubug: a mobile service for rapid bug tracking

Mubug: a mobile service for rapid bug tracking . MOO PAPER. SCIENCE CHINA Information Sciences January 2016, Vol. 59 013101:1 013101:5 doi: 10.1007/s11432-015-5506-4 Mubug: a mobile service for rapid bug tracking Yang FENG, Qin LIU *,MengyuDOU,JiaLIU&ZhenyuCHEN

More information

MOBILE COMPUTING 2/11/18. System Structure. Context as Implicit Input. explicit input. explicit output. explicit input.

MOBILE COMPUTING 2/11/18. System Structure. Context as Implicit Input. explicit input. explicit output. explicit input. MOBILE COMPUTING CSE 40814/60814 Spring 2018 System Structure explicit input explicit output Context as Implicit Input explicit input explicit output Context: state of the user state of the physical environment

More information

EVENT VERIFICATION THROUGH VOICE PROCESS USING ANDROID. Kodela Divya* 1, J.Pratibha 2

EVENT VERIFICATION THROUGH VOICE PROCESS USING ANDROID. Kodela Divya* 1, J.Pratibha 2 ISSN 2277-2685 IJESR/May 2015/ Vol-5/Issue-5/179-183 Kodela Divya et. al./ International Journal of Engineering & Science Research EVENT VERIFICATION THROUGH VOICE PROCESS USING ANDROID ABSTRACT Kodela

More information

Social Navigation Support through Annotation-Based Group Modeling

Social Navigation Support through Annotation-Based Group Modeling Social Navigation Support through Annotation-Based Group Modeling Rosta Farzan 2 and Peter Brusilovsky 1,2 1 School of Information Sciences and 2 Intelligent Systems Program University of Pittsburgh, Pittsburgh

More information

INTERACTIVE 3D ANIMATION SYSTEM BASED ON TOUCH INTERFACE AND EFFICIENT CREATION TOOLS. Anonymous ICME submission

INTERACTIVE 3D ANIMATION SYSTEM BASED ON TOUCH INTERFACE AND EFFICIENT CREATION TOOLS. Anonymous ICME submission INTERACTIVE 3D ANIMATION SYSTEM BASED ON TOUCH INTERFACE AND EFFICIENT CREATION TOOLS Anonymous ICME submission ABSTRACT Recently importance of tablet devices with touch interface increases significantly,

More information

Approach to Increase Accuracy of Multimodal Biometric System for Feature Level Fusion

Approach to Increase Accuracy of Multimodal Biometric System for Feature Level Fusion Approach to Increase Accuracy of Multimodal Biometric System for Feature Level Fusion Er. Munish Kumar, Er. Prabhjit Singh M-Tech(Scholar) Global Institute of Management and Emerging Technology Assistant

More information

A Kinect Sensor based Windows Control Interface

A Kinect Sensor based Windows Control Interface , pp.113-124 http://dx.doi.org/10.14257/ijca.2014.7.3.12 A Kinect Sensor based Windows Control Interface Sang-Hyuk Lee 1 and Seung-Hyun Oh 2 Department of Computer Science, Dongguk University, Gyeongju,

More information