|
[1] R. Aigner, D. Wigdor, H. Benko, M. Haller, D. Lindbauer, A. Ion, S. Zhao, and J. Koh. Understanding mid-air hand gestures: A study of human preferences in usage of gesture types for hci. Microsoft Research TechReport MSR-TR-2012-111, 2012. [2] L. Anthony, Q. Brown, J. Nias, B. Tate, and S. Mohan. Interaction and recogni- tion challenges in interpreting children’s touch and gesture input on mobile devices. ITS’12, pages 225–234. ACM. [3] App Annie. . [4] T. Baba, T. Ushiama, R. Tsuruno, and K. Tomimatsu. Video game that uses skin contact as controller input. SIGGRAPH’07. ACM. [5] A. Biskupski, A. R. Fender, T. M. Feuchtner, M. Karsten, and J. D. Willaredt. Drunken ed: A balance game for public large screen displays. CHI EA’14, pages 289–292. ACM. [6] L. Chan, C.-H. Hsieh, Y.-L. Chen, S. Yang, D.-Y. Huang, R.-H. Liang, and B.-Y. Chen. Cyclops: Wearable and single-piece full-body gesture input device. CHI’15. ACM. [7] L.Chan,R.-H.Liang,M.-C.Tsai,K.-Y.Cheng,C.-H.Su,M.Y.Chen,W.-H.Cheng, and B.-Y. Chen. Fingerpad: Private and subtle interaction using fingertips. UIST’13. ACM. [8] S. Christian, J. Alves, A. Ferreira, D. Jesus, R. Freitas, and N. Vieira. Volcano sal- vation: Interaction through gesture and head tracking. CHI EA’14, pages 297–300. ACM. [9] A. Colaco, A. Kirmani, H. S. Yang, N.-W. Gong, C. Schmandt, and V. K. Goyal. Mime: Compact, low power 3d gesture sensing for interaction with head mounted displays. UIST’13, pages 227–236. ACM. [10] Dollar N Multistroke Recognizer. . [11] J. Epps, S. Lichman, and M. Wu. A study of hand shape use in tabletop gesture interaction. CHI’06, pages 748–753. [12] Epson BT-100 Specs. . [13] Essential Facts About The Computer And Video Game Industry. [14] GameStop. . [15] Google Glass wiki. . [16] D. Grijincu, M. A. Nacenta, and P. O. Kristensson. User-defined interface gestures: Dataset and analysis. ITS’14, pages 25–34. ACM. [17] S. Gustafson, C. Holz, and P. Baudisch. Imaginary phone: Learning imaginary in- terfaces by transferring spatial memory from a familiar device. UIST’11, pages 283– 292. ACM. [18] S. Harada, J. O. Wobbrock, and J. A. Landay. Voice games: Investigation into the use of non-speech voice input for making computer games more accessible. INTER- ACT’11, pages 11–29. Springer-Verlag. [19] C. Harrison, H. Benko, and A. D. Wilson. Omnitouch: Wearable multitouch inter- action everywhere. UIST’11, pages 441–450. ACM. [20] C. Harrison, D. Tan, and D. Morris. Skinput: Appropriating the body as an input surface. CHI’10, pages 453–462. ACM. [21] C.-Y. Hsu, Y.-C. Tung, H.-Y. Wang, S. Chyou, J.-W. Lin, and M. Y. Chen. Glass shooter: Exploring first-person shooter game control with google glass. ICMI’14, pages 70–71. ACM. [22] L. Jing, Z. Cheng, Y. Zhou, J. Wang, and T. Huang. Magic ring: A self-contained gesture input device on finger. MUM’13, pages 39:1–39:4. ACM. [23] Cohen’s kappa - Wikipedia, the free encyclopedia. [24] M. Karam et al. A taxonomy of gestures in human computer interactions. 2005. [25] D. Kim, O. Hilliges, S. Izadi, A. D. Butler, J. Chen, I. Oikonomidis, and P. Olivier. Digits: Freehand 3d interactions anywhere using a wrist-worn gloveless sensor. UIST’12, pages 167–176. ACM. [26] H.-N. Liang, C. Williams, M. Semegen, W. Stuerzlinger, and P. Irani. User-defined surface+motion gestures for 3d manipulation of objects at a distance through a mo- bile device. APCHI’12, pages 299–308. ACM. [27] Google Glass Mini Games. . [28] C. S. Montero, J. Alexander, M. T. Marshall, and S. Subramanian. Would you do that?: Understanding social acceptance of gestural interfaces. MobileHCI’10, pages 275–278. ACM. [29] M. R. Morris. Web on the wall: Insights from a multimodal interaction elicitation study. ITS’12, pages 95–104. ACM. [30] L. E. Nacke, M. Kalyn, C. Lough, and R. L. Mandryk. Biofeedback game design: Using direct and indirect physiological control to enhance game interaction. CHI’11, pages 103–112. ACM. [31] M. Nielsen, M. Storring, T. B. Moeslund, and E. Granum. A procedure for develop- ing intuitive and ergonomic gesture interfaces for hci. In Gesture-Based Communi- cation in Human-Computer Interaction, pages 409–420. Springer, 2004. [32] T.Piumsomboon,A.Clark,M.Billinghurst,andA.Cockburn.User-definedgestures for augmented reality. CHI’13, pages 955–960, New York, NY, USA. ACM. [33] D. Pyryeskin, M. Hancock, and J. Hoey. Comparing elicited gestures to designer- created gestures for selection above a multitouch surface. ITS’12, pages 1–10. ACM. [34] S. Reis. Expanding the magic circle in pervasive casual play. ICEC’12, pages 486– 489. Springer-Verlag. [35] M. Serrano, B. M. Ens, and P. P. Irani. Exploring the use of hand-to-face input for interacting with head-worn displays. CHI’14, pages 3181–3190. ACM. [36] A. J. Sporka, S. H. Kurniawan, M. Mahmud, and P. Slavik. Non-speech input and speech recognition for real-time control of computer games. Assets’06, pages 213– 220. ACM. [37] Steam. . [38] Top 90 Casual Games List. , Analyzed at 2014-08-14. [39] VGChartz. . [40] S. Vickers, H. Istance, and A. Hyrskykari. Performing locomotion tasks in immer- sive computer games with an adapted eye-tracking interface. ACM Trans. Access. Comput., 5(1):2:1–2:33, Sept. 2013. [41] J. R. Williamson, S. Brewster, and R. Vennelakanti. Mo!games: Evaluating mobile gestures in the wild. ICMI’13, pages 173–180. ACM. 33 [42] J. O. Wobbrock, H. H. Aung, B. Rothrock, and B. A. Myers. Maximizing the guess- ability of symbolic input. CHI EA’05, pages 1869–1872. ACM. [43] J. O. Wobbrock, M. R. Morris, and A. D. Wilson. User-defined gestures for surface computing. CHI’09, pages 1083–1092. ACM.
|