帳號:guest(18.118.30.253)          離開系統
字體大小: 字級放大   字級縮小   預設字形  

詳目顯示

以作者查詢圖書館館藏以作者查詢臺灣博碩士論文系統以作者查詢全國書目勘誤回報
作者:李梃煒
作者(英文):Ting-Wei Li
論文名稱:基於手勢辨識之滑鼠操控技術實作與評估
論文名稱(英文):Implementation and Evaluation of Gesture-Based Mouse Control Technology
指導教授:羅壽之
指導教授(英文):Shou-Chih Lo
口試委員:李官陵
張耀中
口試委員(英文):Guan-Ling Lee
Yao-Chung Chang
學位類別:碩士
校院名稱:國立東華大學
系所名稱:資訊工程學系
學號:611021205
出版年(民國):112
畢業學年度:111
語文別:中文
論文頁數:77
關鍵詞:自然用戶介面MediaPipe Hands手勢辨識深度訊息校正滑鼠操控
關鍵詞(英文):natural user interfaceMediaPipe Handsgesture recognitiondepth information adjustmentmouse control
相關次數:
  • 推薦推薦:0
  • 點閱點閱:18
  • 評分評分:系統版面圖檔系統版面圖檔系統版面圖檔系統版面圖檔系統版面圖檔
  • 下載下載:13
  • 收藏收藏:0
隨著科技的不斷進步和社會的發展,計算機已成為人們生活中不可或缺的一部分。人們對於更先進、更直觀的人機互動設備的需求日益增長,尤其是在3D 多媒體系統和元宇宙等領域的應用中。然而,傳統的滑鼠和鍵盤已無法完全滿足這種需求,因此基於視覺的人機互動方式成為了研究者關注的熱門領域。
本研究旨在開發一個基於RGB 攝影機的手勢操控滑鼠系統,以提供使用者更直觀、更自然的控制方式。該系統的目標是解決傳統滑鼠和鍵盤在3D 多媒體系統中存在的操作限制,同時讓使用者能夠輕鬆體驗手勢操控的便利性,而不受深度攝影機的限制。
本研究利用MediaPipe Hands 技術捕捉手部關節點,並提出了一種基於手部特性的校正方法,以校正容易受到遮擋而產生錯誤資料的關節點。我們採用了兩種手勢辨識方法,即rule-based model 和random forest 對手勢進行辨識。同時,我們使用基於秒數的觸發方式對滑鼠進行操控。通過對實驗結果進行分析,我們證明了本研究提出的深度校正方法對手勢辨識提供了有效的幫助,並且在3D 架構的呈現上更貼合真實情況。以外,我們還比較並討論了rule-based model 和random forest 的數據,提出了他們各自的優點和限制。
with the continuous advancement of technology and the development of society,computers have become an indispensable part of people's lives. There is an increasing demand for more advanced and intuitive human-computer interaction devices, especially in the applications of 3D multimedia systems and metaverse. However, traditional mice and keyboards are no longer able to fully meet this demand, leading to a growing interest in vision-based human-computer interaction among researchers.
The objective of this study is to develop a gesture-controlled mouse system based on an RGB camera, providing users with a more intuitive and natural way of control. The system aims to address the operational limitations of traditional mice and keyboards in 3D multimedia systems, allowing users to easily experience the convenience of gesture control without being restricted by depth cameras.
In this research, we employ the MediaPipe Hands technology to capture hand joint points and propose an adjustment method based on hand features to overcome issues caused by occlusion that may result in erroneous data. We utilize two gesture recognition methods, namely rule-based model and random forest, to recognize gestures. Additionally, we employ a time-based triggering mechanism for mouse control. Through the analysis of experimental results, we demonstrate the effectiveness of the proposed depth adjustment method in gesture recognition and its suitability for presenting 3D structures realistically. Furthermore, a comparison and discussion of the data between the rule-based model and random forest are conducted to highlight their respective advantages and limitations.
第1章 前言 1
1-1 研究背景 1
1-2 研究動機與目的 1
1-3 論文綱要 2
第2章 背景知識 3
2-1 人機互動 3
2-1-1 背景 3
2-1-2 現狀與趨勢 4
2-2 手勢 7
2-2-1 手勢種類 7
2-2-2 手勢辨識 7
2-3 MediaPipe 10
2-4 相似論文 17
第3章 研究方法與步驟 21
3-1 資料獲取 21
3-2 資料處理 28
3-3 手勢辨識 33
3-3-1 Rule-Based Model 35
3-3-2 Random Forest 39
3-4 系統運行機制 43
3-4-1 誤差判斷 44
3-4-2 觸發條件 45
第4章 實驗數據與系統實作 47
4-1 實驗環境 47
4-2 實驗方法討論 48
4-2-1 Rule-Based Model實驗結果 51
4-2-2 Random Forest實驗數據 55
4-2-3 深度校正結果 57
4-2-4 實驗探討與辨識方法比較 64
4-3 系統實測討論 66
第5章 結論與未來工作 69
5-1 結論 69
5-2 未來工作 69

CAMSENSE (2018/08/09)。什麼是視覺空間定位?取自https://kknews.cc/zh-tw/tech/6or6vo3.html
HERESY (2011/01/08)。用FAAST把Kinect當Windows鍵盤用! 取自https://kheresy.wordpress.com/2011/01/08/faast_input_key_via_kinect/
Rice Yang (2021/04/12)。用MediaPipe快速搭建Hand Tracking。取自https://u9534056.medium.com/mediapipe-%E7%B0%A1%E5%96%AE%E6%98%93%E7%94%A8%E7%9A%84%E6%B7%B1%E5%BA%A6%E5%AD%B8%E7%BF%92%E6%8E%A8%E7%90%86%E6%A1%86%E6%9E%B6-4898eed9f839
VR領域答主 (2022/02/12)。VR定位技術Outside-in VS Inside-out外向內VS內向外 追蹤技術。取自https://zhuanlan.zhihu.com/p/243300575
李文恩 (2020/01/31)。手勢控制再進化,Glamos帶來虛擬空氣觸控螢幕。取自https://www.techbang.com/posts/75902-gesture-control-reevolves-glamos-brings-virtual-air-touch-screen
李宗翰 (2022/02/27)。從家家有電腦到人人有電腦,英特爾預期PC市場仍存在成長動能,並揭露未來三年登場個人電腦處理器平臺。取自https://www.ithome.com.tw/news/149572
游陳叡(2021)。基於深度學習和MediaPipe的手勢輔助虛擬觸控系統。國立中正大學資訊工程研究所碩士論文,嘉義縣。 取自https://hdl.handle.net/11296/743479
萌繪(2016/06/17)。零基礎漫畫入門-16.手部畫法的解剖。取自https://www.moehui.com/8446.html
黃群翔(2022)。結合 Google MediaPipe實現一手勢辨識控制智能家電之物聯網系統。明志科技大學電子工程系碩士班碩士論文,新北市。 取自https://hdl.handle.net/11296/36ny6a
邱庭毅(2021)。無人機環境感知與階層式手勢控制於人機協作任務應用。﹝碩士論文。國立政治大學﹞臺灣博碩士論文知識加值系統。 https://hdl.handle.net/11296/sh6eby。
Alexander, Andrey, and Karina(2022/09). HaGRID Classification 512p 127k. Retrieved March 27, 2023, form https://www.kaggle.com/datasets/innominate817/hagrid-classification-512p-127k?resource=download
Google-research-datasets (n.d.). Objectron. Retrieved September 8, 2022, from https://github.com/google-research-datasets/Objectron/
International Journal of Mon-Machine Studies (1969). Retrieved from https://dblp.org/db/journals/ijmms/ijmms15.html
Ivan Grishchenko and Valentin Bazarevsky (2020/10/10). MediaPipe Holistic – Simultaneous Face, Hand and Pose Prediction, on Device. Retrieved September 8, 2022, from https://ai.googleblog.com/2020/12/mediapipe-holistic-simultaneous-face.html
Ming Guang Yong, (2019/10/10). Object Detection and Tracking using MediaPipe. Retrieved September 8, 2022, from https://developers.googleblog.com/2019/12/object-detection-and-tracking-using-mediapipe.html
Myoung-Kyu Sohn, Sang-Heon Lee, Dong-Ju Kim, &Hyunduk Kim (2011)。Hand Gesture Key Emulation Toolkit(HandGKET). Retrieved September 8, 2022, from https://sites.google.com/site/kinectapps/handgket
Zhicheng Wang , Genzhi Ye, & MediaPipe team (2020/04/22). MediaPipe KNIFT : Template-based feature matching. Retrieved from https://developers.googleblog.com/2020/04/mediapipe-knift-template-based-feature-matching.html
Ablavatski, A., Vakunov, A., Grishchenko, I., Raveendran, K., & Zhdanovich, M. (2020). Real-time Pupil Tracking from Monocular Video for Digital Puppetry. arXiv preprint arXiv:2006.11341.
Ahmadyan, A., Zhang, L., Ablavatski, A., Wei, J., & Grundmann, M. (2021). Objectron: A large scale dataset of object-centric videos in the wild with pose annotations. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition (pp. 7822-7831).
Bazarevsky, V., Grishchenko, I., Raveendran, K., Zhu, T., Zhang, F., & Grundmann, M. (2020). Blazepose: On-device real-time body pose tracking. arXiv preprint arXiv:2006.10204.
Bazarevsky, V., Kartynnik, Y., Vakunov, A., Raveendran, K., & Grundmann, M. (2019). Blazeface: Sub-millisecond neural face detection on mobile gpus. arXiv preprint arXiv:1907.05047.
Cohen, P. R., & Oviatt, S. L. (1995). The role of voice input for human-machine communication. proceedings of the National Academy of Sciences, 92(22), 9921-9927.
Hasan, H. S., & Kareem, S. A. (2012, November). Human computer interaction for vision based hand gesture recognition: a survey. In 2012 International Conference on Advanced Computer Science Applications and Technologies (ACSAT) (pp. 55-60). IEEE.
Hershey, J. R., Chen, Z., Le Roux, J., & Watanabe, S. (2016, March). Deep clustering: Discriminative embeddings for segmentation and separation. In 2016 IEEE international conference on acoustics, speech and signal processing (ICASSP) (pp. 31-35). IEEE.
Jacob, R. J., & Karn, K. S. (2003). Eye tracking in human-computer interaction and usability research: Ready to deliver the promises. In The mind's eye (pp. 573-605). North-Holland.
Jaimes, A., & Sebe, N. (2007). Multimodal human–computer interaction: A survey. Computer vision and image understanding, 108(1-2), 116-134.
Kartynnik, Y., Ablavatski, A., Grishchenko, I., & Grundmann, M. (2019). Real-time facial surface geometry from monocular video on mobile GPUs. arXiv preprint arXiv:1907.06724.
Kaushik, D., & Jain, R. (2014). Natural user interfaces: Trend in virtual interaction. arXiv preprint arXiv:1405.0101.
Lee, D. L., & You, W. S. (2018). Recognition of complex static hand gestures by using the wristband‐based contour features. IET Image Processing, 12(1), 80-87.
Licklider, J. C. (1960). Man-computer symbiosis. IRE transactions on human factors in electronics, (1), 4-11.
Lin, T. Y., Dollár, P., Girshick, R., He, K., Hariharan, B., & Belongie, S. (2017). Feature pyramid networks for object detection. In Proceedings of the IEEE conference on computer vision and pattern recognition (pp. 2117-2125).
Lin, T. Y., Goyal, P., Girshick, R., He, K., & Dollár, P. (2017). Focal loss for dense object detection. In Proceedings of the IEEE international conference on computer vision (pp. 2980-2988).
Liu, W., Anguelov, D., Erhan, D., Szegedy, C., Reed, S., Fu, C. Y., & Berg, A. C. (2016, October). Ssd: Single shot multibox detector. In European conference on computer vision (pp. 21-37). Springer, Cham.
Majaranta, P., & Bulling, A. (2014). Eye tracking and eye-based human–computer interaction. Advances in physiological computing, 39-65.
Mantiuk, R., Kowalik, M., Nowosielski, A., & Bazyluk, B. (2012, January). Do-it-yourself eye tracker: Low-cost pupil-based eye tracker for computer graphics applications. In International Conference on Multimedia Modeling (pp. 115-125). Springer, Berlin, Heidelberg.
Mesbahi, S. C., Mahraz, M. A., Riffi, J., & Tairi, H. (2018, April). Hand gesture recognition based on convexity approach and background subtraction. In 2018 International Conference on Intelligent Systems and Computer Vision (ISCV) (pp. 1-5). IEEE.
Moscovich, T. (2007). Principles and applications of multi-touch interaction.
Neubeck, A., & Van Gool, L. (2006, August). Efficient non-maximum suppression. In 18th International Conference on Pattern Recognition (ICPR'06) (Vol. 3, pp. 850-855). IEEE.
Ogihara, A., Matsumoto, H., & Shiozaki, A. (2005, December). Hand region extraction by background subtraction with renewable background for hand gesture recognition. In 2006 International Symposium on Intelligent Signal Processing and Communications (pp. 227-230). IEEE.
Rautaray, S. S., & Agrawal, A. (2015). Vision based hand gesture recognition for human computer interaction: a survey. Artificial intelligence review, 43(1), 1-54.
Reddy, V. V., Dhyanchand, T., Krishna, G. V., & Maheshwaram, S. (2020, September). Virtual Mouse Control Using Colored Finger Tips and Hand Gesture Recognition. In 2020 IEEE-HYDCON (pp. 1-5). IEEE.
Shackel, B. (1959). Ergonomics for a computer. Design, 120(1), 36-39.
Shajideen, S. M. S., & Preetha, V. H. (2018, December). Hand Gestures-Virtual Mouse for Human Computer Interaction. In 2018 International Conference on Smart Systems and Inventive Technology (ICSSIT) (pp. 543-546). IEEE.
Shannon, C. E. (1948). A mathematical theory of communication. The Bell system technical journal, 27(3), 379-423.
Tan, P., Han, X., Zou, Y., Qu, X., Xue, J., Li, T., ... & Wang, Z. L. (2022). Self‐Powered Gesture Recognition Wristband Enabled by Machine Learning for Full Keyboard and Multicommand Input. Advanced Materials, 34(21), 2200793.
Tkachenka, A., Karpiak, G., Vakunov, A., Kartynnik, Y., Ablavatski, A., Bazarevsky, V., & Pisarchyk, S. (2019). Real-time hair segmentation and recoloring on mobile gpus. arXiv preprint arXiv:1907.06740.
Wei, J., Ye, G., Mullen, T., Grundmann, M., Ahmadyan, A., & Hou, T. (2019). Instant motion tracking and its applications to augmented reality. arXiv preprint arXiv:1907.06796.
Wilk, M. P., Torres-Sanchez, J., Tedesco, S., & O'Flynn, B. (2018, August). Wearable human computer interface for control within immersive VAMR gaming environments using data glove and hand gestures. In 2018 IEEE Games, Entertainment, Media Conference (GEM) (pp. 1-9). IEEE.
Wolpaw, J. R., McFarland, D. J., Neat, G. W., & Forneris, C. A. (1991). An EEG-based brain-computer interface for cursor control. Electroencephalography and clinical neurophysiology, 78(3), 252-259.
Yao, Y., & Fu, Y. (2014). Contour model-based hand-gesture recognition using the Kinect sensor. IEEE Transactions on Circuits and Systems for Video Technology, 24(11), 1935-1944.
Zhang, F., Bazarevsky, V., Vakunov, A., Tkachenka, A., Sung, G., Chang, C. L., & Grundmann, M. (2020). Mediapipe hands: On-device real-time hand tracking. arXiv preprint arXiv:2006.10214.
Zhao, S., Tan, W., Wu, C., Liu, C., & Wen, S. (2009, June). A novel interactive method of virtual reality system based on hand gesture recognition. In 2009 Chinese Control and Decision Conference (pp. 5879-5882). IEEE.
 
 
 
 
第一頁 上一頁 下一頁 最後一頁 top
* *