帳號:guest(3.133.139.4)          離開系統
字體大小: 字級放大   字級縮小   預設字形  

詳目顯示

以作者查詢圖書館館藏以作者查詢臺灣博碩士論文系統以作者查詢全國書目勘誤回報
作者:Paulo Enrique Linares Otoya
作者(英文):Paulo Enrique Linares Otoya
論文名稱:以頭部姿勢分類和基於臉部特徵點整體學習的姿勢不變人臉識別
論文名稱(英文):Head Pose Classification and Facial Landmark Based Ensemble Learning for Pose-invariant Face Recognition
指導教授:林信鋒
指導教授(英文):Shin-Feng Lin
口試委員:葉家宏
江政欽
口試委員(英文):Chia-Hung Yeh
Cheng-Chin Chiang
學位類別:碩士
校院名稱:國立東華大學
系所名稱:資訊工程學系
學號:611021303
出版年(民國):112
畢業學年度:111
語文別:英文
論文頁數:111
關鍵詞:整體學習臉部特徵點局部特徵描述子姿勢不變人臉識別非線性向量變換
關鍵詞(英文):Ensemble learningfacial landmarkslocal feature descriptorspose-invariant face recognitionnon-linear vector transformations
相關次數:
  • 推薦推薦:0
  • 點閱點閱:16
  • 評分評分:系統版面圖檔系統版面圖檔系統版面圖檔系統版面圖檔系統版面圖檔
  • 下載下載:13
  • 收藏收藏:0
三十多年來,人臉識別因其在生物識別和監控方面的應用而一直是計算機視覺領域的一個活躍研究領域。姿勢不變人臉識別(PIFR)通過分析從不同姿勢捕獲的人臉圖像來解決識別個體的任務。主要目標是通過處理一個人的非正面臉部圖像來識別該人。然而,由於姿勢而導致的臉部外觀的巨大變化意味著在不同姿勢下實現準確的人臉識別面臨著巨大的挑戰。近年來,PIFR 主要關注整體方法。ArcFace、Elastic Face 和FaceNet 等DCNN 已被用來生成人臉圖像映射,隨後用於人臉識別,並取得了可喜的結果。

本論文旨在通過一種新穎的局部方法來解決PIFR。在這種方法中,姿勢不變的人臉識別是通過結合頭部姿勢估計(HPE)、整體學習系統和局部特徵描述子來進行的。在人臉數據庫上為每個受試者訓練一個整體系統。該整體系統包括一個基礎學習器集,其中每個基礎學習器都專門使用從圖庫臉部圖像中特定臉部特徵點周圍的區域提取的特徵向量進行訓練。為了對輸入圖像進行PIFR,實施以下步驟。首先,檢測圖像中的臉部和臉部特徵點。其次,處理臉部特徵點位置以獲得頭部姿勢描述子,用於執行HPE。根據HPE 結果,選擇某些特徵點來對其應用局部特徵描述子,因為其中一些特徵點可能會因頭部姿勢而自遮擋。第三,計算出的描述子被分送給相應的基礎學習器。事實上,一個基礎學習器與特定的臉部特徵點相關聯。第四,合併每個整體系統基礎學習器的輸出,以計算其整體決策支持值。最後,通過選擇具有最高決策支持值的整體系統來計算輸入人臉圖像的身份。

在這項工作中,提出了三種新穎的頭部姿勢描述子(FAV、EFAV、NFCV)與非線性回歸模型相結合來執行HPE。BIWI 和Pointing’04 數據庫的實驗結果表明,所提出的HPE 方法優於使用相同數據庫的幾種最先進的方法。另一方面,提出了一種稱為特定特徵點SIFT (LS-SIFT) 的新型局部特徵描述子,以提高SIFT 針對視點和光照變化的強健性。此外,本論文還開發了三種創新的基礎學習器模型(RD-CS、GMM、Mahalanobis Similarity)。整個PIFR 框架在CMU-PIE、Multi-PIE 和FERET 數據庫上進行了測試。獲得的結果可與最先進的作品相媲美。最後,LS-SIFT 與Mahalanobis Similarity 的結合產生了最好的識別結果。
Face recognition has been an active research field in computer vision for over three decades, due to its applications on biometric authentication, and surveillance. Pose-invariant face recognition (PIFR) addresses the task of identifying individuals by analyzing face images captured from diverse poses. The primary objective is to recognize a person by processing a non-frontal view face image of this person. However, the substantial variations in facial appearance due to pose imply significant challenges in achieving accurate face recognition across different poses, making it a challenging task. In recent years, PIFR has predominantly focused on a holistic approach. DCNNs such as ArcFace, Elastic Face, and FaceNet have been utilized to generate face image embeddings, which are subsequently employed for face recognition with promising results.

This thesis aims to address PIFR from a novel local approach. In this approach, pose-invariant face recognition is conducted by combining head pose estimation (HPE), ensemble learning systems, and local feature descriptors. One ensemble system is trained for each subject on a face database. This ensemble system comprises a base learner set, where each base learner is exclusively trained with feature vectors extracted from the region surrounding an specific facial landmark within a gallery face image. In order to perform PIFR on an input image the following steps are carried out. First, the face and facial landmarks are detected in the image. Second, the facial ladmarks locations are processed to obtain a head pose descriptor, which is employed to perform HPE. According to the HPE results, some landmarks are selected to apply local feature descriptors on them, given that some of them might be self-occluded by the head pose. Third, the computed descriptors are distributed to their corresponding base learners. Indeed, one base learner is linked to a specific facial landmark. Fourth, the base learners’ outputs are combined for each ensemble system to compute its ensemble decision support value. Finally, the identity of the input face image is computed by selecting the ensemble system with the highest decision support.

In this work, three novel head pose descriptors (FAV, EFAV, NFCV) are proposed in conjunction with a non-linear regression model to perform HPE. Experimental results on the BIWI, and Pointing’04 databases showed that the proposed HPE approach outperforms several state-of-the-art works using the same databases. On the other hand, a novel local feature descriptor, called Landmark-specific SIFT (LS-SIFT) is proposed to improve the robustness of SIFT against changes in viewpoint and illumination. Additionally, three innovative base learner models (RD-CS, GMM, Mahalanobis Similarity) are developed in this thesis. The whole PIFR framework is tested on the CMU-PIE, Multi-PIE, and FERET databases. The obtained results are comparable to the state-of-the-art works. At the end, the combination of LS-SIFT with Mahalanobis Similarity yielded the best recognition results.
Chapter 1 Introduction 1
1.1 Motivation 1
1.2 Thesis Organization 3

Chapter 2 Backgrounds 4
2.1 Local feature descriptors 4
2.2 Facial landmark detection 8
2.3 Head Pose Estimation 13
2.4 Ensemble learning 15
2.5 Non-linear vector transformations via Neural Networks 16
2.6 Face recognition 18

Chapter 3 Related work 22
3.1 Works on head pose estimation using facial landmarks 22
3.1.1 Head Pose Estimation in the Wild Assisted by Facial Landmarks Based on Convolutional Neural Networks 22
3.1.2 Web-Shaped Model for Head Pose Estimation: An Approach for Best Exemplar Selection 24
3.2 Works on face recognition using ensemble learning 26
3.2.1 Research on face recognition based on ensemble learning 26
3.2.2 Ensemble of deep convolutional neural networks with gabor face representations for face recognition 27

Chapter 4 The proposed method 29
4.1 Head pose description, estimation, and classification 29
4.2 Learned non-linear mapping for enhancing the robustness of facial landmark descriptors against pose and illumination variability (LS-SIFT) 35
4.3 Facial landmark description 40
4.4 Pose-invariant face recognition using ensemble systems 41
4.4.1 Base learners 43
4.4.2 Ensemble face recognition models 47

Chapter 5 Experimental results and performance analysis 54
5.1 Head pose estimation databases 54
5.1.1 BIWI database 54
5.1.2 Pointing’04 database 55
5.2 Face recognition databases 56
5.2.1 CMU-PIE database 56
5.2.2 CMU Multi-PIE database 56
5.2.3 FERET database 58
5.3 Experiments on head pose estimation and classification 59
5.4 Experiments on Pose-Invariant Face Recognition 70

Chapter 6 Conclusions 85

Appendices 88
A Appendix: Published papers 88
A.1 Large Pose Detection and Facial Landmark Description for Pose-invariant Face Recognition 88
A.2 Pose-invariant Face Recognition via Facial Landmark based Ensemble Learning 89
B Appendix: Obtained Awards 91

References 92
[1] V. Kazemi and J. Sullivan, “One millisecond face alignment with an ensemble of regression
trees,” in 2014 IEEE Conference on Computer Vision and Pattern Recognition (CVPR 2014),
pp. 1867–1874, 2014.
[2] A. Bulat and G. Tzimiropoulos, “How far are we from solving the 2d & 3d face alignment
problem? (and a dataset of 230,000 3d facial landmarks),” in International Conference on
Computer Vision, 2017.
[3] R. Polikar, “Ensemble based systems in decision making,” IEEE Circuits and Systems Mag-
azine, vol. 6, no. 3, pp. 21–45, 2006.
[4] J. Xia, L. Cao, G. Zhang, and J. Liao, “Head pose estimation in the wild assisted by facial
landmarks based on convolutional neural networks,” IEEE Access, vol. 7, pp. 48470–48483,
2019.
[5] Y. Feng, X. An, and S. Li, “Research on face recognition based on ensemble learning,” in
2018 37th Chinese Control Conference (CCC), pp. 9078–9082, 2018.
[6] J. Y. Choi and B. Lee, “Ensemble of deep convolutional neural networks with gabor face rep-
resentations for face recognition,” IEEE Transactions on Image Processing, vol. 29, pp. 3270–
3281, 2020.
[7] Z. Cheng, X. Zhu, and S. Gong, “Surveillance face recognition challenge,” CoRR, 2018.
[8] R. Gonzalez and R. Woods, Digital Image Processing. Pearson, 4 ed., 2018.
[9] F. Perronnin, J. Sánchez, and T. Mensink, “Improving the fisher kernel for large-scale image
classification,” in Computer Vision – ECCV 2010 (K. Daniilidis, P. Maragos, and N. Paragios,
eds.), pp. 143–156, Springer Berlin Heidelberg, 2010.
[10] R. G. Cinbis, J. Verbeek, and C. Schmid, “Segmentation driven object detection with fisher
vectors,” in 2013 IEEE International Conference on Computer Vision, pp. 2968–2975, 2013.
[11] C. Qi, C. Shi, J. Xu, C. Wang, and B. Xiao, “Spatial weighted fisher vector for image retrieval,”
in 2017 IEEE International Conference on Multimedia and Expo (ICME), pp. 463–468, 2017.
[12] M. Okawa, “Vector of locally aggregated descriptors with kaze features for offline signature
verification,” in 2016 IEEE 5th Global Conference on Consumer Electronics, pp. 1–5, 2016.
[13] P. Zhang, Y. Wu, and B. Liu, “Leveraging local and global descriptors in parallel to search
correspondences for visual localization,” CoRR, 2020.
[14] R. Szeliski, Computer vision algorithms and applications. Springer, 2 ed., 2022.
[15] D. G. Lowe, “Distinctive image features from scale-invariant keypoints,” International Journal
of Computer Vision, vol. 60, pp. 91–110, 11 2004.
[16] E. Rublee, V. Rabaud, K. Konolige, and G. Bradski, “Orb: an efficient alternative to sift or
surf,” in Proceedings of the IEEE International Conference on Computer Vision, pp. 2564–
2571, 11 2011.
[17] S. Leutenegger, M. Chli, and R. Y. Siegwart, “Brisk: Binary robust invariant scalable key-
points,” in 2011 International Conference on Computer Vision, pp. 2548–2555, 2011.
[18] M. Sushama and E. Rajinikanth, “Face recognition using drlbp and sift feature extraction,” in
2018 International Conference on Communication and Signal Processing (ICCSP), pp. 994–
999, 2018.
[19] I. M. Al-Bahri, S. O. Fageeri, A. M. Said, and G. M. A. Sagayee, “A comparative study
between pca and sift algorithm for static face recognition,” in 2020 International Conference
on Computer, Control, Electrical, and Electronics Engineering (ICCCEEE), pp. 1–5, 2021.
[20] X. Han, T. Leung, Y. Jia, R. Sukthankar, and A. C. Berg, “Matchnet: Unifying feature and
metric learning for patch-based matching,” in 2015 IEEE Conference on Computer Vision
and Pattern Recognition (CVPR), pp. 3279–3286, 2015.
[21] E. Simo-Serra, E. Trulls, L. Ferraz, I. Kokkinos, P. Fua, and F. Moreno-Noguer, “Discrimi-
native learning of deep convolutional feature point descriptors,” in 2015 IEEE International
Conference on Computer Vision (ICCV), pp. 118–126, 2015.
[22] S. Zagoruyko and N. Komodakis, “Learning to compare image patches via convolutional neural
networks,” in 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR),
pp. 4353–4361, 2015.
[23] Z. Luo, T. Shen, L. Zhou, S. Zhu, R. Zhang, Y. Yao, T. Fang, and L. Quan, “Geodesc:
Learning local descriptors by integrating geometry constraints,” in Proceedings of the European
Conference on Computer Vision (ECCV), September 2018.
[24] S. D. Lin and P. Linares Otoya, “Large pose detection and facial landmark description for
pose-invariant face recognition,” in 2022 IEEE 5th International Conference on Knowledge
Innovation and Invention (ICKII ), pp. 143–148, 2022.
[25] S. D. Lin and P. E. Linares Otoya, “Pose-invariant face recognition via facial landmark based
ensemble learning,” IEEE Access, vol. 11, pp. 44221–44233, 2023.
[26] M. Taskiran, N. Kahraman, and C. E. Erdem, “Face recognition: Past, present and future (a
review),” Digital Signal Processing, vol. 106, 2020.
[27] X. Zhu, Z. Lei, X. Liu, H. Shi, and S. Z. Li, “Face alignment across large poses: A 3d solution,”
in 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 146–155,
2016.
[28] J. Deng, A. Roussos, G. G. Chrysos, E. Ververas, I. Kotsia, J. Shen, and S. Zafeiriou, “The
menpo benchmark for multi-pose 2d and 3d facial landmark localisation and tracking,” In-
ternational Journal of Computer Vision, vol. 127, pp. 599–624, 2018.
[29] Z. Zhang, L. Wang, Q. Zhu, S.-K. Chen, and Y. Chen, “Pose-invariant face recognition using
facial landmarks and weber local descriptor,” Knowledge-Based Systems, vol. 84, 04 2015.
[30] C. Alvarez Casado and M. Bordallo Lopez, “Real-time face alignment: evaluation methods,
training strategies and implementation optimization,” Journal of Real-Time Image Processing,
vol. 18, pp. 2239–2267, 12 2021.
[31] Y. Kartynnik, A. Ablavatski, I. Grishchenko, and M. Grundmann, “Real-time facial surface
geometry from monocular video on mobile gpus,” CoRR, vol. abs/1907.06724, 2019.
[32] V. Bazarevsky, Y. Kartynnik, A. Vakunov, K. Raveendran, and M. Grundmann, “Blazeface:
Sub-millisecond neural face detection on mobile gpus,” CoRR, vol. abs/1907.05047, 2019.
[33] A. Newell, K. Yang, and J. Deng, “Stacked hourglass networks for human pose estimation,” in
European Conference on Computer Vision (ECCV) 2016, pp. 483–499, Springer International
Publishing, 2016.
[34] A. Bulat and G. Tzimiropoulos, “Binarized convolutional landmark localizers for human pose
estimation and face alignment with limited resources,” in 2017 IEEE International Conference
on Computer Vision (ICCV), pp. 3726–3734, 2017.
[35] K. Khan, R. U. Khan, R. Leonardi, P. Migliorati, and S. Benini, “Head pose estimation: A
survey of the last ten years,” Signal Processing: Image Communication, vol. 99, p. 116479,
2021.
[36] S. Afroze and M. Hoque, Internet of Things and Connected Technologies, ch. Head Pose
Classification Based on Deep Convolution Networks, pp. 458–469. Springer, 05 2021.
[37] G. Guo, Y. Fu, C. R. Dyer, and T. S. Huang, “Head pose estimation: Classification or
regression?,” in 2008 19th International Conference on Pattern Recognition, pp. 1–4, 2008.
[38] S. Li, L. Sun, X. Ning, Y. Shi, and X. Dong, “Head pose classification based on line portrait,”
in 2019 International Conference on High Performance Big Data and Intelligent Systems
(HPBD&IS), pp. 186–189, 2019.
[39] R. Polikar, Ensemble Machine Learning: Methods and Applications, ch. Ensemble Learning.
Springer, 2012.
[40] K. Hornik, M. Stinchcombe, and H. White, “Multilayer feedforward networks are universal
approximators,” Neural Networks, vol. 2, no. 5, pp. 359–366, 1989.
[41] Y. Jia, E. Shelhamer, J. Donahue, S. Karayev, J. Long, R. Girshick, S. Guadarrama, and
T. Darrell, “Caffe: Convolutional architecture for fast feature embedding,” arXiv preprint
arXiv:1408.5093, 2014.
[42] M. Abadi, A. Agarwal, P. Barham, E. Brevdo, Z. Chen, C. Citro, G. S. Corrado, A. Davis,
J. Dean, M. Devin, S. Ghemawat, I. Goodfellow, A. Harp, G. Irving, M. Isard, Y. Jia, R. Joze-
fowicz, L. Kaiser, M. Kudlur, J. Levenberg, D. Mané, R. Monga, S. Moore, D. Murray, C. Olah,
M. Schuster, J. Shlens, B. Steiner, I. Sutskever, K. Talwar, P. Tucker, V. Vanhoucke, V. Va-
sudevan, F. Viégas, O. Vinyals, P. Warden, M. Wattenberg, M. Wicke, Y. Yu, and X. Zheng,
“TensorFlow: Large-scale machine learning on heterogeneous systems,” 2015. Software avail-
able from tensorflow.org.
[43] F. Pedregosa, G. Varoquaux, A. Gramfort, V. Michel, B. Thirion, O. Grisel, M. Blondel,
P. Prettenhofer, R. Weiss, V. Dubourg, J. Vanderplas, A. Passos, D. Cournapeau, M. Brucher,
M. Perrot, and E. Duchesnay, “Scikit-learn: Machine learning in Python,” Journal of Machine
Learning Research, vol. 12, pp. 2825–2830, 2011.
[44] C. Petpairote, S. Madarasmi, and K. Chamnongthai, “2d pose-invariant face recognition using
single frontal-view face database,” Wireless Personal Communications, vol. 118, pp. 2015–
2031, 2021.
[45] F. Wu, X.-Y. Jing, X. Dong, R. Hu, D. Yue, L. Wang, Y.-M. Ji, R. Wang, and G. Chen,
“Intraspectrum discrimination and interspectrum correlation analysis deep network for mul-
tispectral face recognition,” IEEE Transactions on Cybernetics, vol. 50, no. 3, pp. 1009–1022,
2020.
[46] C. Ding, J. Choi, D. Tao, and L. S. Davis, “Multi-directional multi-level dual-cross patterns
for robust face recognition,” CoRR, vol. abs/1401.5311, 2014.
[47] C. Ding, C. Xu, and D. Tao, “Multi-task pose-invariant face recognition,” IEEE Transactions
on Image Processing, vol. 24, no. 3, pp. 980–993, 2015.
[48] M. Gunther, A. Costa-Pazo, C. Ding, E. Boutellaa, G. Chiachia, H. Zhang, M. de As-
sis Angeloni, V. Struc, E. Khoury, E. Vazquez-Fernandez, D. Tao, M. Bengherabi, D. Cox,
S. Kiranyaz, T. de Freitas Pereira, J. Zganec-Gros, E. Argones-Rua, N. Pinto, M. Gabbouj,
F. Simoes, S. Dobrisek, D. Gonzalez-Jimenez, A. Rocha, M. U. Neto, N. Pavesic, A. Falcao,
R. Violato, and S. Marcel, “The 2013 face recognition evaluation in mobile environment,” in
2013 International Conference on Biometrics (ICB), pp. 1–7, 2013.
[49] T. Q. Chung, H. C. Huyen, and D. V. Sang, “A novel generative model to synthesize face
images for pose-invariant face recognition,” in 2020 International Conference on Multimedia
Analysis and Pattern Recognition (MAPR), pp. 1–6, 2020.
[50] P. Barra, S. Barra, C. Bisogni, M. De Marsico, and M. Nappi, “Web-shaped model for head
pose estimation: An approach for best exemplar selection,” IEEE Transactions on Image
Processing, vol. 29, pp. 5457–5468, 2020.
[51] K. Sundararajan and D. L. Woodard, “Head pose estimation in the wild using approximate
view manifolds,” in 2015 IEEE Conference on Computer Vision and Pattern Recognition
Workshops (CVPRW), pp. 50–58, 2015.
[52] Google, “Tensorflow api documentation.” Online (accessed on 13/06/2023), June 2023. Avail-
able at: https://www.tensorflow.org/api docs.
[53] Keras, “Keras api documentation.” Online (accessed on 13/06/2023), June 2023. Available
at: https://keras.io/api/.
[54] R. Tibshirani, “Regression shrinkage and selection via the lasso,” Journal of the Royal Sta-
tistical Society. Series B (Methodological), vol. 58, no. 1, pp. 267–288, 1996.
[55] Intel, Open Source Computer Vision (OpenCV) documentation, 4.5.5 ed., November 2022. See
https://docs.opencv.org/4.5.5/.
[56] L. I. Kuncheva, Combining Pattern Classifiers: Methods and Algorithms, ch. 3. Wiley Pub-
lishing, 2nd ed., 2014.
[57] N. C. Oza and K. Tumer, “Input decimation ensembles: Decorrelation through dimensionality
reduction,” in Multiple Classifier Systems, pp. 238–247, Springer Berlin Heidelberg, 2001.
[58] J. Lu, “A survey on bayesian inference for gaussian mixture model.” Online, 2021. Available
at: https://arxiv.org/abs/2108.11753.
[59] G. Fanelli, M. Dantone, J. Gall, A. Fossati, and L. Van Gool, “Random forests for real time
3d face analysis,” International Journal of Computer Vision, vol. 101, 02 2013.
[60] Y. Wang, W. Liang, J. Shen, Y. Jia, and L.-F. Yu, “A deep coarse-to-fine network for head
pose estimation from synthetic data,” Pattern Recognition, vol. 94, pp. 196–206, 2019.
[61] N. Gourier and J. Crowley, “Estimating face orientation from robust detection of salient facial
structures,” FG Net Workshop on Visual Observation of Deictic Gestures, 01 2004.
[62] T. Sim, S. Baker, and M. Bsat, “The cmu pose, illumination, and expression database,”
IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 25, pp. 1615 – 1618,
December 2003.
[63] R. Gross, I. Matthews, J. Cohn, T. Kanade, and S. Baker, “Multi-pie,” in 2008 8th IEEE
International Conference on Automatic Face & Gesture Recognition, pp. 1–8, 2008.
[64] P. Phillips, H. Moon, P. Rauss, and S. Rizvi, “The feret evaluation methodology for face-
recognition algorithms,” in Proceedings of IEEE Computer Society Conference on Computer
Vision and Pattern Recognition (CVPR 1997), pp. 137–143, 1997.
[65] V. Drouard, R. Horaud, A. Deleforge, S. O. Ba, and G. D. Evangelidis, “Robust head-pose
estimation based on partially-latent mixture of linear regression,” CoRR, vol. abs/1603.09732,
2016.
[66] N. Ruiz, E. Chong, and J. M. Rehg, “Fine-grained head pose estimation without keypoints,”
CoRR, vol. abs/1710.00925, 2017.
[67] H.-W. Hsu, T.-Y. Wu, S. Wan, W. H. Wong, and C.-Y. Lee, “Quatnet: Quaternion-based
head pose estimation with multiregression loss,” IEEE Transactions on Multimedia, vol. 21,
no. 4, pp. 1035–1046, 2019.
[68] J. Wang, F. Ullah, Y. Cai, and J. Li, “Non-stationary representation for continuity aware
head pose estimation via deep neural decision trees,” IEEE Access, vol. 7, pp. 181947–181958,
2019.
[69] M. T. Vo, T. Nguyen, and T. Le, “Robust head pose estimation using extreme gradient
boosting machine on stacked autoencoders neural network,” IEEE Access, vol. 8, pp. 3687–
3694, 2020.
[70] E. A. Mostafa and A. A. Farag, “Dynamic weighting of facial features for automatic pose-
invariant face recognition,” in 2012 Ninth Conference on Computer and Robot Vision, pp. 411–
416, 2012.
[71] A. Moeini and H. Moeini, “Real-world and rapid face recognition toward pose and expres-
sion variations via feature library matrix,” IEEE Transactions on Information Forensics and
Security, vol. 10, no. 5, pp. 969–984, 2015.
[72] L. A. Cament, F. J. Galdames, K. W. Bowyer, and C. A. Perez, “Face recognition under pose
variation with local gabor features enhanced by active shape and statistical models,” Pattern
Recognition, vol. 48, no. 11, pp. 3371–3384, 2015.
[73] L.-F. Zhou, Y.-W. Du, W.-S. Li, J.-X. Mi, and X. Luan, “Pose-robust face recognition with
huffman-lbp enhanced by divide-and-rule strategy,” Pattern Recognition, vol. 78, pp. 43–55,
2018.
[74] Y. Tai, J. Yang, Y. Zhang, L. Luo, J. Qian, and Y. Chen, “Face recognition with pose
variations and misalignment via orthogonal procrustes regression,” IEEE Transactions on
Image Processing, vol. 25, no. 6, pp. 2673–2683, 2016.
[75] H. Lin, H. Ma, W. Gong, and C. Wang, “Non-frontal face recognition method with a side-
face-correction generative adversarial networks,” in 2022 3rd International Conference on
Computer Vision, Image and Deep Learning & International Conference on Computer Engi-
neering and Applications (CVIDL & ICCEA), pp. 563–567, 2022.
[76] Z. Zhu, P. Luo, X. Wang, and X. Tang, “Deep learning identity-preserving face space,” in
2013 IEEE International Conference on Computer Vision, pp. 113–120, 2013.
[77] A. Li, S. Shan, and W. Gao, “Coupled bias–variance tradeoff for cross-pose face recognition,”
IEEE Transactions on Image Processing, vol. 21, no. 1, pp. 305–315, 2012.
[78] J. Yan, Y. Mei, X. Liu, C. Dai, and T. Yu, “Patch-wise normalization for pose-invariant face
recognition from single sample,” in 2018 IEEE International Conference on Internet of Things
(iThings) and IEEE Green Computing and Communications (GreenCom) and IEEE Cyber,
Physical and Social Computing (CPSCom) and IEEE Smart Data (SmartData), pp. 712–715,
2018.
 
 
 
 
第一頁 上一頁 下一頁 最後一頁 top
* *