Date of Award
12-2021
Document Type
Dissertation
Degree Name
Doctor of Philosophy (PhD)
Department
Computer Engineering and Sciences
First Advisor
Eraldo Ribeiro
Second Advisor
Munevver Mine Subasi
Third Advisor
Debasis Mitra
Fourth Advisor
Marius Silaghi
Abstract
Head-pose estimation from facial images is an important research topic in computer-vision. It has many applications in detecting the focus of attention, monitoring driver behavior, and human-computer interaction. As with other computer-vision topics, recent research on head-pose estimation has been focused on using deep convolutional neural networks (CNNs). Although deeper networks improve prediction accuracy, they suffer from dependency on expensive hardware such as GPUs to perform real-time inference. As a result, CNN model compression becomes an important concept. In this work, we propose a novel CNN compression method by combing weight pruning and knowledge distillation. Additionally, we improve the state-of-the-art headpose estimation model with image-augmentation and transfer-learning. We apply our compression method to a baseline head-pose estimation model and validate the performance of the compression by creating validation scenarios. Additionally, we test our compression method on different CNN architectures and classification tasks to show the effectiveness of our compression method.
Recommended Citation
Aghli, Nima, "Explainable Transfer-Learning and Knowledge Distillation for Fast and Accurate Head-Pose Estimation" (2021). Theses and Dissertations. 629.
https://repository.fit.edu/etd/629
Comments
Copyright held by author