- Author
- Pieter Van Molle (UGent) , Cedric De Boom, Tim Verbelen (UGent) , Bert Vankeirsbilck (UGent) , Jonas De Vylder (UGent) , Bart Diricx, Pieter Simoens (UGent) and Bart Dhoedt (UGent)
- Organization
- Abstract
- Deep neural networks have achieved state-of-the-art performance in image classification. Due to this success, deep learning is now also being applied to other data modalities such as multispectral images, lidar and radar data. However, successfully training a deep neural network requires a large reddataset. Therefore, transitioning to a new sensor modality (e.g., from regular camera images to multispectral camera images) might result in a drop in performance, due to the limited availability of data in the new modality. This might hinder the adoption rate and time to market for new sensor technologies. In this paper, we present an approach to leverage the knowledge of a teacher network, that was trained using the original data modality, to improve the performance of a student network on a new data modality: a technique known in literature as knowledge distillation. By applying knowledge distillation to the problem of sensor transition, we can greatly speed up this process. We validate this approach using a multimodal version of the MNIST dataset. Especially when little data is available in the new modality (i.e., 10 images), training with additional teacher supervision results in increased performance, with the student network scoring a test set accuracy of 0.77, compared to an accuracy of 0.37 for the baseline. We also explore two extensions to the default method of knowledge distillation, which we evaluate on a multimodal version of the CIFAR-10 dataset: an annealing scheme for the hyperparameter alpha and selective knowledge distillation. Of these two, the first yields the best results. Choosing the optimal annealing scheme results in an increase in test set accuracy of 6%. Finally, we apply our method to the real-world use case of skin lesion classification.
- Keywords
- DIAGNOSIS, MELANOMA, deep learning, knowledge distillation, cross-modal distillation, sensor, upgrade, skin lesion classification, multispectral imaging
Downloads
-
7992.pdf
- full text (Published version)
- |
- open access
- |
- |
- 819.54 KB
Citation
Please use this url to cite or link to this publication: http://hdl.handle.net/1854/LU-8730930
- MLA
- Van Molle, Pieter, et al. “Data-Efficient Sensor Upgrade Path Using Knowledge Distillation.” SENSORS, vol. 21, no. 19, 2021, doi:10.3390/s21196523.
- APA
- Van Molle, P., De Boom, C., Verbelen, T., Vankeirsbilck, B., De Vylder, J., Diricx, B., … Dhoedt, B. (2021). Data-efficient sensor upgrade path using knowledge distillation. SENSORS, 21(19). https://doi.org/10.3390/s21196523
- Chicago author-date
- Van Molle, Pieter, Cedric De Boom, Tim Verbelen, Bert Vankeirsbilck, Jonas De Vylder, Bart Diricx, Pieter Simoens, and Bart Dhoedt. 2021. “Data-Efficient Sensor Upgrade Path Using Knowledge Distillation.” SENSORS 21 (19). https://doi.org/10.3390/s21196523.
- Chicago author-date (all authors)
- Van Molle, Pieter, Cedric De Boom, Tim Verbelen, Bert Vankeirsbilck, Jonas De Vylder, Bart Diricx, Pieter Simoens, and Bart Dhoedt. 2021. “Data-Efficient Sensor Upgrade Path Using Knowledge Distillation.” SENSORS 21 (19). doi:10.3390/s21196523.
- Vancouver
- 1.Van Molle P, De Boom C, Verbelen T, Vankeirsbilck B, De Vylder J, Diricx B, et al. Data-efficient sensor upgrade path using knowledge distillation. SENSORS. 2021;21(19).
- IEEE
- [1]P. Van Molle et al., “Data-efficient sensor upgrade path using knowledge distillation,” SENSORS, vol. 21, no. 19, 2021.
@article{8730930, abstract = {{Deep neural networks have achieved state-of-the-art performance in image classification. Due to this success, deep learning is now also being applied to other data modalities such as multispectral images, lidar and radar data. However, successfully training a deep neural network requires a large reddataset. Therefore, transitioning to a new sensor modality (e.g., from regular camera images to multispectral camera images) might result in a drop in performance, due to the limited availability of data in the new modality. This might hinder the adoption rate and time to market for new sensor technologies. In this paper, we present an approach to leverage the knowledge of a teacher network, that was trained using the original data modality, to improve the performance of a student network on a new data modality: a technique known in literature as knowledge distillation. By applying knowledge distillation to the problem of sensor transition, we can greatly speed up this process. We validate this approach using a multimodal version of the MNIST dataset. Especially when little data is available in the new modality (i.e., 10 images), training with additional teacher supervision results in increased performance, with the student network scoring a test set accuracy of 0.77, compared to an accuracy of 0.37 for the baseline. We also explore two extensions to the default method of knowledge distillation, which we evaluate on a multimodal version of the CIFAR-10 dataset: an annealing scheme for the hyperparameter alpha and selective knowledge distillation. Of these two, the first yields the best results. Choosing the optimal annealing scheme results in an increase in test set accuracy of 6%. Finally, we apply our method to the real-world use case of skin lesion classification.}}, articleno = {{6523}}, author = {{Van Molle, Pieter and De Boom, Cedric and Verbelen, Tim and Vankeirsbilck, Bert and De Vylder, Jonas and Diricx, Bart and Simoens, Pieter and Dhoedt, Bart}}, issn = {{1424-8220}}, journal = {{SENSORS}}, keywords = {{DIAGNOSIS,MELANOMA,deep learning,knowledge distillation,cross-modal distillation,sensor,upgrade,skin lesion classification,multispectral imaging}}, language = {{eng}}, number = {{19}}, pages = {{15}}, title = {{Data-efficient sensor upgrade path using knowledge distillation}}, url = {{http://doi.org/10.3390/s21196523}}, volume = {{21}}, year = {{2021}}, }
- Altmetric
- View in Altmetric
- Web of Science
- Times cited: