Accelerating Machine Learning Research Using Transfer Learning
DOI:
https://doi.org/10.17010/ijcs/2018/v3/i2/123212Keywords:
Accelerate
, Machine Learning, Research, Transfer LearningJanuary 4
, 2018, revised January 30, accepted February 6, Date of publication, March 6, 2018.Abstract
Research and development in machine learning involves a continuous cycle of training, testing and tuning. This is generally a time and computation intensive process, and hence leads to a slow experiment cycle. We intended to accelerate this cycle by using the transfer learning technique. Using transfer learning we can produce the same results in a matter of minutes, which previously took weeks to generate. This technique is also highly beneficial when the training data is scarce and computing infrastructure is limited. Transfer learning has been applied to many fields ranging from image classification, medical diagnosis of tumors, diabetic retinopathy, generating captions, self-driving cars, etc. Transfer learning is considered to be the next driver of machine learning success. In this paper, we introduce transfer learning, describe its implementation, and talk about its various applications and benefits.Downloads
Downloads
Published
How to Cite
Issue
Section
References
R. Rabiser and R. Torkar, "Guest Editors' Introduction : Special Section on Software Eng. and Advanced Appl.," Inform. and Software Technol., 2015, vol. 67, pp. 236. DOI: https://doi.org/10.1016/j.infsof.2015.06.005
L. Fei-Fei, and O. Russakovsky, "Anal. of large-scale visual recognition," Bay Area Vision Meeting, October, 2013.
K. Weiss, T. M. Khoshgoftaar, and D. Wang, "A survey of transfer learning," J. of Big Data, vol. 3, no. 9, 2016. https://doi.org/10.1186/s40537-016-0043-6
S. J. Pan and Q. Yang, "A Survey on Transfer Learning," IEEE Trans. on Knowledge and Data Eng., vol. 22, no. 10, pp. 1345-1359, 2010. doi: 10.1109/TKDE.2009.191
A. Krizhevsky, I. Sutskever, and G. E. Hinton, "ImageNet classification with deep convolutional neural networks," Commun. of the ACM, vol. 60, no. 6, pp. 84-90, 2017. DOI: https://doi.org/10.1145/3065386
Q. Yang, "When Deep Learning Meets Transfer Learning," In Proc. of the 2017 ACM on Conf. on Inform. and Knowledge Manage. (CIKM '17)," ACM, New York, NY, USA, pp. 5-5, 2017. DOI: https://doi.org/10.1145/3132847.3137175
X. Chen, J. Chen, D. Z. Chen, and X. S. Hu, "Optimizing memory efficiency for convolution kernels on Kepler GPUs," In Proc. of the 54th Annu. Design Automation Conf. 2017 (DAC '17). ACM, New York, NY, USA, Article 68, 6 pages, 2017. DOI: https://doi.org/10.1145/3061639.3062297 N. P. Jouppi et. al., "In-datacenter performance anal. of a tensor process. unit," Proc. of the 44th Annu. Int. Symp. on Comput. Architecture, 2017. DOI: 10.1145/3079856.3080246
C. Szegedy, V. Vanhoucke, S. Ioffe, J. Shlens, and Z. Wojna, "Rethinking the inception architecture for computer vision," Comput. Vision and Pattern Recognition (CVPR), 2016. DOI: 10.1109/CVPR.2016.308
A. G. Howard, M. Zhu, B. Chen, D. Kalenichenko, W. Wang, T. Weyand, M. Andreetto, and H. Adam, "Mobilenets: Efficient convolutional neural networks for mobile vision applications," arXiv preprint arXiv:1704.04861, 2017.
Abadi et. al., "TensorFlow: A system for large-scale mach. learning," In Proc. of the 12th USENIX Conf. on Operating Syst. Design and Implementation (OSDI'16), USENIX Assoc., Berkeley, CA, USA, pp. 265-283, 2016.