TY - GEN
T1 - Benchmark of DNN Model Search at Deployment Time
AU - Zhou, Lixi
AU - Jain, Arindam
AU - Wang, Zijie
AU - Das, Amitabh
AU - Yang, Yingzhen
AU - Zou, Jia
N1 - Publisher Copyright: © 2022 ACM.
PY - 2022/7/6
Y1 - 2022/7/6
N2 - Deep learning has become the most popular direction in machine learning and artificial intelligence. However, the preparation of training data, as well as model training, are often time-consuming and become the bottleneck of the end-to-end machine learning lifecycle. Reusing models for inferring a dataset can avoid the costs of retraining. However, when there are multiple candidate models, it is challenging to discover the right model for reuse. Although there exist a number of model sharing platforms such as ModelDB, TensorFlow Hub, PyTorch Hub, and DLHub, most of these systems require model uploaders to manually specify the details of each model and model downloaders to screen keyword search results for selecting a model. We are lacking a highly productive model search tool that selects models for deployment without the need for any manual inspection and/or labeled data from the target domain. This paper proposes multiple model search strategies including various similarity-based approaches and non-similarity-based approaches. We design, implement and evaluate these approaches on multiple model inference scenarios, including activity recognition, image recognition, text classification, natural language processing, and entity matching. The experimental evaluation showed that our proposed asymmetric similarity-based measurement, adaptivity, outperformed symmetric similarity-based measurements and non-similarity-based measurements in most of the workloads.
AB - Deep learning has become the most popular direction in machine learning and artificial intelligence. However, the preparation of training data, as well as model training, are often time-consuming and become the bottleneck of the end-to-end machine learning lifecycle. Reusing models for inferring a dataset can avoid the costs of retraining. However, when there are multiple candidate models, it is challenging to discover the right model for reuse. Although there exist a number of model sharing platforms such as ModelDB, TensorFlow Hub, PyTorch Hub, and DLHub, most of these systems require model uploaders to manually specify the details of each model and model downloaders to screen keyword search results for selecting a model. We are lacking a highly productive model search tool that selects models for deployment without the need for any manual inspection and/or labeled data from the target domain. This paper proposes multiple model search strategies including various similarity-based approaches and non-similarity-based approaches. We design, implement and evaluate these approaches on multiple model inference scenarios, including activity recognition, image recognition, text classification, natural language processing, and entity matching. The experimental evaluation showed that our proposed asymmetric similarity-based measurement, adaptivity, outperformed symmetric similarity-based measurements and non-similarity-based measurements in most of the workloads.
UR - http://www.scopus.com/inward/record.url?scp=85137686816&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85137686816&partnerID=8YFLogxK
U2 - 10.1145/3538712.3538725
DO - 10.1145/3538712.3538725
M3 - Conference contribution
T3 - ACM International Conference Proceeding Series
BT - Scientific and Statistical Database Management - 34th International Conference, SSDBM 2022 - Proceedings
A2 - Pourabbas, Elaheh
A2 - Zhou, Yongluan
A2 - Li, Yuchen
A2 - Yang, Bin
PB - Association for Computing Machinery
T2 - 34th International Conference on Scientific and Statistical Database Management, SSDBM 2022
Y2 - 6 July 2022
ER -