Logo des Repositoriums
 

Continuous Training and Deployment of Deep Learning Models

dc.contributor.authorPrapas, Ioannis
dc.contributor.authorDerakhshan, Behrouz
dc.contributor.authorMahdiraji, Alireza Rezaei
dc.contributor.authorMarkl, Volker
dc.date.accessioned2022-01-27T13:27:55Z
dc.date.available2022-01-27T13:27:55Z
dc.date.issued2021
dc.description.abstractDeep Learning (DL) has consistently surpassed other Machine Learning methods and achieved state-of-the-art performance in multiple cases. Several modern applications like financial and recommender systems require models that are constantly updated with fresh data. The prominent approach for keeping a DL model fresh is to trigger full retraining from scratch when enough new data are available. However, retraining large and complex DL models is time-consuming and compute-intensive. This makes full retraining costly, wasteful, and slow. In this paper, we present an approach to continuously train and deploy DL models. First, we enable continuous training through proactive training that combines samples of historical data with new streaming data. Second, we enable continuous deployment through gradient sparsification that allows us to send a small percentage of the model updates per training iteration. Our experimental results with LeNet5 on MNIST and modern DL models on CIFAR-10 show that proactive training keeps models fresh with comparable—if not superior—performance to full retraining at a fraction of the time. Combined with gradient sparsification, sparse proactive training enables very fast updates of a deployed model with arbitrarily large sparsity, reducing communication per iteration up to four orders of magnitude, with minimal—if any—losses in model quality. Sparse training, however, comes at a price; it incurs overhead on the training that depends on the size of the model and increases the training time by factors ranging from 1.25 to 3 in our experiments. Arguably, a small price to pay for successfully enabling the continuous training and deployment of large DL models.de
dc.identifier.doi10.1007/s13222-021-00386-8
dc.identifier.pissn1610-1995
dc.identifier.urihttp://dx.doi.org/10.1007/s13222-021-00386-8
dc.identifier.urihttps://dl.gi.de/handle/20.500.12116/38049
dc.publisherSpringer
dc.relation.ispartofDatenbank-Spektrum: Vol. 21, No. 3
dc.relation.ispartofseriesDatenbank-Spektrum
dc.subjectContinuous deployment
dc.subjectContinuous training
dc.subjectDeep learning
dc.subjectModel deployment
dc.titleContinuous Training and Deployment of Deep Learning Modelsde
dc.typeText/Journal Article
gi.citation.endPage212
gi.citation.startPage203

Dateien