Abstract | ||
---|---|---|
We present an orchestration scheme for Deep Neural Network (DNN) model serving, capable of computation distribution over the device-to-cloud continuum and low-latency inference. Our system allows automated layer-wise splitting of DNN structures and their adaptive distribution over compute hosts, providing an execution environment for collaborative inference. Model deployment and its self-adaptatio... |
Year | DOI | Venue |
---|---|---|
2021 | 10.1109/IC2E52221.2021.00046 | 2021 IEEE International Conference on Cloud Engineering (IC2E) |
Keywords | DocType | ISSN |
Deep learning,Adaptation models,Runtime,Adaptive systems,Computational modeling,Conferences,Inference algorithms | Conference | 2373-3845 |
ISBN | Citations | PageRank |
978-1-6654-4970-0 | 0 | 0.34 |
References | Authors | |
0 | 3 |
Name | Order | Citations | PageRank |
---|---|---|---|
Matthias Reisinger | 1 | 0 | 0.34 |
Pantelis A. Frangoudis | 2 | 13 | 4.26 |
Schahram Dustdar | 3 | 9347 | 575.71 |