Mostrar el registro sencillo del ítem

dc.contributor.authorFresa, Andrea 
dc.contributor.authorChampati, Jaya Prakash 
dc.date.accessioned2023-07-19T10:00:01Z
dc.date.available2023-07-19T10:00:01Z
dc.date.issued2023-04-19
dc.identifier.issn1045-9219es
dc.identifier.urihttps://hdl.handle.net/20.500.12761/1731
dc.description.abstractWith the emergence of edge computing, the problem of offloading jobs between an Edge Device (ED) and an Edge Server (ES) received significant attention in the past. Motivated by the fact that an increasing number of applications are using Machine Learning (ML) inference from the data samples collected at the EDs, we study the problem of offloading inference jobs by considering the following novel aspects: 1) in contrast to a typical computational job, the processing time of an inference job depends on the size of the ML model, and 2) recently proposed Deep Neural Networks (DNNs) for resource-constrained devices provide the choice of scaling down the model size by trading off the inference accuracy. Considering that multiple ML models are available at the ED, and a powerful ML model is available at the ES, we formulate an Integer Linear Programming (ILP) problem with the objective of maximizing the total inference accuracy of n data samples at the ED subject to a time constraint T on the makespan. Noting that the problem is NP-hard, we propose an approximation algorithm Accuracy Maximization using LP-Relaxation and Rounding (AMR 2 ) and prove that it results in a makespan at most 2T and achieves a total accuracy that is lower by a small constant from the optimal total accuracy implying that AMR 2 is asymptotically optimal. Further, if the data samples are identical we propose Accuracy Maximization using Dynamic Programming (AMDP), an optimal pseudo-polynomial time algorithm. Furthermore, we extend AMR 2 for the case of multiple ESs, where each ES is equipped with a powerful ML model. As proof of concept, we implemented AMR 2 on a Raspberry Pi, equipped with MobileNets, that is connected to a server equipped with ResNet, and studied the total accuracy and makespan performance of AMR 2 for image classification.es
dc.description.sponsorshipJaya Prakash Champaties
dc.language.isoenges
dc.publisherIEEEes
dc.titleOffloading Algorithms for Maximizing Inference Accuracy on Edge Device in an Edge Intelligence Systemes
dc.typejournal articlees
dc.journal.titleIEEE Transactions on Parallel and Distributed Systemses
dc.type.hasVersionVoRes
dc.rights.accessRightsopen accesses
dc.volume.number34es
dc.issue.number7es
dc.identifier.doi10.1109/TPDS.2023.3267458es
dc.page.final2039es
dc.page.initial2025es
dc.subject.keywordEdge Intelligencees
dc.subject.keywordEdge Computinges
dc.subject.keywordIoTes
dc.subject.keywordData modelses
dc.subject.keywordComputational modelinges
dc.subject.keywordInference algorithmses
dc.subject.keywordCostses
dc.subject.keywordServerses
dc.subject.keywordApproximation algorithmses
dc.subject.keywordSchedulinges
dc.subject.keywordcomputational complexityes
dc.subject.keyworddeep learning (artificial intelligence)es
dc.subject.keyworddynamic programminges
dc.subject.keywordedge computinges
dc.subject.keywordinference mechanismses
dc.subject.keywordinteger programminges
dc.subject.keywordlinear programminges
dc.subject.keywordresource allocationes
dc.description.refereedTRUEes
dc.description.statuspubes


Ficheros en el ítem

Este ítem aparece en la(s) siguiente(s) colección(ones)

Mostrar el registro sencillo del ítem