Mostrar el registro sencillo del ítem

dc.contributor.authorBehera, Adarsh Prasad 
dc.contributor.authorDaubaris, Paulius
dc.contributor.authorBravo Aramburu, Iñaki 
dc.contributor.authorGallego Delgado, José 
dc.contributor.authorMorabito, Roberto
dc.contributor.authorWidmer, Joerg 
dc.contributor.authorChampati, Jaya Prakash 
dc.date.accessioned2025-09-04T15:52:25Z
dc.date.available2025-09-04T15:52:25Z
dc.date.issued2025-07
dc.identifier.urihttps://hdl.handle.net/20.500.12761/1958
dc.description.abstractOn-device inference offers significant benefits in edge ML systems, such as improved energy efficiency, responsiveness, and privacy, compared to traditional centralized approaches. However, the resource constraints of embedded devices limit their use to simple inference tasks, creating a trade-off between efficiency and capability. In this context, the Hierarchical Inference (HI) system has emerged as a promising solution that augments the capabilities of the local ML by offloading selected samples to an edge server/cloud for remote ML inference. Existing works, primarily based on simulations, demonstrate that HI improves accuracy. However, they fail to account for the latency and energy consumption in real-world deployments, nor do they consider three key heterogeneous components that characterize ML-enabled IoT systems: hardware, network connectivity, and models. To bridge this gap, this paper systematically evaluates HI against standalone on-device inference by analyzing accuracy, latency, and energy trade-offs across five devices and three image classification datasets. Our findings show that, for a given accuracy requirement, the HI approach we designed achieved up to 73% lower latency and up to 77% lower device energy consumption than an on-device inference system. Despite these gains, HI introduces a fixed energy and latency overhead from on-device inference for all samples. To address this, we propose a hybrid system called Early Exit with HI (EE-HI) and demonstrate that, compared to HI, EE-HI reduces the latency up to 59.7% and lowers the device’s energy consumption up to 60.4%. These findings demonstrate the potential of HI and EE-HI to enable more efficient ML in IoT systems.es
dc.description.sponsorshipMinisterio de Asuntos Económicos y Transformación Digitales
dc.description.sponsorshipEuropean Uniones
dc.description.sponsorshipMinisterio de Trabajo y Economía Sociales
dc.language.isoenges
dc.publisherIEEEes
dc.titleExploring the Boundaries of On-Device Inference: When Tiny Falls Short, Go Hierarchicales
dc.typejournal articlees
dc.journal.titleIEEE Internet of Things Journales
dc.rights.accessRightsopen accesses
dc.identifier.doi10.1109/JIOT.2025.3583477es
dc.relation.projectIDinfo:eu-repo/grantAgreement/EC/H2020-MSCA-2021-PF-01/101062011es
dc.relation.projectNameRISC-6G (Reconfigurable Intelligent Surfaces and Low-power Technologies for Communication and Sensing in 6G Mobile Networks)es
dc.relation.projectNameMAP-6G (Machine Learning-based Privacy Preserving Analytics for 6G Mobile Networks)es
dc.relation.projectNameDIME (Distributed Inference for Energy-efficient Monitoring at the Network Edge Note)es
dc.relation.projectNamePrograma Investigoes
dc.description.refereedTRUEes
dc.description.statuspubes


Ficheros en el ítem

Este ítem aparece en la(s) siguiente(s) colección(ones)

Mostrar el registro sencillo del ítem