Publication:
Profiling Concurrent Vision Inference Workloads on NVIDIA Jetson
| cris.virtual.department | #PLACEHOLDER_PARENT_METADATA_VALUE# | |
| cris.virtual.department | #PLACEHOLDER_PARENT_METADATA_VALUE# | |
| cris.virtual.department | #PLACEHOLDER_PARENT_METADATA_VALUE# | |
| cris.virtual.department | #PLACEHOLDER_PARENT_METADATA_VALUE# | |
| cris.virtual.orcid | 0009-0007-7993-6686 | |
| cris.virtual.orcid | 0000-0001-5817-7886 | |
| cris.virtual.orcid | 0000-0002-1428-0301 | |
| cris.virtual.orcid | 0000-0003-4408-6523 | |
| cris.virtualsource.department | cd809ef4-6c63-4775-8d82-298a275c14a9 | |
| cris.virtualsource.department | c914e7c0-7efb-4c2b-87b4-ae881ddf37db | |
| cris.virtualsource.department | 891de1ef-83e1-4ca0-ae39-c3daab198fe5 | |
| cris.virtualsource.department | 48554e7b-ff43-44b9-9f84-0dcbd96416d7 | |
| cris.virtualsource.orcid | cd809ef4-6c63-4775-8d82-298a275c14a9 | |
| cris.virtualsource.orcid | c914e7c0-7efb-4c2b-87b4-ae881ddf37db | |
| cris.virtualsource.orcid | 891de1ef-83e1-4ca0-ae39-c3daab198fe5 | |
| cris.virtualsource.orcid | 48554e7b-ff43-44b9-9f84-0dcbd96416d7 | |
| dc.contributor.author | Chakraborty, Abhinaba | |
| dc.contributor.author | Tavernier, Wouter | |
| dc.contributor.author | Kourtis, Akis | |
| dc.contributor.author | Pickavet, Mario | |
| dc.contributor.author | Oikonomakis, Andreas | |
| dc.contributor.author | Colle, Didier | |
| dc.date.accessioned | 2026-03-30T08:18:31Z | |
| dc.date.available | 2026-03-30T08:18:31Z | |
| dc.date.createdwos | 2025-09-26 | |
| dc.date.issued | 2025 | |
| dc.description.abstract | The necessity of processing real-time data at the network edge is growing. Low-power AI accelerators, especially edge GPUs, help meet this demand by mitigating cloud-related latency and bandwidth issues. However, GPUs remain underutilised, even in heavy workloads, due to a limited understanding of resource sharing in edge computing. This work analyses key GPU metrics: utilisation, memory, streaming multiprocessors (SMs), and tensorcores on NVIDIA Jetson devices under concurrent vision-inference workloads. Our findings show that while GPU utilisation can reach 100 % with optimisations, SMs and tensor cores often run at only 15-30 % capacity. | |
| dc.description.wosFundingText | The research work presented in this article has been supported by the European Commission under the Horizon Europe Programme and the OASEES project (no. 101092702). | |
| dc.identifier.doi | 10.1109/ISPASS64960.2025.00043 | |
| dc.identifier.isbn | 979-8-3315-0295-9 | |
| dc.identifier.issn | 2994-9513 | |
| dc.identifier.uri | https://imec-publications.be/handle/20.500.12860/58954 | |
| dc.language.iso | eng | |
| dc.provenance.editstepuser | greet.vanhoof@imec.be | |
| dc.publisher | IEEE COMPUTER SOC | |
| dc.source.beginpage | 359 | |
| dc.source.conference | IEEE International Symposium on Performance Analysis of Systems and Software (ISPASS) | |
| dc.source.conferencedate | 2025-05-11 | |
| dc.source.conferencelocation | Gent | |
| dc.source.endpage | 361 | |
| dc.source.journal | 2025 IEEE INTERNATIONAL SYMPOSIUM ON PERFORMANCE ANALYSIS OF SYSTEMS AND SOFTWARE, ISPASS | |
| dc.source.numberofpages | 3 | |
| dc.title | Profiling Concurrent Vision Inference Workloads on NVIDIA Jetson | |
| dc.type | Proceedings paper | |
| dspace.entity.type | Publication | |
| imec.internal.crawledAt | 2025-10-22 | |
| imec.internal.source | crawler | |
| Files | ||
| Publication available in collections: |