Abstract
Multi-camera real-time vision at the Edge is facilitated by low-latency distributed data stores. In this paper, we take the position that latency criticality in the challenging operating conditions at the Edge can only be attained through application specific designs incorporating autonomous computing techniques. In our initial prototype, we implement a key-value Edge data store that autonomously monitors run-time conditions to maintain latency-criticality of one class of data (feature vectors), while sacrificing the latency and accuracy of another class of data (keyframes). Early results show a median latency improvement of 84.8% over non-autonomous operation, for videos with large scene dynamics, and operational conditions of intermittent wireless channel interference.
Original language | English |
---|---|
State | Published - 2018 |
Externally published | Yes |
Event | 1st USENIX Workshop on Hot Topics in Edge Computing, HotEdge 2018, co-located with USENIX ATC 2018 - Boston, United States Duration: Jul 10 2018 → … |
Conference
Conference | 1st USENIX Workshop on Hot Topics in Edge Computing, HotEdge 2018, co-located with USENIX ATC 2018 |
---|---|
Country/Territory | United States |
City | Boston |
Period | 07/10/18 → … |