-
Notifications
You must be signed in to change notification settings - Fork 1
/
Copy pathintro.tex
9 lines (6 loc) · 2.68 KB
/
intro.tex
1
2
3
4
5
6
7
8
9
\section{Introduction}
\label{sec:intro}
In coalition operations, mission commanders face the challenge of gaining adequate situational understanding to support decision making. In such operations, analytic and computational resources may be spread across different units. The resources may be limited in terms of compute power and energy usage. Communication between services is challenging, as networks are formed dynamically and will be both noisy and severely bandwidth-constrained. Analytic tools must therefore be implemented around these constraints, while still providing useful information to analysts and mission commanders.
Deep neural networks (DNNs) have emerged as a powerful tool for a broad variety of analytical tasks including classification, object recognition, anomaly detection, and time-series prediction~\cite{Goodfellow:2016}. As the current state-of-the-art tool in these areas, their deployment to aid coalition situational understanding is inevitable. The above-mentioned constraints on computation and energy use at edge devices in a coalition network will severely limit the capabilities of edge-deployed DNNs, which tend to perform better with increasing depth and thus increasing resource requirements. However, sending data back to the cloud for processing will usually not be possible due to network transfer limitations, so running the DNNs at the edge is still preferable.
We propose a new approach to running DNNs on coalition networks that is designed both to be compatible with the various implementation constraints and to allow appropriately deep networks for good task-performance. We build on ideas developed by Panda et al., who propose executing a subset of a full DNN's layers until adequate confidence is reached, rather than passing the input data through every network layer~\cite{panda2016conditional, panda2017energy}. This approach can reduce computation and energy consumption without reducing task performance. We propose extending this idea to evaluate different DNN layers on different edge devices depending on their availability and computational capabilities.
In the next section, we outline the motivating scenario behind this proposal and describe at a high level how DNN evaluation could be split across edge devices. We provide a detailed account of the splitting and evaluation methods in section \ref{sec:conditional}. These ideas are then extended to the case of spiking neural networks in section \ref{sec:spiking}, as these offer a promising way to further reduce the energy consumption and data transfer requirements of the proposed paradigm. Finally we provide some discussion of the approach, raise potential implementation issues, and propose avenues for further research.