Sensor Control for Information Gain in Dynamic, Sparse and Partially Observed Environments
JB Burns, A Sundaresan, P Sequeira… - arXiv preprint arXiv …, 2022 - arxiv.org
arXiv preprint arXiv:2211.01527, 2022•arxiv.org
We present an approach for autonomous sensor control for information gathering under
partially observable, dynamic and sparsely sampled environments that maximizes
information about entities present in that space. We describe our approach for the task of
Radio-Frequency (RF) spectrum monitoring, where the goal is to search for and track
unknown, dynamic signals in the environment. To this end, we extend the Deep Anticipatory
Network (DAN) Reinforcement Learning (RL) framework by (1) improving exploration in …
partially observable, dynamic and sparsely sampled environments that maximizes
information about entities present in that space. We describe our approach for the task of
Radio-Frequency (RF) spectrum monitoring, where the goal is to search for and track
unknown, dynamic signals in the environment. To this end, we extend the Deep Anticipatory
Network (DAN) Reinforcement Learning (RL) framework by (1) improving exploration in …
We present an approach for autonomous sensor control for information gathering under partially observable, dynamic and sparsely sampled environments that maximizes information about entities present in that space. We describe our approach for the task of Radio-Frequency (RF) spectrum monitoring, where the goal is to search for and track unknown, dynamic signals in the environment. To this end, we extend the Deep Anticipatory Network (DAN) Reinforcement Learning (RL) framework by (1) improving exploration in sparse, non-stationary environments using a novel information gain reward, and (2) scaling up the control space and enabling the monitoring of complex, dynamic activity patterns using hybrid convolutional-recurrent neural layers. We also extend this problem to situations in which sampling from the intended RF spectrum/field is limited and propose a model-based version of the original RL algorithm that fine-tunes the controller via a model that is iteratively improved from the limited field sampling. Results in simulated RF environments of differing complexity show that our system outperforms the standard DAN architecture and is more flexible and robust than baseline expert-designed agents. We also show that it is adaptable to non-stationary emission environments.
arxiv.org
Showing the best result for this search. See all results