A Note on the Reward Function for PHD Filters with Sensor Control
MetadataShow full item record
The context is sensor control for multi-object Bayes filtering in the framework of partially observed Markov decision processes (POMDPs). The current information state is represented by the multi-object probability density function (pdf), while the reward function associated with each sensor control (action) is the information gain measured by the alpha or Rényi divergence. Assuming that both the predicted and updated state can be represented by independent identically distributed (IID) cluster random finite sets (RFSs) or, as a special case, the Poisson RFSs, this work derives the analytic expressions of the corresponding Rényi divergence based information gains. The implementation of Rényi divergence via the sequential Monte Carlo method is presented. The performance of the proposed reward function is demonstrated by a numerical example, where a moving range-only sensor is controlled to estimate the number and the states of several moving objects using the PHD filter.
Showing items related by title, author, creator and subject.
Woon, Siew Fang (2009)Optimal control problems arise in many applications, such as in economics, finance, process engineering, and robotics. Some optimal control problems involve a control which takes values from a discrete set. These problems ...
Wong, Kiing-Ing; Barsoum, Nader; Myint, Cho Zin (2010)The body sensor nodes armed with a MSP430 microcontroller, a IEEE 802.15.4 radio chip, a memory flash and an electronic amplifier circuits is proposed as an educational platform for electronic design and embedded system ...
Ristic, B.; Vo, Ba-Ngu (2010)The problem addressed in this paper is information theoretic sensor control for recursive Bayesian multi-object state-space estimation using random finite sets. The proposed algorithm is formulated in the framework of ...