Goto

Collaborating Authors

 calculation cost


Reviews: Sequential Context Encoding for Duplicate Removal

Neural Information Processing Systems

This paper proposes a new Duplicate Removal method based on RNN. Based on each candidate area, informative features are extracted by using appearance feature, position and ranking information in addition to the score. Then, they are treated as series data and are input into the RNN-based model to improve the final accuracy by capturing global information. The number of candidate regions is enormous to the number of objects that are to be left. Therefore, this paper proposes to reduce the box gradually by dividing it into two stages. In the two stages, the RNN model of the same structure was used. In stage I, to remove simple boxes the model is trained by using NMS results as a teaching signal. In stage II, to remove difficult boxes, the model is trained by using the grand-truth boxes. Experiments showed that mAP is increased in the SOTA object detection methods (FPN, Mask R - CNN, PANet with DCN) with the proposed method.


Fast Inverse Reinforcement Learning with Interval Consistent Graph for Driving Behavior Prediction

Shimosaka, Masamichi (Tokyo Institute of Technology) | Sato, Junichi (The University of Tokyo) | Takenaka, Kazuhito (Denso Corporation) | Hitomi, Kentarou (Denso Corporation)

AAAI Conferences

In contrast, Inverse reinforcement learning (IRL), inverse optimal control, a discrete approach guarantees global optimality once and imitation learning(Ng and Russell 2000; Abbeel proper discrete state space is given, hence it is more suitable and Ng 2004) are modeling frameworks for acquiring rewards for driving behavior modeling. In a discrete approach, (or cost) of a certain environment by using the optimal the calculation cost of MaxEnt IRL is O( S A), where S path under a possibly different environment as training is the number of states and A is the number of actions data. In particular, in human behavior modeling, it is (Ziebart and others 2008). That is, the key for fast prediction shown that human-centered rewards can be obtained with is suppressing the increase of S depending on dimensions maximum entropy inverse reinforcement learning (MaxEnt and preparing a necessary and sufficient action set, A, IRL)(Ziebart and others 2008), which allows suboptimal for representing driving behavior. As examples of existing training data (Huang et al. 2015; Vernaza and Bagnell 2012; discretization schemes, there are mesh grid representation Dragan and Srinivasa 2012; Walker, Gupta, and Hebert (Shimosaka, Kaneko, and Nishi 2014) and random graph 2014). For instance, Ziebart et al. (Ziebart et al. 2008) modeled based representation connected with neighbors (Byravan et the driving behavior of expert taxi drivers and enabled al. 2015). In these approaches, however, A for general dynamic driving behavior prediction based on the experts' very own systems is not trivial. This is because neighbors on experience or knowledge. MaxEnt IRL based driving behavior state space defined by Euclidean distance do not necessarily prediction, which balances safety, comfort, and economic correspond to the transition area of general dynamics performance, is very promising.