T2FPV: Constructing High-Fidelity First-Person View Datasets From Real-World Pedestrian Trajectories

09/22/2022
by   Benjamin Stoler, et al.
0

Predicting pedestrian motion is essential for developing socially-aware robots that interact in a crowded environment. While the natural visual perspective for a social interaction setting is an egocentric view, the majority of existing work in trajectory prediction has been investigated purely in the top-down trajectory space. To support first-person view trajectory prediction research, we present T2FPV, a method for constructing high-fidelity first-person view datasets given a real-world, top-down trajectory dataset; we showcase our approach on the ETH/UCY pedestrian dataset to generate the egocentric visual data of all interacting pedestrians. We report that the bird's-eye view assumption used in the original ETH/UCY dataset, i.e., an agent can observe everyone in the scene with perfect information, does not hold in the first-person views; only a fraction of agents are fully visible during each 20-timestep scene used commonly in existing work. We evaluate existing trajectory prediction approaches under varying levels of realistic perception – displacement errors suffer a 356 information setting. To promote research in first-person view trajectory prediction, we release our T2FPV-ETH dataset and software tools.

READ FULL TEXT

Please sign up or login with your details

Forgot password? Click here to reset