Egocentric Basketball Motion Planning from a Single First-Person Image

2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition(2018)

引用 30|浏览29
暂无评分
摘要
We present a model that uses a single first-person image to generate an egocentric basketball motion sequence in the form of a 12D camera configuration trajectory, which encodes a player's 3D location and 3D head orientation throughout the sequence. To do this, we first introduce a future convolutional neural network (CNN) that predicts an initial sequence of 12D camera configurations, aiming to capture how real players move during a one-on-one basketball game. We also introduce a goal verifier network, which is trained to verify that a given camera configuration is consistent with the final goals of real one-on-one basketball players. Next, we propose an inverse synthesis procedure to synthesize a refined sequence of 12D camera configurations that (1) sufficiently matches the initial configurations predicted by the future CNN, while (2) maximizing the output of the goal verifier network. Finally, by following the trajectory resulting from the refined camera configuration sequence, we obtain the complete 12D motion sequence. Our model generates realistic basketball motion sequences that capture the goals of real players, outperforming standard deep learning approaches such as recurrent neural networks (RNNs), long short-term memory networks (LSTMs), and generative adversarial networks (GANs).
更多
查看译文
关键词
egocentric basketball motion planning,single first-person image,egocentric basketball motion sequence,12D camera configuration trajectory,3D head orientation,12D camera configurations,goal verifier network,refined camera configuration sequence,complete 12D motion sequence,realistic basketball motion sequences,recurrent neural networks,short-term memory networks,generative adversarial networks,convolutional neural network
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要