Video Google: A Text Retrieval Approach to Object Matching in Videos

ICCV(2003)

引用 8464|浏览788
暂无评分
摘要
We describe an approach to object and scene retrievalwhich searches for and localizes all the occurrences of auser outlined object in a video. The object is represented bya set of viewpoint invariant region descriptors so that recognitioncan proceed successfully despite changes in viewpoint,illumination and partial occlusion. The temporalcontinuity of the video within a shot is used to track theregions in order to reject unstable regions and reduce theeffects of noise in the descriptors.The analogy with text retrieval is in the implementationwhere matches on descriptors are pre-computed (using vectorquantization), and inverted file systems and documentrankings are used. The result is that retrieval is immediate,returning a ranked list of key frames/shots in the manner ofGoogle.The method is illustrated for matching on two full lengthfeature films.
更多
查看译文
关键词
text retrieval approach,scene retrievalwhich search,inverted file system,text retrieval,video google,partial occlusion,viewpoint invariant region,full lengthfeature film,key frame,unstable region,manner ofgoogle,machine vision,vector quantization,computer vision,noise reduction
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要