征稿已开启

查看我的稿件

注册已开启

查看我的门票

已截止
活动简介

One of the major success stories of Computer Vision over the 1990s and 2000s was the development of systems that can simultaneously build a map of the environment and localize a camera with respect to that environment. Within a batch framework this is usually known as structure from motion or multi-view reconstruction, and systems that can reconstruct city-scale (or even wider-scale) environments with accurate geometry are now common, whether from video or from a less well ordered set of images. Within the robotics and real-time vision communities this problem is often referred to as visual SLAM, Simultaneous Localization and Mapping. The current state-of-the-art visual SLAM systems can reconstruct large areas either densely or semi-densely (i.e. with a depth estimate at all or many pixels) with high accuracy using just a single camera, in real-time. As impressive as these systems and algorithms are, they have no understanding of the scenes they observe – at best they provide a dense geometric point-cloud – and fall well short of the ability of the human visual system to assimilate high-level visual information. In contrast to much of the work in multi-view analysis, those working with single images have made significant progress in applications such as segmentation and object and place recognition, so that an image can be labelled with high level designations of its content, explaining not just the geometry, but also the semantic content of the image. Nevertheless these algorithms are often slow, require offline learning that does not necessarily adapt or transfer between environments, and importantly have no temporal component. Within the last few years a number of researchers in computer vision and robotics have recognised the benefits that applying this semantic level analysis to image sequences acquired by a moving camera, proposing a shift from purely geometric reconstruction and mapping, to semantic level descriptions of scenes involving objects, surfaces, attributes and scene relations that together capture an understanding of the scene. This shift will enable recognition of long-term change through maps that are more versatile, informative, and compact and so it is likely that through advances in this area the quest for long-term autonomy can make greatest gains. This workshop will bring together interested researchers in multi-view geometry, scene understanding and robotic vision with the common goal of discussing and presenting the state-of-the-art in the use of semantics in structure from motion and robotic vision, and in considering the most fruitful and challenging areas for the development of the field. The workshop will cmpirse invited talks, and a limited number of submitted talks, selected for presentation by the organising committee.

征稿信息
留言
验证码 看不清楚,更换一张
全部留言
重要日期
  • 会议日期

    06月11日

    2015

    06月15日

    2015

  • 06月15日 2015

    注册截止日期

主办单位
IEEE Computer Society
移动端
在手机上打开
小程序
打开微信小程序
客服
扫码或点此咨询