Abstract:
Embodiments herein present a fully convolutional approach for instance-aware semantic segmentation. Embodiments herein, for a given visual image, locate predefined categories of objects therein and produce pixel masks for each object instance. One method embodiment includes receiving an image and generating a pixel-wise score map for a given region of interest within the received image for each pixel cell present therein. For each pixel cell within the region of interest, the method may detect whether the pixel cell belongs to an object to obtain a detection result and determine whether the pixel cell is inside an object instance boundary to obtain a segmentation result. The method may then fuse the results to obtain a result of inside or outside for each pixel cell and form at least one mask based on those values.
Abstract:
본 발명에 따른 논리적 화면 표시장치를 생성하는 시스템은 하나 이상의 콘텐츠를 하나 이상의 디바이스로부터 수신하는 통신모듈, 상기 콘텐츠를 상기 논리적 화면 표시장치에 출력시키기 위한 프로그램이 저장된 메모리 및 상기 메모리에 저장된 프로그램을 실행시키는 프로세서를 포함한다. 이때, 상기 프로세서는 상기 프로그램을 실행시킴에 따라, 물리적 화면 표시장치의 출력 범위 내의 2차원 또는 3차원의 사물공간에 대하여 하나 이상의 논리적 화면 표시객체의 영역을 설정하고, 상기 설정된 논리적 화면 표시객체의 영역에 상기 수신한 콘텐츠를 출력시키는 하나 이상의 논리적 화면 표시장치를 생성하되, 상기 물리적 화면 표시장치와 상기 논리적 화면 표시객체는 L 대 M(L, M≥1)의 상호 관계를 갖고, 상기 논리적 화면 표시객체와 상기 논리적 화면 표시장치는 M 대 N(M, N≥1)의 상호 관계를 갖는다.
Abstract:
A combination video surveillance system and physical deterrent device is disclosed. At least one camera module of the video surveillance system defines a first field of view and is operable to generate image data corresponding to the first field of view. A video analytics module is configured to detect a foreground visual object falling within the first field of view, classify the visual object, and determine an appearance of the visual object. A positioning module is configured to determine a physical location of the visual object. A deterrence device controller is configured to receive the determined physical location of the visual object, to control a deterrence device to be aimed at the physical location of the visual object, and to control the deterrence device to selectively emit the physical effect.
Abstract translation:公开了一种组合视频监视系统和物理威慑装置。 视频监视系统的至少一个相机模块定义第一视野并且可操作以生成对应于第一视野的图像数据。 视频分析模块被配置为检测落入第一视野内的前景视觉对象,分类视觉对象并确定视觉对象的外观。 定位模块被配置为确定视觉对象的物理位置。 威慑装置控制器被配置为接收所确定的视觉对象的物理位置,以控制威慑装置瞄准视觉对象的物理位置,并且控制威慑装置选择性地发出物理效果。 p >
Abstract:
The disclosed embodiments include methods, apparatuses, systems, and UAVs configured to an interactive and automatic initialization of the tracking systems. The disclosed embodiments observe an object of interest in a surrounding of the movable object and detect a feature of the object of interest, which acts as a trigger for automatically initializing the tracking system. As a result, the disclosed embodiments may provide efficiency and reliability to initializing a robotic system.
Abstract:
Provided are methods, apparatuses, and computer-readable medium for automatic tamper detection that are both robust and low in complexity. The described methods can detect many different tamper types while also reducing false positives. In various implementations, the described methods include comparing an input frame against a background picture, using tiles of pixels taken from both the input frame and the background picture. When enough differences are detected between the tiles of pixels from the input frame and the tiles of pixels from the background picture, tamper is suspected. When tamper is suspected, further checking of the frame may be enabled, including checking for blur and/or checking a downsampled version of the whole frame. The additional checks can confirm that an actual tamper of the camera has occurred.
Abstract:
Techniques and systems are provided for processing video data. For example, techniques and systems are provided for performing content-adaptive blob filtering. A number of blobs generated for a video frame is determined. A size of a first blob from the blobs is determined, the first blob including pixels of at least a portion of a first foreground object in the video frame. The first blob is filtered from the plurality of blobs when the size of the first blob is less than a size threshold. The size threshold is determined based on the number of the plurality of blobs generated for the video frame.
Abstract:
A system and method are provided for automatically estimating a repair cost for a vehicle. A method includes: receiving, at a server computing device over an electronic network, one or more images of a damaged vehicle from a client computing device; performing image processing operations on each of the one or more images to detect external damage to a first set of parts of the vehicle; inferring internal damage to a second set of parts of the vehicle based on the detected external damage; and, calculating an estimated repair cost for the vehicle based on the detected external damage and inferred internal damage based on accessing a parts database that includes repair and labor costs for each part in the first and second sets of parts.
Abstract:
Described is a robotic system for detecting obstacles reliably with their ranges by a combination of two-dimensional and three-dimensional sensing. In operation, the system receives an image from a monocular video and range depth data from a range sensor of a scene proximate a mobile platform. The image is segmented, into multiple object regions of interest and time-to-contact (TTC) value are calculated by estimating motion field and operating on image intensities. A two-dimensional (2D) TTC map is then generated by estimating average TTC values over the multiple object regions of interest. A three-dimensional TTC map is then generated by fusing the range depth data with image. Finally, a range-fused TTC map is generated by averaging the 2D TTC map and the 3D TTC map,
Abstract:
A method for extracting object image data from a 3D image comprising the following steps: obtaining background image data in an image coordinate system (x i , y i ,Z i ) of a 3D image vision system, based on a 3D image of a background scene (100); transforming the background image data related to the image coordinate system into transformed background image data related to a fixed coordinate system (X F , y F , Z F ) (200); storing the transformed background image data and storing location data related to the fixed coordinate system, said location data being associated with the location of the 3D image vision system when obtaining the 3D image (300); placing an object in the background scene and obtaining combined image data related to the image coordinate system, based on a 3D image of the object and the background scene (400); transforming the combined image data related to the image coordinate system to transformed combined image data related to the fixed coordinate system (500), and obtaining extracted object image data by subtracting the transformed background image data from the transformed combined image data (600). Also described is a 3D image processing system (13) for extracting object image data, comprising a 3D image vision system (5), and a computer unit (15)