Optical Flow Accounting for Image Haze
    71.
    发明申请
    Optical Flow Accounting for Image Haze 有权
    图像雾度的光流会计

    公开(公告)号:US20140254943A1

    公开(公告)日:2014-09-11

    申请号:US13794408

    申请日:2013-03-11

    CPC classification number: G06T5/003 G06T5/50 G06T7/269 G06T2207/10016

    Abstract: In embodiments of optical flow accounting for image haze, digital images may include objects that are at least partially obscured by a haze that is visible in the digital images, and an estimate of light that is contributed by the haze in the digital images can be determined The haze can be cleared from the digital images based on the estimate of the light that is contributed by the haze, and clearer digital images can be generated. An optical flow between the clearer digital images can then be computed, and the clearer digital images refined based on the optical flow to further clear the haze from the images in an iterative process to improve visibility of the objects in the digital images.

    Abstract translation: 在考虑图像雾度的光学流量的实施例中,数字图像可以包括由数字图像中可见的雾度至少部分地模糊的对象,并且可以确定由数字图像中的雾度贡献的光的估计 可以基于由雾度贡献的光的估计,从数字图像中清除雾度,并且可以产生更清晰的数字图像。 然后可以计算更清晰的数字图像之间的光流,并且基于光流改进更清晰的数字图像,以在迭代过程中进一步清除来自图像的雾度,以提高数字图像中的对象的可视性。

    Metadata-driven method and apparatus for constraining solution space in image processing techniques
    72.
    发明授权
    Metadata-driven method and apparatus for constraining solution space in image processing techniques 有权
    用于在图像处理技术中约束解空间的元数据驱动方法和装置

    公开(公告)号:US08675988B2

    公开(公告)日:2014-03-18

    申请号:US13683966

    申请日:2012-11-21

    Abstract: Methods and apparatus for constraining solution space in image processing techniques may use the metadata for a set of images to constrain an image processing solution to a smaller solution space. In one embodiment, a process may require N parameters for processing an image. A determination may be made from metadata that multiple images were captured with the same camera/lens and with the same settings. A set of values may be estimated for the N parameters from data in one or more of the images. The process may then be applied to each of images using the set of values. In one embodiment, a value for a parameter of a process may be estimated for an image. If the estimated value deviates substantially from a value for the parameter in the metadata, the metadata value is used in the process instead of the estimated value.

    Abstract translation: 在图像处理技术中约束解空间的方法和装置可以使用一组图像的元数据来将图像处理解决方案约束到较小的解空间。 在一个实施例中,过程可能需要用于处理图像的N个参数。 可以从元数据确定使用相同的相机/镜头并以相同的设置捕获多个图像。 可以根据一个或多个图像中的数据为N个参数估计一组值。 然后可以使用该组值将该过程应用于每个图像。 在一个实施例中,可以针对图像估计处理的参数的值。 如果估计值实质上偏离元数据中的参数的值,则在过程中使用元数据值而不是估计值。

    IMAGE CAPTIONING UTILIZING SEMANTIC TEXT MODELING AND ADVERSARIAL LEARNING

    公开(公告)号:US20180373979A1

    公开(公告)日:2018-12-27

    申请号:US15630604

    申请日:2017-06-22

    Abstract: The present disclosure includes methods and systems for generating captions for digital images. In particular, the disclosed systems and methods can train an image encoder neural network and a sentence decoder neural network to generate a caption from an input digital image. For instance, in one or more embodiments, the disclosed systems and methods train an image encoder neural network (e.g., a character-level convolutional neural network) utilizing a semantic similarity constraint, training images, and training captions. Moreover, the disclosed systems and methods can train a sentence decoder neural network (e.g., a character-level recurrent neural network) utilizing training sentences and an adversarial classifier.

    Font replacement based on visual similarity

    公开(公告)号:US10007868B2

    公开(公告)日:2018-06-26

    申请号:US15269492

    申请日:2016-09-19

    Abstract: Font replacement based on visual similarity is described. In one or more embodiments, a font descriptor includes multiple font features derived from a visual appearance of a font by a font visual similarity model. The font visual similarity model can be trained using a machine learning system that recognizes similarity between visual appearances of two different fonts. A source computing device embeds a font descriptor in a document, which is transmitted to a destination computing device. The destination compares the embedded font descriptor to font descriptors corresponding to local fonts. Based on distances between the embedded and the local font descriptors, at least one matching font descriptor is determined. The local font corresponding to the matching font descriptor is deemed similar to the original font. The destination computing device controls presentations of the document using the similar local font. Computation of font descriptors can be outsourced to a remote location.

    Font Attributes for Font Recognition and Similarity

    公开(公告)号:US20180114097A1

    公开(公告)日:2018-04-26

    申请号:US15853120

    申请日:2017-12-22

    Abstract: Font recognition and similarity determination techniques and systems are described. In a first example, localization techniques are described to train a model using machine learning (e.g., a convolutional neural network) using training images. The model is then used to localize text in a subsequently received image, and may do so automatically and without user intervention, e.g., without specifying any of the edges of a bounding box. In a second example, a deep neural network is directly learned as an embedding function of a model that is usable to determine font similarity. In a third example, techniques are described that leverage attributes described in metadata associated with fonts as part of font recognition and similarity determinations.

    Learning image categorization using related attributes

    公开(公告)号:US09953425B2

    公开(公告)日:2018-04-24

    申请号:US14447296

    申请日:2014-07-30

    CPC classification number: G06T7/33 G06K9/627 G06N3/0454

    Abstract: A first set of attributes (e.g., style) is generated through pre-trained single column neural networks and leveraged to regularize the training process of a regularized double-column convolutional neural network (RDCNN). Parameters of the first column (e.g., style) of the RDCNN are fixed during RDCNN training. Parameters of the second column (e.g., aesthetics) are fine-tuned while training the RDCNN and the learning process is supervised by the label identified by the second column (e.g., aesthetics). Thus, features of the images may be leveraged to boost classification accuracy of other features by learning a RDCNN.

Patent Agency Ranking