Visual mode image comparison
    3.
    发明授权

    公开(公告)号:US11854287B2

    公开(公告)日:2023-12-26

    申请号:US17456197

    申请日:2021-11-23

    摘要: A method, a computer program product, and a computer system compare images for content consistency. The method includes receiving a first image including a first document and a second image including a second document. The method includes performing a visual classification analysis on the first image and the second image. The visual classification analysis generates an overlap of the first image with the second image. The method includes determining whether a region of the overlap is indicative of a content inconsistency. As a result of the region of the overlap being indicative of a content inconsistency, the method includes performing a character recognition analysis on a first area of the first image and a second area of the second image corresponding to the region of the overlap to verify the content inconsistency.

    VISUAL MODE IMAGE COMPARISON
    5.
    发明公开

    公开(公告)号:US20230162521A1

    公开(公告)日:2023-05-25

    申请号:US17456197

    申请日:2021-11-23

    IPC分类号: G06K9/00 G06K9/34 G06K9/62

    摘要: A method, a computer program product, and a computer system compare images for content consistency. The method includes receiving a first image including a first document and a second image including a second document. The method includes performing a visual classification analysis on the first image and the second image. The visual classification analysis generates an overlap of the first image with the second image. The method includes determining whether a region of the overlap is indicative of a content inconsistency. As a result of the region of the overlap being indicative of a content inconsistency, the method includes performing a character recognition analysis on a first area of the first image and a second area of the second image corresponding to the region of the overlap to verify the content inconsistency.

    DEPLOYING PARALLELIZABLE DEEP LEARNING MODELS BY ADAPTING TO THE COMPUTING DEVICES

    公开(公告)号:US20220351020A1

    公开(公告)日:2022-11-03

    申请号:US17245541

    申请日:2021-04-30

    IPC分类号: G06N3/04

    摘要: In an approach to deploying parallelizable deep learning models by adapting to the computing devices, a deep learning model is split into a plurality of slices, where each slice can exchange data with related slices. Virtual models are created from the plurality of slices, where the virtual models are based on capabilities of a plurality of devices on which the one or more virtual models are to be deployed, and further where each virtual model contains each slice of the plurality of slices. The one or more virtual models are stored in a cache. Responsive to determining that the deep learning model is to be deployed on one or more devices, a candidate model is selected from the virtual models in the cache, where the selection is based on information from a device monitor about the devices.

    ADAPTIVELY COMPRESSING A DEEP LEARNING MODEL

    公开(公告)号:US20230103149A1

    公开(公告)日:2023-03-30

    申请号:US17449652

    申请日:2021-09-30

    IPC分类号: G06N3/04

    摘要: An approach is provided for adaptively compressing a deep learning model. An original deep learning model for different Internet of Things (IoT) devices is determined. Device information is collected from the IoT devices. Based on the device information, multiple recommendation engines are selected from a set of recommendation engines. Compression factor combinations are determined by using the multiple recommendation engines. Compression ratios and model accuracies for the compression factor combinations are determined. Based on the compression ratios and the model accuracies, an optimal compression factor combination is selected from the compression factor combinations. A compressed deep learning model is generated by compressing the original deep learning model by using the optimal compression factor. The compressed deep learning model is deployed to the IoT devices.

    Automatic image annotations
    8.
    发明授权

    公开(公告)号:US11615618B2

    公开(公告)日:2023-03-28

    申请号:US17225165

    申请日:2021-04-08

    IPC分类号: G06K9/62 G06V20/20 G06V10/46

    摘要: A computer-implemented method for annotating images is disclosed. The computer-implemented method includes generating a saliency map corresponding to an input image, wherein the input image is an image that requires annotation, generating a behavior saliency map, wherein the behavior saliency map is a saliency map formed from an average of a plurality of objects contained within respective bounding boxes of a plurality of sample images, generating a historical saliency map, wherein the historical saliency map is a saliency map formed from an average of a plurality of tagged objects in the plurality of sample images, fusing the saliency map corresponding to the input image, the behavior saliency map, and the historical saliency map to form a fused saliency map, and generating, based on the fused saliency map, a bounding box around an object in the input image.

    AUTOMATIC IMAGE ANNOTATIONS
    9.
    发明申请

    公开(公告)号:US20220327312A1

    公开(公告)日:2022-10-13

    申请号:US17225165

    申请日:2021-04-08

    IPC分类号: G06K9/00 G06K9/46 G06K9/62

    摘要: A computer-implemented method for annotating images is disclosed. The computer-implemented method includes generating a saliency map corresponding to an input image, wherein the input image is an image that requires annotation, generating a behavior saliency map, wherein the behavior saliency map is a saliency map formed from an average of a plurality of objects contained within respective bounding boxes of a plurality of sample images, generating a historical saliency map, wherein the historical saliency map is a saliency map formed from an average of a plurality of tagged objects in the plurality of sample images, fusing the saliency map corresponding to the input image, the behavior saliency map, and the historical saliency map to form a fused saliency map, and generating, based on the fused saliency map, a bounding box around an object in the input image.