Abstract:
Apparatuses and a method are provided for providing feedback to a user, who may be visually-impaired. In one implementation, a method is provided for providing feedback to a visually impaired user. The method comprises receiving from a mobile image sensor real time image data that includes a representation of an object in an environment of the visually impaired user. The mobile image sensor is configured to be connected to glasses worn by the visually impaired user. Further, the method comprises receiving a signal indicating a desire of the visually impaired user to obtain information about the object. The method also includes accessing a database holding information about a plurality of objects, and comparing information derived from the received real time image data with information in the database. The method comprises providing the visually impaired user with nonvisual feedback that the object is not locatable in the database.
Abstract:
An apparatus and method are provided for performing one or more actions based on triggers detecting within captured image data. In one implementation, a method is provided for audibly reading text retrieved from a captured image. According to the method, real-time image data is captured from an environment of a user, and an existence of a trigger is determined within the captured image data. In one aspect, the trigger may be associated with a desire of the user to hear text read aloud, and the trigger identifies an intermediate portion of the text a distance from a level break in the text. The method includes performing a layout analysis on the text to identify the level break associated with the trigger, and reading aloud text beginning from the level break associated with the trigger.
Abstract:
According to an example, whether a picture of a web page is a two-dimensional code picture is determined when the picture is triggered, a user is prompted to determine whether to identify the two-dimensional code picture when the picture of the web page is the two-dimensional code picture, the two-dimensional code picture is parsed after receiving an instruction of identifying the two-dimensional code picture from the user, two-dimensional code information is obtained, and processing is performed according to the two-dimensional code information.
Abstract:
The display of text by a first device, that is extracted and used by a second communication device, based upon a determined context of the text, is provided. The image displayed for the first device is captured by an image capture element of a second communication device, which also has a recognition module and an extraction module. The image includes the text and a context element representing a context of the text.
Abstract:
A computer-implemented method of storing image data comprising images of a subject generated by a medical imaging device is disclosed. The method comprises: a) capturing the image data; b) receiving subject identification metadata; c) analysing at least one selected element of the image data to detect features identifying the subject and modifying the or each selected element of the image data by removing or obscuring any such detected features; and d) storing a subject record comprising the or each modified selected element of the image data and the subject identification metadata.
Abstract:
A method includes tracking an object in each of a plurality of frames of video data to generate a tracking result. The method also includes performing object processing of a subset of frames of the plurality of frames selected according to a multi-frame latency of an object detector or an object recognizer. The method includes combining the tracking result with an output of the object processing to produce a combined output.
Abstract:
A method for recognizing a text block in an object is disclosed. The text block includes a set of characters. A plurality of images of the object are captured and received. The object in the received images is then identified by extracting a pattern in one of the object images and comparing the extracted pattern with predetermined patterns. Further, a boundary of the object in each of the object images is detected and verified based on predetermined size information of the identified object. Text blocks in the object images are identified based on predetermined location information of the identified object. Interim sets of characters in the identified text blocks are generated based on format information of the identified object. Based on the interim sets of characters, a set of characters in the text block in the object is determined.
Abstract:
Detecting a static graphic object (such as a logo, title, or sub-title) in a sequence of video frames may be accomplished by analyzing each selected one of a plurality of pixels in a video frame of the sequence of video frames. Basic conditions for the selected pixel may be tested to determine whether the selected pixel is a static pixel. When the selected pixel is a static pixel, a static similarity measure and a forward motion similarity measure may be determined for the selected pixel. A temporal score for the selected pixel may be determined based at least in part on the similarity measures. Finally, a static graphic object decision for the selected pixel may be made based at least in part on the temporal score.
Abstract:
A system and method for an adaptive threshold Web Page segmenting is disclosed. In one embodiment, a method performed by a physical computing system having one or more processors for segmenting a Web page including a plurality of nodes includes parsing content in the Web page into the plurality of nodes using the physical computing system, obtaining feature values between each pair of nodes using the physical computing system, estimating an adaptive threshold value using the obtained feature values using the physical computing system, and segmenting the Web page by comparing the feature values associated with each pair of nodes with the estimated adaptive threshold value.