Abstract:
A method includes receiving video data representative of a video frame. The video frame includes a first representation of a first object. The method also includes determining whether to replace the first representation in the video frame with a second representation of a second object based on an object matching rule. In response to determining that the first representation is to be replaced with the second representation, the method includes manipulating a three-dimensional model of the second object to generate the second representation of the second object so that a first visual property of the second representation matches a second visual property of the first representation. The method further includes generating a modified version of the video frame that includes the second representation in place of the first representation.
Abstract:
Methods for obfuscating an image of a subject in a captured media are disclosed. For example, a method receives a communication from an endpoint device of a subject indicating that the image of the subject is to be obfuscated in a captured media. The communication may include a feature set associated with the subject, where the feature set contains facial features of the subject and motion information associated with the subject. The method then detects the image of the subject in the captured media. For example, the image of the subject is detected by matching the facial features of the subject to the image of the subject in the captured media and matching the motion information associated with the subject to a trajectory of the image of the subject in the captured media. The method then obfuscates the image of the subject in the captured media.
Abstract:
Speaker content generated in an audio conference is selectively visually represented. A profile for each audience member who listen to an audio conference is obtained. Speaker content from audio conference participants who speak in the audio conference is monitored. The speaker content from each of the audio conference participants is analyzed. Based on the analyzing and on the profiles for each of the plurality of audience members, visual representations of the speaker content to present to the audience members are identified. Visual representations of the speaker content are generated based on the analyzing. Different visual representations of the speaker content are presented to different audience members based on the analyzing and identifying.
Abstract:
Methods, systems, and computer program calibrate a vision system. An image of a human gesture is received that frames a display device. A boundary defined by the human gesture is computed, and gesture area defined by the boundary is also computed. The gesture area is then mapped to pixels in the display device.
Abstract:
Methods, systems, and computer program calibrate a vision system. An image of a human gesture is received that frames a display device. A boundary defined by the human gesture is computed, and gesture area defined by the boundary is also computed. The gesture area is then mapped to pixels in the display device.
Abstract:
A method includes receiving image data corresponding to an image and detecting an object represented within the image. The method further includes selecting a portion of the image data that corresponds to the object and determining object quality values based on the portion of the image data. The method also includes determining an object category corresponding to the object and accessing object category metrics associated with the object category. The method includes performing a comparison of the object quality values to the object category metrics associated with the object category and initiating an action based on the comparison.
Abstract:
A method includes receiving image data corresponding to an image and detecting an object represented within the image. The method further includes selecting a portion of the image data that corresponds to the object and determining object quality values based on the portion of the image data. The method also includes determining an object category corresponding to the object and accessing object category metrics associated with the object category. The method includes performing a comparison of the object quality values to the object category metrics associated with the object category and initiating an action based on the comparison.
Abstract:
A method includes receiving, at a server computing device, a request to stream media content to a user device. The request is associated with data identifying a personal quality profile. The method includes retrieving, via the server computing device, the personal quality profile associated with the data. The personal quality profile indicates media quality preferences. The method includes determining, via a processor of the server computing device, a version of the media content to send to the user device based on the media quality preferences of the personal quality profile and based on a delivery bit rate. The method also includes sending the version from the server computing device to the user device in response to the request.
Abstract:
Disclosed herein are systems, methods, and non-transitory computer-readable storage media for performing speaker verification. A system configured to practice the method receives a request to verify a speaker, generates a text challenge that is unique to the request, and, in response to the request, prompts the speaker to utter the text challenge. Then the system records a dynamic image feature of the speaker as the speaker utters the text challenge, and performs speaker verification based on the dynamic image feature and the text challenge. Recording the dynamic image feature of the speaker can include recording video of the speaker while speaking the text challenge. The dynamic feature can include a movement pattern of head, lips, mouth, eyes, and/or eyebrows of the speaker. The dynamic image feature can relate to phonetic content of the speaker speaking the challenge, speech prosody, and the speaker's facial expression responding to content of the challenge.
Abstract:
A method and apparatus for displaying a plurality of secondary images based on image data of a primary image include displaying a primary image and a plurality of secondary images based on the image data of the primary image. Each of the plurality of secondary images is displayed having a size based on a relevance of each respective one of the plurality of secondary images to the primary image. The primary image can be selected from a plurality of images or a frame of a video. Subsets of secondary images can be displayed in clusters including images having content similar to particular content of the primary image.