-
公开(公告)号:US11342002B1
公开(公告)日:2022-05-24
申请号:US16210545
申请日:2018-12-05
Applicant: Amazon Technologies, Inc.
Inventor: Prabhakar Gupta , Shaktisingh P Shekhawat , Kumar Keshav
IPC: H04N9/80 , G11B27/19 , H04N9/87 , G11B27/036 , H04N5/92
Abstract: An automated solution to determine suitable time ranges or timestamps for captions is described. In one example, a content file includes subtitle data with captions for display over respective timeframes of video. Audio data is extracted from the video, and the audio data is compared against a sound threshold to identify auditory timeframes in which sound is above the threshold. The subtitle data is also parsed to identify subtitle-free timeframes in the video. A series of candidate time ranges is then identified based on overlapping ranges of the auditory timeframes and the subtitle-free timeframes. In some cases, one or more of the candidate time ranges can be merged together or omitted, and a final series of time ranges or timestamps for captions is obtained. The time ranges or timestamps can be used to add additional non-verbal and contextual captions and indicators, for example, or for other purposes.
-
公开(公告)号:US11910073B1
公开(公告)日:2024-02-20
申请号:US17819918
申请日:2022-08-15
Applicant: Amazon Technologies, Inc.
Inventor: Mayank Sharma , Prabhakar Gupta , Honey Gupta , Kumar Keshav
IPC: H04N21/8549 , H04N21/466 , H04N21/472
CPC classification number: H04N21/8549 , H04N21/466 , H04N21/47217
Abstract: A respective set of features, including emotion-related features, are extracted from segments of a video for which a preview is to be generated. A subset of the segments is chosen using the features and filtering criteria including at least one emotion-based filtering criterion. Respective weighted preview-suitability scores are assigned to the segments of the subset using at least a metric of similarity between individual segments and a plot summary of the video. The scores are used to select and combine segments to form a preview for the video.
-
公开(公告)号:US20240223872A1
公开(公告)日:2024-07-04
申请号:US18411720
申请日:2024-01-12
Applicant: Amazon Technologies, Inc.
Inventor: Mayank Sharma , Prabhakar Gupta , Honey Gupta , Kumar Keshav
IPC: H04N21/8549 , H04N21/466 , H04N21/472
CPC classification number: H04N21/8549 , H04N21/466 , H04N21/47217
Abstract: A respective set of features, including emotion-related features, are extracted from segments of a video for which a preview is to be generated. A subset of the segments is chosen using the features and filtering criteria including at least one emotion-based filtering criterion. Respective weighted preview-suitability scores are assigned to the segments of the subset using at least a metric of similarity between individual segments and a plot summary of the video. The scores are used to select and combine segments to form a preview for the video.
-
公开(公告)号:US12189683B1
公开(公告)日:2025-01-07
申请号:US17547727
申请日:2021-12-10
Applicant: Amazon Technologies, Inc.
Inventor: Mayank Sharma , Anil Kumar Nelakanti , Prabhakar Gupta , Kumar Keshav
Abstract: Described herein is a computer-implemented method for extracting and identifying an audio song. An audio file can be accessed by a computing device. A set of audio categories and a set of probabilities associated with the set of audio categories can be determined for a first audio clip. A subset of the set of audio categories can be determined based on a subset of the set of probabilities. Each audio category of the subset of the set of audio categories can correspond to an audio class label. Whether the first audio clip is part of a song can be determined. The song can be defined by combining the first audio clip with other audio clips.
-
公开(公告)号:US10936827B1
公开(公告)日:2021-03-02
申请号:US16169327
申请日:2018-10-24
Applicant: Amazon Technologies, Inc.
Inventor: Prabhakar Gupta , Shaktisingh P. Shekhawat , Kumar Keshav
IPC: G06F40/51 , G06F40/30 , G06F40/58 , G06F40/289 , G06F40/45
Abstract: Disclosed are various embodiments for evaluating the accuracy of a translation of a source text. Word embeddings from a first language and a second language are aligned in a shared vector space. Word pairs from the sourced text and translated text are then identified. Subsequently, similarity scores between respective word embeddings for the words in the word pair are calculated. Word pairs are then selected based on the similarity scores. The accuracy of the translation is then evaluated based at least in part on the selected word pairs.
-
-
-
-