-
公开(公告)号:US20240346950A1
公开(公告)日:2024-10-17
申请号:US18399891
申请日:2023-12-29
发明人: Jing-Jing GUO , Steve Shu LIU
IPC分类号: G09B19/04 , G10L15/02 , G10L21/0208 , G10L25/51
CPC分类号: G09B19/04 , G10L15/02 , G10L21/0208 , G10L25/51 , G10L2015/025
摘要: A speaking practice system with redundant pronunciation correction is shown, which provides a goodness of pronunciation (GOP) evaluation system running on a data processing server to detect redundant pronunciation in an audio recording. The audio recording is recorded when the user reads a practice text aloud. According to the detected redundant pronunciations, the user is informed to make corrections.
-
公开(公告)号:US12008921B2
公开(公告)日:2024-06-11
申请号:US18152625
申请日:2023-01-10
申请人: 617 Education Inc.
发明人: Tom Dillon
CPC分类号: G09B7/04 , G06F3/167 , G09B19/04 , G10L15/02 , G10L15/063 , G10L15/22 , G10L25/18 , G10L25/30 , G10L2015/025 , G10L2015/225
摘要: Systems and methods are described for grapheme-phoneme correspondence learning. In an example, a display of a device is caused to output a grapheme graphical user interface (GUI) that includes a grapheme. Audio data representative of a sound made by the human user is received based on the grapheme shown on the display. A grapheme-phoneme model can determine whether the sound made by the human corresponds to a phoneme for the displayed grapheme based on the audio data. The grapheme-phoneme model is trained based on augmented spectrogram data. A speaker is caused to output a sound representative of the phoneme for the grapheme to provide the human with a correct pronunciation of the grapheme in response to the grapheme-phoneme model determining that the sound made by the human does not correspond to the phoneme for the grapheme.
-
公开(公告)号:US11955026B2
公开(公告)日:2024-04-09
申请号:US16583339
申请日:2019-09-26
发明人: Cheng-Fang Lin , Ching-Chun Liu , Ting-Chieh Yu , Yu-Siang Chen , Ryan Young
CPC分类号: G09B19/04 , G06N3/08 , G10L15/22 , G10L17/02 , G10L17/18 , G10L2015/227 , G10L2015/228
摘要: A method, computer program product, and computer system for public speaking guidance is provided. A processor retrieves speaker data regarding a speech made by a user. A processor separates the speaker data into one or more speaker modalities. A processor extracts one or more speaker features from the speaker data for the one or more speaker modalities. A processor generates a performance classification based on the one or more speaker features. A processor sends to the user guidance regarding the speech based on the performance classification.
-
公开(公告)号:US11935425B2
公开(公告)日:2024-03-19
申请号:US17008053
申请日:2020-08-31
发明人: Manato Ono
摘要: Pronunciation learning processing is performed, in which evaluation scores on pronunciation for respective words are acquired from a pronunciation test that uses multiple words, the acquired evaluation scores are summated for each combination of consecutive pronunciation components in the words, and learning information based on the result of summation is output.
-
公开(公告)号:US20230368687A1
公开(公告)日:2023-11-16
申请号:US18315608
申请日:2023-05-11
摘要: In an approach to assessment of phonological awareness, one or more images are displayed on a Graphical User Interface (GUI), wherein the one or more images represent one or more words; a response is received from a subject as to whether the subject knows the specific word for each of the one or more images; responsive to the subject knows the specific word for any of the one or more images, the specific word is added to a list of known words for the subject; one or more known item images are displayed on the GUI, wherein the one or more known item images are based on the list of known words for the subject; a picture of phonological awareness is determined that is disentangled from word familiarity based on the one or more known item images.
-
公开(公告)号:US11817005B2
公开(公告)日:2023-11-14
申请号:US16176551
申请日:2018-10-31
发明人: Roxana Monge Nunez
摘要: Approaches presented herein enable delivery of real-time internet of things (IoT) feedback to optimize a public speaking performance. More specifically, a set of data representing a speaking performance of a user is captured and analyzed to generate a speaking performance profile of the user. This profile is compared to a reference speaking performance profile and, based on the comparison, a set of performance improvement strategies for the user is generated. A performance improvement strategy is selected from the set of performance improvement strategies based on an identification of an availability of a set of IoT devices for delivery of at least one of the strategies. Instructions are then communicated, responsive to the captured speaking performance associated with the user, to an available IoT device to deliver the selected performance improvement strategy to the user through an output user interface of the available IoT device during the speaking performance.
-
公开(公告)号:US11813222B2
公开(公告)日:2023-11-14
申请号:US17464406
申请日:2021-09-01
CPC分类号: A61H23/008 , A61H23/02 , G09B19/04 , H04W4/80 , A61H23/0236 , A61H23/0245 , A61H23/0254 , A61H2201/5012 , A61H2201/5015 , A61H2201/5058 , A61H2201/5071 , A61H2201/5076 , A61H2205/04 , A61H2230/405
摘要: An apparatus includes a vibrational transducer, a placement band, a driver module and a control module. The placement band is configured to hold the vibrational transducer adjacent to the skin surface overlying the cricoid cartilage and trachea region of a patient's neck. The driver module is configured to apply a drive signal to the vibrational transducer. The control module is configured to receive at least one input configured to provide vibrational operating information and control the driver module to cause the vibrational transducer to apply a vibratory stimulation in an amount determined, at least in part, by the vibrational operating information.
-
公开(公告)号:US11636406B2
公开(公告)日:2023-04-25
申请号:US17645568
申请日:2021-12-22
发明人: Martin L. Cohen , Edward G. Brown
IPC分类号: G09B5/00 , G06Q10/06 , G09B5/06 , G09B5/10 , G09B7/00 , G09B7/02 , G09B7/04 , G09B7/06 , G09B7/08 , G09B9/00 , G09B15/00 , G09B17/00 , G09B19/00 , G09B19/06 , G09B19/18 , G09B23/28 , G06Q10/0639 , G06Q99/00 , G09B5/04 , G06T13/80 , G09B19/02 , G09B19/04 , G06T11/00 , H04L67/02
摘要: The present invention is directed to interactive training, and in particular, to methods and systems for computerized interactive skill training. An example embodiment provides a method and system for providing skill training using a computerized system. The computerized system receives a selection of a first training subject. Several related training components can be invoked, such as reading, watching, performing, and/or reviewing components. In addition, a scored challenge session is provided, wherein a training challenge is provided to a user via a terminal, optionally in video form.
-
公开(公告)号:US20220327956A1
公开(公告)日:2022-10-13
申请号:US17754265
申请日:2020-09-10
发明人: Andrew Butler , Vera Blau-McCandliss , Carey Lee
IPC分类号: G09B19/04 , G09B19/06 , G09B5/06 , G10L15/187 , G06F3/01 , G06V40/20 , G06V40/16 , G02B27/01
摘要: A set of machines functions as a language teaching lab. Configured by suitable hardware, software, accessories, or any suitable combination thereof, such a language teaching lab accesses multiple sources and types of data, such as video streams, audio streams, thermal imaging data, eye tracker data, breath anemometer data, biosensor data, accelerometer data, depth sensor data, or any suitable combination thereof. From the accessed data, the language teaching lab detects that the user is pronouncing, for example, a word, a phrase, or a sentence, and then causes presentation of a reference pronunciation of that word, phrase, or sentence. Other apparatus, systems, and methods are also disclosed.
-
公开(公告)号:US11335360B2
公开(公告)日:2022-05-17
申请号:US16578307
申请日:2019-09-21
IPC分类号: G10L25/63 , G10L15/28 , G10L15/26 , G10L17/26 , G10L25/72 , G10L21/10 , G09B19/04 , G06F3/04847 , G06F3/0482
摘要: In one aspect, a device includes at least one processor and storage accessible to the at least one processor. The storage includes instructions executable by the at least one processor to analyze the decibel levels of audio of a user's speech. The instructions are executable to, based on the analysis, enhance a transcript of the user's speech with indications of particular words from the user's speech as being associated with one or more emotions of the user.
-
-
-
-
-
-
-
-
-