MULTIPLE SIMULTANEOUS FRAMING ALTERNATIVES USING SPEAKER TRACKING

    公开(公告)号:US20200068172A1

    公开(公告)日:2020-02-27

    申请号:US16665386

    申请日:2019-10-28

    Abstract: In one embodiment, a video conference endpoint may detect a one or more participants within a field of view of a camera of the video conference endpoint. The video conference endpoint may determine one or more alternative framings of an output of the camera of the video conference endpoint based on the detected one or more participants. The video conference endpoint may send the output of the camera of the video conference endpoint to one or more far-end video conference endpoints participating in a video conference with the video conference endpoint. The video conference endpoint may send data descriptive of the one or more alternative framings of the output of the camera to the far-end video conference endpoints. The far-end video conference endpoints may utilize the data to display one of the one or more alternative framings.

    AUTOMATED VIDEO EDITING BASED ON ACTIVITY IN VIDEO CONFERENCE
    16.
    发明申请
    AUTOMATED VIDEO EDITING BASED ON ACTIVITY IN VIDEO CONFERENCE 有权
    基于视频会议活动的自动视频编辑

    公开(公告)号:US20160359941A1

    公开(公告)日:2016-12-08

    申请号:US14733485

    申请日:2015-06-08

    Abstract: In one embodiment, a method includes receiving at a network device, video and activity data for a video conference, automatically processing the video at the network device based on the activity data, and transmitting edited video from the network device. Processing comprises identifying active locations in the video and editing the video to display each of the active locations before a start of activity at the location and switch between the active locations. An apparatus and logic are also disclosed herein.

    Abstract translation: 在一个实施例中,一种方法包括在网络设备处接收视频会议的视频和活动数据,基于活动数据在网络设备处自动处理视频,以及从网络设备发送编辑的视频。 处理包括识别视频中的活动位置并编辑视频以在该位置开始活动之前显示每个活动位置并在活动位置之间切换。 本文还公开了一种装置和逻辑。

    Use of face and motion detection for best view framing in video conference endpoint
    17.
    发明授权
    Use of face and motion detection for best view framing in video conference endpoint 有权
    使用脸部和运动检测来实现视频会议终端中的最佳视图构图

    公开(公告)号:US09307200B2

    公开(公告)日:2016-04-05

    申请号:US14249443

    申请日:2014-04-10

    CPC classification number: H04N7/147 G06K9/00255 H04N5/23219 H04N7/15

    Abstract: A video conference endpoint detects faces at associated face positions in video frames capturing a scene. The endpoint frames the video frames to a view of the scene encompassing all of the detected faces. The endpoint detects that a previously detected face is no longer detected. In response, a timeout period is started and independently of detecting faces, motion is detected across the view. It is determined if any detected motion (i) coincides with the face position of the previously detected face that is no longer detected, and (ii) occurs before the timeout period expires. If conditions (i) and (ii) are met, the endpoint restarts the timeout period and repeats the independently detecting motion and the determining. Otherwise, the endpoint reframes the view to encompass the remaining detected faces.

    Abstract translation: 视频会议终端检测拍摄场景的视频帧中相关联的脸部位置处的脸部。 端点将视频帧框架到包含所有检测到的面部的场景视图。 端点检测到先前检测到的脸部不再被检测到。 作为响应,开始超时时段并且独立于检测面,在整个视图中检测到运动。 确定任何检测到的运动(i)是否与不再检测到的先前检测到的面部的面部位置一致,并且(ii)在超时时段到期之前发生。 如果满足条件(i)和(ii),端点重新启动超时时间段,并重复独立检测运动和确定。 否则,端点重新构造视图以包含剩余的检测到的面。

    Group and conversational framing for speaker tracking in a video conference system

    公开(公告)号:US10708544B2

    公开(公告)日:2020-07-07

    申请号:US16287191

    申请日:2019-02-27

    Abstract: In one embodiment, a method is provided to intelligently frame groups of participants in a meeting. This gives a more pleasing experience with fewer switches, better contextual understanding, and more natural framing, as would be seen in a video production made by a human director. Furthermore, in accordance with another embodiment, conversational framing techniques are provided. During speaker tracking, when two local participants are addressing each other, a method is provided to show a close-up framing showing both participants. By evaluating the direction participants are looking and a speaker history, it is determined if there is a local discussion going on, and an appropriate framing is selected to give far-end participants the most contextually rich experience.

    Multiple simultaneous framing alternatives using speaker tracking

    公开(公告)号:US10516852B2

    公开(公告)日:2019-12-24

    申请号:US15981299

    申请日:2018-05-16

    Abstract: In one embodiment, a video conference endpoint may detect a one or more participants within a field of view of a camera of the video conference endpoint. The video conference endpoint may determine one or more alternative framings of an output of the camera of the video conference endpoint based on the detected one or more participants. The video conference endpoint may send the output of the camera of the video conference endpoint to one or more far-end video conference endpoints participating in a video conference with the video conference endpoint. The video conference endpoint may send data descriptive of the one or more alternative framings of the output of the camera to the far-end video conference endpoints. The far-end video conference endpoints may utilize the data to display one of the one or more alternative framings.

Patent Agency Ranking