AUTOMATED CONFIGURATION OF BEHAVIOR OF A TELEPRESENCE SYSTEM BASED ON SPATIAL DETECTION OF TELEPRESENCE COMPONENTS
A system that automatically configures the behavior of the display devices of a video conference endpoint. The controller may detect, at a microphone array having a predetermined physical relationship with respect to a camera, audio emitted from one or more loudspeakers, each loudspeaker having a predetermined physical relationship with respect to at least one of one or more display devices in a conference room. The controller may then generate data representing a spatial relationship between the one or more display devices and the camera based on the detected audio. Finally, the controller may assign video sources received by the endpoint to each of the one or more display devices based on the data representing the spatial relationship and the content of each received video source, and may also assign outputs from multiple video cameras to an outgoing video stream based on the on the data representing the spatial relationship.
The present disclosure relates to configuring components of a video conference endpoint in a conference room based on spatial detection of the components.
BACKGROUNDVideo conference endpoints are deployed in conference rooms. The conference rooms can differ in size and configuration which affects the, layout/placement of the video conference endpoint components in the conference room, and use of the conference room. The placement of the components within the conference room, such as the relationship and placement of the display screens with respect to the camera(s), affects the experience of participants of a conference session within the conference room. Because no two conference rooms are the same size and shape, a standard layout for a conference room is not possible. This results in different placements of the camera(s) with respect to the display screens of a conference room. Typically, an operator has to manually select which display screen is to receive specific video sources, including which display screen, or portion of a display screen, is to display the live video stream of the participants of the conference session that are present at another video conference endpoint. Such manual selection is cumbersome and inconvenient, and often does not place the live video stream of participants of the conference session from another video conference endpoint at a position that maximizes eye contact between participants at separate video conference endpoints.
Techniques presented herein relate to automatically configuring the one or more display devices of a video conference endpoint based on spatial detection of the components of the video conference endpoint and the content of the video sources received by the video conference endpoint. The video conference endpoint may include one or more display devices, one or more loudspeakers having a predetermined physical relationship with respect to at least one of the one or more display devices, at least one camera, a microphone array having a predetermined physical relationship with respect to the camera, and a controller. The controller may be configured to detect, at a microphone array having a predetermined physical relationship with respect to a camera, the audio emitted from each of one or more loudspeakers, each loudspeaker having a predetermined physical relationship with respect to at least one of one or more display devices in a conference room. The controller may further be configured to generate data representing a spatial relationship between each of the one or more display devices and the camera based on the detected audio.
Example EmbodimentsWith reference to
Each video conference endpoint 104 may include at least one video camera (VC) 112, at least one display device 114, a loudspeaker (LDSPKR) 116 coupled to or integrated with the display device 114, one or more microphones arrays (MIC) 118 coupled to or integrated with the camera 112, and an endpoint controller 120 configured to control the video camera(s) 112, at least one display device 114, the loudspeaker 116, and the one or more microphone arrays 118. In a transmit direction, endpoints 104 capture audio/video from their local participants 106 with video camera 112/microphone array 118, encode the captured audio/video into data packets, and transmit the data packets to other endpoints or to the conference server 102. In a receive direction, endpoints 104 decode audio/video from data packets received from the conference server 102 or other endpoints and present the audio/video to their local participants 106 via display device 114/loudspeaker 116.
Referring now to
In some forms, the display devices may be separate from one or more cameras, and the microphone arrays may be separate from the display devices and one or more cameras. For example, an end user may use his/her own display devices, and some cameras available in the market are configured to attach to a microphone stand supporting the microphone array. However, even in that situation, the camera will, once attach, have a known predetermined physical relationship with respect to the microphone array. In summary, the various components of an endpoint may be integrated together when sold, or may be configured after purchase to be physically attached to each other so as to have a predetermined physical relationship. Furthermore, the loudspeakers 116(1)-116(4) may have a predetermined physical relationship with respect to the display devices 114(1)-114(4), respectively. In some embodiments, the loudspeakers 116(1)-116(4) may be integrated with the display devices 114(1)-114(4), respectively. While
As depicted in the example of
The video conference endpoint 104 further includes an endpoint user control device 204 disposed within the conference room 200. The endpoint user control device 204 may be movable within the room 200. The endpoint user control device 204 may be a tablet computer, smartphone or other similar device on which an endpoint controller application is installed. The endpoint user control device 204 may be configured to manage each of the display devices 114(1)-114(4), including, but not limited to, the content displayed on each of the display devices 114(1)-114(4). The endpoint user control device 204 may also be configured control the pan, tilt, and zoom the video cameras 112(1)-112(4) (in the mechanical or digital domain) as necessary to capture video of different views that encompass one or more of participants 106.
Video conference endpoint 104 uses (i) audio detection techniques to detect audio sources, i.e., loudspeakers 116(1)-116(4), by the microphone arrays 118(1)-118(4) and to determine the spatial relationship between the cameras 112(1)-112(4), display devices 114(1)-114(4), loudspeakers 116(1)-116(4), and microphone arrays 118(1)-118(4); (ii) face detection techniques to detect faces and associated positions thereof of participants 106 around the table 202; and (iii) object detection techniques to detected the shape of specific and known objects, e.g., the table 202.
In accordance with techniques presented herein, video conference endpoint 104 defines/establishes the spatial relationship between cameras 112(1)-112(4) and display devices 114(1)-114(4), and automatically determines which display device 114(1)-114(4) will display certain video feeds received by the video conference endpoint 104. In support of this, video conference endpoint 104 also defines the probability that an audio source detected by the microphone array 118(1)-118(4) is disposed above, below, to the right of, or to the left of the respective cameras 112(1)-112(4) and the respective microphone arrays 118(1)-118(4), and thus also defines the probability that a display device 114(1)-114(4) is disposed above, below, to the right of, or to the left of the respective cameras 112(1)-112(4) and respective microphone arrays 118(1)-118(4). In certain cases described below, endpoint 104 automatically determines which display device 114(1)-114(4) to display a live video feed of remote participants 106 located at a remote video conference endpoint 104.
Referring now to
The first video conference endpoint 104(1) includes a plurality of display devices 114(1)-114(4) positioned around the conference room 200(1). Display devices 114(1)-114(3) may be screens configured to display content from video sources, while display device 114(4) may be a user-interactive digital display device (e.g., a whiteboard or touch screen). Display devices 114(1)-114(4) may contain a camera 112(1)-112(4), respectively, and a microphone array 118(1)-118(4), respectively, integrated with the cameras 112(1)-112(4), respectively. Cameras 112(1)-112(4) are each operated under control of endpoint 104(1) to capture video of different views or scenes of multiple participants 106 seated around a table 202(1) opposite from or facing (i.e., in front of) the cameras 112(1)-112(4) (and display devices 114(1)-114(4)). Furthermore, display devices 114(1)-114(4) may contain an integrated loudspeaker 116(1)-116(4), respectively.
The second video conference endpoint 104(2) includes a plurality of display devices 114(5)-114(8) positioned around the conference room 200(2). Display devices 114(5)-114(7) may be screens configured to display content from video sources, while display device 114(8) may be a user-interactive digital display device (e.g., a whiteboard or touch screen). Display devices 114(5)-114(8) may contain a camera 112(5)-112(8), respectively, and a microphone array 118(5)-118(8), respectively, integrated with the cameras 112(5)-112(8), respectively. Cameras 112(5)-112(8) are each operated under control of endpoint 104(2) to capture video of different views or scenes of multiple participants 106 seated around a table 202(2) opposite from or facing (i.e., in front of) the cameras 112(5)-112(8) (and display devices 114(5)-114(8)). Furthermore, display devices 114(5)-114(8) may contain an integrated loudspeaker 116(5)-116(8), respectively.
As illustrated in
As described herein, the video conference endpoint 104(1) may be configured to use data representing the spatial relationship of video conference components generated according to the techniques presented herein to assign video sources contained in an incoming video feed received from video conference endpoint 104(2) to display devices in conference room 200(1), and to assign outputs from a plurality of cameras in conference room 200(1) in an outgoing video feed to be sent to video conference endpoint 104(2) in conference room 200(2). Similarly, video conference endpoint 104(2) may be configured to use data representing the spatial relationship of video conference components generated according to the techniques presented herein to assign video sources contained in an incoming video feed received from video conference endpoint 104(1) to display devices in conference room 200(2), and to assign outputs from a plurality of cameras in conference room 200(2) in an outgoing video feed to be sent to video conference endpoint 104(2) in conference room 200(1).
With reference to
As illustrated in
Furthermore,
With reference to
Even with multiple display devices 114(1), 114(2), the video conference endpoint 104 is configured to determine on which screen 310(1), 310(2) to display the second video content or live video feed 314, as well as the positioning on the selected screen 310(1), 310(2) such that the live video feed 314 is positioned proximate to the camera 112 to enable better “eye contact” between participants 106 at different endpoints 104. As illustrated in
As illustrated in
As
Reference is now made to
Processor 500 may take the form of a collection of microcontrollers and/or microprocessors, for example, each configured to execute respective software instructions stored in the memory 504. The collection of microcontrollers may include, for example: a video controller to receive, send, and process video signals related to display 114 and video camera 112; an audio processor to receive, send, and process audio signals related to loudspeaker 116 and microphone array 118; and a high-level controller to provide overall control. Portions of memory 504 (and the instruction therein) may be integrated with processor 500. As used herein, the terms “audio” and “sound” are synonymous and interchangeable.
In a distributed processor embodiment, endpoint controller 120 is a distributed processor, including, but not limited to, (i) an audio processor for the microphone array 118 to determine audio angle of arrival of a sound source (as discussed below), and (ii) a video coder/decoder (i.e., codec) that is also configured to analyze the content of the video sources received by the endpoint 104.
The memory 504 may include read only memory (ROM), random access memory (RAM), magnetic disk storage media devices, optical storage media devices, flash memory devices, electrical, optical, or other physical/tangible (e.g., non-transitory) memory storage devices. Thus, in general, the memory 504 may comprise one or more computer readable storage media (e.g., a memory device) encoded with software comprising computer executable instructions and when the software is executed (by the processor 500) it is operable to perform the operations described herein. For example, the memory 504 stores or is encoded with software instructions for Source Display Positioning Module 506 to perform operations described herein for determining the spatial relationship between camera 112 and display devices 114 and determining which display device 114 will display a live video feed. Source Display Positioning Module 506 also includes an Audio Analysis Module 508 and an Image Analysis Module 510. Audio Analysis Module 508 may determine the angle of arrival of a sound source as received by the microphone array 118. Image Analysis Module 510 may evaluate the content of video sources received by the video conference endpoint 104 and determine which display device 114 will display a received video source based on the information acquired by the Audio Analysis Module 508.
With reference to
With reference to
With reference to
At 815, the endpoint controller 120 determines whether the audio output detected by the microphone array 118 originates from a location (e.g., a loudspeaker 116) that is less than a predetermined distance (e.g., three meters) away from the microphone array 118. If it is determined at 815 that the loudspeaker 116 is less than the predetermined distance (three meters) away from the microphone array 118, then the endpoint controller 120 continues to determine the azimuth and elevation angles of the detected audio output with respect to the microphone array 118 at 820 and 850. However, if, at 815, the endpoint controller 120 determines that the detected audio output is not less than the predetermined distance (three meters) away from the microphone array 118, then the endpoint controller 120 skips determining the azimuth and elevation angles of the detected audio, and, at 895, does not update the picture in picture positioning of the live video feed 314. When the detected audio output originates more than the predetermined distance (three meters) from the microphone array 118, the positioning of the live video feed 314 may not be updated because the live video feed 314 may already be disposed in an optimized position. However, other examples of detected audio outputs that originated from more than three meters from the microphone array 118 could include audio outputs that originate from external sources (e.g., talking participants 106, participant 106 devices, etc.), detected the audio outputs that originated from a loudspeaker 116 but reflected off of the floor and/or walls, etc.
After determining that the detected audio output originates from a location less than the predetermined distance (three meters) away from the microphone array 118 (e.g., from a loudspeaker 116 disposed less than three meters from the microphone array 118), then at 820, the endpoint controller 120 calculates whether the audio output detected by the microphone array has an elevation angle φ greater than 75 degrees. If, at 820, the determined elevation angle φ is greater than 75 degrees, then, at 825, the endpoint controller 120 increases a probability value that the loudspeaker 116 is above the camera 112 (P(above)). If, at 820, the determined elevation angle φ is not greater than 75 degrees, then, at 830, the endpoint controller 120 decreases the probability that the loudspeaker 116 is above the camera 112 (P(above)).
If the endpoint controller 120 decreases the probability that the loudspeaker 116 is above the camera (P(above)), then, at 835, the endpoint controller 120 calculates whether the audio output detected by the microphone array has an elevation angle φ less than −75 degrees. If, at 835, the determined elevation angle φ is less than −75 degrees, then, at 840, the endpoint controller 120 increases the probability that the loudspeaker 116 is below the camera 112 (P(below)). If, at 835, the determined elevation angle φ is not less than −75 degrees, then, at 845, the endpoint controller 120 decreases the probability that the loudspeaker 116 is below the camera 112 (P(below)).
At 850, the endpoint controller 120 calculates whether the audio output detected by the microphone array has an azimuth angle θ greater than 75 degrees. If, at 850, the determined azimuth angle θ is greater than 75 degrees, then, at 855, the endpoint controller 120 increases the probability that the loudspeaker 116 is disposed to the right of the camera 112 (P(right)). If, at 850, the determined azimuth angle θ is not greater than 75 degrees, then, at 860, the endpoint controller 120 decreases the probability that the loudspeaker 116 is disposed to the right of the camera 112 (P(right)).
If the endpoint controller 120 decreases the probability that the loudspeaker 116 is disposed to the right of the camera 112 (P(right)), then, at 865, the endpoint controller 120 calculates whether the audio output detected by the microphone array 118 has an azimuth angle θ less than −75 degrees. If, at 865, the determined azimuth angle θ is less than −75 degrees, then, at 870, the endpoint controller 120 increases the probability that the loudspeaker 116 is disposed to the left of the camera 112 (P(left)). If, at 865, the determined azimuth angle θ is not less than −75 degrees, then, at 875, the endpoint controller 120 decreases the probability that the loudspeaker 116 is disposed to the left of the camera 112 (P(left)).
After all of the probabilities P(above), P(below), P(right), P(left) have been calculated and it is verified that the loudspeaker 116 is less than 3 meters away from the microphone array 118, the endpoint controller 120, at 880, determines a spatial relationship value S between the loudspeaker 116 generating the audio output and the microphone array 118 by determining which calculated probability P(above), P(below), P(right), P(left) has the largest value. In one embodiment, the endpoint controller 120 may also disregard any of the probabilities P(above), P(below), P(right), P(left) with lower values. At 885, the endpoint controller 120 then determines whether or not the spatial relationship value S is greater than a predetermined threshold value. If, at 885, the endpoint controller 120 determines that the spatial relationship value S is greater than the predefined threshold, then, at 890, the endpoint controller 120 updates the picture in picture positioning of the live video feed 314 so that the live video feed 314 is positioned proximate to the camera 112 as illustrated in
In another example, with reference to the conference rooms 200, 200(1), and 200(2) depicted in
With continued reference to
With reference to
Once all of the spatial relationship and placement data has been collected, the endpoint controller, at 915, builds an internal model of the placement and relationship of the display devices 114 and cameras 112 in the conference room 200. At 920, the endpoint controller 120 then configures the display device roles and rules for presenting video and audio content based on the derived model and the content of the video and audio sources. For example, if, based on the derived model, the camera 112 is adjacent to (e.g., above, below, to the side) a display device 114 and the content of one of the video and audio sources is a live video feed 314 of another endpoint 104, then the endpoint controller 120 may determine that that display device 114 should receive and display the live video feed 314. In another example, if, based on the derived model, the camera 112 is disposed opposite of the display device 114 (e.g., the camera 112 is across the conference room 200 from a user-interactive display device 114), the endpoint controller 120 may determine that that camera 112 should be utilized to display the presentation presented on that display device 114. Finally, at 925, the endpoint controller 120 presents the three dimensional model to the conference room 200 through the display 600 of the endpoint user control device 204, illustrated in
With reference to
At 1005, the endpoint controller 120 utilizes facial detection software, in conjunction with the camera 112 of a conference room 200 of an endpoint 104, to detect the faces of participants 106 positioned around a table 202 within a conference room 200. Facial detection techniques are well known in the art and are not described in detail herein. At 1010, the endpoint controller 120 analyzes the image captured by the camera 112 and plots a line passing through the detected faces around the table 202, as shown in
After the endpoint controller 120 plots a line passing through the detected faces of the participants 106 within a conference room 200, the endpoint controller 120, at 1015, determines whether or not the value of “a” in the equation y=ax2+bx+c is greater than or equal to zero. If, at 1015, the value of “a” is found to be greater than or equal to zero, then, at 1020, the endpoint controller 120 can make the determination that the camera 112 is disposed at a lower height within the conference room 200 and is thus likely disposed below the display device 114 in the conference room 200.
However, if, at 1015, the value of “a” is found to be less than zero, then, at 1025, the endpoint controller 120 can make the determination that the camera 112 is disposed at a higher height within the conference room 200 and is thus likely disposed above the display device 114 in the conference room 200.
In some embodiments, in order to rely on the facial detection method the detected faces of the participants 106 disposed in the middle of the field of view of the camera 112 need to be smaller than the detected faces of the participants 106 faces disposed on the sides of the field of view of the camera 112. As illustrated in
Returning back to
At 1035, once the information has been collected regarding whether the camera 112 is mounted above or below the display device 114, the endpoint controller 120 uses the information to update the layout of the screen 310 of the display device 114. As previously explained and illustrated with regard to
With reference to
After determining the spatial relationship between the camera 112 and the display devices 114, the endpoint controller 120 may then assign video sources to each of the plurality of display devices based on the data representing the spatial relationship, the content of the video sources, and the use of the camera 112. For example, if it is determined that a camera is disposed adjacent to a display device 114 (e.g., mounted directly above or directly below the display device 114), then that display device 114 may receive a live video feed of another remote video conference endpoint operated by other participants 106 while the camera 112 records a live video feed of the conference room 200 in which it is disposed. The live video feed of the conference room 200 may be sent to the remote video conference endpoint for viewing by participants at that remote video conference endpoint. In another example, as previously explained, if it is determined that another display device, such as a user-interactive display device, is disposed opposite of a camera 112, that camera 112 may be used to show the participant 106 presenting or collaborating on the user-interactive display device 114.
Techniques presented herein automatically determine the roles of the cameras and the display devices of a video conference endpoint within a conference room when participating in a video conference. The detection of the placement of components of a video conference endpoint/system in a room is automated using spatial detection of audio signals emitted by the components using a microphone array and image analysis to optimize screen usage and visualization of the room for simpler control. No additional equipment is needed. Rather, the equipment to be installed as part of the video conference endpoint is used for the process. This makes it easy to add and remove components, such as a digital whiteboard to/from that room. In addition to automatic setup, these techniques can be useful in suggesting layouts and room configurations in a semi-manual wizard-guided type of setup procedure.
Specifically, generated audio from the each of the loudspeakers of the one or more display devices is detected by the microphone arrays of each of the cameras to determine the spatial relationship between each of the cameras and each of the loudspeakers within a conference room. The determined spatial relationship, along with the content of the video and audio sources of the video conference may be used by the controller of the video conference endpoint to automatically determine the roles of the one or more display devices and each of the cameras of the conference room. This eliminates a need to manually set up each video conference room, and eliminates the need to have participants of the video conference manually switch the roles of the display devices and the cameras during a video conference.
In summary, in one form, a method is provided comprising: detecting, at a microphone array having a predetermined physical relationship with respect to a camera, the audio emitted from each of one or more loudspeakers, each loudspeaker having a predetermined physical relationship with respect to at least one of one or more display devices in a conference room; and generating data representing a spatial relationship between each of the one or more display devices and the camera based on the detected audio.
In another form, an apparatus is provided comprising: a camera configured to capture video of a field of view; a microphone array having a predetermined physical relationship with respect to the camera, the microphone array configured to transduce audio received at the microphone array; and a processor to control the camera and the microphone array to: cause the microphone array to detect audio emitted from one or more loudspeakers having a predetermined physical relationship with respect to at least one of one or more display devices in a conference room; and generate data representing a spatial relationship between each of the one or more display devices and the camera based on the detected audio.
In yet another form, a (non-transitory) processor readable medium is provided. The medium stores instructions that, when executed by a processor, cause the processor to: detect, at a microphone array having a predetermined physical relationship with respect to a camera, the audio emitted from each of one or more loudspeakers, each loudspeaker having a predetermined physical relationship with respect to at least one of one or more display devices in a conference room; and generate data representing a spatial relationship between each of the one or more display devices and the camera based on the detected audio.
As described herein, the data representing the spatial relationship may be used to assign one or more video sources of an incoming video feed from a remote conference room to corresponding ones of the one or more display devices. Similarly, the data representing the spatial relationship may be used to assign video outputs from a plurality of cameras in a conference room to an outgoing video feed to be sent to a remote conference room.
The above description is intended by way of example only. Various modifications and structural changes may be made therein without departing from the scope of the concepts described herein and within the scope and range of equivalents of the claims.
Claims
1. A method comprising:
- detecting, at a microphone array having a predetermined physical relationship with respect to a plurality of cameras, audio emitted from each of one or more loudspeakers, each loudspeaker having a predetermined physical relationship with respect to at least one of one or more display devices in a conference room;
- generating data representing a spatial relationship between each of the one or more display devices and each of the plurality of cameras based on the detected audio, where generating data representing the spatial relationship includes determining azimuth and elevation angles between each of the plurality of cameras and each of the one or more display devices; and
- assigning one or more video sources of an incoming video feed from a remote conference room to corresponding ones of the one or more display devices based on the data representing the spatial relationship and content of the one or more video sources.
2. (canceled)
3. (canceled)
4. The method of claim 1, wherein generating data representing the spatial relationship further comprises:
- determining, based on the determined azimuth and elevation angles, at least one of: a first probability that a first camera of the plurality of cameras is disposed above one of the display devices, a second probability that the first camera is disposed below one of the display devices, a third probability that the first camera is disposed right of one of the display devices, or fourth probability that the first camera is disposed left of one of the display devices.
5. The method of claim 4, wherein assigning comprises:
- assigning a video source of the incoming video feed to be displayed on a top side of a screen, a bottom side of the screen, a right side of the screen, or a left side of the screen of one of the display devices based on where the first probability, second probability, third probability and fourth probability indicate the first camera is disposed with respect to the one of the display devices.
6. (canceled)
7. The method of claim 27, further comprising:
- tagging the respective video outputs from the plurality of cameras with data indicative of a respective field of view of each of the plurality of cameras in the conference room.
8. The method of claim 27, wherein one of the display devices is a user-interactive display device, and assigning further comprises:
- assigning, based on the data representing the spatial relationship, a particular one of the plurality of cameras that is positioned in the conference room opposite the user-interactive display device to capture a participant presenting on the user-interactive display device and a surrounding area of the user-interactive display device.
9. An apparatus comprising:
- a plurality of cameras configured to capture video within a conference room;
- a microphone array having a predetermined physical relationship with respect to the plurality of cameras, the microphone array configured to transduce audio received at the microphone array; and
- a processor configured to control the plurality of cameras and the microphone array, wherein the processor is configured to: cause the microphone array to detect audio emitted from one or more loudspeakers having a predetermined physical relationship with respect to at least one of one or more display devices in the conference room; generate data representing a spatial relationship between each of the one or more display devices and each of the plurality of cameras based on the detected audio by determining azimuth and elevation angles between each of the plurality of cameras and each of the one or more display devices; and assign one or more video sources of an incoming video feed from a remote conference room to corresponding ones of the one or more display devices based on the data representing the spatial relationship and content of the one or more video sources.
10. (canceled)
11. (canceled)
12. The apparatus of claim 9, wherein the processor, when generating data representing the spatial relationship, is further configured to:
- determine, based on the determined azimuth and elevation angles, at least one of: a first probability that a first camera of the plurality of cameras is disposed above one of the display devices, a second probability that the first camera is disposed below one of the display devices, a third probability that the first camera is disposed right of one of the display devices, or fourth probability that the first camera is disposed left of one of the display devices.
13. The apparatus of claim 12, wherein the processor is further configured to:
- assign a video source of the incoming video feed to be displayed on a top side of a screen, a bottom side of the screen, a right side of the screen, or a left side of the screen of one of the display devices based on where the first probability, second probability, third probability and fourth probability indicate the first camera is disposed with respect to the one of the display devices.
14. (canceled)
15. The apparatus of claim 28, wherein the processor is further configured to:
- tag the respective video outputs from the plurality of cameras with data indicative of a respective field of view of each of the plurality of cameras in the conference room.
16. The apparatus of claim 28, wherein one of the display devices is a user-interactive display device and the processor is further configured to:
- assign, based on the data representing the spatial relationship, a particular one of the plurality of cameras that is positioned in the conference room opposite the user-interactive display device to capture a participant presenting on the user-interactive display device and a surrounding area of the user-interactive display device.
17. One or more non-transitory computer readable storage media, the computer readable storage media being encoded with software comprising computer executable instructions, and when the software is executed, operable to:
- detect, at a microphone array having a predetermined physical relationship with respect to a plurality of cameras, audio emitted from each of one or more loudspeakers, each loudspeaker having a predetermined physical relationship with respect to at least one of one or more display devices in a conference room;
- generate data representing a spatial relationship between each of the one or more display devices and each of the plurality of cameras based on the detected audio by determining azimuth and elevation angles between each of the plurality of cameras and each of the one or more display devices; and
- assign one or more video sources of an incoming video feed from a remote conference room to corresponding ones of the one or more display devices based on the data representing the spatial relationship and content of the one or more video sources.
18. (canceled)
19. (canceled)
20. (canceled)
21. The non-transitory computer-readable storage media of claim 29, wherein the instructions are further operable to:
- tag the respective video outputs from the plurality of cameras with data indicative of a respective field of view of each of the plurality of cameras in the conference room.
22. The non-transitory computer-readable storage media of claim 17, wherein the instructions, when generating data representing the spatial relationship, are further operable to:
- determine, based on the determined azimuth and elevation angles, at least one of: a first probability that a first camera of the plurality of cameras is disposed above one of the display devices, a second probability that the first camera is disposed below one of the display devices, a third probability that the first camera is disposed right of one of the display devices, or fourth probability that the first camera is disposed left of one of the display devices.
23. The non-transitory computer-readable storage media of claim 22, wherein the instructions are further operable to:
- assign a video source of the incoming video feed to be displayed on a top side of a screen, a bottom side of the screen, a right side of the screen, or a left side of the screen of one of the display devices based on where the first probability, second probability, third probability and fourth probability indicate the first camera is disposed with respect to the one of the display devices.
24. The method of claim 1, further comprising:
- determining, from the detected audio of each of the one or more loudspeakers, whether each of the one or more loudspeakers is within a predetermined distance from the microphone array.
25. The apparatus of claim 9, wherein the processor is further configured to:
- determine, from the detected audio of each of the one or more loudspeakers, whether each of the one or more loudspeakers is within a predetermined distance from the microphone array.
26. The non-transitory computer-readable storage media of claim 17, wherein the instructions are further operable to:
- determine, from the detected audio of each of the one or more loudspeakers, whether each of the one or more loudspeakers is within a predetermined distance from the microphone array.
27. The method of claim 1, further comprising:
- assigning video outputs from the plurality of cameras to an outgoing video feed based on the data representing the spatial relationship, the outgoing video feed to be sent from the conference room to a remote conference room.
28. The apparatus of claim 9, wherein the processor is further configured to:
- assign video outputs from the plurality of cameras to an outgoing video feed based on the data representing the spatial relationship, the outgoing video feed to be sent from the conference room to a remote conference room.
29. The non-transitory computer readable storage media of claim 17, wherein the instructions are further operable to:
- assign video outputs from the plurality of cameras to an outgoing video feed based on the data representing the spatial relationship, the outgoing video feed to be sent from the conference room to a remote conference room.
Type: Application
Filed: Oct 31, 2016
Publication Date: May 3, 2018
Inventors: Glenn R. G. Aarrestad (Hovik), Lennart Burenius (Oslo), Jochen Christof Schirdewahn (Stabekk), Johan Ludvig Nielsen (Oslo)
Application Number: 15/338,676