COMMUNICATION SYSTEM, COMMUNICATION APPARATUS AND COMMUNICATION METHOD
A communication apparatus includes: an area dividing unit that divides, of an output image which is an image that the communication apparatus has output and includes at least a shot image obtained by shooting the surroundings of another communication apparatus which communicates with the communication apparatus, a display area indicating an area where the shot image is displayed into as many unit areas as the number of people captured in the shot image; a transmission control unit that controls transmitting coordinate information which indicates, of the output image, a unit area corresponding to a position pointed by a user to the other communication apparatus; and an output control unit that, when the communication apparatus has received output information including a voice subjected to directivity control according to, of the shot image, an area corresponding to the coordinate information from the other communication apparatus, controls outputting the received output information,
Latest RICOH COMPANY, LTD. Patents:
- REBOOTING SYSTEM, REBOOTING METHOD, AND RECORDING MEDIUM
- Information processing apparatus, information processing system, and information processing method
- Image reading device determining transmission destination
- Image forming apparatus
- Information processing apparatus, information processing system, and information processing method
The present application claims priority under 35 U.S.C. §119 to Japanese Patent Application No. 2015-212270 filed Oct. 28, 2015. The contents of which are incorporated. herein by reference in their entirety.
BACKGROUND GE THE INVENTION1. Field of the Invention
The present invention relates to a communication system, a communication apparatus, and a communication method.
2. Description of the Related Art
As a form of a communication system for performing data transmission/reception among multiple communication apparatuses, there is known a video (television) conference system that realizes a remote conference by using a communication network. This video conference system can conduct a conference between remote locations almost like a face-to-face meeting in such a way that by using a communication apparatus (a terminal device) of the video conference system set up in a conference room where one of the parties participating in the remote conference, an image of the conference room such as conference participants and a voice, such as speaker's voice, are converted into digital data, and the digital data is transmitted to a communication apparatus of the other party to display the image on a display and output the voice from a speaker in a conference room of the other party.
In such a video conference system, a microphone is used to acquire conference participants' voice, and a camera is used to acquire their image. The camera has an angle of view, and therefore cannot capture an image of a conference participant who is outside the angle of view. To solve this problem, there is known a method of using a panoramic camera capable of capturing a 360-degree panoramic image. On the other hand, a general microphone is omnidirectional, and therefore picks up the ambient sound besides participant's speech. To solve this problem, there is known a method of using a microphone array to control the directivity, i.e., increase the sensitivity of a microphone that picks up the sound from a particular direction, thereby suppressing the ambient sound from being picked up and enabling participant's speech to be heard clearly. For example, Japanese Unexamined Patent Application Publication No. 2007-274463 has disclosed a method of how a conference terminal forms a beam of sound picked up by a microphone array according to a selected placement pattern of participants. Furthermore, Japanese Patent No. 5028944 has disclosed a technology to detect the direction of a speaker by use of a microphone array composed of an array of a plurality of microphones and make the shooting direction of a camera follow the detected direction.
There can be considered a method of realizing a more realistic video conference with a combination of a panoramic camera and a microphone array, namely, by shooting a whole conference room with the panoramic camera and, if there is a speaker, setting the sound pickup area of a microphone toward at least the speaker. However, this combination can set the sound pickup area of a microphone toward a speaker, though a communication apparatus of the other party may not want output focused on the speaker. Therefore, there is a problem that this combination does not enable an output intended by each communication apparatus composing a communication system.
In view of the above, there is a need to provide a communication system, a communication apparatus, and a communication method that enable one's intended output.
SUMMARY OF THE INVENTIONAccording to exemplary embodiments of the present invention, there is provided a communication system comprising a first communication apparatus and a second communication apparatus that transmits/receives data to/from the first communication apparatus, wherein the first communication apparatus includes: an area dividing unit that divides, of an output image which is an image that the first communication apparatus has output and includes at least a shot image obtained by shooting the surroundings of the second communication apparatus, a display area indicating an area where the shot image is displayed into as many unit areas as the number of people captured in the shot image; a first transmission control unit that performs control of transmitting coordinate information which indicates, of the output image, a unit area corresponding to a position pointed by a user to the second communication apparatus; and an output control unit that, when the first communication apparatus has received output information including a voice subjected to directivity control according to, of the shot image, an area corresponding to the coordinate information from the second communication apparatus, performs control of outputting the received output information, and the second communication apparatus includes; an acquiring unit that acquires the shot image; an identifying unit that, when the second communication apparatus has received the coordinate information from the first communication apparatus, identifies, of the shot image acquired by the acquiring unit, an area corresponding to the received coordinate information based on correspondence information indicating a correspondence relationship between the coordinates of the output image and the coordinates of the shot image; a directivity control unit that controls the directivity of a microphone array including a plurality of microphones to increase the sensitivity of, of the microphones installed in the second communication apparatus, a microphone corresponding to the area identified by the identifying unit; and a second transmission control unit that performs control of transmitting output information including at least a voice subjected to directivity control by the directivity control unit to the first communication apparatus.
Exemplary embodiments of the present invention also provide a communication apparatus comprising: an area dividing unit that divides, of an output image which is an image that the communication apparatus has output and includes at least a shot image obtained by shooting the surroundings of another communication apparatus which communicates with the communication apparatus, a display area indicating an area where the shot image is displayed into as many unit areas as the number of people captured in the shot image; a transmission control unit that performs control of transmitting coordinate information which indicates, of the output image, a unit area corresponding to a position pointed by a user to the other communication apparatus; and an output control unit that, when the communication apparatus has received output information including a voice subjected to directivity control according to, of the shot image, an area corresponding to the coordinate information from the other communication apparatus, performs control of outputting the received output information.
Exemplary embodiments of the present invention also provide a communication method for a communication system including a first communication apparatus and a second communication apparatus that transmits/receives data to/from the first communication apparatus, the communication method comprising: dividing, by the first communication apparatus, of an output image which is an image that the first communication apparatus has output and includes at least a shot image obtained by shooting the surroundings of the second communication apparatus, a display area indicating an area where the shot image is displayed into as many unit areas as the number of people captured in the shot image; performing, by the first communication apparatus, control of transmitting coordinate information which indicates, of the output image, a unit area corresponding to a position pointed by a user to the second communication apparatus; when having received output information including a voice subjected to directivity control according to, of the shot image, an area corresponding to the coordinate information from the second communication apparatus, performing, by the first communication apparatus, control of outputting the received output information; acquiring, by the second communication apparatus, the shot image; when having received the coordinate information from the first communication apparatus, identifying, by the second communication apparatus, of the shot image acquired at the acquiring, an area corresponding to the received coordinate information on the basis of correspondence information indicating a correspondence relationship between the coordinates of the output image and the coordinates of the shot image; controlling, by the second communication apparatus, the directivity of a microphone array including a plurality of microphones to increase the sensitivity of, of the microphones installed in the second communication apparatus, a microphone corresponding to the area identified at the identifying; and performing, by the second communication apparatus, control of transmitting output information including at least a voice subjected to directivity control at the controlling to the first communication apparatus.
Exemplary embodiments of the present invention also provide a communication method for a communication apparatus, the communication method comprising: dividing, of an output image which is an image that the communication apparatus has output and includes at least a shot image obtained by shooting the surroundings of another communication apparatus which communicates with the communication apparatus, a display area indicating an area where the shot image is displayed into as many unit areas as the number of people captured in the shot image; performing control of transmitting coordinate information which indicates, of the output image, a unit area corresponding to a position pointed by a user to the other communication apparatus; and when the communication apparatus has received output information including a voice subjected to directivity control according to, of the shot image, an area corresponding to the coordinate information from the other communication apparatus, performing control of outputting the received output information.
The accompanying drawings are intended to depict exemplary embodiments of the present invention and should not be interpreted to limit the scope thereof. Identical or similar reference numerals designate identical or similar components throughout the various drawings.
DESCRIPTION OF THE EMBODIMENTSThe terminology used herein is for the purpose of describing particular embodiments only and is not intended to be limiting of the present invention.
As used herein, the singular forms “a”, “an” and “the” are intended to include the plural forms as well, unless the context clearly indicates otherwise.
In describing preferred embodiments illustrated in the drawings, specific terminology may be employed for the sake of clarity. However, the disclosure of this patent specification is not intended to be limited to the specific terminology so selected, and it is to be understood that each specific element includes all technical equivalents that have the same function, operate in a similar manner, and achieve a similar result.
An exemplary embodiment of a communication system, a communication apparatus, and a communication method according to the present invention will be described in detail below with reference to accompanying drawings.
The server 4 monitors whether each conference terminal 2 is connected to the server 4, and performs control required at the time of a conference, such as control of calling the conference terminals 2 at the start of the conference. When one conference terminal 2 transmits data during the conference, the conference terminal 2 transmits image and voice data to the server 4, and the server 4 transmits the image and voice data to the other conference terminal 2 on the side of the other party. When one conference terminal 2 receives data, the conference terminal 2 receives image and voice data of the other conference terminal 2 on the side of the other party through the server 4. For example, when a conference is conducted in Locations A and B, data that a conference terminal 2 of Location A has transmitted is transmitted to a conference terminal 2 of Location B through the server 4, and is not transmitted to the other conference terminals 2 (conference terminals 2 not participating in the conference). Likewise, data that the conference terminal 2 of Location B has transmitted is transmitted to the conference terminal 2 of Location A participating in the conference through the server 4, and is not transmitted to the other conference terminals 2 not participating in the conference. By performing the control described above, a conference can be conducted among multiple conference terminals 2 (in multiple locations).
Subsequently, a configuration of the conference terminals 2 is explained. Incidentally, the conference terminals 2 set up in Locations A and B have the same configuration; therefore, in the following description, one conference terminal 2 is cited.
The panoramic camera 10 is an example of an “image shooting unit,” and generates a shot image obtained by shooting an image. In this example, the panoramic camera 10 generates a panoramic image (an example of a shot image) obtained by shooting a 360-degree panorama around the panoramic camera 10 (which can be considered as a 360-degree panorama around the conference terminal 2), and transmits the generated panoramic image to the CPU 14. For example, the panoramic camera 10 is composed of a known omnidirectional camera or the like. By shooting a 360-degree panorama around the panoramic camera 10, a shot image in which all conference participants existing around the panoramic camera 10 are captured can be generated. The term “panoramic image” here means an image generated by combining multiple images (shot images) taken by moving one camera with a plurality of imaging sensors or by using multiple cameras with a plurality of imaging sensors. Incidentally, in this example, the area shot by the panoramic camera 10 is a 360-degree panorama around the panoramic camera 10; however, the area can be smaller than this.
The display unit 11 has a function of displaying image data received from the CPU 14 on a screen. In this example, the display unit 11 is composed of a liquid crystal display device or the like.
The microphone array 12 includes a plurality of microphones installed dispersively in the conference terminal 2, and has a function of acquiring conference participant's voice and transmitting the acquired voice to the CPU 14. The term “microphone array” is composed of a plurality of omnidirectional microphones, and forms the directivity by means of a beamformer. In general, the term “beamformer” is a technique to form the directivity by use of differences in the time for sound to reach microphones. The directivity formed by the beamformer can freely set the target direction, such as a vertical direction or a horizontal direction, according to the arrangement microphones. Therefore, the area where the microphone array wants to pick up sound (the target area) can be arbitrarily changed.
Returning to the explanation of
The storage device 15 stores therein various control programs (such as a video conference control program) executed by the CPU 14, a conversion table to be described later, etc. The storage device 15 is, for example, a non-volatile storage medium such as a flash memory or an HDD.
The memory 16 is used for unfolding of a program executed by the CPU 14 and temporary storage of operation data. The memory 16 is, for example, a volatile memory such as a DDR memory. The LAN I/F unit 17 connects the conference terminal 2 to another conference terminal 2 via the network 3, and transmits/receives data (image and voice data) to/from the other conference terminal 2. The LAN I/F unit 17 is, for example, a wired LAN that is compatible with 10BASE-T, 100BASE-TX, or 1000BASE-T and is connected to an Ethernet(trademark) network or a wireless LAN compatible with 802.11a/b/g/n/ac.
The operation unit 18 is a device used in various operations (various operations related to control over devices of the conference terminal 2) made by a user, and includes, for example, a keyboard, buttons, etc.
The projector 300 has a function of performing wireless communication with each of the stylus pen 400 and the conference terminal 2; the stylus pen 400 has a function of performing wireless communication with the projector 300. The projector 300 performs wireless communication with the stylus pen 400, thereby can acquire information indicating the position at which the stylus pen 400 is pointing on the projection plane from the stylus pen 400, and therefore can be always aware of the position at which the stylus pen 400 is pointing (which can be considered as the position at which a user is pointing) on the projection plane. Furthermore, the conference terminal 2, also, performs wireless communication with the projector 300, thereby can be always aware of the position at which the stylus pen 400 is pointing on the projection plane.
In this example, a pair of the same projector 300 and its dedicated stylus pen 400 is set up in each of Locations A and B. Here, a conference terminal 2 and a projector 300 connected to the conference terminal 2 correspond to a “first communication apparatus” or a “second communication apparatus.” Incidentally, the type of an output device connected to the conference terminal 2 is optional; for example, an interactive whiteboard can be connected to the conference terminal 2.
Furthermore, for example, a single device having both the functions of the conference terminal 2 and the functions of the projector 300 can be set up in each location. In this configuration, the single device corresponds to the “first communication apparatus” or the “second communication apparatus.”
As shown in
The network unit 100 transmits/receives data to/from another conference terminal 2 that is the other party of a conference.
The control unit 101 is a part that performs various controls and arithmetic operations. Functions that the control unit 101 has will be described in detail later.
The display control unit 102 has a function of controlling display (projection of an image on the projection plane) performed by the projector 300.
The wireless communication unit 109 performs wireless communication with the projector 300, and acquires position information that indicates the position at which the stylus pen 400 is pointing on the projection plane from the projector 300. The wireless communication unit 109 can notify the control unit 101 of the position information acquired from the projector 300.
A panoramic image generated by the panoramic camera 10 is sent to the camera I/F receiver 111. The camera I/F receiver 111 is assumed to be a high-speed serial I/F, such as V-by-One (trademark) or HDMI (trademark). The distortion-correction processing unit 113 corrects distortion of the panoramic image subjected to serial/parallel conversion by the camera I/F receiver 111, and outputs the corrected panoramic image to the control unit 101. The lens-characteristics holding unit 112 stores therein conversion parameters for correcting distortion according to lens characteristics, and the distortion-correction processing unit 113 can use a conversion parameter to correct distortion of a panoramic image.
The CPU 311 executes a program stored in the storage unit 312, and controls the operation of the projector 300 overall. The storage unit 312 is composed of a ROM or HDD for storing therein a program executed by the CPU 311 and data required to execute the program, a RAM that serves as a work area of the CPU 311, etc. The input unit 313 is used to perform various inputs to the projector 300, and includes a touch panel, key switches, etc. The communication I/F 314 is an interface for communicating with the stylus pen 400 and the conference terminal 2. The projecting unit 315 projects image data to be projected on the projection plane, such as a screen. The projecting unit 315 includes a projection optical system, such as a projection lens. Functions that the projector 300 has will be described later.
Also in Location 2, a conference terminal 2 is put on a desk. Just like in Location 1, since the conference terminal 2 is equipped with the panoramic camera 10, the conference terminal 2 is assumed to be put on the center of the desk, and is equipped with the microphone array 12 including the microphones a to f. In Location 2, it is shown that three people, Persons A to C, are participating in the video conference. Just like in Location 1, a projector 300 is connected to the conference terminal 2 set up in Location 2 via a video output cable; an image including a panoramic image on the side of Location 1 (an image in which a panoramic image on the side of Location 1 is displayed) is projected on a projection plane in Location 2. Below a projected image 2, a panoramic image in which all people participating in the conference in Location 1 (in this example, Persons D to H) are captured is displayed. A speaker is normally displayed in a cutout image; however, for example, if any one of the conference participants in Location 2 has specified an area enclosed by a dotted line with the stylus pen 400 as shown in
How to specify a cutout area of a projected image is explained with
As will be described later, in the present embodiment, of a projected image, a display area indicating an area where a panoramic image (a panoramic image on the side of the other party of a conference) is displayed is divided into as many unit areas as the number of people captured in the panoramic image; when a user has performed an operation of pointing at the center point (which is not necessarily the center point) of any one of the people captured in the display area as the center point of a cutout with the stylus pen 400 or the like, a unit area including the center point of the cutout is specified as a cutout area. This can simplify the operation of specifying, of a display area of a projected image, an area in which an object to be focused on is captured (the operation of specifying a cutout area), and therefore, it is possible to improve the user-friendliness.
Then, upon receipt of pressing on an Exit icon 320 to instruct to end the cutout-area specification as shown in
Here, as shown in
The conference terminal 2 on the side of Location 1 generates output information including a cutout image which is, of the panoramic image on the side of Location 1, an area corresponding to the coordinate information received from the conference terminal 2 on the side of Location 2 and a voice subjected to directivity control according to the area, and transmits the generated output information to the conference terminal 2 on the side of Location 2. In the example shown in
Then, the conference terminal 2 of the Location 1 side transmits the output information including the cutout image cut out as described above and the voice subjected to directivity control to the conference terminal 2 of the Location 2 side. The conference terminal 2 of the Location 2 side outputs the output information received from the conference terminal 2 of the Location 1 side.
Detailed contents of respective functions that a projector 300 and the control unit 101 of a conference terminal 2 have are explained below. First, functions that a projector 300 has are explained.
The projection control unit 321 performs control of projecting an image input from the conference terminal 2 on a projection plane under the control of the display control unit 102.
The area dividing unit 322 divides, of a projected image indicating an image that the the projector 300 has projected on the projection plane, a display area indicating an area where a panoramic image (a panoramic image on another conference terminal 2 that is the other party of a conference) obtained by shooting the surroundings of the other conference terminal 2 of the other party is displayed into as many unit areas as the number of people captured in the panoramic image. In the present embodiment, the area dividing unit 322 equally divides the display area into as many unit areas as the number of users operations performed to point at people captured in the display area (which can be considered as the number of positions in the display area indicated by the user's operations), thereby can obtain a plurality of unit areas.
For example, a projected image output by the projector 300 is assumed. to be as shown in
Returning to the explanation of
The coordinate-information-transmission control unit 324 performs control of transmitting coordinate information indicating a unit area selected by the unit-area selecting unit 323 to the conference terminal 2 connected to the projector 300.
The above-described functions that the projector 300 has (the projection control unit 321, the area dividing unit 322, the unit-area selecting unit 323, and the coordinate-information-transmission control unit 324) can be realized by the CPU 311 executing a program stored in the storage unit 312 or the like, or at least some of these functions can be realized by a dedicated hardware circuitry (such as a semiconductor integrated circuit).
Subsequently, functions that the control unit 101 of a conference terminal 2 has are explained.
As shown in
The first transmission control unit 121 performs, when having received coordinate information from a projector 300 connected to the conference terminal 2, control of transmitting the received coordinate information to another conference terminal 2 that is the other party of a conference. That is, the first transmission control unit 121 performs control of transmitting coordinate information indicating, of a projected image, a unit area corresponding to the position pointed by a user to another conference terminal 2 that is the other party of a conference.
The acquiring unit 122 acquires a panoramic image obtained by the panoramic camera 10 shooting the surroundings of the conference terminal 2. In this example, the acquiring unit 122 acquires a corrected panoramic image input from the distortion-correction processing unit 113.
When the conference terminal 2 has received from another conference terminal 2 coordinate information indicating, of a display area (an area where a panoramic image of the conference terminal 2 side is displayed) of a projected image that the other conference terminal 2 has output, a unit area corresponding to the position pointed by a user of the other conference terminal 2, the identifying unit 123 identifies, of a panoramic image acquired by the acquiring unit 122, an area corresponding to the received coordinate information based on correspondence information indicating a correspondence relationship between the coordinates of the projected image (which can be considered as the coordinates of an area on which an image is projected on a projection plane) and the coordinates of the panoramic image. In this example, the correspondence information has been stored in the storage device 15 in advance. Furthermore, a general video conference system enables a user to freely change the layout (change the display mode), such as to project only an image of the user's party or to project only an image of the other party; therefore, the relationship between the coordinates of a projected image and the coordinates of a panoramic image is not always in one-to-one correspondence. Accordingly, the correspondence information in this example associates the coordinates of a projected image with the coordinates of a panoramic image with respect to each display mode (layout information) of the projector 300.
The cutting-out unit 124 cuts out, of a panoramic image acquired by the acquiring unit 122, an image of an area identified by the identifying unit 123 as a cutout image.
The directivity control unit 125 controls the directivity of the microphone array 12 to increase the sensitivity of, of a plurality of microphones installed dispersively in the conference terminal 2, a microphone corresponding to an area identified by the identifying unit 123 (in this example, an area within a panoramic image). The directivity control unit 125 can determine the microphone corresponding to the coordinates of the area identified by the identifying unit 123 based on position information indicating a relationship between the positions of the microphones included in the microphone array 12 and the coordinate of the panoramic image. The position information can be stored in, for example, the storage device 15 or the like in advance.
The second transmission control unit 126 performs control of transmitting output information including at least a voice subjected to directivity control by the directivity control unit 125 to another conference terminal 2. In the present embodiment, the second transmission control unit 126 performs control of transmitting output information including a voice subjected to directivity control by the directivity control unit 125 and a cutout image cut out by the cutting-out unit 124 to another conference terminal 2. More specifically, the second transmission control unit 126 performs control of transmitting output information including a panoramic image acquired by the acquiring unit 122, a voice subjected to directivity control by the directivity control unit 125, and a cutout image cut out by the cutting-out unit 124 to another conference terminal 2. Incidentally, the output information may include at least a voice subjected to directivity control by the directivity control unit 125 (a voice subjected to directivity control according to, of a shot image acquired by the acquiring unit 122, an area corresponding to coordinate information received from another conference terminal 2). Furthermore, the control unit 101 can be configured to not include, for example, the cutting-out unit 124.
Furthermore, if the conference terminal 2 has not received coordinate information from another conference terminal 2, the second transmission control unit 126 performs control of transmitting general conference information including a panoramic image acquired by the acquiring unit 122, a cutout image that is a close-up of a speaker who is one of conference participants captured in the panoramic image, and voice data picked up by the microphone array 12 to the other conference terminal 2.
The output control unit 127 performs control of outputting an image and voice received from another conference terminal 2. The output control unit 127 performs control of instructing the display control unit 102 to cause the projector 300 to output (project) an image received from another conference terminal 2 onto a projection plane and outputting a voice received from the other conference terminal 2 from the speaker 13. When the conference terminal 2 has received output information from another conference terminal 2, the output control unit 127 in the present embodiment performs control of outputting the received output information. More specifically, the output control unit 127 performs control of instructing the display control unit 102 to output a composite image of a cutout image and a panoramic image that are included in the received output information and outputting a voice included in the received output information from the speaker 13.
Furthermore, when the conference terminal 2 has received general conference information from another conference terminal 2, the output control unit 127 performs control of outputting the received general conference information.
The above-described functions that the control unit 101 has (the first transmission control unit 121, the acquiring unit 122, the identifying unit 123, the cutting-out unit 124, the directivity control unit 125, the second transmission control unit 126, and the output control unit 127) can be realized by the CPU 14 executing a program stored in the storage device 15 or the like, or at least some of the functions that the control unit 101 has can be realized by a dedicated hardware circuitry (such as a semiconductor integrated circuit).
Furthermore, in the above example, the panoramic camera 10 and the speaker 13 are included in the conference terminal 2; however, the configuration of the conference terminals 2 is not limited to this, and, for example, the panoramic camera 10 and the speaker 13 can be provided outside of the conference terminals 2.
As described above, upon receipt of coordinate information from another conference terminal 2 that is the other party of a conference, the conference terminal 2 in the present embodiment identifies, of a panoramic image acquired from the panoramic camera 10 of the conference terminal 2, an area corresponding to the received coordinate information based on correspondence information, and cuts out an image of the identified area as a cutout image. Furthermore, the conference terminal 2 controls the directivity of the microphone array 12 to increase the sensitivity of, of a plurality of microphones installed dispersively in the conference terminal 2, a microphone corresponding to the identified area. Then, the conference terminal 2 transmits output information including the cutout image and a voice subjected to the directivity control to the other conference terminal 2, and the other conference terminal 2 outputs the received output information. Accordingly, the other conference terminal 2 can perform an intended output. Incidentally, in the present embodiment, the shooting range of the panoramic camera is a 360-degree panorama. However, the essentials of the present embodiment are to specify a portion of a shot image as a cutout image and control the directivity of the microphone array to increase the sensitivity of a microphone corresponding to the cutout image. Therefore, as the shooting range, the angle of view of the camera can be below 360 degrees, and, for example, can be about 80 degrees.
The embodiment according to the present invention is explained above; however, the present invention is not limited to the embodiment as is described above, and, in the practical phase, components can be modified without departing from the scope of the invention. Furthermore, various inventions can be formed by appropriate combinations of several of the components described in the above embodiment. Moreover, for example, some of the components described in the above embodiment can be eliminated.
In the above embodiment, the area dividing unit 322 equally divides the display area into as many unit areas as the number of user's operations performed to point at people captured in the display area, thereby obtaining a plurality of unit areas; however, for example, we assume a case where the positions of people captured in the display area are converged on one side as shown in
Accordingly, for example, the area dividing unit 322 can be configured to divide the display area into unit areas that correspond one-to-one to multiple positions in the display area indicated by user's operations on the basis of a relative positional relationship between four vertices of a projected image and the multiple positions in the display area indicated by the user's operations. Specific contents of this are explained below.
When the projector 300 has gone into the calibration mode, the area dividing unit 322 performs control of outputting information (which can be an image or a voice) prompting a user to point at four vertices of a projected image. Incidentally, for example, if an output device connected to the conference terminal 2 is an interactive whiteboard, the area dividing unit 322 does not have to perform this control because respective coordinates of four vertices of an image (an example of an output image) that the interactive whiteboard displays thereon are recognized in advance.
Then, the area dividing unit 322 performs control of outputting information prompting the user to point at the center point of each of the people captured in the display area, and detects multiple positions in the display area indicated by user's operations. By figuring out a relative positional relationship between the multiple positions in the display area indicated by the user's operations (which may be referred to as the “indicated positions” in the following description) and four vertices of the display area, the positions of cutout lines (cutout lines extending vertically) for cutting out multiple unit areas that correspond one-to-one to the indicated positions can be found.
Here we explain this focusing on one of the indicated positions. In the following description, the focused indicated position is referred to as the “focused position.” First, we assume a case where there are the indicated positions adjacent to both sides of the focused position. In this case, the lateral position of a right-hand cutout line for cutting out a unit area including the focused position can be found as the position at a distance from the left side of the display area in a lateral direction of, for example, a difference between the lateral length of the display area and the distance from the right side of the display area to the lateral position of the focused position plus half the distance between the lateral position of the focused position and the lateral position of the indicated position adjacent to the right side of the focused position. Likewise, the lateral position of a left-hand cutout line for cutting out the unit area including the focused position can be found as the position at a distance from the left side of the display area in the lateral direction of, for example, the difference between the lateral length of the display area and the distance from the right side of the display area to the lateral position of the focused position minus half the distance between the lateral position of the focused position and the lateral position of the indicated position adjacent to the left side of the focused position.
Next, we assume a case where there is no indicated position adjacent to the left side of the focused position, and there is the indicated position adjacent to the right side of the focused position. In this case, the lateral position of a right-hand cutout line for cutting out a unit area including the focused position can be found in the same manner as the first case. On the other hand, the lateral position of a left-hand cutout line for cutting out the unit area including the focused position can be found as the position at a distance from the left side of the display area in the lateral direction of, for example, a difference between the lateral length of the display area and the distance from the right side of the display area to the lateral position of the focused position minus half the distance between the lateral position of the focused position and the lateral position of the indicated position adjacent to the right side of the focused position.
Furthermore, we assume a case where there is the indicated position adjacent to the left side of the focused position, and there is no indicated position adjacent to the right side of the focused position. In this case, the lateral position of a left-hand cutout line for cutting oat a unit area including the focused position can be found in the same manner as the first case. On the other hand, the lateral position of a right-hand cutout line for cutting out the unit area including the focused position can be found as the position at a distance from the left side of the display area in the lateral direction of, for example, a difference between the lateral length of the display area and the distance from the right side of the display area to the lateral position of the focused position plus half the distance between the lateral position of the focused position and the lateral position of the indicated position adjacent to the left side of the focused position.
For example, we assume a projected image shown in
For example, as shown in
Furthermore, the following is how to find cutout lines for cutting out a unit area including the indicated position corresponding to Person H that has the adjacent indicated position on the left thereof and has no adjacent indicated position on the right thereof. The lateral position of a right-hand cutout line for cutting out the unit area including the indicated position corresponding to Person H can be found as the position at a distance of (b1−a1)−(b1−h1)+((h1−g1)/2) from the left side of the display area in the x-direction. On the other hand, the lateral position of a left-hand cutout line for cutting out the unit area including the indicated position corresponding to Person H can be found as the position at a distance of (b1−a1)−(b1−h1)−((h1−g1)/2) from the left side of the display area in the x-direction.
Moreover, the following is how to find cutout lines for cutting out a unit area including the indicated position corresponding to Person D that has no adjacent indicated position on the left thereof and has the adjacent indicated position on the right thereof. The lateral position of a right-hand cutout line for cutting out the unit area including the indicated position corresponding to Person D can be found as the position at a distance of (b1−a1)−(b1−d1)−((e1−d1)/2) from the left side of the display area in the x-direction. On the other hand, the lateral position of a left-hand cutout line for cutting out the unit area including the indicated position corresponding to Person D can be found as the position at a distance of (b1−a1)−(b1−d1)−((e1−d1)/2) from the left side of the display area in the x-direction.
Programs executed by the conference terminal 2 or the projector 300 can be provided in such a manner that each program is recorded on a computer-readable recording medium, such as a CD-ROM, a flexible disk (FD), a CD-R, a digital versatile disk (DVD), or a universal serial bus (DSP) flash drive, in an installable or executable file format, or can be provided or distributed via a network such as the Internet. Furthermore, the programs can be provided in such a manner that each program is built into a ROM or the like in advance.
According to exemplary embodiments of the present invention, it is possible to provide a communication system, a communication apparatus, and a communication method that enable one's intended output.
The above-described embodiments are illustrative and do not limit the present invention. Thus, numerous additional modifications and variations are possible in light of the above teachings. For example, at least one element of different illustrative and exemplary embodiments herein may be combined with each other or substituted for each other within the scope of this disclosure and appended claims. Further, features of components of the embodiments, such as the number, the position, and the shape are not limited the embodiments and thus may be preferably set. It is therefore to be understood that within the scope of the appended claims, the disclosure of the present invention may be practiced otherwise than as specifically described herein.
The method steps, processes, or operations described herein are not to be construed as necessarily requiring their performance in the particular order discussed or illustrated, unless specifically identified as an order of performance or clearly identified through the context. It is also to be understood that additional or alternative steps may be employed.
Further, any of the above-described apparatus, devices or units can be implemented as a hardware apparatus, such as a special-purpose circuit or device, or as a hardware/software combination, such as a processor executing a software program.
Further, as described above, any one of the above-described and other methods of the present invention may be embodied in the form of a computer program stored in any kind of storage medium. Examples of storage mediums include, but are not limited to, flexible disk, hard disk, optical discs, magneto-optical discs, magnetic tapes, nonvolatile memory, semiconductor memory, read-only-memory (ROM), etc.
Alternatively, any one of the above-described and other methods of the present invention may be implemented by an application specific integrated circuit (ASIC), a digital signal processor (DSP) or a field programmable gate array (FPGA), prepared by interconnecting an appropriate network of conventional component circuits or by a combination thereof with one or more conventional general purpose microprocessors or signal processors programmed accordingly.
Each of the functions of the described embodiments may be implemented by one or more processing circuits or circuitry. Processing circuitry includes a programmed processor, as a processor includes circuitry. A processing circuit also includes devices such as an application specific integrated circuit (ASIC), digital signal processor (DSP), field programmable gate array (FPGA) and conventional circuit components arranged to perform the recited functions.
Claims
1. A communication system comprising a first communication apparatus and a second communication apparatus that transmits/receives data to/from the first communication apparatus, wherein
- the first communication apparatus includes: an area dividing unit that divides, of an output image which is an image that the first communication apparatus has output and includes at least a shot image obtained by shooting the surroundings of the second communication apparatus, a display area indicating an area where the shot image is displayed into as many unit areas as the number of people captured in the shot image; a first transmission control unit that performs control of transmitting coordinate information which indicates, of the output image, a unit area corresponding to a position pointed by a user to the second communication apparatus; and an output control unit that, when the first communication apparatus has received output information including a voice subjected to directivity control according to, of the shot image, an area corresponding to the coordinate information from the second communication apparatus, performs control of outputting the received output information, and
- the second communication apparatus includes: an acquiring unit that acquires the shot image; an identifying unit that, when the second communication apparatus has received the coordinate information from the first communication apparatus, identifies, of the shot image acquired by the acquiring unit, an area corresponding to the received coordinate information based on correspondence information indicating a correspondence relationship between the coordinates of the output image and the coordinates of the shot image; a directivity control unit that controls the directivity of a microphone array including a plurality of microphones to increase the sensitivity of, of the microphones installed in the second communication apparatus, a microphone corresponding to the area identified by the identifying unit; and a second transmission control unit that performs control of transmitting output information including at least a voice subjected to directivity control by the directivity control unit to the first communication apparatus.
2. The communication system according to claim 1, wherein
- the second communication apparatus further includes a cutting-out unit that cuts out, of the shot image acquired by the acquiring unit, an image of the area identified by the identifying unit as a cutout image, and
- the second transmission control unit performs control of transmitting output information including the cutout image cut out by the cutting-out unit and the voice subjected to directivity control by the directivity control unit to the first communication apparatus.
3. The communication system according to claim 1, wherein
- the area dividing unit equally divides the display area into as many unit areas as the number of user's operations performed to point at people captured in the display area, thereby obtaining multiple unit areas.
4. The communication system according to claim 1, wherein
- the area dividing unit divides the display area into multiple unit areas that correspond one-to-one to multiple positions in the display area indicated by user's operations on the basis of a relative positional relationship between four vertices of the output image and the multiple positions in the display area indicated by the user's operations.
5. A communication apparatus comprising:
- an area dividing unit that divides, of an output image which is an image that the communication apparatus has output and includes at least a shot image obtained by shooting the surroundings of another communication apparatus which communicates with the communication apparatus, a display area indicating an area where the shot image is displayed into as many unit areas as the number of people captured in the shot image;
- a transmission control unit that performs control of transmitting coordinate information which indicates, of the output image, a unit area corresponding to a position pointed by a user to the other communication apparatus; and
- an output control unit that, when the communication apparatus has received output information including a voice subjected to directivity control according to, of the shot image, an area corresponding to the coordinate information from the other communication apparatus, performs control of outputting the received output information.
6. A communication method for a communication system including a first communication apparatus and a second communication apparatus that transmits/receives data to/from the first communication apparatus, the communication method comprising:
- dividing, by the first communication apparatus, of an output image which is an image that the first communication apparatus has output and includes at least a shot image obtained by shooting the surroundings of the second communication apparatus, a display area indicating an area where the shot image is displayed into as many unit areas as the number of people captured in the shot image;
- performing, by the first communication apparatus, control of transmitting coordinate information which indicates, of the output image, a unit area corresponding to a position pointed by a user to the second communication apparatus;
- when having received output information including a voice subjected to directivity control according to, of the shot image, an area corresponding to the coordinate information from the second communication apparatus, performing, by the first communication apparatus, control of outputting the received output information;
- acquiring, by the second communication apparatus, the shot image;
- when having received the coordinate information from the first communication apparatus, identifying, by the second communication apparatus, of the shot image acquired at the acquiring, an area corresponding to the received coordinate information on the basis of correspondence information indicating a correspondence relationship between the coordinates of the output image and the coordinates of the shot image;
- controlling, by the second communication apparatus, the directivity of a microphone array including a plurality of microphones to increase the sensitivity of, of the microphones installed in the second communication apparatus, a microphone corresponding to the area identified at the identifying; and
- performing, by the second communication apparatus, control of transmitting output information including at least a voice subjected to directivity control at the controlling to the first communication apparatus.
7. A communication method for a communication apparatus, the communication method comprising:
- dividing, of an output image which is an image that the communication apparatus has output and includes at least a shot image obtained by shooting the surroundings of another communication apparatus which communicates with the communication apparatus, a display area indicating an area where the shot image is displayed into as many unit areas as the number of people captured in the shot image;
- performing control of transmitting coordinate information which indicates, of the output image, a unit area corresponding to a position pointed by a user to the other communication apparatus; and
- when the communication apparatus has received output information including a voice subjected to directivity control according to, of the shot image, an area corresponding to the coordinate information from the other communication apparatus, performing control of outputting the received output information.
Type: Application
Filed: Oct 19, 2016
Publication Date: May 4, 2017
Applicant: RICOH COMPANY, LTD. (Tokyo)
Inventors: Masato TAKAHASHI (Tokyo), Koji Kuwata (Kanagawa), Kiyoto Igarashi (Kanagawa), Tomoyuki Goto (Kanagawa), Kazuki Kitazawa (Kanagawa), Nobumasa Gingawa (Kanagawa), Miku Hakamatani (Kanagawa)
Application Number: 15/297,334