Image processing apparatus, image processing method, and recording medium
An image processing apparatus for coupling to an imaging apparatus that generates a captured image covering substantially a 360-degree field of view and for transmitting an output image to an image forming apparatus includes a setting unit configured to select a type of a polyhedron that is to be constructed by folding a development printed on a medium according to the output image, a converter configured to convert coordinates in the captured image into coordinates in the development that is to be printed on the medium and folded to construct the polyhedron, such that a zenith in the captured image is printed at a topmost point of the polyhedron constructed by folding the development printed on the medium, and an image generator configured to generate the output image based on the converted coordinates.
Latest Ricoh Company, Ltd. Patents:
- COMMUNICATION MANAGEMENT SYSTEM, COMMUNICATION SYSTEM, COMMUNICATION MANAGEMENT DEVICE, IMAGE PROCESSING METHOD, AND NON-TRANSITORY COMPUTER-READABLE MEDIUM
- IMAGE PROCESSING DEVICE, IMAGE FORMING APPARATUS, AND EDGE DETECTION METHOD
- IMAGE FORMING APPARATUS
- IMAGE READING DEVICE, IMAGE FORMING APPARATUS, AND IMAGE READING METHOD
- PRINT MANAGEMENT SYSTEM, PRINT MANAGEMENT METHOD, AND NON-TRANSITORY COMPUTER-EXECUTABLE MEDIUM
The present application claims priority under 35 U.S.C. § 119 to Japanese Patent Applications No. 2016-133774, filed on Jul. 5, 2016, No. 2017-018970, filed on Feb. 3, 2017, and No. 2017-043209, filed on Mar. 7, 2017, the contents of which are incorporated herein by reference in their entirety.
BACKGROUND OF THE INVENTION 1. Field of the InventionThe disclosures herein generally relate to an image processing apparatus, an image processing method, and a recording medium.
2. Description of the Related ArtMethods known in the art for creating a three-dimensional paper model or the like involve printing a development drawing on a two-dimensional paper sheet or the like based on the intended three-dimensional shape, cutting out the development, and then building the model.
For example, a computer may first identify the coordinates of vertexes constituting the three-dimensional shape. The computer next identifies each face constituted by three or more vertexes and associated with these vertexes. The computer then identifies each edge that constitutes one of the edges of two faces adjacent to each other and that is associated with these two adjacent faces. The computer further converts the coordinates of the vertexes constituting a selected one of the two identified adjacent faces into two-dimensional coordinates to develop the selected face, followed by selecting a face to be next developed based on the edges of the developed face. A method for producing a development drawing of a three-dimensional shape in this manner is known in the art (see Patent Document 1, for example).
SUMMARY OF THE INVENTIONIn one embodiment, an image processing apparatus for coupling to an imaging apparatus that generates a captured image covering substantially a 360-degree field of view and for transmitting an output image to an image forming apparatus includes a setting unit configured to select a type of a polyhedron that is to be constructed by folding a development printed on a medium according to the output image, a converter configured to convert coordinates in the captured image into coordinates in the development that is to be printed on the medium and folded to construct the polyhedron, such that a zenith in the captured image is printed at a topmost point of the polyhedron constructed by folding the development printed on the medium, and an image generator configured to generate the output image based on the converted coordinates.
Other objects and further features of embodiments will be apparent from the following detailed description when read in conjunction with the accompanying drawings, in which:
There may be a case in which an image (i.e., captured image) taken by an imaging device such as an omnidirectional camera is printed on a paper sheet or the like, followed by folding the sheet having the image printed thereon to a polyhedron shape. When a conventional method is used in such a case, the topmost point of the polyhedron may not coincide with the zenith of the captured image.
It may be an object of embodiments of the present disclosures to provide an image processing apparatus for which a polyhedron produced by folding a paper sheet or the like having a captured image printed thereon has the topmost point thereof corresponding to the zenith of the captured image.
In the following, a description will be given of embodiments of the present invention.
<Examples of Usage of Image Processing Apparatus>
The omnidirectional camera 1 may be a camera having a plurality of optical systems, for example. The omnidirectional camera 1 produce a captured image covering a large visual field such as the entire sphere around the omnidirectional camera 1 based on images captured by the plurality of optical systems. Such a captured image may cover a 360-degree field of view in the horizontal plane, for example.
The captured image is not limited to an image showing the full 360 degrees. The captured image may be an image covering a limited portion of the full 360 degrees, an image covering a hemisphere, or may be an image having a portion thereof being removed.
The omnidirectional camera 1 supplies a captured image to the PC 2. The PC 2 processes the captured image to produce an output image. As the PC supplies an output image to the printer 3, the printer 3 forms an image on a medium such as a paper or plastic sheet or the like based on the output image.
The printed image is cut out of the paper sheet or the like having the image formed thereon. The cut-out sheet is then folded to construct a polyhedron such as a tetrahedron, a hexahedron, an octahedron, a dodecahedron, an icosahedron, a cylinder, a cone or pyramid, or a prism. The polyhedron is not limited to these shapes, and may have any shape.
In the example, the omnidirectional camera 1, the PC 2, and the printer 3 are coupled through a wired connection or a wireless connection. The PC 2 may download from the omnidirectional camera 1 the data of a captured image such as a full-sphere image produced by the omnidirectional camera 1. The connection may alternatively be provided through a network or through an intervening server, for example.
The configuration that the image processing apparatus utilizes is not limited to the configuration illustrated in
The apparatuses used in this system may be as follows.
<Example of Imaging Apparatus>
The omnidirectional camera 1 includes a front-face imaging device 1H1, a rear-face imaging device 1H2, and a switch 1H3. In this example, the optical systems of the front-face imaging device 1H1 and the rear-face imaging device 1H2 are used to take a picture. Based on the images captured by use of these optical systems, the omnidirectional camera 1 produces a captured image covering the full sphere, for example.
The switch 1H3 is a shutter button, which is an example of an input device that is used by a user to instruct the omnidirectional camera 1 to take a picture.
As illustrated in
The image captured by the front-face imaging device 1H1 covers a wide field of view, for example, a view angle of 180 degrees, in front of the omnidirectional camera 1 as illustrated in
The view angle of each optical system may preferably be larger than or equal to 180 degrees and lower than or equal to 200 degrees. In the case of the view angle exceeding 180 degrees, the hemisphere image illustrated in
The image captured by the rear-face imaging device 1H2 covers a wide field of view, for example, a view angle of 180 degrees, in rear of the omnidirectional camera 1 as illustrated in
The omnidirectional camera 1 performs distortion correction and image stitching to produce the captured image illustrated in
The above-noted example is directed to a case in which the user presses the switch 1H3 to control a shutter-button operation. Alternatively, the shutter button may be remotely controlled by a display apparatus such as the PC 2.
<Example of Hardware Configuration of Imaging Apparatus>
The imaging unit 1H4 includes the front-face imaging device 1H1 and the rear-face imaging device 1H2. A lens 1H5 and a lens 1H6 are disposed in one-to-one correspondence with the front-face imaging device 1H1 and the rear-face imaging device 1H2, respectively.
The front-face imaging device 1H1 and the rear-face imaging device 1H2 are camera units. Specifically, each of the front-face imaging device 1H1 and the rear-face imaging device 1H2 includes an optical sensor such as a CMOS (complementary metal oxide semiconductor) or a CCD (charge coupled device). The front-face imaging device 1H1 converts incident light entering through the lens 1H5 into image data indicative of a hemisphere image or the like. Similarly, the rear-face imaging device 1H2 converts incident light entering through the lens 1H6 into image data indicative of a hemisphere image or the like.
The imaging unit 1H4 then supplies to the image processing unit 1H7 the image data generated by the front-face imaging device 1H1 and the image data generated by the rear-face imaging device 1H2. The supplied image data may represent the front-side hemisphere image illustrated in
The front-face imaging device 1H1 and the rear-face imaging device 1H2 may further include additional optical elements such as a diaphragm and/or a low-pass filter in addition to the lens for the purpose of capturing high-quality images. Moreover, the front-face imaging device 1H1 and the rear-face imaging device 1H2 may perform defective-pixel correction, image stabilization, and the like for the purpose of capturing high-quality images.
The image processing unit 1H7 produces a full-sphere image as illustrated in
The imaging control unit 1H8 is a control device for controlling the hardware of the omnidirectional camera 1.
The CPU 1H9 serves as an arithmetic device for performing arithmetic operations and data processing for required processes, and also serves as a control device for controlling hardware. For example, the CPU 1H9 performs processing based on the programs that are installed in advance.
The ROM 1H10, the SRAM 1H11, and the DRAM 1H12 are examples of memory devices. Specifically, the ROM 1H10 stores programs, data, parameters, and/or the like for causing the CPU 1H9 to perform processing, for example. The SRAM 1H11 and the DRAM 1H12 store programs for causing the CPU 1H9 to perform processing based on the programs, and also store data used by the programs, data generated by the programs, etc., for example. The omnidirectional camera 1 may further include an auxiliary memory device such as a hard-disk drive.
The operation I/F 1H13, which is coupled to an input device such as the switch 1H3, serves as an interface for receiving an instruction from the user operating on the omnidirectional camera 1. The operation I/F 1H13 may be an input device such as a switch, a connector and/or cable for coupling an input device, a circuit for processing signals supplied from the input device, a driver, a control device, etc. The operation I/F 1H13 may further include an output device such as a display. The operation I/F 1H13 may be a touchscreen panel serving as both an input device and an output device. The operation I/F 1H13 may include an interface such as a USB (universal serial bus) to couple a memory medium such as a flash memory to the omnidirectional camera 1. With such an arrangement, the operation I/F 1H13 may allow the omnidirectional camera 1 to exchange data with the memory medium.
The operation I/F 1H13 may include a power switch or a parameter inputting switch for operations other than the shutter operation.
The network I/F 1H14, the wireless I/F 1H15, and the antenna 1H16 allow the omnidirectional camera 1 to be coupled to an external apparatus through a wired or wireless connection. The omnidirectional camera 1 may be coupled to a network through the network I/F 1H14 to transmit data to the PC 2 (see
The bus 1H17 is used to exchange data between the hardware units of the omnidirectional camera 1. The bus 1H17 is an internal bus. Specifically, the bus 1H17 may be a PCI Express (peripheral component interconnect bus express) or the like.
The omnidirectional camera 1 is not limited to the configuration having two imaging devices. The omnidirectional camera 1 may have three or more imaging devices. The omnidirectional camera 1 may change the viewing direction of a single imaging device to capture a plurality of partial images. The optical systems of the omnidirectional camera 1 are not limited to the configuration using a fisheye lens. A wide-angle lens may alternatively be used.
Processing performed in the omnidirectional camera 1 may alternatively be performed by another device. For example, part or all of the processing may be performed by the PC 2 or another information processing apparatus coupled through the network upon transmitting data, parameters, and the like from the omnidirectional camera 1. In this manner, the image processing system may include a plurality of information processing apparatuses for performing distributed processing in a redundant or parallel manner.
<Example of Hardware Configuration of Image Processing Apparatus>
The auxiliary memory device 2H1 stores data, parameters, programs, and/or the like. Specifically, the auxiliary memory device 2H1 may be a hard-disc drive, a flash SSD (solid state drive), or the like. The data stored in the auxiliary memory device 2H1 may be stored partially or entirely in a file server or the like coupled through the network I/F 2H6 in a redundant manner, or alternatively in a proxy fashion.
The main memory device 2H2 serves as a work memory providing a memory area used by programs for performing processing. The main memory device 2H2 stores data, programs, parameters, and/or the like. Specifically, the main memory device 2H2 may be an SRAM or a DRAM, for example. The main memory device 2H2 may further include a control device for write and read operations.
The input device 2H3 serves to receive an instruction from a user operating the input device 2H3, and may also serve as an output device for displaying images, processed data, and/or the like. Specifically, the input device 2H3 may be an input device such as a keyboard and mouse. The input device 2H3 may receive an instruction from a user operating on a GUI (graphical user interface) or the like displayed thereon.
The output device 2H4 may be a display or the like for displaying images, processed data, and/or the like to a user. The input device 2H3 and the output device 2H4 may be implemented as a single consolidated device such as a touchscreen panel.
The CPU 2H5 serves as an arithmetic device for performing arithmetic operations and data processing for required processes, and also serves as a control device for controlling hardware. The CPU 2H5 may include a plurality of CPUs, devices, or cores for performing parallel, redundant, or distributed processing. The PC 2 may further include an auxiliary device for processing such as a GPU (graphical processing unit) in the inside or outside thereof.
The network I/F 2H6 is coupled to an external apparatus through a network in a wired or wireless manner. Specifically, the network I/F 2H6 may be an antenna, a peripheral circuit, a driver, and the like for transmitting and receiving data and/or the like. For example, the PC 2 uses the CPU 2H5 and the network I/F 2H6 to receive image data from the omnidirectional camera 1 (see
The image processing apparatus is not limited to a PC. Namely, the image processing apparatus may be a computer other than a PC. For example, the image processing apparatus may be a smartphone, a PDA (personal digital assistance), a tablet, a portable phone, or the like, or a combination thereof. The image processing apparatus may not be limited to a single apparatus, and may be implemented as a plurality of apparatuses.
<Example of Captured Image>
It may be noted that the process of generating a full-sphere image is not limited to the process involving equidistant cylindrical projection. For example, the arrangement of pixels of the hemisphere image illustrated in
The process of generating a full-sphere image may include distortion correction or the like for correcting distortion with respect to the hemisphere images illustrated in
Performing the processes as described above for generating a full-sphere image, the omnidirectional camera 1 produces a full-sphere image from a plurality of captured hemisphere images. Different processes may alternatively be employed to generate a full-sphere image.
<Example of Processing of Image Processing Apparatus>
<Example of Acquisition of Captured Image>
In step S101, the PC 2 acquires a captured image. For example, the processes performed by the omnidirectional camera as illustrated in
<Example of Setting Type of Polyhedron>
In step S102, the PC 2 sets a type of a polyhedron. For example, the PC 2 displays a setting screen to a user, and receives an instruction from the user to select a type of a polyhedron.
In the example illustrated in
By referring to
<Example of Setting Placement Position>
In step S104, the PC 2 sets the placement position of the polyhedron. For example, the PC 2 displays a setting screen to a user, and receives an instruction from the user to select the placement position of the polyhedron.
Use of different placement positions causes the positions of the topmost point to differ even when the same polyhedron is used. The topmost point may change as follows depending on the placement position.
By referring to
By referring to
<Example of Coordinate Conversion>
In step S106, the PC 2 converts coordinates. For example, the PC 2 calculates two-dimensional coordinates in a captured image corresponding to three-dimensional coordinates as described in the following. In the following, three-dimensional coordinates are expressed as (X, Y, Z), and two-dimensional coordinates in a captured image is expressed as (x, y. The Z axis coincides with the vertical direction, i.e., the direction of gravity. The Y axis represents the depth direction. The X axis represents a direction perpendicular to the depth direction and to the vertical direction.
In step S201 in
In this formula (1), normalized X, Y, and Z are denoted as tmpx, tmpy, and tmpz, respectively.
The PC 2 then performs step S202 based on the X coordinate and the Y coordinate, i.e., (tmpx, tmpy), among the normalized three-dimensional coordinates.
In step S202, the PC 2 checks whether the conditions of tmpx=0 and tmpy=0 are satisfied. Upon determining that both tmpx=0 and tmpy=0 are satisfied (YES in step S202), the PC 2 proceeds to step S206. Upon determining that at least one of tmpx=0 and tmpy=0 is not satisfied (NO in step S202), the PC 2 proceeds to step S203.
In step S203, the PC 2 performs calculation by use of the following formula (2).
In step S204, the PC 2 checks whether tmpy2 is greater than 0. Upon determining that tmpy2 is greater than 0 (YES in step S204), the PC 2 proceeds to step S205. Upon determining that tmpy2 is not greater than 0 (NO in step S204), the PC 2 proceeds to step S206.
In step S205, the PC 2 performs calculation by use of the following formula (3).
In step S206, the PC 2 performs calculation by use of the following formula (4). When step S206 is to be performed upon selecting “YES” in step S202, tempx2 has not been calculated by use of the formula (2). In such a case, the PC 2 performs calculation by use of the following formula (4) by treating tmpx as tmpx2 (i.e., tmpx2=tmpx).
In this manner, the PC 2 calculates “x” of the two-dimensional coordinates by use of the above-noted formula (3) or (4).
In step S207, the PC 2 performs calculation by use of the following formula (5).
In the above-noted formulas (3), (4), and (5), the “width” represents the size of a captured image in the width direction thereof, i.e., in the x-axis direction. Further, the “height” represents the size of the captured image in the height direction thereof, i.e., in the y-axis direction.
Through the calculations described above, the PC 2 converts three-dimensional coordinates (X, Y, Z) into two-dimensional coordinates (x, y). In this manner, the PC 2 identifies two-dimensional coordinates (x, Y) in the captured image with respect to an image point at three-dimensional coordinates (X, Y, Z) on a polyhedron.
By referring to
As illustrated in
The PC 2 generates an output image representing a development by use of data as follows.
<Example of Data Structure>
The PC 2 has a data table prepared in advance as in TABLE 1 shown below.
The table data of TABLE 1 specifies types of polyhedrons registered in the PC 2. Among these shapes registered in TABLE 1, the user selects a polyhedron type on the first setting screen PN1 illustrated in
Each shape specified in TABLE 1 is associated with vertexes specified in TABLE 2 shown below, for example.
TABLE 2 provides a correspondence between the vertexes of each polyhedron specified in TABLE 1 and three-dimensional coordinates of these vertexes. The table data of TABLE 2 are stored in the PC 2 in advance. In TABLE 2, the “vertex ID” is a number or the like that identifies each vertex. The “vertex No.” is a number that identifies each one of the vertexes of each shape. The “shape ID” is a number or the like that identifies each shape specified in TABLE 1. The “coordinates” show the three-dimensional coordinates of each vertex. In the example shown in TABLE 2, the rows for which the “shape ID” is “001” correspond to the entry in TABLE 1 for which the “shape ID” is “001”. Namely, the rows for which the “vertex ID” is “001” through “004” in TABLE 2 indicate the vertexes of a tetrahedron In this case, thus, the number of data items representing vertexes is equal to four.
Placement positions are determined based on the table data of TABLE 3 shown below, for example.
TABLE 3 specifies placement positions with respect to each shape specified in TABLE 1. The table data of TABLE 3 are stored in the PC 2 in advance. The “shape ID” in TABLE 3 is a number or the like that identifies each shape specified in TABLE 1. The “placement ID” is a number specifying the type of a placement position. The “placement name” indicates the name of each placement position. The “coordinates of topmost point” represent the coordinates of a topmost point defined for each placement position in which the polyhedron is placed according to the “placement ID”. The “placement image” indicates the name of a file containing an image showing a placement position. Based on TABLE 3, the second setting display PN2 is displayed as illustrated in
The faces of each development and the vertexes of each face are determined for each shape based on the table data of TABLE 4 as shown below, for example. The table data of TABLE 4 are stored in the PC 2 in advance.
The table data of TABLE 4 provides a correspondence between a shape as defined in TABLE 1 and the faces of each polyhedron of such a shape. The case of the “shape ID” being “001”, for example, corresponds to the polyhedron type that is a tetrahedron as shown in TABLE 1. In the case of a tetrahedron, a polyhedron has four faces. TABLE 4 thus uses the “face No.” that is “1” through “4” to specify the faces of the tetrahedron. Namely, the four faces of a tetrahedron are each given a different “face No.” in TABLE 4.
In this manner, the “shape ID” in TABLE 4 corresponds to the “shape ID” specified in TABLE 1. The “face No.” is a number or the like that identifies each face of a polyhedron. The “face vertex No.” is a number or the like that identifies the vertexes of each face. In the case of a tetrahedron, each face is a triangle, and has three vertexes. In TABLE 4, thus, each face has three “face vertex No.” entries assigned thereto. The “coordinates” show the two-dimensional coordinates of each vertex in a development.
Each vertex of each polyhedron has the two-dimensional coordinates associated with the three-dimensional coordinates according to table data as in TABLE 5 shown below.
The table data of TABLE 5 provides a correspondence between vertexes as defined in TABLE 4 and vertexes as defined in TABLE 2. To be more specific, TABLE 5 links a vertex identified by “face vertex No.” with a vertex identified by a “vertex No.” The “face vertex No.” is a number or the like in TABLE 4 that identifies two-dimensional coordinates, i.e., coordinates in a development drawing. On the other hand, the “vertex No.” is a number or the like in TABLE 2 that identifies three-dimensional coordinates on a constructed polyhedron.
By use of these data, the image processing apparatus finds a match between three-dimensional coordinates on a constructed polyhedron and two-dimensional coordinates on a development. The conversion process illustrated in
More specifically, the image processing apparatus first determines whether a point of interest on a two-dimensional plane on which a development is to be formed falls within the boundaries of a development. TABLE 4 provides two-dimensional coordinates of the vertexes of each face that constitutes the development. The image processing apparatus can thus determine whether the point of interest on the plane is situated within the boundaries of any given face of the development. For any point inside the boundaries of the development, the image processing apparatus needs to acquire the three-dimensional coordinates of this point on the polyhedron constructed by folding the development.
In order to do so, the image processing apparatus may identify a face in which this point is situated on the two-dimensional plane, followed by interpolating this point within the identified face. For example, the position of this point may be expressed on the plane as a weighted sum of two vectors that extend from a given start-point vertex of the identified face toward the two adjacent vertexes of this face. With respect to these three vertexes (i.e., one start-point vertex and two endpoint vertexes) on the two-dimensional plane, the image processing apparatus may identify three corresponding vertexes in three-dimensional space and the three-dimensional coordinates thereof by use of TABLE 4, TABLE 5, and TABLE 2. The image processing apparatus can thus find two vectors in the three-dimensional space that correspond to the two noted vectors on the two-dimensional plane. In the three-dimensional space, the image processing apparatus may then identify the three-dimensional coordinates of the point of interest as the end point of a weighted sum of the two corresponding vectors extending from the start-point vertex.
When the three-dimensional coordinates of the topmost point for the selected placement position are different from (0, 0, Z1) (Z1: any proper value), the image processing apparatus may need to correct the three-dimensional coordinates of the point of interest through three-dimensional rotational transformation. This rotational transformation is defined as rotation around an axis perpendicular to both the vector from the origin to the topmost point and the vector from the origin to (0, 0, Z1) such as to bring the topmost point to (0, 0, Z1). Here, the origin is positioned at the center of the polyhedron.
The image processing apparatus may obtain such a rotation axis by calculating the outer product between the vector from the origin to the topmost point and the vector from the origin to (0, 0, Z1). The angle of rotation to bring the topmost point to (0, 0, Z1) may be obtained by calculating the inner product between these two vectors. In this manner, the image processing apparatus is able to identify rotational transformation that brings the topmost point to (0, 0, Z1). The image processing apparatus applies the same rotational transformation to the three-dimensional coordinates of the point of interest in order to perform zenith correction.
Finally, the image processing apparatus performs the calculations illustrated in
It should be noted that the three-dimensional coordinates as provided in TABLE 2 may be provided with respect to each different placement position of each polyhedron. For example, TABLE 2 may list the three-dimensional coordinates of vertexes of a tetrahedron placed in the face placement, the three-dimensional coordinates of vertexes of a tetrahedron placed in the point placement, and three-dimensional coordinates of vertexes of a tetrahedron placed in the edge placement. In such a case, there is no need to perform rotational transformation as described above to make a zenith correction. It suffices for the image processing apparatus to use the three-dimensional coordinates corresponding to the selected placement position in order to find proper three-dimensional coordinates of the previously-noted point of interest.
<Variation 1>
The shape of a development may be selected by a user as follows.
As illustrated in
<Variation 2>
The PC 2 may be configured such that a predetermined image different from a captured image is printed on a given face according to a user setting. The following example will be directed to a case in which a QR code (registered trademark) serving as an example of the predetermined image is printed.
The given face may preferably be a bottom face BF. Namely, an image such as a QR code (registered trademark) is preferably printed on the bottom face BF as illustrated in
Selecting “no” on the image setting screen PN5 illustrated in
<Variation 3>
The development drawing may be provided with tabs for use in gluing at the time of folding and constructing, i.e., provided with glue tabs.
The output image illustrated in
The output image illustrated in
The provision of tabs or faces for bonding purposes as described above allows the image processing apparatus to produce an output image representing a development that is easy to construct. Further, tabs or faces may be formed on a development so as to serve as a handling portion of a polyhedron.
<Variation 4>
The image forming apparatus may reduce the size of, or make a parallel shift of, a captured image as described in the following, for example.
As illustrated in
The image processing apparatus may make a parallel shift of a captured image as described in the following.
In consideration of this, the image processing apparatus makes a parallel shift of the input image IMG1 toward the zenith to produce the parallel-shift image IMG3 as illustrated in
In this manner, the image processing apparatus is able to produce an output image such that the objects of interest or the like appearing in the captured image are printed in the viewable areas of a constructed polyhedron.
In other cases, a predetermined object may appear in a lower portion of an input image. In the case of an omnidirectional camera being used for an image capture, a finger or the like of a user may be captured as a predetermined object. It may be noted that the image processing apparatus is able to detect a predetermined object such as a finger by checking whether a flesh color occupies more than a predetermined area size, for example. In the case of a predetermined object appearing in the lower portion of an input image as described above, the image processing apparatus may reduce the size of, or make a parallel shift of, the image such that the predetermined object is printed on the bottom face.
<Variation 5>
The coordinates of a zenith may be selected in a captured image. In the following, a description will be given of an example in which the image illustrated in
In step S301, the PC 2 selects the coordinates of a new zenith. In the following, a description will be given of the method of selecting coordinates with reference to a drawing.
By referring to
In the above-noted formula (6), the “width” represents the size of a captured image in the width direction thereof, i.e., in the x-axis direction, similarly to the case in the formulas (3), (4), and (5). Further, the “height” represents the size of the captured image in the height direction thereof, i.e., in the y-axis direction. Further, (x, y) represents the two-dimensional coordinates of the selected point in the captured image. Three-dimensional coordinates (X, Y, Z) calculated by use of the formula (6) then serve as the coordinates of a new zenith specified by (x, y).
After the processes described above, the PC is able to treat the coordinates of a point selected by the user in the captured image as the zenith of the captured image.
<Example of Functional Configuration>
The setting unit 2F1 serves to set a type of a polyhedron. The setting unit 2F1 may be implemented by use of the input device 2H3 (see
The conversion unit 2F2 converts coordinates in a captured image into coordinates in a development that is printed on a medium and folded to construct a polyhedron such that a zenith in the captured image is printed at the position of a topmost point of the polyhedron constructed by folding the development printed on the medium. The conversion unit 2F2 may be implemented by use of the CPU 2H5 (see
The output image generating unit 2F3 generates an output image IMGOUT based on the coordinates converted by the conversion unit 2F2. The output image generating unit 2F3 may be implemented by use of the CPU 2H5 (see
The selecting unit 2F4 serves to select (specify) the coordinates of a new zenith. The selecting unit 2F4 may be implemented by use of the input device 2H3 (see
The development is a drawing that is folded into a polyhedron. There are many types of polyhedrons, and polyhedrons are placed in various placement positions. The PC 2 uses the setting unit 2F1 to display a setting screen that allows a user to select a type of a polyhedron and a placement position of a polyhedron as illustrated in
As illustrated in
Moreover, the PC 2 is able to use the process by the conversion unit 2F2, e.g., the process illustrated in
The output image IMGOUT is used to produce an image of a development on a medium. The PC 2 is able to produce the output image IMGOUT by finding a match between two-dimensional coordinates in the development and three-dimensional coordinates in accordance with TABLE 5, for example. In this manner, the PC 2 produces an output image such that a polyhedron constructed by folding a paper sheet or the like having a captured image printed thereon has the topmost point thereof corresponding to the zenith of the captured image.
The functional configuration that includes the selecting unit 2F4 allows the coordinates of a new zenith to be selected as illustrated in
The embodiment of the present invention may be implemented by use of programs written in programming languages or the like. Namely, the embodiment of the present embodiment may be implemented as a program for causing a computer such as an image processing apparatus to perform an image processing method. Such a program may be stored and distributed in a computer-readable recording medium such as a flash memory, an SD (registered trademark) card, an optical disk, or the like. The program may also be distributed through electrical communication lines such as the Internet.
In the embodiment of the present invention, part or all of the processing may alternatively be implemented by use of a programmable device (PD) such as a field-programmable gate array (FPGA), for example. In the embodiment of the present invention, further, part or all of the processing may alternatively be implemented by use of an ASIC (application specific integrated circuit).
According to at least one embodiment, an image processing apparatus is provided for which a polyhedron produced by folding a paper sheet or the like having a captured image printed thereon has the topmost point thereof corresponding to the zenith of the captured image.
Although the preferred embodiments of the present invention have heretofore been described, the present invention is not limited to such particular embodiments. Various modifications and changes may be made without departing from the scope of the invention.
RELATED-ART DOCUMENTS Patent Document
- [Patent Document 1] Japanese Unexamined Patent Application Publication No. 2002-328959
Claims
1. An image processing apparatus for coupling to an imaging apparatus that generates a captured image covering substantially a 360-degree field of view and for transmitting an output image to an image forming apparatus, the image processing apparatus comprising:
- a memory; and
- a processor coupled to the memory and configured to perform functions of:
- a setting unit configured to select a type of a polyhedron that is to be constructed by folding a development printed on a medium according to the output image;
- a converter configured to perform a conversion process to convert two-dimensional coordinates in the captured image into two-dimensional coordinates in the development that is to be printed on the medium and folded to construct the polyhedron, such that a zenith in the captured image is printed at a topmost point of the polyhedron constructed by folding the development printed on the medium; and
- an image generator configured to generate the output image based on the converted coordinates,
- wherein the conversion process includes:
- calculating conversion between the two-dimensional coordinates in the captured image and three-dimensional coordinates on the polyhedron to be constructed; and
- calculating conversion between the three-dimensional coordinates on the polyhedron and the two-dimensional coordinates in the development, and
- wherein the three-dimensional coordinates on the polyhedron vary in response to a placement position of the polyhedron.
2. The image processing apparatus according to claim 1, wherein the placement position of the polyhedron is specified in response to a user input, and the converter is configured to convert the two-dimensional coordinates in the captured image into the two-dimensional coordinates in the development in accordance with the placement position.
3. The image processing apparatus according to claim 2, wherein the placement position is selectable from placement positions, the placement positions including a placement position in which one of a plurality of faces of the polyhedron is in contact with a ground, a placement position in which one of a plurality of vertexes of the polyhedron is in contact with the ground, and a placement position in which one of a plurality of edges of the polyhedron is in contact with the ground.
4. The image processing apparatus according to claim 1, wherein the converter uses data that provides a correspondence between three-dimensional coordinates of vertexes of the polyhedron and two-dimensional coordinates of the vertexes in the development.
5. The image processing apparatus according to claim 1, wherein a predetermined image different from the captured image is printed on one of a plurality of faces of the polyhedron.
6. The image processing apparatus according to claim 1, wherein the development includes tabs or faces for use in bonding faces of the polyhedron.
7. The image processing apparatus according to claim 1, wherein the captured image is reduced in size to generate the output image such that an entirety of the captured image is printed on faces of the polyhedron other than a bottom face.
8. The image processing apparatus according to claim 1, wherein the captured image is reduced in size to generate the output image such that a predetermined object appearing in the captured image is printed on a bottom face of the polyhedron.
9. The image processing apparatus according to claim 1, wherein the captured image is subjected to a parallel shift to generate the output image such that an entirety of the captured image is printed on faces of the polyhedron other than a bottom face.
10. The image processing apparatus according to claim 1, further comprising a selecting unit configured to select coordinates of a point to become the zenith.
11. A method for processing an image in an image processing apparatus for coupling to an imaging apparatus that generates a captured image covering substantially a 360-degree field of view and for transmitting an output image to an image forming apparatus, the method comprising:
- selecting a type of a polyhedron that is to be constructed by folding a development printed on a medium according to the output image;
- performing a conversion process of converting two-dimensional coordinates in the captured image into two-dimensional coordinates in the development that is to be printed on the medium and folded to construct the polyhedron, such that a zenith in the captured image is printed at a topmost point of the polyhedron constructed by folding the development printed on the medium; and
- generating the output image based on the converted coordinates,
- wherein the conversion process includes:
- calculating conversion between the two-dimensional coordinates in the captured image and three-dimensional coordinates on the polyhedron to be constructed; and
- calculating conversion between the three-dimensional coordinates on the polyhedron and the two-dimensional coordinates in the development, and
- wherein the three-dimensional coordinates on the polyhedron vary in response to a placement position of the polyhedron.
12. A non-transitory recording medium having a program recorded thereon for causing a computer to perform an image processing method, the computer being for coupling to an imaging apparatus that generates a captured image covering substantially a 360-degree field of view and for transmitting an output image to an image forming apparatus, the program comprising:
- selecting a type of a polyhedron that is to be constructed by folding a development printed on a medium according to the output image;
- performing a conversion process of converting two-dimensional coordinates in the captured image into two-dimensional coordinates in the development that is to be printed on the medium and folded to construct the polyhedron, such that a zenith in the captured image is printed at a topmost point of the polyhedron constructed by folding the development printed on the medium; and
- generating the output image based on the converted coordinates,
- wherein the conversion process includes:
- calculating conversion between the two-dimensional coordinates in the captured image and three-dimensional coordinates on the polyhedron to be constructed; and
- calculating conversion between the three-dimensional coordinates on the polyhedron and the two-dimensional coordinates in the development, and
- wherein the three-dimensional coordinates on the polyhedron vary in response to a placement position of the polyhedron.
6201882 | March 13, 2001 | Tanaka |
9501857 | November 22, 2016 | Itoh et al. |
20050089770 | April 28, 2005 | Liu |
20140085412 | March 27, 2014 | Hayashi |
20150170335 | June 18, 2015 | Otsuka et al. |
20170039909 | February 9, 2017 | Itoh et al. |
20170150007 | May 25, 2017 | Snowball |
2002-328959 | November 2002 | JP |
2017-041089 | February 2017 | JP |
2015/093553 | June 2015 | WO |
- Anonymous: “Flexify 2 plug-in (version 2.84) used in Adobe Photoshop CS6”, Jun. 16, 2016 (Jun. 16, 2016), pp. 1-23, XP055383376, Retrieved from the Internet: URL:http://web.archive.org/web/20160616162642/http://www.flamingpear.com/free-trials.html [retrieved on Jun. 20, 2017] the whole document.
- Anonymous: “Free trials—Flaming Pear Software”, Jun. 16, 2016 (Jun. 16, 2016), XP055382840, Retrieved from the Internet: URL:http://web.archive.org/web/20160616162642/http://www.flamingpear.com/free-trials.html [retrieved on Jun. 19, 2017] the whole document.
- C. A. Furuti: “Map Projections—Polyhedral Maps. Introduction, common projections, tetrahedral globes”, Mar. 5, 2016 (Mar. 5, 2016), XP055381824, Retrieved from the Internet: URL:https://web.archive.org/web/20160305093629/http://www.progonos.com/furuti/MapProj/Normal/ProjPoly/projPoly.html [retrieved on Jun. 14, 2017] Sec. “Introduction”.
- C. A. Furuti: “Map Projections—Polyhedral Maps. Cubic pseudoglobes, gnomonic projection”, Mar. 8, 2016 (Mar. 8, 2016), XP055381821, Retrieved from the Internet: URL:https://web.archive.org/web/20160308173631/http://www.progonos.com/furuti/MapProj/Normal/ProjPoly/Foldout/Cube/cube.html [retrieved on Jun. 14, 2017] Sec. “High-Resolution Maps”, maps 5, 7, and 8.
- C. A. Furuti: “Map Projections—Polyhedral Maps. Icosahedral pseudoglobes, gnomonic projection. Map Projections Resolution—3 Maps”, May 21, 2016 (May 21, 2016), XP055381819, Retrieved from the Internet: URL:https://web.archive.org/web/20160521175153/http://www.progonos.com:80/furuti/MapProj/Normal/ProjPoly/projPoly.html [retrieved on Jun. 14, 2017] Sec. “High-Resolution Maps”, maps 4, 6, and 7.
- Extended European Search Report for 17160014.1 dated Jul. 4, 2017.
Type: Grant
Filed: Mar 8, 2017
Date of Patent: May 15, 2018
Patent Publication Number: 20180011669
Assignee: Ricoh Company, Ltd. (Tokyo)
Inventors: Aiko Ohtsuka (Tokyo), Atsushi Funami (Tokyo)
Primary Examiner: Moustapha Diaby
Application Number: 15/452,887
International Classification: G06F 3/12 (20060101); H04N 1/00 (20060101);