Method, medium, and apparatus encoding and/or decoding multichannel audio signals

- Samsung Electronics

A method, medium, and apparatus encoding and/or decoding a multichannel audio signal. The method includes detecting the type of spatial extension data included in an encoding result of an audio signal, if the spatial extension data is data indicating a core audio object type related to a technique of encoding core audio data, detecting the core audio object type; decoding core audio data by using a decoding technique according to the detected core audio object type, if the spatial extension data is residual coding data, decoding the residual coding data by using the decoding technique according to the core audio object type, and up-mixing the decoded core audio data by using the decoded residual coding data. According to the method, the core audio data and residual coding data may be decoded by using an identical decoding technique, thereby reducing complexity at the decoding end.

Skip to: Description  ·  Claims  ·  References Cited  · Patent History  ·  Patent History
Description
CROSS-REFERENCE TO RELATED APPLICATIONS

This application is a continuation application of U.S. patent application Ser. No. 11/907,398 filed on Oct. 11, 2007, which claims the priority benefits of Korean Patent Application No. 10-2006-0101580, filed on Oct. 18, 2006, and Korean Patent Application No. 10-2007-0088315, filed on Aug. 31, 2007, in the Korean Intellectual Property Office, the disclosures of each of which are incorporated herein in their entirety by reference.

BACKGROUND

1. Field

One or more embodiments of the present invention relate to a method, medium, and apparatus encoding and/or decoding multichannel audio signals, and more particularly, to a method, medium, and apparatus encoding and/or decoding a residual signal used to up-mix an audio signal.

2. Description of the Related Art

A moving picture experts group (MPEG) surround encoding technique is used to compress audio data in relation to spatial sources. The MPEG surround encoding technique allows an audio signal, compressed according to MPEG audio layer-3 (MP3), MPEG-4 advanced audio coding (AAC), or MPEG-4 high efficiency (HE)-AAC, to be converted into an encoded multichannel surround audio signal. The MPEG surround encoding technique has advantages over other encoding techniques in that this technique maintains backward compatibility to existing stereo equipment, and can be used to reduce bitrates, i.e., a transmission speed, desired for high quality multichannel audio compression while using existing equipment.

According to MPEG surround encoding standards, a core audio signal is conventionally encoded by using any one encoding technique from among bit sliced arithmetic coding (BSAC), AAC, and MP3, while corresponding residual signals are encoded only according to AAC.

Accordingly, when such a core audio signal is encoded with an encoding technique other than AAC, according to the MPEG surround standards, the core audio signal and a residual signal would be encoded by using different encoding techniques. Accordingly, at the decoding end, the core audio signal and the residual signal should be decoded through different decoding techniques. Briefly, herein, the use of the terms encoding technique and encoding method are used interchangeably, with the particular discussion below using the term ‘technique’ for simplicity of discussion to distinguish a method of the present invention from such encoding methods or techniques.

Thus, the inventors of the present invention have discovered that that there is a desire for a method, medium, and apparatus to attempt to overcome such drawbacks and/or problems potentially resulting from such conventionally required different encoding techniques.

SUMMARY

One or more embodiments of the present invention provide a method, medium, and apparatus decoding a multichannel audio signal, capable of reducing complexity at the decoding end when a residual signal is decoded.

One or more embodiments of the present invention further provide a method, medium, and apparatus encoding a multichannel audio signal, capable of reducing complexity at the encoding end when a residual signal is encoded.

Additional aspects and/or advantages will be set forth in part in the description which follows and, in part, will be apparent from the description, or may be learned by practice of the invention.

According to an aspect of the present invention, there is provided a method of decoding a multichannel audio signal, the method including: detecting a type of spatial extension data included in an encoding result of an audio signal; if the spatial extension data includes data indicating a core audio object type related to a method of encoding core audio data, detecting the core audio object type; decoding the core audio data by using a decoding method according to the detected core audio object type; if the spatial extension data includes residual coding data, decoding the residual coding data by using the decoding method according to the core audio object type; and up-mixing the decoded core audio data by using the decoded residual coding data.

According to another aspect of the present invention, there is provided a computer readable recording medium having embodied thereon a computer program for executing a method of decoding a multichannel audio signal, wherein the method includes: detecting a type of spatial extension data included in an encoding result of an audio signal; if the spatial extension data includes data indicating a core audio object type related to a method of encoding core audio data, detecting the core audio object type; decoding the core audio data by using a decoding method according to the detected core audio object type; if the spatial extension data includes residual coding data, decoding the residual coding data by using the decoding method according to the core audio object type; and up-mixing the decoded core audio data by using the decoded residual coding data.

According to another aspect of the present invention, there is provided an apparatus for decoding a multichannel audio signal, the apparatus including: a spatial extension data type detecting unit detecting a type of spatial extension data included in an encoding result of an audio signal; a core audio object type detecting unit, if the spatial extension data includes data indicating a core audio object type related to a method of encoding core audio data, detecting the core audio object type; a core audio data decoding unit decoding the core audio data by using a decoding method according to the detected core audio object type; a residual coding data decoding unit, if the spatial extension data includes residual coding data, decoding the residual coding data by using the decoding method according to the core audio object type; and an up-mixing unit up-mixing the decoded core audio data by using the decoded residual coding data.

According to another aspect of the present invention, there is provided a method of encoding a multichannel audio signal, the method including: generating core audio data and residual data by down-mixing an input audio signal; encoding the core audio data by using a predetermined encoding method; encoding the residual data by using the predetermined encoding method according to a core audio object type related to the method by which the core audio data is encoded; and outputting the encoded core audio data and the encoded residual data as an encoding result of the audio signal.

According to another aspect of the present invention, there is provided an apparatus encoding a multichannel audio signal, the apparatus including: a down-mixing unit generating core audio data and residual data by down-mixing an input audio signal; a core audio data encoding unit encoding the core audio data by using a predetermined encoding method; a residual data encoding unit encoding the residual data by using the predetermined encoding method according to a core audio object type related to the method by which the core audio data is encoded; and a multiplexing unit outputting the encoded core audio data and the encoded residual data as an encoding result of the audio signal.

BRIEF DESCRIPTION OF THE DRAWINGS

These and/or other aspects and advantages will become apparent and more readily appreciated from the following description of the embodiments, taken in conjunction with the accompanying drawings of which:

FIG. 1 illustrates an apparatus decoding a multichannel audio signal, according to an embodiment of the present invention;

FIG. 2 illustrates a syntax file for detecting a spatial extension data type, according to an embodiment of the present invention;

FIG. 3 illustrates a table including assigned values corresponding to “bsSacExtType” illustrated in FIG. 2, according to an embodiment of the present invention;

FIG. 4 illustrates a syntax file for reading a core audio object type, according to an embodiment of the present invention;

FIG. 5 illustrating a syntax file for decoding residual coding data, according to an embodiment of the present invention;

FIG. 6 illustrates a syntax file for decoding arbitrary down-mix residual data, according to an embodiment of the present invention;

FIG. 7 illustrates a method of decoding a multichannel audio signal, according to an embodiment of the present invention;

FIG. 8 illustrates an apparatus encoding a multichannel audio signal, according to an embodiment of the present invention; and

FIG. 9 illustrates a method of encoding a multichannel audio signal, according to an embodiment of the present invention.

DETAILED DESCRIPTION OF EMBODIMENTS

Reference will now be made in detail to embodiments, examples of which are illustrated in the accompanying drawings, wherein like reference numerals refer to like elements throughout. In this regard, embodiments of the present invention may be embodied in many different forms and should not be construed as being limited to embodiments set forth herein. Accordingly, embodiments are merely described below, by referring to the figures, to explain aspects of the present invention.

FIG. 1 illustrates an apparatus decoding a multichannel audio signal, according to an embodiment of the present invention. Herein, the term apparatus should be considered synonymous with the term system, and not limited to a single enclosure or all described elements embodied in single respective enclosures in all embodiments, but rather, depending on embodiment, is open to being embodied together or separately in differing enclosures and/or locations through differing elements, e.g., a respective apparatus/system could be a single processing element or implemented through a distributed network, noting that additional and alternative embodiments are equally available.

Referring to FIG. 1, the apparatus decoding a multichannel audio signal, according to an embodiment, may include a demultiplexing unit 100, a spatial extension data type detecting unit 110, a core audio object type detecting unit 120, a core audio data decoding unit 130, a residual coding decoding unit 140, an arbitrary down-mix residual coding data decoding unit 150, a spatial extension data decoding unit 160, and an up-mixing unit 170, for example. Here, up-mixing is a concept that includes generating plural signals, e.g., stereo signals, of two or more channels from a single signal, e.g., a mono signal. Similarly, down-mixing is a corresponding concept that includes encoding plural signals, e.g., stereo signals, of two or more channels into a single channel, e.g., a mono channel.

Thus, here, the demultiplexing unit 100 may receive a bitstream, e.g., from an encoding end through an input terminal IN, and demultiplex the bitstream.

FIG. 2 illustrates an example syntax file for detecting a spatial extension data type, according to an embodiment of the present invention. Further, for example, FIG. 3 illustrates a table showing assignment of values corresponding to “bsSacExtType” illustrated in FIG. 2, according to an embodiment of the present invention. Thus, according to one embodiment, an operation of the spatial extension data type detecting unit 110 will now be further explained in greater detail with reference to FIGS. 1 through 3.

The spatial extension data type detecting unit 110 may detect the type of spatial extension data, e.g., in a header, of data which is demultiplexed by the demultiplexing unit 100. More specifically, the spatial extension data type detecting unit 110 may detect the type of the spatial extensional data in the header of the demultiplexed data according to a function SpatialExtensionConfig(), illustrated in FIG. 2, for example. Here, in the illustrated function SpatialExtensionConfig(), “bsSacExtType” indicates the type of spatial extension data.

Referring to FIG. 3, in this embodiment, if “bsSacExtType” is a “0”, spatial extension data may be indicated as being residual coding data; if “bsSacExtType” is “1”, spatial extension data may be indicated as being arbitrary down-mix residual coding data; and if “bsSacExtType” is “12”, spatial extension data may be indicated as being a core audio object type of moving picture experts group (MPEG)-4 audio, for example. Here, the core audio object type is defined as an audio object type for correspondingly encoding a signal which is down-mixed at an encoding end. However, these particular indications and audio object types are just for one or more embodiments of the present invention, noting that a person of ordinary skill in the art of the present invention should understand that alternate embodiments are equally available.

In other words, if 0 is assigned to “bsSacExtType”, the spatial extension data type detecting unit 110 may determine that the type of spatial extension data is residual coding data. If 1 is assigned to “bsSacExtType”, the spatial extension data type detecting unit 110 may determine that the type of spatial extension data is arbitrary down-mix residual coding data, and if 12 is assigned to “bsSacExtType”, the spatial extension data type detecting unit 110 may determine that the type of spatial extension data is data indicating the core audio object type of MPEG-4 audio.

An operation of an apparatus for decoding an audio signal according to a spatial extension data type detected by the spatial extension data type detecting unit 110 will now be explained in greater detail with further reference to FIG. 4.

First, the case where the spatial extension data type detected by the spatial extension data type detecting unit 110 is data indicating the core audio object type of MPEG-4 audio will be explained, i.e., “bsSacExtType” is 12, according to the above indication examples.

FIG. 4 illustrates a syntax file, for example, for reading a core audio object type, according to an embodiment of the present invention. Accordingly, according to an embodiment, an operation of the core audio object type detecting unit 120 will now be explained with reference to FIGS. 1 and 4.

As a result of detecting the type of spatial extension data in the spatial extension data type detecting unit 110, if it is determined that the spatial extension data is data indicating the core audio object type of MPEG-4 audio, the core audio object type detecting unit 120 may detect the core audio object type.

More specifically, the core audio object type detecting unit 120 may read the core audio object type by using a function “SpatialExtensionConfigData(12)”, for example, illustrated in FIG. 4. Here, “coreAudioObjectType” indicates the core audio object type of MPEG-4 audio.

Referring again to FIG. 1, the core audio data decoding unit 130 may decode core audio data, as demultiplexed by the demultiplexing unit 100. More specifically, the core audio data decoding unit 130 may decode the demultiplexed core audio data according to the core audio object type detected by the core audio object type detecting unit 120, for example.

As described above, the core audio object “type” is defined as an audio object type that is used for encoding a signal during a down-mixing at an encoding end. Here, the core audio data can be encoded by using any one encoding technique from among a variety of encoding techniques, such as bit sliced arithmetic coding (BSAC), (MP3), advanced audio coding (AAC), and MPEG audio layer-3 (MP3), at the encoding end, for example. Here, the referenced BSAC, AAC, and MP3 encoding techniques are just some of the available encoding techniques available in embodiments of the present invention, and a person of ordinary skill in the art of the present invention should understand that core audio data can be encoded by using a variety of encoding techniques.

Secondly, the case where the spatial extension data type detected by the spatial extension data type detecting unit 110 is residual coding data will now be explained, i.e., “bsSacExtType” is 0, according to the above indication examples.

FIG. 5 illustrating a syntax file, for example, for decoding residual coding data, according to an embodiment of the present invention. Accordingly, according to an embodiment, an operation of the residual coding data decoding unit 140 will now be explained with reference to FIGS. 1 and 5.

The residual coding data decoding unit 140 may include a first core audio object type determining unit 141, a first BSAC decoding unit 142, and a first AAC decoding unit 143, for example, and may decode residual coding data, according to an embodiment of the present invention.

As a result of the detecting of the type of spatial extension data in the spatial extension data type detecting unit 110, for example, if it is determined that the spatial extension data is residual coding data, the first core audio object type determining unit 141 may further determine whether the core audio object type is the ‘BSAC’ type.

Referring to FIG. 5, in this example, since the value/variable of “22” is assigned as the core audio object type of ‘BSAC’, the first core audio object type determining unit 141 may determine whether “coreAudioObjecType”, detected by the core audio object type detecting unit 120, corresponds to “22”.

As a result of the determination in the first core audio object type determining unit 141, if the core audio object type corresponds to ‘BSAC’, the first BSAC decoding unit 142 may decode a residual signal according to a ‘BSAC’ decoding technique. For example, in an embodiment, the first BSAC decoding unit 142 can be executed according to an operation indicated by reference numeral 500 or 520 in the syntax illustrated in FIG. 5. Here, in this operation indicated by the reference numeral 500 or 520, the first BSAC decoding unit 142 decodes residual coding data according to a function bsac_raw_data_block()defined in MPEG-4 ER BSAC. Here, further, in this embodiment, “nch” of bsac_raw_data_block()may always desirably be set as 1. In this case, “nch” indicates the number of channels.

If it is determined by the first core audio object type determining unit 141 that the core audio object type does not correspond to the ‘BSAC’ type, the first AAC decoding unit 143 may decode residual coding data according to an AAC decoding technique. For example, in this embodiment, the first AAC decoding unit 143 can be executed according to an operation indicated by reference numeral 510 or 530 illustrated in FIG. 5. Here, in this operation indicated by the reference numeral 510 or 530, the first AAC decoding unit 143 decodes residual coding data according to individual_channel_stream(0) defined in “MPEG-2 AAC low complexity profile bitstream syntax” described in subclause 6.3 of ISO/IEC 13818-7, for example.

However, this described AAC technique is just one embodiment for the first AAC decoding unit 143, noting that alternative embodiments are equally available.

Thus, if it is determined by the first core audio object type determining unit 141 that the core audio object type does not correspond to the ‘BSAC’ type, residual coding data can be decoded in the first AAC decoding unit 143 according to a decoding technique corresponding to the core audio object type detected by the first core audio object type determining unit 141. For example, if the core audio object type detected by the first core audio object type determining unit 141 is ‘MP3’, residual coding data may be decoded by ‘MP3’ in the first AAC decoding unit 143.

Thus, core audio data decoded in the core audio data decoding unit 130 can be up-mixed to a multichannel signal, by using residual coding data decoded in the first BSAC decoding unit 142 or the first AAC decoding unit 143.

Thirdly, the case where the spatial extension data type, e.g., detected by the spatial extension data type detecting unit 110 is an arbitrary down-mix residual coding data will now be explained, i.e., “bsSacExtType” is 1, according to the above indication examples.

FIG. 6 illustrates a syntax file, for example, for decoding arbitrary down-mix residual data, according to an embodiment of the present invention. According to an embodiment, an operation of the arbitrary down-mix residual coding data decoding unit 150 will now be explained with reference to FIGS. 1 and 6.

The arbitrary down-mix residual coding data decoding unit 150 may include a second core audio object type determining unit 151, a second BSAC decoding unit 152, and a second AAC decoding unit 153, for example, and decode arbitrary down-mix residual coding data, according to an embodiment of the present invention.

As a result of an example determination by the second core audio object type determining unit 151, if the core audio object type corresponds to the ‘BSAC’ type, the second BSAC decoding unit 152 may decode arbitrary down-mix residual coding data according to a ‘BSAC’ decoding technique. For example, the second BSAC decoding unit 152 may be executed according to at least one of operations indicated by reference numerals 600, 620, 640, and 660 of the syntax illustrated in FIG. 6. In at least one of the operations indicated by the reference numerals 600, 620, 640, and 660, for example, the second BSAC decoding unit 152 may decode arbitrary down-mix residual coding data according to a function bsac_raw_data_block()defined in MPEG-4 ER BSAC. Here, in such an embodiment, “nch” of bsac_raw_data_block()may always desirably be set as 1. In this case, “nch” indicates the number of channels.

If it is determined by the first core audio object type determining unit 151 that the core audio object type does not correspond to the ‘BSAC’ type, the second AAC decoding unit 152 may decode arbitrary down-mix residual coding data according to an ‘AAC’ decoding technique. For example, the second AAC decoding unit 153 may be executed by at least one of the operations indicated by the reference numerals 600, 620, 640, and 660. Here, in this example, in the operation indicated by the reference numeral 610 or 650, the second AAC decoding unit 153 may decode arbitrary down-mix residual coding data according to individual_channel_stream(0) defined in “MPEG-2 AAC low complexity profile bitstream syntax” described in subclause 6.3 of ISO/IEC 13818-7, for example. Further, in the operation indicated by the reference numeral 630 or 670, the second AAC decoding unit 153 may decode arbitrary down-mix residual coding data according to channel_pair_element()defined in “MPEG-2 AAC low complexity profile bitstream syntax” described in subclause 6.3 of ISO/IEC 13818-7, for example. Here, the parameter “common_window” may desirably be set as 1.

However, similar to above, the referenced AAC is just one embodiment of the second AAC decoding unit 153. If it is determined by the second core audio object type determining unit 151 that the core audio object type does not correspond to the ‘BSAC’ type, arbitrary down-mix residual coding data may be decoded in the second AAC decoding unit 153 according to a decoding technique corresponding to the core audio object type detected by the second core audio object type determining unit 151. For example, if the core audio object type detected by the second core audio object type determining unit 151 is ‘MP3’, arbitrary down-mix residual coding data may be decoded by ‘MP3’ in the second AAC decoding unit 153, again noting that alternative embodiments are equally available.

Thus, again, core audio data decoded in the core audio data decoding unit 130 can be up-mixed to a multichannel signal, by using arbitrary down-mix residual coding data decoded in the second BSAC decoding unit 152 or the second AAC decoding unit 153, for example.

Fourthly, the case where the spatial extension data type, e.g., as detected by the spatial extension data type detecting unit 110, is none of data indicating the core audio object type of MPEG-4 audio, residual coding data, or arbitrary down-mix residual coding data, will now be explained.

The spatial extension data decoding unit 160 may perform decoding by a technique corresponding to the type of spatial extension data detected by the spatial extension data type detecting unit 110. Thus, core audio data decoded in the core audio data decoding unit 130 may be up-mixed to a multichannel signal, by using data decoded in the spatial extension data decoding unit 160, for example.

The up-mixing unit 170, thus, may further up-mix the core audio data decoded in the core audio data decoding unit 130, to a multichannel signal, by using the result decoded in the first and second BSAC decoding units 142 and 152, the first and second ACC decoding units 143 and 153, or the spatial extension data decoding unit 160, for example.

FIG. 7 illustrates a method of decoding a multichannel audio signal, according to an embodiment of the present invention.

As only one example, such an embodiment may correspond to example sequential processes of the example apparatus illustrated in FIG. 1, but is not limited thereto and alternate embodiments are equally available. Regardless, this embodiment will now be briefly described in conjunction with FIG. 1, with repeated descriptions thereof being omitted.

In operation 700, the type of spatial extension data included/represented in an encoded audio signal may be detected, e.g., by the spatial extension data type detecting unit 110, for example.

In operation 710, if spatial extension data is data indicating the core audio object type, related to the encoding technique for the corresponding core audio data of the encoded audio signal, the core audio object type may be detected, e.g., by the core audio object type detecting unit 1210, for example.

In operation 720, core audio data may be decoded by using a corresponding decoding technique according to the detected core audio object type, e.g., by the core audio data decoding unit 130, for example.

In operation 730, if spatial extension data is residual coding data, residual coding data may be decoded by using a corresponding decoding technique according to the detected core audio object type, e.g., by the residual coding data decoding unit 140, for example.

In operation 740, the decoded core audio data may then be up-mixed by using residual coding data, e.g., by the up-mixing unit 170, for example.

Here, in an embodiment, if the spatial extension data is arbitrary down-mixed residual coding data, the method of decoding an audio signal may further include an operation for decoding arbitrary down-mix residual coding data by using a decoding technique according to a core audio object type. In this case, the up-mixing unit 170 may, thus, up-mix the decoded core audio data by using decoded residual coding data and decoded arbitrary down-mix residual coding data.

In addition, in an embodiment, if the spatial extension data is data other than data indicating a core audio object type, residual coding data, and arbitrary down-mix residual coding data, the technique of decoding the audio signal may further include an operation for decoding spatial extension data by a decoding technique according to the spatial extension data type. In this case, the up-mixing unit 170 may, thus, up-mix the decoded core audio data by using decoded residual coding data, decoded arbitrary down-mix residual coding data, and decoded spatial extension data.

FIG. 8 illustrates an apparatus encoding a multichannel audio signal, according to an embodiment of the present invention.

Referring to FIG. 8, the apparatus for encoding a multichannel audio signal may include a down-mixing unit 800, a core audio data encoding unit 810, a residual data encoding unit 820, an arbitrary down-mix residual data encoding unit 830, and a multiplexing unit 840, for example.

The down-mixing unit 800 may down-mix an input signal (IN). Here, the input signal (IN) may be a pulse code modulation (PCM) signal, for example, obtained through modulation of an audio signal or an analog voice signal, noting that alternatives are equally available. As noted above, the down-mixing may include the generating of a mono signal of one channel from a stereo signal of two or more channels. By performing such down-mixing, the amount of bits assigned in an encoding process can be reduced.

The core audio data encoding unit 810 may encode core audio data, e.g., as output from the down-mixing unit 800, according to a predetermined encoding technique. Here, the core audio data can be encoded by using any one of a variety of example encoding techniques such as BSAC, AAC, and MP3. Briefly, as noted above, BSAC, AAC, and MP3 are just some embodiments of the present invention, and a person of ordinary skill in the art of the present invention should understand that the core audio data can be encoded by using a variety of encoding techniques, depending on embodiment.

The residual data encoding unit 820 may include a first core audio object type determining unit 821, a first BSAC encoding unit 822, and a first AAC encoding unit 823, for example, and encode residual data.

The first core audio object type 821 may determine a core audio object type related to the encoding technique used in encoding the core audio data, e.g., in the core audio data encoding unit 810, thereby determining the encoding technique for the residual data. For example, if an encoded core audio object type is ‘BSAC’, the first core audio object type determining unit 821 may determine the encoding technique for the residual data to be a ‘BSAC’ encoding technique, and if the encoded core audio object type is ‘AAC’, the first core audio object type determining unit 821 may determine the encoding technique for the residual data to be an ‘AAC’ encoding technique.

If the determination result of the first core audio object type determining unit 821 indicates that a core audio object type is the ‘BSAC’ type, the first BSAC encoding unit 822 may encode residual data by the ‘BSAC’ technique. In this way, the core audio data and the residual data may be encoded by using an identical encoding technique, thereby reducing the complexity at the encoding end compared to conventional systems.

If the determination result of the first core audio object type determining unit 821 indicates that a core audio object type is the ‘AAC’ type, the first AAC encoding unit 823 may encode residual data by the ‘AAC’ technique. In this way, the core audio data and the residual data may be encoded by using an identical encoding technique, thereby reducing the complexity at the encoding end compared to conventional system.

However, similar to that discussed above, the ‘AAC’ technique in the first AAC encoding unit 823 is just one embodiment, and if it is determined by the first core audio object type determining unit 821 that a core audio object type does not correspond to the ‘BSAC’ type, residual data can be encoded in the first AAC encoding unit 823 by an encoding technique corresponding to a core audio object type detected by the first core audio object type determining unit 821. For example, if the core audio object type detected by the first core audio object type determining unit 821 is an ‘MP3’ type, residual data can be encoded in the first AAC encoding unit 823 by such an ‘MP3’ encoding technique.

The arbitrary down-mix residual data encoding unit 830 may include a second core audio object type determining unit 831, a second BSAC encoding unit 832, and a second AAC encoding unit 833, for example, and encode residual data, according to an embodiment of the present invention.

The second core audio object type 831 may determine a core audio object type related to the encoding technique used for the encoded core audio data in the core audio data encoding unit 810, thereby determining the encoding technique for the residual data. For example, if a core audio object type is the ‘BSAC’ type, the second core audio object type determining unit 831 may determine the encoding technique for the residual data to be a ‘BSAC’ encoding technique, and if a core audio object type is the ‘AAC’ type, the first core audio object type determining unit 821 may determine the encoding technique for the residual data to be an ‘AAC’ encoding technique.

If the determination result of the second core audio object type determining unit 831 indicates that a core audio object type is the ‘BSAC’ type, the second BSAC encoding unit 832 may encode residual data by the ‘BSAC’ encoding technique. In this way, the core audio data and the residual data may be encoded by using an identical encoding technique, thereby reducing complexity at the encoding end compared to conventional systems.

If the determination result of the second core audio object type determining unit 831 indicates that the core audio object type is the ‘AAC’ type, the second AAC encoding unit 833 may encode the residual data by the ‘AAC’ encoding technique. In this way, the core audio data and the residual data may be encoded by using an identical encoding technique, thereby reducing complexity at the encoding end compared to conventional systems.

However, similar to above, ‘AAC’ in the second AAC encoding unit 833 is just one embodiment, and if it is determined by the second core audio object type determining unit 831 that a core audio object type does not correspond to the ‘BSAC’ type, residual data can be encoded in the second AAC encoding unit 833 by an encoding technique corresponding to a core audio object type detected by the second core audio object type determining unit 831. For example, if the core audio object type detected by the second core audio object type determining unit 831 is an ‘MP3’ type, residual data can be encoded in the second AAC encoding unit 833 by using an ‘MP3’ technique.

The multiplexing unit 840 may generate a bitstream, for example, by multiplexing encoded results of the core audio data encoding unit 810, encoded results of the first and second BSAC encoding units 822 and 832, and encoded results of the first and second MC encoding units 823 and 833, and output the example bitstream to an output terminal (OUT).

FIG. 9 illustrates a method of encoding a multichannel audio signal, according to an embodiment of the present invention.

As only one example, such an embodiment may correspond to example sequential processes of the example apparatus illustrated in FIG. 8, but is not limited thereto and alternate embodiments are equally available. Regardless, this embodiment will now be briefly described in conjunction with FIG. 8, with repeated descriptions thereof being omitted.

In operation 900, an input audio signal may be down-mixed, e.g., by the down-mixing unit 800, thereby generating core audio data and residual data, for example.

In operation 910, the core audio data may be encoded according to a predetermined encoding technique, e.g., by the core audio data encoding unit 810, for example.

In operation 920, the residual data may be encoded by a predetermined encoding technique based on a core audio object type related to the encoding technique used in encoding the core audio data, e.g., by the residual data encoding unit 820, for example.

In operation 930, the encoded core audio data and the encoded residual data may be multiplexed and a result of the multiplexing may be output as the encoded audio signal, e.g., by the multiplexing unit 840, for example.

Above, through operation 900, core audio data, residual data, and arbitrary down-mix residual data can be generated by down-mixing the input audio signal.

Here, based upon the above, in this case, the method of encoding an audio signal, according to an embodiment, may further include an operation of encoding the arbitrary down-mix residual data by using a predetermined encoding technique according to a core audio object type. In this case, the multiplexing unit 940, for example, may multiplex the encoded core audio data, the encoded residual data, and the encoded arbitrary down-mix residual data, and output the result of the multiplexing as the encoding result of the audio signal.

In addition to the above described embodiments, embodiments of the present invention can also be implemented through computer readable code/instructions in/on a recording medium, e.g., a computer readable medium, to control at least one processing element to implement any above described embodiment. The medium can correspond to any medium/media permitting the storing and/or transmission of the computer readable code.

The computer readable code can be recorded/transferred on a medium in a variety of ways, with examples of the medium including recording media, such as magnetic storage media (e.g., ROM, floppy disks, hard disks, etc.) and optical recording media (e.g., CD-ROMs, or DVDs), and transmission media such as media carrying or including carrier waves, as well as elements of the Internet, for example. Thus, the medium may be such a defined and measurable structure including or carrying a signal or information, such as a device carrying a bitstream, for example, according to embodiments of the present invention. The media may also be a distributed network, so that the computer readable code is stored/transferred and executed in a distributed fashion. Still further, as only an example, the processing element could include a processor or a computer processor, and processing elements may be distributed and/or included in a single device.

According to one or more embodiments of the present invention, the decoding method may include: detecting the type of spatial extension data included in an encoding result of an audio signal; if the spatial extension data is data indicating a core audio object type related to a technique for encoding core audio data, detecting the core audio object type; decoding core audio data by a decoding technique according to the detected core audio object type; if the spatial extension data is residual coding data, decoding the residual coding data by the decoding technique according to the core audio object type; and up-mixing the decoded core audio data by using the decoded residual coding data. In this way, the core audio data and the residual coding data may be decoded by an identical decoding technique, thereby reducing complexity at the decoding end compared to conventional systems.

According to one or more embodiments of the present invention, the encoding method may include: generating core audio data and residual data by down-mixing an input audio signal; encoding the core audio data by a predetermined encoding technique; encoding the residual data by the predetermined encoding technique according to a core audio object type related to the technique by which the core audio data is encoded; and outputting the encoded core audio data and the encoded residual data as the encoding result of the audio signal. In this way, the core audio data and the residual data may be encoded by using an identical encoding technique, thereby reducing complexity at the encoding end compared to conventional systems.

While aspects of the present invention has been particularly shown and described with reference to differing embodiments thereof, it should be understood that these exemplary embodiments should be considered in a descriptive sense only and not for purposes of limitation. Any narrowing or broadening of functionality or capability of an aspect in one embodiment should not considered as a respective broadening or narrowing of similar features in a different embodiment, i.e., descriptions of features or aspects within each embodiment should typically be considered as available for other similar features or aspects in the remaining embodiments.

Thus, although a few embodiments have been shown and described, it would be appreciated by those skilled in the art that changes may be made in these embodiments without departing from the principles and spirit of the invention, the scope of which is defined in the claims and their equivalents.

Claims

1. A multi-channel decoding apparatus comprising:

a decoding unit configured:
to decode a mono down-mixed signal;
to decode a residual signal, based on information indicating whether a residual coding is applied; and
to decode side information for generating a plurality of channel signals from the decoded mono down-mixed signal, based on the information indicating whether the residual coding is applied; and
a upmixing unit configured to reconstruct the plurality of channel signals by upmixing the decoded mono down-mixed signal using the decoded side information and the decoded residual signal.

2. The multi-channel decoding apparatus of claim 1, wherein the mono down-mixed signal and the residual signal are decoded using an identical decoding technique.

Referenced Cited
U.S. Patent Documents
6456966 September 24, 2002 Iwabuchi
6502069 December 31, 2002 Grill et al.
6529604 March 4, 2003 Park et al.
7734473 June 8, 2010 Schuijers et al.
7761303 July 20, 2010 Pang et al.
7801735 September 21, 2010 Thumpudi et al.
7991495 August 2, 2011 Pang et al.
8116459 February 14, 2012 Disch et al.
20040049379 March 11, 2004 Thumpudi et al.
20040247035 December 9, 2004 Schroder et al.
20050058304 March 17, 2005 Baumgarte et al.
20060013405 January 19, 2006 Oh et al.
20060190247 August 24, 2006 Lindblom
20060235678 October 19, 2006 Kim et al.
20060239473 October 26, 2006 Kjorling et al.
20070233296 October 4, 2007 Kim et al.
20070236858 October 11, 2007 Disch et al.
20080046237 February 21, 2008 Zopf et al.
20090125314 May 14, 2009 Hellmuth et al.
20090225991 September 10, 2009 Oh et al.
20110013790 January 20, 2011 Hilpert et al.
20110022402 January 27, 2011 Engdegard et al.
Foreign Patent Documents
10-2006-0077832 July 2006 KR
2006/103581 October 2006 WO
Other references
  • Office Action for U.S. Appl. No. 11/907,398 mailed Oct. 7, 2010.
  • Office Action for U.S. Appl. No. 11/907,398 mailed Jun. 15, 2011.
  • Notice of Allowance for U.S. Appl. No. 11/907,398 mailed Aug. 21, 2012.
  • Office Action for U.S. Appl. No. 11/907,398 mailed Jan. 31, 2013.
  • Notice of Allowance for U.S. Appl. No. 11/907,398 mailed Jun. 21, 2013.
  • M. Walters, et al.; A Closer Look into MPEG-4 High Efficiency AAC, Audio Engineering Society, Convention Paper Presented at the 115 Convention, Oct. 2003, pp. 1-16.
  • ISO/IEC FDIS 23003-1:2006(E), “Information technology—MPEG audio technologies—Part 1:MPEG Surround”, Jul. 2006, pp. i-vi, 1-283.
  • Creusere, C.D.; “Understanding Perceptual Distortion in MPEG Scalable Audio Coding”, Speech and Audio Processing, III Transactions on vol. 13, No. 3, pp. 422-431, May 2005.
  • “Text of Second Working Draft for MPEG Surround”, ISO/IEC JTC 1/SC 29/WG 11, No. N7387, Jul. 29, 2005, 140 pp.
  • J. Breebaart, et al. “MPEG Spatial Audio Coding/MPEG Surround: Overview and Current Status”, Proc. 119th AES Convention, New York, USA, Oct. 2005, Preprint 6447.
  • L. Villemoes, eta l.; “MPEG Surround: The Forthcoming ISO Standard for Spatial Audio Coding”, 28th AES Int. Conference, Pitea, Sweden, 2006.
  • U.S. Appl. No. 11/907,398, filed Oct. 11, 2007, Kim et al., Samsung Electronics Co., Ltd.
  • Korean Office Action mailed Oct. 22, 2013 in corresponding Korean Application No. 10-2007-0088315.
Patent History
Patent number: 8849678
Type: Grant
Filed: Oct 28, 2013
Date of Patent: Sep 30, 2014
Patent Publication Number: 20140052455
Assignee: Samsung Electronics Co., Ltd. (Suwon-Si)
Inventors: Jung-hoe Kim (Seoul), Eun-mi Oh (Seongnam-si)
Primary Examiner: Edgar Guerra-Erazo
Application Number: 14/065,073