Method and apparatus for visual lossless image syntactic encoding
A visual perception threshold unit for image processing identifies a plurality of visual perception threshold levels to be associated with the pixels of a video frame, wherein the threshold levels define contrast levels above which a human eye can distinguish a pixel from among its neighboring pixels of the video frame. The present invention also includes a method of generating visual perception thresholds by analysis of the details of the video frames, estimating the parameters of the details, and defining a visual perception threshold for each detail in accordance with the estimated detail parameters. The present invention further includes a method of describing images by determining which details in the image can be distinguished by the human eye and which ones can only be detected by it.
This application is a continuation application of U.S. Ser. No. 10/121,685, filed Apr. 15, 2002, now U.S. Pat. No. 6,952,500, which is a continuation application of U.S. Ser. No. 09/524,618, filed Mar. 14, 2000, issued as U.S. Pat. No. 6,473,532, which patents are incorporated herein by reference.
FIELD OF THE INVENTIONThe present invention relates generally to processing of video images and, in particular, to syntactic encoding of images for later compression by standard compression techniques.
BACKGROUND OF THE INVENTIONThere are many types of video signals, such as digital broadcast television (TV), video conferencing, interactive TV, etc. All of these signals, in their digital form, are divided into frames, each of which consists of many pixels (image elements), each of which requires 8-24 bits to describe them. The result is megabits of data per frame.
Before storing and/or transmitting these signals, they typically are compressed, using one of many standard video compression techniques, such as JPEG, MPEG, H-compression, etc. These compression standards use video signal transforms and intra- and inter-frame coding which exploit spatial and temporal correlations among pixels of a frame and across frames.
However, these compression techniques create a number of well-known, undesirable and unacceptable artifacts, such as blockiness, low resolution and wiggles, among others. These are particularly problematic for broadcast TV (satellite TV, cable TV, etc.) or for systems with very low bit rates (video conferencing, videophone).
Much research has been performed to try and improve the standard compression techniques. The following patents and articles discuss various prior art methods to do so:
U.S. Pat. Nos. 5,870,501, 5,847,766, 5,845,012, 5,796,864, 5,774,593, 5,586,200, 5,491,519, 5,341,442;
Raj Talluri et al, “A Robust, Scalable, Object-Based Video Compression Technique for Very Low Bit-Rate Coding,” IEEE Transactions of Circuit and Systems for Video Technology, vol. 7, No. 1, February 1997;
AwadKh. Al-Asmari, “An Adaptive Hybrid Coding Scheme for HDTV and Digital Sequences,” IEEE Transactions on Consumer Electronics, vol. 42, No. 3, pp. 926-936, August 1995;
Kwok-tung Lo and Jian Feng, “Predictive Mean Search Algorithms for Fast VQ Encoding of Images,” IEEE Transactions On Consumer Electronics, vol. 41, No. 2, pp. 327-331, May 1995;
James Goel et al. “Pre-processing for MPEG Compression Using Adaptive Spatial Filtering”, IEEE Transactions On Consumer Electronics, vol. 41, No. 3, pp. 687-698, August 1995;
Jian Feng et al. “Motion Adaptive Classified Vector Quantization for ATM Video Coding”, IEEE Transactions on Consumer Electronics, vol. 41, No. 2, p. 322-326, May 1995;
Austin Y. Lan et al., “Scene-Context Dependent Reference—Frame Placement for MPEG Video Coding,” IEEE Transactions on Circuits and Systems for Video Technology, vol. 9, No.3, pp. 478-489, April 1999;
Kuo-Chin Fan, Kou-Sou Kan, “An Active Scene Analysis-Based approach for Pseudoconstant Bit-Rate Video Coding”, IEEE Transactions on Circuits and Systems for Video Technology, vol. 8 No.2, pp. 159-170, April 1998;
Takashi Ida and Yoko Sambansugi, “Image Segmentation and Contour Detection Using Fractal Coding”, IEEE Transactions on Circuits and Systems for Video Technology, vol. 8, No. 8, pp. 968-975, December 1998;
Liang Shen and Rangaraj M. Rangayyan, “A Segmentation-Based Lossless Image Coding Method for High-Resolution Medical Image Compression,” IEEE Transactions on Medical Imaging, vol. 16, No. 3, pp. 301-316, June 1997;
Adrian Munteanu et al., “Wavelet-Based Lossless Compression of Coronary Angiographic Images”, IEEE Transactions on Medical Imaging, vol. 18, No. 3, p. 272-281, March 1999; and
Akira Okumura et al., “Signal Analysis and Compression Performance Evaluation of Pathological Microscopic Images,” IEEE Transactions on Medical Imaging, vol. 16, No. 6, pp. 701-710, December 1997.
SUMMARY OF THE INVENTIONAn object of the present invention is to provide a method and apparatus for video compression which is generally lossless vis-à-vis what the human eye perceives.
There is therefore provided, in accordance with a preferred embodiment of the present invention, a visual perception threshold unit for image processing. The threshold unit identifies a plurality of visual perception threshold levels to be associated with the pixels of a video frame, wherein the threshold levels define contrast levels above which a human eye can distinguish a pixel from among its neighboring pixels of the video frame.
There is also provided, in accordance with a preferred embodiment of the present invention, the visual perception threshold unit which includes a parameter generator and a threshold generator. The parameter generator generates a multiplicity of parameters that describe at least some of the information content of the processed frame. From the parameters, the threshold generator generates a plurality of visual perception threshold levels to be associated with the pixels of the video frame. The threshold levels define contrast levels above which a human eye can distinguish a pixel from among its neighboring pixels of the frame.
Moreover, in accordance with a preferred embodiment of the present invention, the parameter generator includes a volume unit, a color unit, an intensity unit or some combination of the three. The volume unit determines the volume of information in the frame, the color unit determines the per pixel color and the intensity unit determines a cross-frame change of intensity.
There is also provided, in accordance with a preferred embodiment of the present invention, a method for generating visual perception thresholds. The method includes analysis of the details of the frames of a video signal, estimating the parameters of the details, and defining a visual perception threshold for each detail in accordance with the estimated detail parameters.
There is also provided, in accordance with a preferred embodiment of the present invention, a method for describing images. The method includes determining which details in the image can be distinguished by the human eye and which ones can only be detected by it.
Moreover, in accordance with a preferred embodiment of the present invention, the method also includes providing one bit to describe a pixel which can only be detected by the human eye, and providing three bits to describe a pixel which can be distinguished by the human eye.
Further, in accordance with a preferred embodiment of the present invention, the method also includes smoothing the data of less-distinguished details.
Finally, in accordance with a preferred embodiment of the present invention, the step of determining details also includes identifying areas of high contrast and areas whose details have small dimensions.
The present invention will be understood and appreciated more fully from the following detailed description taken in conjunction with the appended drawings in which:
Applicants have realized that there are different levels of image detail in an image and that the human eye perceives these details in different ways. In particular, Applicants have realized the following:
-
- 1. Picture details whose detection mainly depends on the level of noise in the image occupy approximately 50-80% of an image.
- 2. A visual perception detection threshold for image details does not depend on the shape of the details in the image.
- 3. A visual perception threshold THD depends on a number of picture parameters, including the general brightness of the image. It does not depend on the noise spectrum.
The present invention is a method for describing, and then encoding, images based on which details in the image can be distinguished by the human eye and which ones can only be detected by it.
Reference is now made to
The human eye can distinguish most of the birds of the image. However, there is at least one bird, labeled 14, which the eye can detect but cannot determine all of its relative contrast details. Furthermore, there are large swaths of the image (in the background) which have no details in them.
The present invention is a method and system for syntactic encoding of video frames before they are sent to a standard video compression unit. The present invention separates the details of a frame into two different types, those that can only be detected (for which only one bit will suffice to describe each of their pixels) and those which can be distinguished (for which at least three bits are needed to describe the intensity of each of their pixels).
Reference is now made to
Modulator 26 modulates the reduced volume bit stream and transmits it to a receiver 30, which, as in the prior art, includes a demodulator 32 and a decoder 34. Demodulator 32 demodulates the transmitted signal and decoder 34 decodes and decompresses the demodulated signal. The result is provided to a monitor 36 for display.
It will be appreciated that, although the compression ratios are high in the present invention, the resultant video displayed on monitor 36 is not visually degraded. This is because encoder 20 attempts to quantify each frame of the video signal according to which sections of the frame are more or less distinguished by the human eye. For the less-distinguished sections, encoder 20 either provides pixels of a minimum bit volume, thus reducing the overall bit volume of the frame or smoothes the data of the sections such that video compression encoder 24 will later significantly compress these sections, thus resulting in a smaller bit volume in the compressed frame. Since the human eye does not distinguish these sections, the reproduced frame is not perceived significantly differently than the original frame, despite its smaller bit volume.
Reference is now made to
It is noted that frames are composed of pixels, each having luminance Y and two chrominance Cr and Cb components, each of which is typically defined by eight bits. VLS encoder 20 generally separately processes the three components. However, the bandwidth of the chrominance signals is half as wide as that of the luminance signal. Thus, the filters (in the x direction of the frame) for chrominance have a narrower bandwidth. The following discussion shows the filters for the luminance signal Y.
Frame analyzer 42 comprises a spatial-temporal analyzer 50, a parameter estimator 52, a visual perception threshold determiner 54 and a subclass determiner 56. Details of these elements are provided in
As discussed hereinabove, details which the human eye distinguishes are ones of high contrast and ones whose details have small dimensions. Areas of high contrast are areas with a lot of high frequency content. Thus, spatial-temporal analyzer 50 generates a plurality of filtered frames from the current frame, each filtered through a different high pass filter (HPF), where each high pass filter retains a different range of frequencies therein.
In particular, the filters of
-
- HPF-R3: 1−cos10 x
- HPF-R2: 1−cos6 x
- HPF-R1: 1−cos2 x
- HPF-C2: 1-cos4 y
- HPF-C1: 1−cos2 y
The high pass filters can also be considered as digital equivalents of optical apertures. The higher the cut-off frequency, the smaller the aperture. Thus, filters HPF-R1 and HPF-C1 retain only very small details in the frame (of 1-4 pixels in size) while filter HPF-R3 retains much larger details (of up to 11 pixels).
In the following, the filtered frames will be labeled by the type of filter (HPF-X) used to create them.
Returning to
Parameter estimator 52 takes the current frame and the filtered and difference frames and generates a set of parameters that describe the information content of the current frame. The parameters are determined on a pixel-by-pixel basis or on a per frame basis, as relevant. It is noted that the parameters do not have to be calculated to great accuracy as they are used in combination to determine a per pixel, visual perception threshold THDi.
At least some of the following parameters are determined:
Signal to noise ratio (SNR): this parameter can be determined by generating a difference frame between the current frame and the frame before it, high pass filtering of the difference frame, summing the intensities of the pixels in the filtered frame, normalized by both the number of pixels N in a frame and the maximum intensity IMAX possible for the pixel. If the frame is a television frame, the maximum intensity is 255 quanta (8 bits). The highs frequency filter selects only those intensities lower than 3σ, where σ indicates a level less than which the human eye cannot perceive noise. For example, σ can be 46 dB, equivalent to a reduction in signal strength of a factor of 200.
Normalized NΔi: this measures the change Δi, per pixel i, from the current frame to its previous frame. This value is then normalized by the maximum intensity IMAX possible for the pixel.
Normalized volume of intraframe change NIXY: this measures the volume of change in a frame IXY (or how much detail there is in a frame), normalized by the maximum possible amount of information MAXINFO within a frame (i.e. 8 bits per pixel x N pixels per frame). Since the highest frequency range indicates the amount of change in a frame, the volume of change IXY is a sum of the intensities in the filtered frame having the highest frequency range, such as filtered frame HPF-R1.
Normalized volume of interframe changes NIF: this measures the volume of changes IF between the current frame and its previous frame, normalized by the maximum possible amount of information MAXINFO within a frame. The volume of interframe changes IF is the sum of the intensities in the difference frame.
Normalized volume of change within a group of frames NIGOP: this measures the volume of changes IGOP over a group of frames, where the group is from 2 to 15 frames, as selected by the user. It is normalized by the maximum possible amount of information MAXINFO within a frame and by the number of frames in the group.
Normalized luminance level NYi: Yi is the luminance level of a pixel in the current frame. It is normalized by the maximum intensity IMAX possible for the pixel.
Color saturation pI: this is the color saturation level of the ith pixel and it is determined by:
where Cr,i and Cb,i are the chrominance levels of the ith pixel.
Hue hi: this is the general hue of the ith pixel and is determined by:
Alternatively, hue hi can be determined by interpolating Table 1, below.
Response to hue Ri(hi): this is the human vision response to a given hue and is given by Table 1, below. Interpolation is typically used to produce a specific value of the response R(h) for a specific value of hue h.
Visual perception threshold determiner 54 determines the visual perception threshold THDI per pixel as follows:
Subclass determiner 56 compares each pixel i of each high pass filtered frame HPF-X to its associated threshold THDi to determine whether or not that pixel is significantly present in each filtered frame, where “significantly present” is defined by the threshold level and by the “detail dimension” (i.e. the size of the object or detail in the image of which the pixel forms a part). Subclass determiner 56 then defines the subclass to which the pixel belongs.
For the example provided above, if the pixel is not present in any of the filtered frames, the pixel must belong to an object of large size or the detail is only detected but not distinguished. If the pixel is only found in the filtered frame of HPF-C2 or in both frames HPF-C1 and HPF-C2, it must be a horizontal edge (an edge in the Y direction of the frame). If it is found in filtered frames HPF-R3 and HPF-C2, it is a single small detail. If the pixel is found only in filtered frames HPF-R1, HPF-R2 and HPF-R3, it is a very small vertical edge. If, in addition, it is also found in filtered frame HPF-C2, then the pixel is a very small, single detail.
The above logic is summarized and expanded in Table 2.
The output of subclass determiner 56 is an indication of the subclass to which each pixel of the current frame belongs. Intra-frame processor 44 performs spatial filtering of the frame, where the type of filter utilized varies in accordance with the subclass to which the pixel belongs.
In accordance with a preferred embodiment of the present invention, intra-frame processor 44 filters each subclass of the frame differently and according to the information content of the subclass. The filtering limits the bandwidth of each subclass which is equivalent to sampling the data at different frequencies. Subclasses with a lot of content are sampled at a high frequency while subclasses with little content, such as a plain background area, are sampled at a low frequency.
Another way to consider the operation of the filters is that they smooth the data of the subclass, removing “noisiness” in the picture that the human eye does not perceive. Thus, intra-frame processor 44 changes the intensity of the pixel by an amount less than the visual distinguishing threshold for that pixel. Pixels whose contrast is lower than the threshold (i.e. details which were detected only) are transformed with non-linear filters. If desired, the data size of the detected only pixels can be reduced from 8 bits to 1 or 2 bits, depending on the visual threshold level and the detail dimension for the pixel. For the other pixels (i.e. the distinguished ones), 3 or 4 bits is sufficient.
Intra-frame processor 44 comprises a controllable filter bank 60 and a filter selector 62. Controllable filter bank 60 comprises a set of low pass and non-linear filters, shown in
Controllable filter bank 60 also includes time aligners (TA) which add any necessary delays to ensure that the pixel currently being processed remains at its appropriate location within the frame.
The low pass filters (LPF) are associated with the high pass filters used in analyzer 50. Thus, the cutoff frequencies of the low pass filters are close to those of the high pass filters. The low pass filters thus pass that which their associated high pass filters ignore.
Table 3 lists the type of filters activated per subclass, where the header for the column indicates both the type of filter and the label of the switch SW-X of
The output of intra-frame processor 44 is a processed version of the current frame which uses fewer bits to describe the frame than the original version.
Reference is now made to
The embodiments of
Summer 68 takes the difference of the processed current frame, produced by processor 44, and the previous frame, stored in either intermediate memory 84 (
In the first track, the low pass filter is used. Each pixel of the filtered frame is compared to a general, large detail, threshold THD-LF which is typically set to 5% of the maximum expected intensity for the frame. Thus, the pixels which are kept are only those which changed by more than 5% (i.e. those whose changes can be “seen” by the human eye).
) In the second track, the difference frame is high pass filtered. Since high pass filtering retains the small details, each pixel of the high pass filtered frame is compared to the particular threshold THD, for that pixel, as produced by threshold determiner 54. If the difference pixel has an intensity above the threshold THDi (i.e. the change in the pixel is significant for detailed visual perception), it is allowed through (i.e. switch 80 is set to pass the pixel).
Summer 82 adds the filtered difference pixels passed by switches 78 and/or 80 with the pixel of the previous frame to “produce the new pixel”. If switches 78 and 80 did not pass anything, the new pixel is the same as the previous pixel. Otherwise, the new pixel is the sum of the previous pixel and the low and high frequency components of the difference pixel.
Reference is now briefly made to
It is noted that the present invention can be implemented with a field programmable gate array (FPGA) and the frame memory can be implemented with SRAM or SDRAM.
The methods and apparatus disclosed herein have been described without reference to specific hardware or software. Rather, the methods and apparatus have been described in a manner sufficient to enable persons of ordinary skill in the art to readily adapt commercially available hardware and software as may be needed to reduce any of the embodiments of the present invention to practice without undue experimentation and using conventional techniques.
It will be appreciated by persons skilled in the art that the present invention is not limited by what has been particularly shown and described herein above. Rather the scope of the invention is defined by the claims that follow:
Claims
1. A visual perception threshold unit for image processing, the threshold unit comprising:
- a parameter generator to generate a multiplicity of parameters that describe at least some of the information content of at least one video frame to be processed; and
- a threshold generator to generate from said parameters, a plurality of visual perception threshold levels to be associated with the pixels of the at least one video frame,
- wherein said threshold levels define contrast levels above which a human eye can distinguish a pixel from among its neighboring pixels of said at least one video frame.
2. A unit according to claim 1 and, wherein said parameter generator comprises at least one of the following units:
- a volume unit which determines the a volume of information in said at least one video frame;
- a color unit which determines a per pixel color; and
- an intensity unit which determines a cross-frame change of intensity.
3. A method of generating visual perception thresholds for image processing implemented by one or more elements of a video encoding device, the method comprising: wherein said estimating comprises at least one of the following:
- analyzing details of frames of a video signal;
- estimating parameters of said details; and
- defining a visual perception threshold for each of said details in accordance with said estimated detail parameters,
- determining a per-pixel signal intensity change between a current frame and a previous frame, normalized by a maximum intensity;
- determining a normalized volume of intraframe change by high frequency filtering of said current frame, summing the intensities of said filtered frame and normalizing the resultant sum by the a maximum possible amount of information within a frame;
- generating a volume of inter-frame changes between a said current frame and its said previous frame normalized by said maximum possible amount of information volume within a frame;
- generating a normalized volume of inter-frame changes for a group of pictures frames from the output of said previous step of generating;
- evaluating a signal-to-noise ratio by high pass filtering a difference frame between said current frame and its said previous frame by selecting those intensities of said difference frame lower than a threshold defined as three times a noise level under which noise intensities are not perceptible to the human eye, summing the intensities of the pixels in the filtered difference frame and normalizing said sum by said maximum intensity and by the a total number of pixels in a frame;
- generating a normalized intensity value per-pixel;
- generating a per-pixel color saturation level;
- generating a per-pixel hue value; and
- determining a per-pixel response to said hue value.
4. A method for describing an image implemeneted by one or more elements of a video encoding device, the method comprising
- determining which details in said image can be distinguished by the human eye and which ones can only be detected by it;
- providing one bit to describe a pixel which can only be detected by the human eye; and
- providing three bits to describe a pixel which can be distinguished by the human eye.
5. A video compression system comprising:
- a parameter generator to generate one or more parameters that describe information content of a video frame; and
- a threshold generator to generate, from at least one of the parameters, a plurality of visual perception threshold levels to be associated with pixels of the video frame, wherein said threshold levels define contrast levels above which a pixel of the video frame can be visually distinguished from its neighboring pixels of the video frame.
6. A video compression system according to claim 5, wherein the parameter generator comprises a volume unit configured to determine a volume of information in the video frame.
7. A video compression system according to claim 5, wherein the parameter generator comprises a color unit configured to determine a per pixel color.
8. A video compression system according to claim 5, wherein the parameter generator comprises an intensity unit configured to determine a cross-frame change of intensity.
9. A video compression system according to claim 5, wherein the parameter generator and the threshold generator are implemented in a field programmable gate array (FPGA).
10. A video encoder comprising:
- a parameter generator to generate multiple parameters that describe information content of a video frame; and
- a threshold generator to generate, from at least one of the multiple parameters, a plurality of visual perception threshold levels to be associated with pixels of the video frame, wherein said threshold levels define contrast levels above which a pixel of the video frame can be visually distinguished from its neighboring pixels of the video frame.
11. A video encoder according to claim 10, wherein the parameter generator comprises a volume unit configured to determine a volume of information in the video frame.
12. A video encoder according to claim 10, wherein the parameter generator comprises a color unit configured to determine a per pixel color.
13. A video encoder according to claim 10, wherein the parameter generator comprises an intensity unit configured to determine a cross-frame change of intensity.
14. A video encoder according to claim 10, wherein the video encoder is embodied in a field programmable gate array (FPGA).
15. A video encoder according to claim 10, wherein the video encoder comprises a visual lossless syntactic encoder.
16. A system comprising:
- means for generating one or more parameters that describe information content of a video frame; and
- means for generating, from at least one of the parameters, a plurality of visual perception threshold levels to be associated with pixels of the video frame, wherein said threshold levels define contrast levels above which a pixel of the video frame can be visually distinguished from its neighboring pixels of the video frame.
17. A system according to claim 16, wherein the one or more parameters are associated with at least one of:
- a volume of information in the video frame;
- a cross-frame change of intensity; or
- a per pixel color.
18. A system according to claim 16, wherein the system is embodied in a field programmable gate array (FPGA).
19. A video compression system comprising:
- means for analyzing one or more details associated with one or more frames of a video signal;
- means for estimating parameters of individual analyzed details; and
- means for defining a visual perception threshold for individual analyzed details in accordance with at least one of the estimated parameters,
- wherein said means for estimating comprises at least one of: means for determining a per-pixel signal intensity change between a current frame and a previous frame, normalized by a maximum intensity; means for determining a normalized volume of intraframe change by high frequency filtering of said current frame, summing the intensities of said filtered frame and normalizing the resultant sum by a maximum possible amount of information within a frame; means for generating a volume of inter-frame changes between said current frame and said previous frame normalized by said maximum possible amount of information within a frame; means for generating a normalized volume of inter-frame changes within a group of frames normalized by said maximum possible amount of information within a frame and by a number of frames comprising said group of frames; means for evaluating a signal-to-noise ratio by high pass filtering a difference frame between said current frame and said previous frame by selecting intensities of said difference frame lower than a threshold defined as three times a noise level under which noise intensities are not visually perceptible, summing the intensities of pixels in the filtered difference frame and normalizing said sum by said maximum intensity and by the total number of pixels in a frame; means for generating a normalized intensity value per-pixel; means for generating a per-pixel color saturation level; means for generating a per-pixel hue value; or means for determining a per-pixel response to said hue value.
20. A video compression system according to claim 19, embodied in a field programmable gate array (FPGA).
21. A method implemented by one or more elements of a video encoding device comprising:
- identifying one or more distinguishable details in an image, individual distinguishable details being associated with a contrast level at which a pixel can be visually distinguished from among its neighboring pixels;
- using a plurality of bits to describe individual identified distinguishable details; and
- using less than said plurality of bits to describe one or more individual details in the image not identified as distinguishable.
22. A method according to claim 21, further comprising identifying the one or more of the individual details in the image not identified as distinguishable as being visually detectable.
23. A method according to claim 21, wherein using the plurality of bits comprises using three bits.
24. A method according to claim 21, further comprising performing the identifying, the using the plurality of bits, and the using less than said plurality of bits by the one or more elements embodied in a field programmable gate array (FPGA).
25. A system comprising:
- means for identifying one or more distinguishable details in an image, individual distinguishable details being associated with a contrast level at which a pixel can be visually distinguished from among its neighboring pixels;
- means for using a plurality of bits to describe individual identified distinguishable details; and
- means for using less than said plurality of bits to describe one or more individual details in the image not identified as distinguishable.
26. A system according to claim 25, wherein one or more of the individual details in the image not identified as distinguishable are identified as being visually detectable.
27. A system according to claim 25, wherein the plurality of bits comprises three bits.
28. A system according to claim 25 comprising part of a field programmable gate array (FPGA).
29. A visual perception threshold unit according to claim 1, wherein one or both of the parameter generator or the threshold generator are implemented in a video encoder.
30. A visual perception threshold unit according to claim 29, wherein the video encoder comprises a visual lossless syntactic encoder.
31. A visual perception threshold unit according to claim 1 comprising part of a field programmable gate array (FPGA).
2697758 | August 1950 | Little, Jr. |
3961133 | June 1, 1976 | Bennett |
4855825 | August 8, 1989 | Haikonen et al. |
4947255 | August 7, 1990 | Jackson et al. |
5012333 | April 30, 1991 | Lee |
5126847 | June 30, 1992 | Kori |
5194943 | March 16, 1993 | Tomita et al. |
5245445 | September 14, 1993 | Fujisawa |
5301016 | April 5, 1994 | Gehrmann |
5339171 | August 16, 1994 | Fujisawa |
5341442 | August 23, 1994 | Barrett |
5384601 | January 24, 1995 | Yamashita et al. |
5404174 | April 4, 1995 | Sugahara |
5428398 | June 27, 1995 | Faroudja |
5467404 | November 14, 1995 | Vuylsteke |
5488675 | January 30, 1996 | Hanna |
5491514 | February 13, 1996 | Fukuda |
5491517 | February 13, 1996 | Kreitman et al. |
5491519 | February 13, 1996 | Kim |
5510824 | April 23, 1996 | Nelson |
5537510 | July 16, 1996 | Kim |
5539475 | July 23, 1996 | Sadjadian et al. |
5542008 | July 30, 1996 | Sugahara et al. |
5555557 | September 10, 1996 | Mailloux |
5557340 | September 17, 1996 | Millward |
5565921 | October 15, 1996 | Sasaki et al. |
5566251 | October 15, 1996 | Hanna et al. |
5586200 | December 17, 1996 | Devaney et al. |
5592226 | January 7, 1997 | Lee et al. |
5613035 | March 18, 1997 | Kim |
5614937 | March 25, 1997 | Nelson |
5627580 | May 6, 1997 | Nelson |
5627937 | May 6, 1997 | Kim |
5648801 | July 15, 1997 | Beardsley et al. |
5653234 | August 5, 1997 | Kim et al. |
5694492 | December 2, 1997 | Kim |
5717463 | February 10, 1998 | Brailean et al. |
5774593 | June 30, 1998 | Zick |
5787203 | July 28, 1998 | Lee et al. |
5790195 | August 4, 1998 | Ohsawa |
5796864 | August 18, 1998 | Callahan |
5799111 | August 25, 1998 | Guissin |
5828776 | October 27, 1998 | Lee et al. |
5838835 | November 17, 1998 | Martens |
5844607 | December 1, 1998 | Gebler et al. |
5844614 | December 1, 1998 | Chong et al. |
5845012 | December 1, 1998 | Jung |
5847766 | December 8, 1998 | Peak |
5847772 | December 8, 1998 | Wells |
5850294 | December 15, 1998 | Apostolopoulos et al. |
5852475 | December 22, 1998 | Gupta et al. |
5870501 | February 9, 1999 | Kim |
5881174 | March 9, 1999 | Kim |
5883983 | March 16, 1999 | Lee et al. |
5901178 | May 4, 1999 | Lee et al. |
5914748 | June 22, 1999 | Parulski |
5974159 | October 26, 1999 | Lubin et al. |
5982926 | November 9, 1999 | Kuo et al. |
5991464 | November 23, 1999 | Hsu et al. |
5995656 | November 30, 1999 | Kim |
6005626 | December 21, 1999 | Ding |
6014172 | January 11, 2000 | Niesen |
6037986 | March 14, 2000 | Gouliardon et al. |
6055340 | April 25, 2000 | Nagao |
6094511 | July 25, 2000 | Metcalfe et al. |
6097848 | August 1, 2000 | Salvati |
6100625 | August 8, 2000 | Yamamoto |
6130723 | October 10, 2000 | Medin |
6191772 | February 20, 2001 | Mical et al. |
6229925 | May 8, 2001 | Alexandre et al. |
6236751 | May 22, 2001 | Farrell |
6259489 | July 10, 2001 | Flannaghan et al. |
6282299 | August 28, 2001 | Tewfik et al. |
6320676 | November 20, 2001 | Yoshidome |
6366705 | April 2, 2002 | Chiu et al. |
6385647 | May 7, 2002 | Willis et al. |
6404460 | June 11, 2002 | Chen et al. |
6463173 | October 8, 2002 | Tretter |
6466912 | October 15, 2002 | Johnston |
6473532 | October 29, 2002 | Sheraizin |
6509158 | January 21, 2003 | Schwartzz |
6522425 | February 18, 2003 | Yoshidome |
6554181 | April 29, 2003 | Dammers et al. |
6559826 | May 6, 2003 | Mendelson et al. |
6567116 | May 20, 2003 | Aman et al. |
6580825 | June 17, 2003 | Bhaskar |
6610256 | August 26, 2003 | Schwartz |
6628327 | September 30, 2003 | Aoki et al. |
6643398 | November 4, 2003 | Moriwaki |
6707487 | March 16, 2004 | Aman |
6728317 | April 27, 2004 | Demos |
6753929 | June 22, 2004 | Sheraizin |
6757449 | June 29, 2004 | Loce |
6782287 | August 24, 2004 | Grzeszczuk |
6835693 | December 28, 2004 | Bourdelais et al. |
6845181 | January 18, 2005 | Dupin et al. |
6847391 | January 25, 2005 | Kassatly |
6873442 | March 29, 2005 | Gindele et al. |
6940545 | September 6, 2005 | Ray et al. |
6940903 | September 6, 2005 | Zhao et al. |
6970506 | November 29, 2005 | Kim et al. |
7003174 | February 21, 2006 | Kryukov et al. |
7049074 | May 23, 2006 | Schwartz |
7075993 | July 11, 2006 | O'Brien |
7087021 | August 8, 2006 | Paternostro |
7110601 | September 19, 2006 | Yanagisawa |
7133451 | November 7, 2006 | Kim et al. |
7139425 | November 21, 2006 | Takahashi |
7164717 | January 16, 2007 | Katsavounidis |
7184071 | February 27, 2007 | Chellappa et al. |
7221706 | May 22, 2007 | Zhao et al. |
7221805 | May 22, 2007 | Bachelder |
7526142 | April 28, 2009 | Sheraizin et al. |
7639892 | December 29, 2009 | Sheraizin |
7742108 | June 22, 2010 | Sheraizin et al. |
20010003545 | June 14, 2001 | Hong |
20020015508 | February 7, 2002 | Hannigan et al. |
20020104854 | August 8, 2002 | Lan |
20020122494 | September 5, 2002 | Sheraizin |
20020181598 | December 5, 2002 | Vetro et al. |
20030107681 | June 12, 2003 | Masayuki et al. |
20030122969 | July 3, 2003 | Higuchi |
20030152283 | August 14, 2003 | Kagumi |
20040091145 | May 13, 2004 | Kohashi |
20040184673 | September 23, 2004 | Watanabe |
20040190789 | September 30, 2004 | Liu et al. |
20050013485 | January 20, 2005 | Masaki |
20050259185 | November 24, 2005 | Kim et al. |
20060013503 | January 19, 2006 | Kim |
20090161754 | June 25, 2009 | Sheraizin et al. |
0502615 | September 1992 | EP |
0729117 | August 1996 | EP |
1503612 | March 1978 | GB |
55-71363 | May 1980 | JP |
6-133221 | July 1985 | JP |
01206775 | August 1989 | JP |
4-83480 | March 1992 | JP |
08-191440 | July 1996 | JP |
WO00/19726 | April 2000 | WO |
- “Notice of Allowance”, U.S. Appl. No. 12/316,168, (Jun. 1, 2010), 9 pages.
- “Non Final Office Action”, U.S. Appl. No. 10/898,557, (Jun. 8, 2010), 5 pages.
- “Notice of Allowability”, U.S. Appl. No. 12/316,168, (Jun. 29, 2010), 7 pages.
- “Non Final Office Action”, U.S. Appl. No. 12/316,168, (Jun. 24, 2009), 11 pages.
- “Non Final Office Action”, U.S. Appl. No. 10/898,557, (Jul. 8, 2009), 6 pages.
- “Notice of Allowance/Base Issue Fee”, U.S. Appl. No. 11/027,674, (Jul. 23, 2009), 6 pages.
- “Notice of Allowance”, U.S. Appl. No. 11/027,674, (Feb. 24, 2009), 16 pages.
- “Restriction Requirement”, U.S. Appl. No. 10/851,190, (May 19, 2009),8 pages.
- “Notice of Allowance”, U.S. Appl. No. 10/898,557, (Jan. 27, 2010), 6 pages.
- “Notice of Allowance”, U.S. Appl. No. 12/316,168, (Jan. 29, 2010), 8 pages.
- “Notice of Allowance”, U.S. Appl. No. 10/851,190, (Feb. 8, 2010), 4 pages.
- “Non Final Office Action”, U.S. Appl. No. 10/851,190, (Sep. 1, 2009), 8 pages.
- Banhom, et al., “Digital Image Restoration”, IEEE Signal Proc., (Mar. 1997),24-41.
- Belkacem-Boussaid, “A New Image Smoothing Method Based on a Simple Model of Spatial Processing in the Earl Stages of Human Vision”, IEEE Trans. on Image Proc., vol. 9, No. 2,(Feb. 2000),220-226.
- Chan, “A practical postprocessing technique for real-time block-based coding sytem”, IEEE trans on circuits and systems for video technology, vol. 8, No. 1(Feb. 1998),4-8.
- Liu, et al., “A new postprocessing technique for the block based DCT coding based on the convex- projection theory”, IEEE trans on consumer electronics, vol. 4, No. 3,(Aug. 1998), 1054-1061.
- Lee, et al., “Efficient algorithm and architecture for post processor in HDTV”, IEEE trans on consumer electronics, vol. 44, No. 1,(Feb. 1998), 16-26.
- Choung, et al., “A fast adaptive image restoration filter for reducing block artifact in compressed images”, IEEE trans on consumer electronics, vol. 44, No. 1,(Nov. 1997),1340-1346.
- Jeon, B et al., “Blocking artifacts reduction in image compression with block boundary discontinuity criterion”, IEEE trans on circuits and systems for video technology, vol. 8, No. 3,(Jun. 1998),34557.
- Zhong, et al., “Derivation of prediction equation for blocking effect reduction”, IEEE trans on circuits and systems for video technology, vol. 9, No. 3,(Apr. 1999),415-418.
- Meier, et al., “Reduction of blocking artifacts in image and video coding”, IEEE trans on cicuits and systems for video technology, (Apr. 1999),490-500.
- Kim, et al., “An advanced contrast enhancement using partially overlapped sub-block histogram equalization”, IEEE Trans on circuits and systems for video technology, vol. 11, No. 4,(Apr. 2001),475-484.
- Stark, Alex J., “Adaptive image contrast enhancement Enhancement using generalizations of histogram equalization”, IEEE trans on image processing, vol. 9, No. 5,(May 2000),889-896.
- Hier, et al., “Real time locally adaptive contrast enhancement; A practical key to overcoming display and human visual system limitation”, SID digest, (1993),491-493.
- Chang, Dah-Chung “Image contrast enhancement based on local standard deviation”, IEEE trans on medical imaging, vol. 17, No. 4,(Aug. 1998), 518-531.
- Kim, Yeong T., “Contrast enhancement using brightness preserving bi-histogram equalization”, IEEE trans on consumer electronics, vol. 43, No. 1,(Feb. 1997),1-8.
- Kim, Tae K., et al., “Contrast enhancement system using spatiallly adaptive histogram equalization with temporal filtering”, IEEE trans on consumer electronics, vol. 44, No. 1,(Feb. 1998),82-87.
- Sakaue, Shigeo et al., “Adaptive gamma processing of the video camera for expansion of the dynamic range”, IEEE trans on consumer electronics, vol. 41, No. 3,(Aug. 1995),555-582.
- Russ, John C., “The image processing handbook”, CRS press Inc., (1995),674.
- Talluri, Raj et al., “A Robust, Scalable, Object-Based Video Compression Technique for Very Low Bit-Rate Coding”, IEEE Transaction of Circuit and Systems for Video Technology, (Feb. 1997),vol. 7, No. 1.
- Awadkh, Al-Asmari “An Adaptive Hybrid Coding Scheme for HDTV and Digital Sequences”, IEEE Transacitons on Consumer Electronics, vol. 42, No. 3,(Aug. 1995),926-936.
- Kwok, Tung Lo “Predictive Mean Search Algorithms for Fast VQ Encoding of Images”, IEEE Transactions On Consumer Electronics, vol. 41, No. 2,(May 1995),327-331.
- Goel, James et al., “Pre-processing for MPEG Compression Using Adaptive Spatial Filtering”, IEEE Transactions On Consumer electronics, vol. 41, No. 3,(Aug. 1995),687-698.
- Feng, Jian et al., “Motion Adaptive Classified Vector Quantization for ATM Video Coding”, IEEE Transactions on Consumer Electronics,, vol. 41, No. 2,(May 1995),322-326.
- Lan, Austin Y., et al., “Scene-Context Dependent Reference—Frame Placement for MPEG Video Coding”, IEEE Transactions on Circuits and Systems for Video Technology, vol. 9, No. 3,(Apr. 1999),478-489.
- Fan, Kuo-Chin et al., “An Active Scene Analysis-Based approach for Pseudo constant Bit-Rate Video Coding”, IEEE Transactions on Circuits and Systems for Video Technology, vol. 8, No. 2,(Apr. 1998),159-170.
- Takashi, Ida et al., “Image Segmentation and Contour Detection Using Fractal Coding”, IEEE Transitions on Circuits and Systems for Video Technology, vol. 8, No. 8,(Dec. 1998),968-975.
- Liang, Shen “A Segmentation-Based Lossless Image Coding Method for High-Resolution Medical Image Compression”, IEEE Transactions on Medical Imaging, vol. 16, No. 3,(Jun. 1997),301-316.
- Munteanu, Adrian et al., “Wavelet-Based Lossless Compression of Coronary Angiographic Images”, IEEE Transactions on Medical Imaging, vol. 18, No. 3,(Mar. 1999),272-281.
- Okumura, Akira et al., “Signal Analysis and Compression performance Evaluation of Pathological Microscopic Images”, IEEE Transactions on Medical Imaging, vol. 16, No. 6,(Dec. 1997),701-710.
- Pappas, et al., “Digital Color Restoration of Old Paintings”, IEEEE Trans. on Image Proc., vol. 9, No. 2,(Feb. 2000),291-294.
- Kim, et al., “Impact of HVS Models on Model-based Halftoning”, IEEE Trans. on Image Proc., vol. 11, No. 3,(Mar. 2002),258-269.
- Hentschel, et al., “Effective Paking Filter and is Implementation on a Programmable Architecture”, IEEE Trans. on Consumer Electronics, vol. 47, No. 1,(Feb. 2001),33-39.
- Yang, et al., “Maximum-Likelihood Parameter Estimation for Image Ringing-Artifact Removal”, IEEE Trans. on Cicuits and Systems for Video Technology, vol. 11, No. 8,(Aug. 2001),963-973.
- Immerkaer, “Use of Blur-Space of Deblurring and Edge- Preserving Noise Smoothing”, IEEE Trans On Image Proc., vol. 10, No. 6,(Jun. 2001),837-840.
- Chan, et al., “The Digital TV Filter and Nonlinear Denoising”, IEEE Trans on Image Proc., vol. 10, No. 2,(Feb. 2001),231-241.
- Liu, et al., “Complexity- Regularized Image Denoising”, IEEE Trans on Image Processing, vol. 10, No. 6,(Jun. 2001),341-351.
- Jostschulte, et al., “Perception Adaptive Temporal TV-noise Reduction Using Contour Preserving Prefilter Techniques”, IEEE on Consumer Electronics, vol. 44, No. 3,(Aug. 1998),1091-1096.
- Brice, Richard “Multimedia and Virtual Reality Engineering”, (1997),1-8, 174-175, 280-283.
- Conway, Lynn et al., “Video mirroring and Iconic Gestures: Enhancing Basic Videophones to Provide Visual Coaching and Visual Control”, IEEE Transactions on Consumer Electronics, vol. 44, No. 2,(May 1998),388-397.
- Cheol, Hong M., et al., “A new adaptive quantization method to reduce blocking effect”, IEEE transaction on consumer electronics, vol. 44, No. 3,(Aug. 1998),768-772.
- Massimo, Mancuso et al., “Advanced pre/ post processing for DCT coded images”, IEEE transactions on consumer electronics, vol. 44, No. 3,(Aug. 1998),1039-1041.
- Tao, Chen “Adaptive postfiltering of transform coefficients for the reduction of blocking artifacts”, IEEE transactions on circuits and systems for video technology, vol. 11, No. 5,(May 2001),594-602.
- Polesel, Andrea et al., “Image Enhancement Via Adaptive Unsharp Masking”, IEEE transactions on image processing, vol. 9, No. 3,(Mar. 2000),505-510.
- Olukayode, A et al., “An algorithm for integrated noise reduction and sharpness enhancement”, IEEE Transactions on Consumer Electronics, vol. 46, No. 3,(Aug. 2000),474-480.
- Sung-Hoon, Hong, “Joint video coding of MPEG-2 video programs for digital broadcasting services”, IEEE transactions on broadcasting, vol. 44, No. 2,(Jun. 1998),153-164.
- Lim, Jae “Two dimensional signal and image processing”, USA Simon & Schuster, (1990),430.
- Sherazain, et al., “U.S. Appl. No. 09/524,618”, filed Mar. 14, 2000.
- Leonard, Eugene “Considerations regarding the use of digital data to generate video backgrounds”, SMPTE journal, vol. 87,,(Aug. 1987),499-504.
- Tescher, Andrew “Multimedia is the message”, IEEE signal processing magazine, vol. 16, No. 1,(Jan. 1999),44-54.
- Mancuso, et al., “A new post-processing algorithm to reduce artifacts in block coded images”, IEEE trans on consumer electronics, vol. 43, No. 3,(Aug. 1997),303-307.
- Yang, J et al., “Noise estimation for blocking artifacts reduction in DCT coded images”, IEEE trans on circuits and systems for video tech nology, vol. 10, No. 7,(Oct. 2000),1116-1120.
- Kuo, et al., “Adaptive postprocessor for block endoded images”, IEEE trans on circuits and systems for video technology, vol. 5, No. 4,(Aug. 1995),298-304.
- Min, et al., “A new adaptive quantization method to reduce blocking effect”, IEEE Trans on consumer electronics, vol. 44, No. 3,(Aug. 1998),768-772.
Type: Grant
Filed: Aug 21, 2008
Date of Patent: Feb 15, 2011
Inventors: Semion Sheraizin (Mazkeret Batya 78604), Vitaly Sheraizin (Mazkeret Batya 78604)
Primary Examiner: Phuoc Tran
Application Number: 12/196,180
International Classification: G06K 9/38 (20060101);