Abstract: The present disclosure relates to image processing apparatus and method intended to be capable of suppressing deterioration in image quality. [Solution] A predicated image is generated using inter-prediction and constrained intra-prediction, the inter-prediction using as a reference image an image of a past frame processed before a current frame, the constrained intra-prediction using no neighboring pixels subjected to inter-prediction, the image of the past frame being encoded, stored in a frame memory, read from the frame memory, and then decoded. The present disclosure is applicable to, in one example, an image processing apparatus, an image coding apparatus, or the like.
Abstract: A method of adding comfort noise to a video sequence comprising setting parameters of a deblocking filter of a video encoder to change values during the video sequence, encoding frames of the video sequence using the parameters of the deblocking filter that are set to change values during the video sequence, thereby introducing comfort noise in the video sequence, and including the encoded frames in a bitstream together with an indication of which parameters of the deblocking filter were used when encoding the frames of the video sequence.
Type:
Grant
Filed:
December 18, 2018
Date of Patent:
November 10, 2020
Assignee:
AXIS AB
Inventors:
Alexander Toresson, Viktor Edpalm, Fredrik Pihl
Abstract: Systems and methods for controlling the recording, storing and transmitting of video content captured by security cameras in order to drastically reduce the overall bandwidth required to remotely monitor a physical space. More specifically, embodiments of the present invention relate to a method of triggering or activating a video camera to record and/or transmit images of a physical environment over a network only when measurements taken by a separate structural sensor camera or depth-sensing device indicate that sufficient physical changes have occurred in the environment to warrant video recording.
Abstract: A video encoding method eliminates the need for back channel feedback by using long-term reference frames for recovering data transmission errors. A videoconferencing endpoint captures image data and designates a first frame as a long term reference (LTR) frame. Subsequent intra frames us the LTR frame for reference. When a new LTR frame is designated, subsequent frames use the newly designated LTR frame for reference only after it is determined that the newly designated LTR frame is fully operational to serve in that role.
Abstract: According to a control device of the present invention, there is provided a center server including a display processing unit that performs control so as to display a panorama image display area in which a panorama image captured with a camera terminal device is displayed and an enlarged image display area in which an enlarged image is displayed, the enlarged image being obtained by enlarging an area corresponding to a position specified on the panorama image, and a control unit that performs control so as to capture with the camera terminal device an image of the area corresponding to the position specified on the panorama image. The display processing unit displays on the enlarged image display area the image captured with the camera terminal device in accordance with control by the control unit, and the panorama image is a circular panorama image.
Abstract: Methods, systems, and apparatus, including computer programs encoded on a storage device, for preventing false alarms due to display images. In one aspect, a monitoring system is disclosed that includes a processor and a computer storage media storing instructions that, when executed by the processor, cause the processor to perform operations. The operations can include obtaining, by the monitoring system, image data that depicts a portion of a property, determining, by the monitoring system, that the image data depicts an object, based on determining, by the monitoring system, that the image data depicts an object, determining, by the monitoring system, whether the depicted object is located within an exclusionary region of the property, and based on determining, by the monitoring system, that the depicted object is not located within an exclusionary region of the property, triggering, by the monitoring system, an event based on the image data.
Abstract: Image processing device for producing in real-time a digital composite image from a sequence of digital images recorded by a camera device, in particular an endoscopic camera device, the image processing device including a selecting unit, a key point detection unit, a transforming unit and a joining unit, wherein the key point detection unit includes a maximum detection unit configured for executing following steps separately for the filter response for the reference image and for the filter response for the further image, wherein a variable threshold is used: i) creating blocks by dividing the respective filter response, ii) calculating the variable threshold for each of the blocks, iii) discarding those blocks of the blocks from further consideration, in which the respective filter response at a reference point of the respective block is less than the respective variable threshold.
Type:
Grant
Filed:
May 30, 2019
Date of Patent:
September 15, 2020
Assignee:
Fraunhofer-Gesellschaft zur Foerderung der angewandten Forschung e.V.
Inventors:
Tobias Bergen, Michaela Benz, Andreas Ernst, Thomas Wittenberg, Christian Muenzenmayer, Frederik Zilly, Malte Avenhaus
Abstract: A method for generating a prediction block for coding a block of a frame using intra prediction. The method includes determining, using a training region, filter coefficients for generating the prediction block, the training region neighbors the block and includes a plurality of reconstructed pixels, the filter coefficients minimize a function of differences, each difference being a respective difference between a pixel in the training region and a prediction of that pixel in the training region, and the predictions use the filter coefficients; and generating the prediction block using the determined filter coefficients.
Abstract: Example implementations described herein are directed to systems and methods for providing a panoramic video connection between one location and another, or between one location and a number of distributed remote viewers, which provides reciprocity in terms of awareness of who is viewing and what they are looking at. In example implementations described herein, radial displays present panoramic video from remote locations, or synthesized views of face shots positioned to indicate the viewing direction of remote viewers.
Type:
Grant
Filed:
June 6, 2019
Date of Patent:
September 15, 2020
Assignee:
FUJI XEROX CO., LTD.
Inventors:
Donald Kimber, James Vaughan, Anthony Dunnigan, Pernilla Quarfordt
Abstract: Innovations in intra-picture prediction with multiple candidate reference lines available are described herein. For example, intra-picture prediction for a current block uses a non-adjacent reference line of sample values to predict the sample values of the current block. This can improve the effectiveness of the intra-picture prediction when the reference line of sample values that is adjacent the current block includes significant capture noise, significant quantization error, or significantly different values (compared to the current block) due to an occlusion.
Abstract: A method for enhancing an accuracy of object distance estimation based on a subject camera by performing pitch calibration of the subject camera more precisely with additional information acquired through V2V communication is provided. And the method includes steps of: (a) a computing device, performing (i) a process of instructing an initial pitch calibration module to apply a pitch calculation operation to the reference image, to thereby generate an initial estimated pitch, and (ii) a process of instructing an object detection network to apply a neural network operation to the reference image, to thereby generate reference object detection information; (b) the computing device instructing an adjusting pitch calibration module to (i) select a target object, (ii) calculate an estimated target height of the target object, (iii) calculate an error corresponding to the initial estimated pitch, and (iv) determine an adjusted estimated pitch on the subject camera by using the error.
Type:
Grant
Filed:
December 23, 2019
Date of Patent:
August 18, 2020
Assignee:
StradVision, Inc.
Inventors:
Kye-Hyeon Kim, Yongjoong Kim, Hak-Kyoung Kim, Woonhyun Nam, SukHoon Boo, Myungchul Sung, Dongsoo Shin, Donghun Yeo, Wooju Ryu, Myeong-Chun Lee, Hyungsoo Lee, Taewoong Jang, Kyungjoong Jeong, Hongmo Je, Hojin Cho
Abstract: A method for encoding at least one image divided into blocks. The method implements, for a current block to be encoded: determining a first prediction type for the current block; determining a prediction procedure associated with the first prediction type; obtaining a predictor block on the basis of the determined prediction procedure; and calculating a residual block that represents the difference between the predictor block obtained and the current block. The method also includes: selecting a prediction procedure from among a plurality of prediction procedures of a second type, different from the first type; and encoding the residual block calculated by using an encoding parameter determined on the basis of the selected prediction procedure.
Abstract: The disclosure provides an image sensor capable of performing a good shading correction on the imaging result regardless of the combination of the used lens module and imaging module. The body module of the image sensor obtains the format information of each module from the attached imaging module and lens module and performs on the image data from the imaging module the shading correction in accordance with the shading correction information associated with the combination of the obtained format information.
Abstract: An object displacement detection method includes capturing n images of an object for obtaining n sets of image dots, where the object corresponds to an ith set of image dots in an ith image of the n images; performing (n?1) difference calculations using the n sets of image dots to obtain (n?1) sets of difference image dots, where a jth set of difference image dots of the (n?1) sets of difference image dots is generated by performing a jth difference calculation of the (n?1) difference calculations using a (j+1)th set of image dots and a jth set of image dots of the n sets of the image dots; and determining the object has displaced when a sum of numbers of the (n?1) sets of difference image dots reaches a first threshold.
Abstract: System and method for synchronizing display of panoramic videos is disclosed. In one embodiment, a method of synchronizing display of panoramic video frames is discloses that includes transmitting a first panoramic image, displaying one or more first perspective frames based on the first panoramic image, determining an alignment parameter in response to a selected orientation of the first panoramic image, wherein the selected orientation has an associated one of the one or more first perspective frames, transmitting a second panoramic image, and displaying one or more second perspective frames based on the second panoramic image, wherein displaying the one or more second perspective frames includes displaying an initial one of the one or more second perspective frames having an orientation based on the alignment parameter.
Abstract: A system and method of defining a plane for planar coding in JVET in which first and second lines can be defined based upon pixels in left-adjacent and top-adjacent coding units. In some embodiments, the least squares method can be employed to define the relevant lines. One point along each of the lines can then be identified and the y-intercepts of the two lines can be averaged to obtain a third point. The three points can then be used to identify and define a plane for planar coding in JVET.
Abstract: In one embodiment, a method determines an encoded plurality of layers for a scalable video stream for a video where each of the plurality of layers includes a respective layer parameter setting. An additional layer for the plurality of layers is determined where the additional layer enhances a base layer in the plurality of layers and the respective layer parameter settings for the encoded plurality of layers do not take into account the additional layer. The method then determines an additional layer parameter setting for the additional layer. The additional layer parameter setting specifies a relationship between the additional layer and at least a portion of the plurality of layers where the additional layer parameter setting is used to decode the additional layer and the at least a portion of the plurality of layers.
Abstract: Provided are picture encoding and decoding methods and apparatuses, and a picture encoding and decoding system. The picture encoding method includes: determining an adjustment factor according to pixel sampling values of a video picture; converting the video picture according to the adjustment factor, and encoding the converted video picture; and writing the adjustment factor into a bitstream obtained by encoding the converted video picture.