Abstract: Apparatus and methods are provided that employ one or more of a variety of techniques for reducing the time required to display high resolution images on a high dynamic range display having a light source layer and a display layer. In one technique, the image resolution is reduced, an effective luminance pattern is determined for the reduced resolution image, and the resolution of the effective luminance pattern is then increased to the resolution of the—display layer. In another technique, the light source layer's point spread function is decomposed into a plurality of components, and an effective luminance pattern is determined for each component. The effective luminance patterns are then combined to produce a total effective luminance pattern. Additional image display time reduction techniques are provided.
Abstract: A method of facilitating the re-joining of a virtual meeting by a participant. The method comprises determining at what time the participant was disconnected from the virtual meeting; and determining whether or not the participant was talking within a time period containing the time at which the participant was disconnected from the virtual meeting. If it is determined that the participant was talking within the time period, then, consequently, the method comprises deciding not to enable the participant to replay stored data representative of a missed portion of the virtual meeting. As well as the method, related apparatuses, systems and computer readable storage media are disclosed.
Abstract: The disclosure relates to handling nuisance in teleconference system. An endpoint device (400) for use in a teleconference includes an acquiring unit (401), a judging unit (402), a controller (403) and a processing unit (404). The acquiring unit acquires a media stream for presentation in the teleconference, and receives information from another device. The information includes a first estimation on whether the media stream is a nuisance to the teleconference. As the nuisance to a teleconference, audio or video signals are perceived by users as actually not relevant to the conference session or causing unpleasant feeling or confusion. The judging unit decides whether the media stream is the nuisance at least based on the information. The controller controls the processing of the media stream to degrade or suppress the presentation of the media stream in case that the media stream is decided as the nuisance. The processing unit processes the media stream under the control of the controller.
Abstract: In a decoder, a processor extracts a control map of false contour filtering from a part of a multi-layer video signal that includes a low dynamic range image mapped from an original high-dynamic range (HDR) image. It determines one or more filter parameters for a sparse finite-impulse-response (FIR) filter, where the one or more filter parameters relate to at least in part on the control map of false contour filtering and a predicted image predicted from the low dynamic range image. It applies the sparse FIR filter to filter pixel values in a portion of the predicted image based at least in part on the control map of false contour filtering, and it reconstructs a version of the original HDR image based at least in part on the portion of the predicted image as filtered by the FIR filter.
Abstract: In some embodiments, a method for processing an audio signal in an audio processing apparatus is disclosed. The method includes receiving an audio signal and a parameter, the parameter indicating a location of an auditory event boundary. An audio portion between consecutive auditory event boundaries constitutes an auditory event. The method further includes applying a modification to the audio signal based in part on an occurrence of the auditory event. The parameter may be generated by monitoring a characteristic of the audio signal and identifying a change in the characteristic.
Abstract: In some embodiments, a method for benchmarking an audio processing algorithm (“APA”) while the APA is executed in a manner simulating expected real time execution by a deployed system. Other embodiments include a method including steps of determining a synthetic APA which corresponds to a counterpart APA (intended for real use by a first deployed system), and benchmarking the synthetic APA while it is executed in a manner simulating expected real time execution of the synthetic APA by a contemplated deployed system. Other aspects include a system or device configured to implement any embodiment of the inventive method, or including a memory which stores data indicative of at least one synthetic APA determined in accordance with, or a benchmark generated by, an embodiment of the inventive method or steps thereof, and a computer readable medium which stores code for implementing any embodiment of the inventive method or steps thereof.
Abstract: Methods for generating encoded audio programs indicative of N channels of discontinuity-corrected, encoded audio content, including by applying discontinuity correction values to multi-channel audio content, and for rendering such a program (e.g., to generate a discontinuity-corrected M-channel mix of content indicated by the program). Other aspects are systems or devices (e.g., encoders or decoders, or rendering systems) configured to implement any of the methods.
Abstract: An image decoding apparatus that includes a motion compensation prediction circuit configured to conduct motion compensation prediction for each of blocks to be decoded by using the reconstructed image, an inverse transformation circuit configured to conduct inverse orthogonal transformation for the data of the blocks to be decoded, and a determination circuit configured to determine a filtering strength and whether or not to conduct filtering, with respect to each of the boundaries. In addition, the determining circuit is configured to determine filtering is conducted when at least one of the two adjacent blocks is intra-coded, and filtering is not conducted when both of the two adjacent blocks are not intra-coded, a non-zero transformation coefficient is not coded in both of the two adjacent blocks, the two adjacent blocks are predicted by the same reference frame, and an absolute value of a difference between motion vectors of the two adjacent blocks is smaller than a specified threshold value.
Abstract: An apparatus may include an inner module, an outer module that substantially surrounds a perimeter of the inner module, a plurality of light emitters, and a light distribution medium. The plurality of light emitters may be positioned under the inner module and project light radially outward. The light distribution medium may transport the light projected from the plurality of light emitters to an edge of the light distribution medium. The edge may include a diffusive surface and traverse a substantial portion of a boundary between the inner module and the outer module.
Abstract: Example embodiments disclosed herein relate to source separation in audio content. A method for separating sources from audio content is disclosed, the audio content being of a multi-channel format based on a plurality of channels. The method comprises performing a component analysis on the audio content for each of the plurality of channels to generate a plurality of components, each of the plurality of components comprising a plurality of time-frequency tiles in full frequency band; generating at least one dominant source with at least one of the time-frequency tiles from the plurality of the components and separating the sources from the audio content by estimating spatial parameters and spectral parameters based on the dominant source. Corresponding system and computer program product are also disclosed.
Abstract: An input media signal that carries input media content is received. The input media content is used to generate output media content in an output media signal. It is determined whether identification-and-timing (IAT) data is to be authored for the output media content. In response to determining that the output IAT data is to be authored for the output media content, output IAT data is authored for the output media content. At least a part of the output IAT data for at least a part of the output media content is encoded, along with the part of the output media content, into the output media signal. In some example scenarios, this output media signal then contains the IAT data and other related data for synchronization of additional media content with the output media content in content rendering/presentation operations.
Abstract: Apparatus and methods for mapping video signal parameters such as tone and color may be applied at various points in a video generation and delivery pipeline. apparatus may be configured to control mappings based on a range of inputs which may include one or more of: ambient conditions, user inputs, control information, adaptation models. Apparatus and methods may be applied to display video or other images so as to preserve a creative intent embodied in video or other image data.
Abstract: A method for encoding a LUT defined as a lattice of vertices is disclosed. At least one value is of each vertex of the lattice. The method comprises for a current vertex: predicting the at least one value of said current vertex from another value which is for example obtained from reconstructed values of neighboring vertices; and encoding in a bitstream at least one residue computed between the at least one value of the current vertex and its prediction in a bitstream.
Abstract: Multi-channel audio content is mixed for a particular loudspeaker setup. However, a consumer's audio setup is very likely to use a different placement of speakers. The present invention provides a method of rendering multi-channel audio that assures replay of the spatial signal components with equal loudness of the signal. A method for obtaining an energy preserving mixing matrix (G) for mixing L1 input audio channels to L2 output channels comprises steps of obtaining a first mixing matrix ?, performing a singular value decomposition on the first mixing matrix ? to obtain a singularity matrix S, processing the singularity matrix S to obtain a processed singularity matrix ?, determining a scaling factor a, and calculating an improved mixing matrix G according to G=a U ? VT. The perceived sound, loudness, timbre and spatial impression of multi-channel audio replayed on an arbitrary loudspeaker setup practically equals that of the original speaker setup.
Abstract: Novel methods and systems for encoding standard dynamic range video to improve the final quality after converting standard dynamic range video into enhanced dynamic range video are disclosed. A dual layer codec structure that amplifies certain codeword ranges can be used to send enhanced information to the decoder in order to achieve an enhanced (higher bit depth) image signal. The enhanced standard dynamic range signal can then be up-converted to enhanced dynamic range video without banding artifacts in the areas corresponding to those certain codeword ranges.
Abstract: Example embodiments disclosed herein relate to audio signal processing. A method of processing an audio signal is disclosed. The method includes detecting, based on a power distribution of the audio signal, a type of content of a frame of the audio signal, generating a first gain based on a sound level of the frame for adjusting the sound level, processing the audio signal by applying the first gain to the frame; and in response to the type of content being detected to be a breath sound, generating a second gain for mitigating the breath sound and processing the audio signal by applying the second gain to the frame. Corresponding system and computer program product are also disclosed.
Abstract: A novel spatial light modulator (SLM) includes a cover glass, and modulation layer, and a plurality of pixel mirrors, and separates unwanted, reflected light from desired, modulated light. In one embodiment, a geometrical relationship exists between the cover glass and the pixel mirrors, such that light that reflects from the cover glass is separated from light that reflects from the pixel mirrors and is transmitted from the SLM. In one example, one of the cover glass or the pixel mirrors is angled with respect to the modulation layer. In another example embodiment, the cover glass has a particular thickness, which introduces destructive interference between light that reflects from the top and bottom surfaces of the cover glass. In another embodiment antireflective coatings are disposed between optical interfaces of the SLM. In another embodiment, light from the SLM is directed through an optical filter to remove unwanted light.
Abstract: A spherical image of a spatial environment is received and contains spherically arranged pixel values indexed by a time value. The spherical image is represented in a content creation coordinate system in reference to a spatial position in the spatial environment. The spatial position is indexed by the time value. A spatial relationship is determined between the content creation coordinate system and a spherical image reference coordinate system. Based at least in part on the spatial relationship and the spherically arranged pixel values, spherical distributions of image metadata are determined for the spherical image.
Abstract: Stereoscopic video data and corresponding depth map data for stereoscopic and auto-stereoscopic displays are coded using a coded base layer and one or more coded enhancement layers. Given a 3D input picture and corresponding input depth map data, a side-by-side and a top-and-bottom picture are generated based on the input picture. Using an encoder, the side-by-side picture is coded to generate a coded base layer Using the encoder and a texture reference processing unit (RPU), the top-and-bottom picture is encoded to generate a first enhancement layer, wherein the first enhancement layer is coded based on the base layer stream, and using the encoder and a depth-map RPU, depth data for the side-by-side picture are encoded to generate a second enhancement layer, wherein the second enhancement layer is coded based on to the base layer. Alternative single, dual, and multi-layer depth map delivery systems are also presented.
June 12, 2013
Date of Patent:
December 25, 2018
Dolby Laboratories Licensing Corporation, Dolby International AB
Gopi Lakshminarayanan, Samir Hulyalkar, Tao Chen, Klaas Heinrich Schueuer, Amit Gulati, Hariharan Ganapathy, Philipp Kraetzer
Abstract: Adjacent blocks are identified in an image. Coding parameters for the adjacent blocks are identified. Deblock filtering between the identified adjacent blocks is skipped if the coding parameters for the identified adjacent blocks are similar and not skipped if the coding parameters for the identified adjacent blocks are substantially different.