Abstract: A coding method implemented by a decoding/encoding device for coding video data comprises coding, in a supplemental enhancement information (SEI) message, a set of sequence parameters. The set of sequence parameters is associated with temporal IDs of the sub-layers of a bitstream or coded layer video sequence (C(L)VS). The coding method further comprises processing the bitstream or C(L)VS by using the set of sequence parameters included in the SEI message. The set of sequence parameters is used for deriving or representing a picture rate of the sequence.
Abstract: A method includes receiving transform coefficients corresponding to a scaled video input signal, the scaled video input signal including a plurality of spatial layers that include a base layer. The method also includes determining a spatial rate factor based on a sample of frames from the scaled video input signal. The spatial rate factor defines a factor for bit rate allocation at each spatial layer of an encoded bit stream formed from the scaled video input signal. The spatial rate factor is represented by a difference between a rate of bits per transform coefficient of the base layer and an average rate of bits per transform coefficient. The method also includes reducing a distortion for the plurality of spatial layers by allocating a bit rate to each spatial layer based on the spatial rate factor and the sample of frames.
Abstract: Motion Vector Management for Decoder Side Motion Vector Refinement is disclosed. A method of video processing including: determining, for a conversion between a first block of video and a bitstream representation of the first block of video, whether and/or how to apply decoder-side motion vector refinement (DMVR) based on a signaled information; and performing the conversion based on the determination.
Type:
Grant
Filed:
September 1, 2021
Date of Patent:
May 21, 2024
Assignees:
BEIJING BYTEDANCE NETWORK TECHNOLOGY CO., LTD, BYTEDANCE INC.
Inventors:
Weijia Zhu, Hongbin Liu, Kai Zhang, Li Zhang, Jizheng Xu, Yue Wang
Abstract: An imaging device can stop displaying a reticle or other information when the imaging device determines that it is being used in a prohibited way. The imaging device can include an image capture device configured to view an image of a scene that includes an object viewed by the imaging device; a display to display an image of a reticle and the image of the scene; a hardware processor; and a memory storing computer-readable instructions, the instructions executable by the hardware processor to perform operations. The operations can include performing image recognition analysis on the image of the scene; determining that the object in the image is a prohibited object; and causing the display to cease displaying the image of the reticle.
Abstract: Implementations of the subject technology provide power efficient object detection for electronic devices. Power efficient object detection can be provided in order to allow extended tracking of objects by portable devices such as battery powered devices. In one or more implementations, a recommendation engine including a machine learning model and/or a rule-based recommendation process may be provided that has been trained to output a sensor recommendation indicating at least one sensor for object tracking. The sensor recommendation can indicate which of several sensors to operate to track a particular object and/or an accuracy level for operating the recommended sensor.
Abstract: According to the disclosure of the present document, a procedure of deriving indexes for inverse luma mapping and/or chroma residual scaling can be simplified, and linear LMCS can be applied. Therefore, the complexity of LMCS can be reduced, and the efficiency of video/image coding can be increased.
Abstract: An image decoding method according to the present invention includes: a step for dividing the current picture into a plurality of tiles; a step for decoding division information indicating a slice defining method; and a step for determining a first slice on the basis of the division information.
Abstract: A three-dimensional data encoding method includes: encoding a first flag indicating whether a node having a parent node different from a parent node of a current node is to be referred to in encoding of the current node included in an n-ary tree structure of three-dimensional points included in three-dimensional data; selecting a coding table from N coding tables according to occupancy states of neighboring nodes of the current node, and performing arithmetic encoding on information of the current node using the coding table selected, when the first flag indicates that the node is to be referred to; and selecting a coding table from M coding tables according to the occupancy states of the neighboring nodes of the current node, and performing arithmetic encoding on information of the current node using the coding table selected, when the first flag indicates that the node is not to be referred to.
Type:
Grant
Filed:
February 6, 2023
Date of Patent:
March 19, 2024
Assignee:
PANASONIC INTELLECTUAL PROPERTY CORPORATION OF AMERICA
Inventors:
Toshiyasu Sugio, Chi Wang, Pongsak Lasang, Chung Dean Han, Noritaka Iguchi
Abstract: An electronic device is provided. The electronic device includes a housing including a first plate, a second plate, and a side member which surrounds a space between the first plate and the second plate, a display, an antenna array disposed in the housing or in part of the housing, an image sensor, a wireless communication device electrically coupled to the antenna array, a processor, and a memory. The memory may store instructions, when executed, causing the processor to obtain and receive at least one image using the image sensor, recognize an object in the at least one image, transmit a sequence of directional beams in at least one second direction, using the antenna array, receive a sequence of reflected waves reflected by the object, using the antenna array, and recognize the object, based at least in part on the recognized object and the sequence of the reflected waves.
Type:
Grant
Filed:
May 13, 2019
Date of Patent:
March 12, 2024
Assignee:
Samsung Electronics Co., Ltd.
Inventors:
Hyunkee Min, Bokun Choi, Jongmu Choi, Sunkee Lee
Abstract: An apparatus comprising an interface and a processor. The interface may be configured to receive pixel data generated by a capture device and an activation signal generated by at least one of a plurality of sensors. The processor may be configured to process the pixel data arranged as video frames, perform computer vision on the video frames to detect objects, perform a classification of the objects, present a control signal in response to receiving the activation signal, determine whether there is a match between the classification and an object class and generate a calibration signal in response to the match. The computer vision may determine which zone that the objects have been detected in. Each zone may comprise a coverage region of one of the sensors. The calibration signal may adjust parameters of the sensors that correspond to the zone with the objects that do not have the match.
Abstract: An endoscope image-capturing device includes: a first case inside of which is sealed; an image sensor arranged inside the first case; an electro-optic conversion element arranged outside the first case and configured to convert an image signal output from the image sensor into an optical signal; and a sealing member sealing the electro-optic conversion element.
Abstract: A point cloud data transmission method according to embodiments comprises the steps of: encoding point cloud data including geometry data and attribute data; and transmitting a bitstream including the point cloud data. A point cloud data reception method according to embodiments comprises the steps of: receiving a bitstream including point cloud data including geometry data and attribute data; and decoding the point cloud data.
Abstract: A system comprises an encoder configured to compress attribute information and/or spatial information for volumetric visual content and/or a decoder configured to decompress compressed attribute and/or spatial information for the volumetric visual content. The encoder is configured to convert a 3D representation of the visual volumetric content into a 2D image based representation. The encoder is further configured to scale the patch in 2D space independent of any scaling in 3D space. Auxiliary information is signaled for use in identifying 2D scaled or unscaled patches in an image frame, mapping the patches into 3D space, and adjusting for any scaling factors applied at the encoder.
Type:
Grant
Filed:
October 14, 2022
Date of Patent:
January 30, 2024
Assignee:
Apple Inc.
Inventors:
Jungsun Kim, Khaled Mammou, Alexandros Tourapis
Abstract: A method for transmitting point cloud data according to embodiments may encode and transmit point cloud data. A method for receiving point cloud data according to embodiments may receive and decode point cloud data.
Abstract: A control apparatus controls one or more image capturing units. The apparatus comprises: an obtaining unit configured to, based on an image of a plurality of objects captured by the image capturing units, obtain positions of the plurality of objects; and a generation unit configured to, based on at least the image, the positions of the plurality of objects and the orientation of the image capturing units, generate a control command for changing the orientation of the image capturing units.
Abstract: Systems and methods include an infrared camera configured to capture an infrared image of a scene, a visible light camera configured to capture a visible light image of the scene, and a logic device configured to simultaneously capture a pair of images of the scene comprising the infrared image of the scene and the visible image of the scene, align the pair of images so that a pixel location in one of the pair of images has a corresponding pixel location in the other image, classify the visible image, annotate the infrared image based, at least in part, on the classification of the visible image, and add the annotated infrared image to a neural network training dataset for use in training a neural network for infrared image classification.
Type:
Grant
Filed:
April 30, 2021
Date of Patent:
January 16, 2024
Assignee:
Teledyne FLIR Commercial Systems, Inc.
Inventors:
Austin A. Richards, Andres Prieto-Moreno
Abstract: According to one aspect, a method for remote monitoring of electrical equipment includes acquiring a set of data points, each data point representing a temperature associated with a piece of electrical equipment or a component thereof, assigning each data point to one or more groups of data points, and defining an alarm metric for each group. Each group's alarm metric may be defined independently of other group's metrics. The defined alarm metrics are used to determine the health of the electrical equipment. The data may be determined from virtual probes within an infrared sensor and/or received from RFID devices containing temperature sensor, which are attached to or near the equipment to be monitored, for example. The methods described herein do not require conversion of sensor data into temperature values, and thus obviate the need for expensive sensors and/or computationally demanding conversion, compensation, and calibration routines.
Abstract: Sensor information is received (S41) at a host device from a mobile device. A pose of the mobile device is then determined by the host device based on the received sensor information. The pose of the mobile device is compared with pose information (S42) of previous poses of the mobile device, each previous pose of the mobile device being associated with a performance level of the system for data transmitted from the host device and received and output at the mobile device having that previous pose.
Abstract: Disclosed are an eye lens, glasses, a head-mounted display, and a VR system. The eye lens comprises: a first lens portion and a second lens portion, the first lens portion is connected to the second lens portion, and the field angle of the first lens portion and the field angle of the second lens portion form an overall field angle of the eye lens. The present invention solves the technical problem of a large number of lenses and narrow field angles caused by coaxial arrangement of multiple lenses.
Abstract: An image decoding method for decoding a bitstream including a coded signal resulting from coding tiles and slices into which an image is partitioned, the method including decoding the coded signal, wherein each of the slices is either a normal slice having, in a header, information used for an other slice or a dependent slice which is decoded using information included in a slice header of another slice, and when the normal slice starts from a position other than a beginning of a first tile, a second tile coded next to the first tile does not start from the dependent slice.
Type:
Grant
Filed:
March 24, 2021
Date of Patent:
January 9, 2024
Assignee:
SUN PATENT TRUST
Inventors:
Semih Esenlik, Matthias Narroschke, Thomas Wedi