Systems and methods are described for capturing, using a forward-facing camera associated with a head-mounted augmented reality (AR) head-mounted display (HMD), images of portions of first and second display devices in an environment, the first and second display devices displaying first and second portions of content related to an AR presentation, and displaying a third portion of content related to the AR presentation on the AR HMD, the third portion determined based upon the images of portions of the first and second display devices captured using the forward-facing camera. Moreover, the first and second display devices may be active stereo display, and the AR HMD may simultaneously function as shutter glasses.
H04N 13/332 - Displays for viewing with the aid of special glasses or head-mounted displays [HMD]
H04N 13/344 - Displays for viewing with the aid of special glasses or head-mounted displays [HMD] with head-mounted left-right displays
H04N 13/361 - Reproducing mixed stereoscopic images; Reproducing mixed monoscopic and stereoscopic images, e.g. a stereoscopic image overlay window on a monoscopic image background
H04N 13/363 - Image reproducers using image projection screens
2.
SYSTEM AND METHOD FOR HYBRID FORMAT SPATIAL DATA DISTRIBUTION AND RENDERING
Systems and methods are described for providing spatial content using a hybrid format. In some embodiments, a client device receives, from a server, surface light field representations of a plurality of scene elements in a 3D scene, including a first scene element. The client device provides to the server an indication of a dynamic behavior of a second scene element different from the first scene element. Further, in response to the indication, the client device receives from the server information defining the first scene element in a 3D asset format. The client device then renders at least the first scene element in the 3D asset format.
Systems and methods are described for display of a depth image (depth plus texture) using multiple focal planes. In one embodiment, a depth image (which may be a frame of a depth video, consisting of a video plus depth sequence) is mapped to a first set of image planes. The depth image (or a subsequent frame of the depth video) is mapped to a second set of image planes. Each image plane in the first and second set has a specified depth, and the first and second set differ in at least one depth. Each of the image planes is displayed in the first set at the respective depth of that image plane, and, subsequently, each of the image planes in the second set is displayed at its respective depth. Display of the first and second sets may be cyclically alternated at rate sufficiently high to avoid perceptible flicker.
An encoding method is presented. Transform coefficients of a block are first obtained. They are then scanned according to a scan pattern for entropy coding. The scan pattern is determined responsive to a shape of the block.
H04N 19/129 - Scanning of coding units, e.g. zig-zag scan of transform coefficients or flexible macroblock ordering [FMO]
H04N 19/13 - Adaptive entropy coding, e.g. adaptive variable length coding [AVLC] or context adaptive binary arithmetic coding [CABAC]
H04N 19/147 - Data rate or code amount at the encoder output according to rate distortion criteria
H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
H04N 19/18 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being a set of transform coefficients
5.
CHROMA QUANTIZATION PARAMETER ADJUSTMENT IN VIDEO ENCODING AND DECODING
A video coding system uses a value of a chroma quantization parameter table to quantize or de-quantize chroma values of a block of the video, the chroma quantization parameter table being directly coded in a set of parameters at the picture level or at the sequence level of the coded video stream. A corresponding encoding method, encoding device, decoding method, decoding device, and non-transitory readable medium are proposed.
H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
H04N 19/186 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being a colour or a chrominance component
H04N 19/51 - Motion estimation or motion compensation
H04N 19/70 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards
Different implementations are described, particularly implementations for determining a set of predictor candidates for affine merge coding mode from neighboring blocks for motion compensation of a picture block based on a motion model. The motion model, may be, e.g., an affine model in a merge mode or AMVP mode for a video content encoder or decoder. The motion model, may be, e.g., an affine model based on top-left/top-right control point motion vectors or an affine model based on top-left/botton-left control point motion vectors. Such affine model may be signaled by a flag. In an embodiment, predictor candidates are sorted in the set based on a criterion such as, e.g., a validity check or a vectors coherence cost. In an embodiment, a predictor candidate is selected from the set based on a motion model for each of the multiple predictor candidates, and may be based on a criterion such as, e.g., a rate distortion cost. The corresponding motion field is determined based on, e.g., one or more corresponding control point motion vectors for the block being encoded or decoded. The corresponding motion field of an embodiment identifies motion vectors used for prediction of sub-blocks of the block being encoded or decoded.
H04N 19/52 - Processing of motion vectors by encoding by predictive encoding
H04N 19/147 - Data rate or code amount at the encoder output according to rate distortion criteria
H04N 19/159 - Prediction type, e.g. intra-frame, inter-frame or bidirectional frame prediction
H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
H04N 19/56 - Motion estimation with initialisation of the vector search, e.g. estimating a good candidate to initiate a search
H04N 19/567 - Motion estimation based on rate distortion criteria
7.
METHOD AND DEVICE FOR CODING THE GEOMETRY OF A POINT CLOUD
The present principles relate to a method and device method for encoding depth values of orthogonally projected points of a point cloud onto a projection plane. The present principles also relate to a method and device for decoding a point cloud, a computer readable program and a video signal.
H04N 19/597 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding specially adapted for multi-view video sequence encoding
H04N 19/103 - Selection of coding mode or of prediction mode
H04N 19/147 - Data rate or code amount at the encoder output according to rate distortion criteria
H04N 19/17 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object
An improvement in coding efficiency is achieved through restrictions on successive divisions of asymmetric splitting in advanced video coding algorithms, which frequently rely on splitting of a block of video data prior to coding using several transform sizes. Successive divisions of asymmetric splitting are forbidden if an equivalent split can be attained using triple splitting. In an embodiment, a video block is split using successive splits, but the second type of split is dependent on the first type of split.
H04N 19/122 - Selection of transform size, e.g. 8x8 or 2x4x8 DCT; Selection of sub-band transforms of varying structure or type
H04N 19/119 - Adaptive subdivision aspects e.g. subdivision of a picture into rectangular or non-rectangular coding blocks
H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
Conventional intra-prediction uses pixels from left and upper neighbour blocks to predict a macroblock (MB). Thus, the MBs must be sequentially processed, since reconstructed left and upper MBs must be available for prediction. In an improved method for encoding Intra predicted MBs, a MB is encoded in two steps: first, a first portion of the MB is encoded independently, without references outside the MB. Pixels of the first portion can be Intra predicted using DC mode. Then, the first portion is reconstructed. The remaining pixels of the MB, being a second portion, are intra predicted from the reconstructed pixels of the first portion and then reconstructed. The first portion comprises at least one column or one row of pixels of the MB. The encoding is applied to at least two Intra predicted MBs per slice, or per picture if no slices are used.
H04N 19/593 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving spatial prediction techniques
10.
A METHOD FOR ENCODING DEPTH VALUES OF A SET OF 3D POINTS ONCE ORTHOGONALLY PROJECTED INTO AT LEAST ONE IMAGE REGION OF A PROJECTION PLANE
At least one embodiment relates to a method comprising encoding a projection mode associated with an image region of a first depth image, said projection mode indicating if said image region stores either the minimum or the maximum depth value of at least one orthogonally projected 3D point of the point cloud along a same projection direction; and encoding said depth minimum and maximum depth values as a function of said projection mode.
A method of predicting or inferring a piece of information for encoding or decoding uses a predictor from a spatial neighbor, or from a reference picture, instead of predicting or inferring the information, such as a flag for a current block, from a motion predictor block. This procedure can be used in several ways, or used in various motion vector prediction modes, such as Frame Rate Up Conversion or various variations of Advanced Motion Prediction. Other embodiments provide other methods and combinations to derive or infer motion information to increase coding efficiency.
H04N 19/52 - Processing of motion vectors by encoding by predictive encoding
H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
H04N 19/184 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being bits, e.g. of the compressed video stream
12.
METHOD FOR RENDERING OF AUGMENTED REALITY CONTENT IN COMBINATION WITH EXTERNAL DISPLAY
Systems and methods are described for improved color rendering. In some embodiments, video is captured of a real-world scene, and a region of screen content is detected in the captured video. A processor selectively applies a screen-content color transformation on the region of screen content to generate processed video, and the processed video is displayed substantially in real time, for example on a video-see-through head-mounted display. The screen-content color processing may be different for different types of external displays. The screen-content color processing may also be determined based at least in part on illumination conditions. In some embodiments, the color of displayed virtual objects is adjusted based on visual parameters of the screen content.
G06Q 10/04 - Forecasting or optimisation specially adapted for administrative or management purposes, e.g. linear programming or "cutting stock problem"
G06F 16/2457 - Query processing with adaptation to user needs
G06V 20/20 - Scenes; Scene-specific elements in augmented reality scenes
G06F 18/2113 - Selection of the most significant subset of features by ranking or filtering the set of features, e.g. using a measure of variance or of feature cross-correlation
G06F 18/21 - Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
Different implementations are described, particularly implementations for selecting a predictor candidate from a set of multiple predictor candidates for motion compensation of a picture block based on a motion model. The motion model, may be, e.g., an affine model in a merge mode for a video content encoder or decoder. In an embodiment, a predictor candidate is selected from the set based on a motion model for each of the multiple predictor candidates, and may be based on a criterion such as, e.g., a rate distortion cost. The corresponding motion field is determined based on, e.g., one or more corresponding control point motion vectors for the block being encoded or decoded. The corresponding motion field of an embodiment to identifies motion vectors used for prediction of sub-blocks of the block being encoded or decoded.
H04N 19/105 - Selection of the reference unit for prediction within a chosen coding or prediction mode, e.g. adaptive choice of position and number of pixels used for prediction
H04N 19/139 - Analysis of motion vectors, e.g. their magnitude, direction, variance or reliability
H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
H04N 19/52 - Processing of motion vectors by encoding by predictive encoding
14.
SYSTEM AND METHOD FOR ADAPTIVE SPATIAL CONTENT STREAMING WITH MULTIPLE LEVELS OF DETAIL AND DEGREES OF FREEDOM
Some embodiments of an example method may include: receiving a manifest file for streaming content, the manifest file identifying one or more degrees of freedom representations of content; tracking bandwidth available; selecting a selected representation from the one or more degrees of freedom representations based on the bandwidth available; retrieving the selected representation; and rendering the selected representation. Some embodiments of the example method may include determining estimated download latency of the one or more degrees of freedom representations. Some embodiments of the example method may include tracking client capabilities. For some embodiments of the example method, selecting the selected representation may be based on the estimated download latency and/or the client capabilities.
H04N 21/2662 - Controlling the complexity of the video stream, e.g. by scaling the resolution or bitrate of the video stream based on the client capabilities
H04N 21/24 - Monitoring of processes or resources, e.g. monitoring of server load, available bandwidth or upstream requests
H04L 65/612 - Network streaming of media packets for supporting one-way streaming services, e.g. Internet radio for unicast
15.
METHOD AND DEVICE FOR INTRA-PREDICTIVE ENCODING/DECODING A CODING UNIT COMPRISING PICTURE DATA, SAID INTRA-PREDICTIVE ENCODING DEPENDING ON A PREDICTION TREE AND A TRANSFORM TREE
Video encoding and decoding using non-rectangular partitioning. An example video decoding method includes: obtaining information identifying a non-rectangular first partitioning of a current block into a plurality of prediction units; determining a second partitioning of the current block into a plurality of transform units, wherein each transform unit is completely spatially contained within a respective one of the prediction units; and decoding the current block according to the non-rectangular first partitioning and the second partitioning. In some embodiments, the scan order of the transform units is determined such that all transform units contained in a same one of the prediction units are scanned before starting to scan transform units that are contained in a different one of the prediction units.
H04N 19/129 - Scanning of coding units, e.g. zig-zag scan of transform coefficients or flexible macroblock ordering [FMO]
H04N 19/119 - Adaptive subdivision aspects e.g. subdivision of a picture into rectangular or non-rectangular coding blocks
H04N 19/593 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving spatial prediction techniques
H04N 19/159 - Prediction type, e.g. intra-frame, inter-frame or bidirectional frame prediction
H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
16.
METHODS AND APPARATUS FOR VIDEO USABILITY INFORMATION (VUI) FOR SCALABLE VIDEO CODING (SVC)
There are provided methods and apparatus for video usability information (VUI) for scalable video coding (SVC). An apparatus includes an encoder (100) for encoding video signal data into a bitstream. The encoder specifies video user information, excluding hypothetical reference decoder parameters, in the bitstream using a high level syntax element. The video user information corresponds to a set of interoperability points in the bitstream relating to scalable video coding (340, 355).
H04N 19/30 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using hierarchical techniques, e.g. scalability
H04N 19/70 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards
17.
SCALAR QUANTIZER DECISION SCHEME FOR DEPENDENT SCALAR QUANTIZATION
When dependent scalar quantization is used, the choice of the quantizer depends on the decoding of the preceding transform coefficient, and the entropy decoding of a transform coefficient depends on quantizer choice. To maintain high throughput in hardware implementations for transform coefficient entropy coding, several decision schemes of the scaler quantizer are proposed. In one implementation, the state transition and the context model selection are based on only regular coded bins. For example, the state transition can be based on the sum of the SIG, gt1 and gt2 flags, the exclusive-or function of the SIG, gt1 and gt2 flags, or based on only the gt1 or gt2 flag. When a block of transform coefficients is coded, the regular mode bins can be coded first in one or more scan passes, and the remaining bypass coded bins are grouped together in another one or more scan passes.
H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
H04N 19/18 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being a set of transform coefficients
H04N 19/91 - Entropy coding, e.g. variable length coding [VLC] or arithmetic coding
18.
METHODS AND APPARATUS FOR SIGNALING INTRA PREDICTION FOR LARGE BLOCKS FOR VIDEO ENCODERS AND DECODERS
Methods and apparatus are provided for signaling intra prediction for large blocks for video encoders and decoders. An apparatus includes a video encoder (400) for encoding picture data for at least one large block in a picture by signaling intra prediction for the at least one large block. The intra prediction is signaled by selecting a basic coding until size and assigning a single spatial intra partition type for the basic coding until size. The single spatial intra partition type is selectable from among a plurality of spatial intra partition types. The at least one large block has a large block size greater than a block size of the basic coding unit. The intra prediction is hierarchical layer intra prediction and is performed for the at least one large block by at least one of splitting from the large block size to the basic coding until size and merging from the basic coding unit size to the large block size.
H04N 19/136 - Incoming video signal characteristics or properties
H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
H04N 19/70 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards
H04N 19/147 - Data rate or code amount at the encoder output according to rate distortion criteria
H04N 19/196 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the adaptation method, adaptation tool or adaptation type used for the adaptive coding being specially adapted for the computation of encoding parameters, e.g. by averaging previously computed encoding parameters
H04N 19/11 - Selection of coding mode or of prediction mode among a plurality of spatial predictive coding modes
H04N 19/463 - Embedding additional information in the video signal during the compression process by compressing encoding parameters before transmission
H04N 19/157 - Assigned coding mode, i.e. the coding mode being predefined or preselected to be further used for selection of another element or parameter
H04N 19/159 - Prediction type, e.g. intra-frame, inter-frame or bidirectional frame prediction
H04N 19/46 - Embedding additional information in the video signal during the compression process
H04N 19/593 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving spatial prediction techniques
Methods and apparatus for performing in-loop filtering in an encoder or a decoder provide regions which use a common set of filter parameters An index can be sent from an encoder to a decoder indicating which set of filter parameters is to be used for a particular region. The in-loop filter can be Sample Adaptive Offset, Adaptive Loop Filter, or any other such filter. An encoder classifies regions of a picture according to blocks using a common set of filter parameter. The classification can be in the form of a map. Filtering blocks use the common set of filter parameters for a region. A decoder parses a bitstream for a set of filter parameters and an index indicative of the filter parameter set for a region being decoded.
H04N 19/82 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals - Details of filtering operations specially adapted for video compression, e.g. for pixel interpolation involving filtering within a prediction loop
H04N 19/117 - Filters, e.g. for pre-processing or post-processing
H04N 19/167 - Position within a video image, e.g. region of interest [ROI]
H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
H04N 19/196 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the adaptation method, adaptation tool or adaptation type used for the adaptive coding being specially adapted for the computation of encoding parameters, e.g. by averaging previously computed encoding parameters
H04N 19/70 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards
20.
METHOD AND APPARATUS FOR VIDEO ENCODING AND DECODING USING BI-PREDICTION
Different implementations are described, particularly implementations for video encoding and decoding using motion compensation with bi-prediction are presented. The encoding method comprises for a picture, obtaining a first predictor for a block of the picture using a first reference picture; obtaining a second predictor for said block of the picture using a second reference picture; using the first predictor and the second predictor for forming a third predictor for the block in bi-prediction inter prediction, wherein the third predictor is obtained as a weighted average of the first predictor and the second predictor; and wherein a weight used in the weighed prediction depend on the position of the sample in the block. Ohers embodiments are presented for implementing block triangle partition prediction, for implementing block partition prediction using multiple patterns and for corresponding motion compensation in decoding method.
H04N 19/119 - Adaptive subdivision aspects e.g. subdivision of a picture into rectangular or non-rectangular coding blocks
H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
H04N 19/186 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being a colour or a chrominance component
H04N 19/577 - Motion compensation with bidirectional frame interpolation, i.e. using B-pictures
H04N 19/182 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being a pixel
H04N 19/105 - Selection of the reference unit for prediction within a chosen coding or prediction mode, e.g. adaptive choice of position and number of pixels used for prediction
21.
CODING AND DECODING METHODS OF A PICTURE BLOCK, CORRESPONDING DEVICES AND DATA STREAM
A method for decoding a picture block is disclosed. The decoding method comprises:—decoding (10) at least one stream S_diff into decoded data and into one information for identifying a re-constructed reference picture in a decoder picture buffer;—reconstructing (12) a special reference picture from at least the identified reconstructed reference picture and from the decoded data;—reconstructing (16) the picture block from at least the special reference picture, wherein the at least the special reference picture is not displayed.
H04N 19/44 - Decoders specially adapted therefor, e.g. video decoders which are asymmetric with respect to the encoder
H04N 21/438 - Interfacing the downstream path of the transmission network originating from a server, e.g. retrieving MPEG packets from an IP network
H04N 21/235 - Processing of additional data, e.g. scrambling of additional data or processing content descriptors
H04N 19/573 - Motion compensation with multiple frame prediction using two or more reference frames in a given prediction direction
H04N 19/30 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using hierarchical techniques, e.g. scalability
H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
H04N 19/50 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
22.
METHOD AND APPARATUS FOR ENCODING AND DECODING THREE-DIMENSIONAL SCENES IN AND FROM A DATA STREAM
Methods and devices are provided to encode and decode a data stream carrying data representative of a three-dimensional scene, the data stream comprising color pictures packed in a color image; depth pictures packed in a depth image; and a set of patch data items comprising de-projection data; data for retrieving a color picture in the color image and geometry data. Two types of geometry data are possible. The first type of data describes how to retrieve a depth picture in the depth image. The second type of data comprises an identifier of a parametric function and a list of parameter values for the identified parametric function.
METHOD AND DEVICE FOR OBTAINING A SECOND IMAGE FROM A FIRST IMAGE WHEN THE DYNAMIC RANGE OF THE LUMINANCE OF THE FIRST IMAGE IS GREATER THAN THE DYNAMIC RANGE OF THE LUMINANCE OF THE SECOND IMAGE
The present disclosure relates to a method and device for obtaining a second image from a first image when the dynamic range of the luminance of the first image is greater than the dynamic range of the luminance of the second image. The disclosure describes deriving at least one component representative of the colors of the second image from the first image, and maximizing at least one derived component according to a maximum value depending on a linear-light luminance component of the first image.
H04N 19/186 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being a colour or a chrominance component
H04N 19/85 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using pre-processing or post-processing specially adapted for video compression
H04N 19/86 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using pre-processing or post-processing specially adapted for video compression involving reduction of coding artifacts, e.g. of blockiness
24.
METHOD AND APPARATUS FOR VIDEO ENCODING AND DECODING USING PATTERN-BASED BLOCK FILTERING
Methods and apparatuses for video coding and decoding are provided. The method of video encoding includes accessing a reconstructed block corresponding to a block in a picture of a video, determining at least one filter pattern based on a property of the block and filtering the reconstructed block according to the at least one filter pattern. The method of video decoding includes accessing a reconstructed block corresponding to a block in a picture of an encoded video, determining at least one filter pattern based on a property of the block and filtering the reconstructed block according to the at least one filter pattern. A bitstream formatted to include encoded data, a computer-readable storage medium and a computer program product are also described.
H04N 19/117 - Filters, e.g. for pre-processing or post-processing
H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
H04N 19/70 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards
H04N 19/80 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals - Details of filtering operations specially adapted for video compression, e.g. for pixel interpolation
H04N 19/174 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a slice, e.g. a line of blocks or a group of blocks
25.
METHOD AND APPARATUS FOR VIDEO ENCODING AND DECODING BASED ON A LINEAR MODEL RESPONSIVE TO NEIGHBORING SAMPLES
Implementations for video encoding and decoding based on a linear model responsive to neighboring samples are described. Accordingly, for a block being encoded or decoded in a picture, at least one spatial neighboring template is determined and at least one linear model parameter is determined based on reconstructed samples of the at least one spatial neighboring template. In a first embodiment, the number N of reconstructed samples corresponds to N=2k with k chosen so that n is the maxim integer smaller than sum of the block width and block height. In a second embodiment, an offset for the relative position of a first sample in the template among samples of a left, respectively top, neighboring line of the block is determined. In a third embodiment, the number of reconstructed samples is set to a higher number in the larger dimension of the block.
H04N 19/593 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving spatial prediction techniques
H04N 19/147 - Data rate or code amount at the encoder output according to rate distortion criteria
H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
H04N 19/186 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being a colour or a chrominance component
26.
METHOD AND SYSTEM FOR THE DETECTION AND AUGMENTATION OF TACTILE INTERACTIONS IN AUGMENTED REALITY
Program actions may be initiated after detection of a pre-defined gesture by a user with a real-world object. Users may interact with their physical environment in an augmented reality by detecting interactions with real objects using a combination of location and motion detection, material identification using wearable sensors, or both. Based on detected sensor data from user interaction with a real-world object, a predefined gesture may be identified and a program action associated with that target interaction for the real-world object may be executed. In some cases, the user experience may be enhanced by providing haptic feedback in response to tactile gestures and resulting events.
The present disclosure relates to a method and device for adapting a video content decoded from elementary streams to the characteristics of a display from at least one type of metadata giving information regarding said elementary streams. Such a method comprises: —obtaining (102) an additional information (HDR DESCR.) indicating the presence of one particular type of metadata; —determining if said video content decoded from elementary streams is displayable on said display (11) from said additional information (HDR DESCR.) and the characteristics of the display (EDID); and —if said video content decoded from elementary streams is determined as being displayable, selecting (105) a process from said additional information and the characteristics of the display and adapting (106) the video content according to the selected process.
H04N 21/43 - Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronizing decoder's clock; Client middleware
H04N 21/434 - Disassembling of a multiplex stream, e.g. demultiplexing audio and video streams or extraction of additional data from a video stream; Remultiplexing of multiplex streams; Extraction or processing of SI; Disassembling of packetised elementary stream
H04N 21/235 - Processing of additional data, e.g. scrambling of additional data or processing content descriptors
28.
METHODS AND APPARATUS FOR DETERMINING QUANTIZATION PARAMETER PREDICTORS FROM A PLURALITY OF NEIGHBORING QUANTIZATION PARAMETERS
Methods and apparatus are provided for determining quantization parameter predictors from a plurality of neighboring quantization parameters. An apparatus includes an encoder for encoding image data for at least a portion of a picture using a quantization parameter predictor for a current quantization parameter to be applied to the image data. The quantization parameter predictor is determined using multiple quantization parameters from previously coded neighboring portions. A difference between the current quantization parameter and the quantization parameter predictor is encoded for signaling to a corresponding decoder.
H04N 19/126 - Quantisation - Details of normalisation or weighting functions, e.g. normalisation matrices or variable uniform quantisers
H04N 19/196 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the adaptation method, adaptation tool or adaptation type used for the adaptive coding being specially adapted for the computation of encoding parameters, e.g. by averaging previously computed encoding parameters
H04N 19/61 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding in combination with predictive coding
H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
H04N 19/157 - Assigned coding mode, i.e. the coding mode being predefined or preselected to be further used for selection of another element or parameter
H04N 19/51 - Motion estimation or motion compensation
H04N 19/593 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving spatial prediction techniques
H04N 19/174 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a slice, e.g. a line of blocks or a group of blocks
A cross-component dependent tool to be used for a chroma block of a picture is enabled responsive to a size of said chroma block and to a size of at least one luma block co-located with the chroma block. Then, the chroma block is decoded responsive to said enablement of said cross-component dependent tool. A encoding and decoding methods are presented as well as encoding and decoding devices.
H04N 19/186 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being a colour or a chrominance component
H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
H04N 19/119 - Adaptive subdivision aspects e.g. subdivision of a picture into rectangular or non-rectangular coding blocks
30.
ENCODING AND DECODING METHODS AND CORRESPONDING DEVICES
A decoding method is disclosed. First, context for a syntax clement associated with a current transform coefficient of a block of a picture is determined. The context is determined based on the area of said block, on the position of the current transform coefficient within the block and on the number of non-zero neighboring transform coefficients in a local template. Second, the syntax element is decoded based at least on the determined context. Advantageously, the local template depends on the shape of said block.
H04N 19/44 - Decoders specially adapted therefor, e.g. video decoders which are asymmetric with respect to the encoder
H04N 19/60 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding
H04N 19/70 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards
H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
31.
METHOD AND SYSTEM FOR NEAR-EYE FOCAL PLANE OVERLAYS FOR 3D PERCEPTION OF CONTENT ON 2D DISPLAYS
Some embodiments of a method may include: identifying two-dimensional (2D) content present in an image of a real-world scene; retrieving metadata comprising depth information associated with the 2D content; generating a plurality of focal plane images using the metadata, the plurality of focal plane images comprising depth cues for the 2D content; and displaying the plurality of focal plane images as a see-through overlay synchronized with the 2D content.
H04N 13/344 - Displays for viewing with the aid of special glasses or head-mounted displays [HMD] with head-mounted left-right displays
H04N 13/395 - Volumetric displays, i.e. systems where the image is built up from picture elements distributed through a volume with depth sampling, i.e. the volume being constructed from a stack or sequence of 2D image planes
H04N 13/139 - Format conversion, e.g. of frame-rate or size
H04N 13/398 - Synchronisation thereof; Control thereof
G02B 30/52 - Optical systems or apparatus for producing three-dimensional [3D] effects, e.g. stereoscopic images the image being built up from image elements distributed over a 3D volume, e.g. voxels the 3D volume being constructed from a stack or sequence of 2D planes, e.g. depth sampling systems
A Local illumination compensation system for video encoding and decoding uses memory for storing illumination compensation parameters and does not require access to reconstructed pixels of neighboring blocks. A set of illumination compensation parameters is stored in a dedicated buffer, which is of limited size, and which is decoupled from the coding unit level storage of information. The buffer contains a set of illumination compensation parameters, which may be, for example, computed (or determined in some other manner) on the fly or determined beforehand (for example for example obtained from the video signal or from a device).
H04N 19/186 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being a colour or a chrominance component
H04N 19/136 - Incoming video signal characteristics or properties
H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
33.
METHODS AND APPARATUS FOR IMPROVED ENTROPY ENCODING AND DECODING
Methods and apparatus are provided for improved entropy encoding and decoding. An apparatus includes a video encoder (200) for encoding at least a block in a picture by transforming a residue of the block to obtain transform coefficients, quantizing the transform coefficients to obtain quantized transform coefficients, and entropy coding the quantized transform coefficients. The quantized transform coefficients are encoded using a flag to indicate that a current one of the quantized transform coefficients being processed is a last non-zero coefficient for the block having a value greater than or equal to a specified value.
H04N 19/60 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding
H03M 7/40 - Conversion to or from variable length codes, e.g. Shannon-Fano code, Huffman code, Morse code
H04N 19/70 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards
H04N 19/46 - Embedding additional information in the video signal during the compression process
H04N 19/13 - Adaptive entropy coding, e.g. adaptive variable length coding [AVLC] or context adaptive binary arithmetic coding [CABAC]
H04N 19/61 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding in combination with predictive coding
34.
METHOD AND APPARATUS FOR SELECTING A CODING MODE USED FOR ENCODING/DECODING A RESIDUAL BLOCK
The present principles relate to a method and device. A method for encoding a residual block comprises: obtaining (500) a first coding mode relative to a first 2D transform when coding the residual blocks according to a coding mode relative to a first 2D transform is enabled; obtaining (510) a second coding mode relative to a second 2D transform when coding the residual blocks according to a coding mode relative to a second 2D transform is enabled; and encoding (530) the residual block according to either said first coding mode or said second coding mode or both; the method is characterized in that enabling or disabling (520) the coding of the residual block according to said second coding mode depends on said first coding mode. The present principles relate also to a method and device for encoding/decoding a picture.
H04N 19/12 - Selection from among a plurality of transforms or standards, e.g. selection between discrete cosine transform [DCT] and sub-band transform or selection between H.263 and H.264
H04N 19/105 - Selection of the reference unit for prediction within a chosen coding or prediction mode, e.g. adaptive choice of position and number of pixels used for prediction
H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
H04N 19/60 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding
35.
CLOSED-LOOP EXTENDED REALITY STIMULUS PRESENTATION AND PERCEPTION FEEDBACK USING MULTI-CHANNEL PHYSIOLOGICAL SIGNALS
In example embodiments, a method includes obtaining a first measurement of at least a first physiological parameter of a user from a time before presentation of an information item to the user. The information item may be, for example, a notification, an advertisement, or an emergency alert, among other possibilities. A second measurement of the first physiological parameter is obtained from a time after a beginning of the presentation of the information item to the user (e.g. during the presentation). Based at least on a comparison between the first and second measurements, a determination is made of whether the user has perceived the information item. The determination may also be based on a corruption coefficient indicating an amount by which an activity of the user is likely to interfere with the first physiological parameter.
The present embodiments relate to a method and device. The method comprises obtaining at least one first point from at least one point of a point cloud by projecting said point of the point cloud onto a projection plane and obtaining at least one other point of the point cloud determined according to said at least one first point; determining and encoding at least one interpolation coding mode for said at least one first point based on at least one reconstructed point obtained from said at least one first point and at least one interpolation point defined by said at least one interpolation coding mode to approximate said at least one other point of the point cloud; and signaling said at least interpolation coding mode as values of image data.
Methods, apparatuses and streams are disclosed for transmitting tiled volumetric video and, at the receiver, for generating an atlas image compatible with a legacy decoder. At the server side, viewport information is obtained and a first list of central tiles and a second list of border tiles are selected. A central tile is a part of an image obtained by projecting the 3D scene onto an image plane according to a central point of view. A border tile is an image comprising dis-occluding patches. Sizes and shapes of border tiles are function of size and shape of central tiles. At the client side, tiles are arranged according to a layout selected in a set of layouts according to the number, sizes and shapes of border tiles.
Methods, apparatuses and streams are disclosed for transmitting tiled volumetric video and, at the receiver, for generating an atlas image compatible with a legacy decoder. At the server side, viewport information is obtained and a first list of central tiles and a second list of border tiles are selected. A central tile is a part of an image obtained by projecting the 3D scene onto an image plane according to a central point of view. A border tile is an image comprising dis-occluding patches. Sizes and shapes of border tiles are function of size and shape of central tiles. At the client side, tiles are arranged according to a layout selected in a set of layouts according to the number, sizes and shapes of border tiles.
FIG. 8.
H04N 13/178 - Metadata, e.g. disparity information
H04N 19/597 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding specially adapted for multi-view video sequence encoding
38.
REFINEMENT MODE PROCESSING IN VIDEO ENCODING AND DECODING
Described are systems and methods for encoding or decoding a picture part of video information can include using a refinement mode per block where the refinement mode can be based on a refinement parameter. The refinement mode can include a cross-component refinement that can be a cross-component chroma refinement. The refinement mode can include enabling selection per block of a refinement parameter where the refinement parameter can include one or more chroma refinement parameters included in a chroma refinement table.
H04N 19/186 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being a colour or a chrominance component
H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
H04N 19/80 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals - Details of filtering operations specially adapted for video compression, e.g. for pixel interpolation
Predicting a component of a current pixel belonging to a current sub-aperture image in a matrix of sub-aperture images captured by a sensor of a type I plenoptic camera can involve, first, determining a location on the sensor based on: a distance from an exit pupil of a main lens of the camera to a micro-lens array of the camera; a focal length of the main lens; a focal length of the micro-lenses of the micro-lens array; and a set of parameters of a model of the camera allowing for a derivation of a two-plane parameterization describing the field of rays corresponding to the pixels of the sensor; and, second, predicting the component based on one reference pixel belonging to a reference sub-aperture image in the matrix and located on the sensor in a neighborhood of the location.
H04N 19/597 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding specially adapted for multi-view video sequence encoding
H04N 19/105 - Selection of the reference unit for prediction within a chosen coding or prediction mode, e.g. adaptive choice of position and number of pixels used for prediction
H04N 19/182 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being a pixel
40.
METHODS AND APPARATUS FOR VIDEO ENCODING AND DECODING BINARY SETS USING ADAPTIVE TREE SELECTION
Methods and apparatus are provided for encoding and decoding binary sets using adaptive tree selection. In one exemplary encoding method embodiment, picture data is encoded for a block in a picture; in which one of a plurality of trees structures is selected to code a binary set of data for indicating coefficient significance for the block. In another exemplary encoding method embodiment, picture data is encoded for a block in a picture, in which one or more trees are used to encode a binary set of data for indicating coefficient significance for the block, the one or more trees each having a plurality of nodes, at least one of the nodes of the one or more trees being modified responsive to at least one parameter.
H04N 19/184 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being bits, e.g. of the compressed video stream
H04N 19/159 - Prediction type, e.g. intra-frame, inter-frame or bidirectional frame prediction
H04N 19/70 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards
H04N 19/147 - Data rate or code amount at the encoder output according to rate distortion criteria
H04N 19/172 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a picture, frame or field
H04N 19/46 - Embedding additional information in the video signal during the compression process
H04N 19/61 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding in combination with predictive coding
H04N 19/91 - Entropy coding, e.g. variable length coding [VLC] or arithmetic coding
H04N 19/64 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding using sub-band based transform, e.g. wavelets characterised by ordering of coefficients or of bits for transmission
H04N 19/117 - Filters, e.g. for pre-processing or post-processing
H04N 19/139 - Analysis of motion vectors, e.g. their magnitude, direction, variance or reliability
H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
H04N 19/625 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding using discrete cosine transform [DCT]
Different implementations are described, particularly implementations for video encoding and decoding are presented including a method for deblocking an image. According to an implementation, in a method for deblocking an image, at least one boundary is determined between a first block of samples and a second block of samples: a boundary strength is determined according to at least one of a prediction mode of the first block and a prediction mode of the second block; and samples of the first and second blocks neighboring the at least one boundary are filtered according to the boundary strength. Advantageously, in case the prediction mode of the first block is a weighted prediction mode, the boundary strength further depends on the relative weight of samples used in predicting the first block of samples according to the weighted prediction mode of the first block and reciprocally for the second block.
H04N 19/117 - Filters, e.g. for pre-processing or post-processing
H04N 19/105 - Selection of the reference unit for prediction within a chosen coding or prediction mode, e.g. adaptive choice of position and number of pixels used for prediction
H04N 19/132 - Sampling, masking or truncation of coding units, e.g. adaptive resampling, frame skipping, frame interpolation or high-frequency transform coefficient masking
H04N 19/159 - Prediction type, e.g. intra-frame, inter-frame or bidirectional frame prediction
H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
H04N 19/82 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals - Details of filtering operations specially adapted for video compression, e.g. for pixel interpolation involving filtering within a prediction loop
H04N 19/189 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the adaptation method, adaptation tool or adaptation type used for the adaptive coding
H04N 19/635 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding using sub-band based transform, e.g. wavelets - characterised by filter definition or implementation details
42.
ILLUMINATION COMPENSATION FLAG IN FRAME RATE UP-CONVERSION WITH TEMPLATE MATCHING
Inferring an illumination compensation flag during encoding or decoding of a video image signal using frame rate up conversion can save one bit and eliminate complexity. The illumination compensation flag can be derived from the corresponding flags of at least one bi-predictive or bi-directional prediction candidates. The flag can also be derived from some function of the flags from those candidates. Alternatively, several flags can be used for respective coding or decoding of blocks if there are more than one prediction candidate using illumination compensation.
H04N 19/46 - Embedding additional information in the video signal during the compression process
H04N 19/159 - Prediction type, e.g. intra-frame, inter-frame or bidirectional frame prediction
H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
H04N 19/186 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being a colour or a chrominance component
43.
QUANTIZATION PARAMETER PREDICTION FOR VIDEO ENCODING AND DECODING
Apparatus and methods for encoding and decoding video data having improved quantization parameter prediction are presented and can include using a combination of quantization parameter prediction values obtained from respective ones of a plurality of quantization parameter prediction methods to encode or decode the video data.
H04N 19/186 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being a colour or a chrominance component
H04N 19/70 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards
A method for decoding a picture block is disclosed. The decoding method comprises: —decoding (10) at least one stream S_diff into decoded data and into one information for identifying a reconstructed reference picture in a decoder picture buffer; —reconstructing (12) a special reference picture from at least the identified reconstructed reference picture and from the decoded data; —reconstructing (16) the picture block from at least the special reference picture, wherein the at least the special reference picture is not displayed.
H04N 19/44 - Decoders specially adapted therefor, e.g. video decoders which are asymmetric with respect to the encoder
H04N 21/438 - Interfacing the downstream path of the transmission network originating from a server, e.g. retrieving MPEG packets from an IP network
H04N 21/235 - Processing of additional data, e.g. scrambling of additional data or processing content descriptors
H04N 19/573 - Motion compensation with multiple frame prediction using two or more reference frames in a given prediction direction
H04N 19/30 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using hierarchical techniques, e.g. scalability
H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
H04N 19/50 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
45.
METHOD AND APPARATUS FOR VIDEO ENCODING AND DECODING WITH PARTIALLY SHARED LUMA AND CHROMA CODING TREES
Different implementations are described, particularly implementations for video encoding and decoding based on wherein the partitioning tree structure is partially shared between luma and chroma blocks are presented. According to an implementation, a single partitioning tree structure is shared between luma and chroma blocks from a root node of the partitioning tree down to a switching node and a dual partitioning tree structure is determined for luma and chroma blocks from the switching node down to a leaf node of the partitioning tree. Thus, this implementation optimizes the coding efficiency by allowing separated trees for smaller blocks, while limiting the complexity of decoding pipeline.
H04N 19/119 - Adaptive subdivision aspects e.g. subdivision of a picture into rectangular or non-rectangular coding blocks
H04N 19/159 - Prediction type, e.g. intra-frame, inter-frame or bidirectional frame prediction
H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
H04N 19/186 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being a colour or a chrominance component
H04N 19/46 - Embedding additional information in the video signal during the compression process
H04N 19/88 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using pre-processing or post-processing specially adapted for video compression involving rearrangement of data among different coding units, e.g. shuffling, interleaving, scrambling or permutation of pixel data or permutation of transform coefficient data among different blocks
Video signal coding and decoding functions can generate lists of potential candidates to use in coding and decoding, for example, predictors. Video signal coding component candidate undergo operations before potential inclusion in candidate lists. The candidates are checked after being modified by the operations to see if other equal candidates are already in the candidate list. If equal candidates are not in the list, the modified candidates are added to the candidate list. If equal candidates are already in the list, the modified candidates are not added to the list. Operations that can be performed comprise rounding and clipping.
H04N 19/105 - Selection of the reference unit for prediction within a chosen coding or prediction mode, e.g. adaptive choice of position and number of pixels used for prediction
H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
H04N 19/46 - Embedding additional information in the video signal during the compression process
H04N 19/52 - Processing of motion vectors by encoding by predictive encoding
47.
Methods and systems of automatic calibration for dynamic display configurations
Systems and methods are described for capturing, using a forward-facing camera associated with a head-mounted augmented reality (AR) head-mounted display (HMD), images of portions of first and second display devices in an environment, the first and second display devices displaying first and second portions of content related to an AR presentation, and displaying a third portion of content related to the AR presentation on the AR HMD, the third portion determined based upon the images of portions of the first and second display devices captured using the forward-facing camera. Moreover, the first and second display devices may be active stereo display, and the AR HMD may simultaneously function as shutter glasses.
H04N 13/363 - Image reproducers using image projection screens
H04N 13/361 - Reproducing mixed stereoscopic images; Reproducing mixed monoscopic and stereoscopic images, e.g. a stereoscopic image overlay window on a monoscopic image background
H04N 13/332 - Displays for viewing with the aid of special glasses or head-mounted displays [HMD]
48.
MANAGING CODING TOOLS COMBINATIONS AND RESTRICTIONS
Video coding tools can be controlled by including syntax in a video bitstream that makes better use of video decoding resources. An encoder inserts syntax into a video bitstream to enable a decoder to parse the bitstream and easily control which tools combinations are enabled, which combinations are not permitted, and which tools are activated for various components in a multiple component bitstream, leading to potential parallelization of bitstream decoding.
H04N 19/70 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards
H04N 19/103 - Selection of coding mode or of prediction mode
H04N 19/157 - Assigned coding mode, i.e. the coding mode being predefined or preselected to be further used for selection of another element or parameter
H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
H04N 19/503 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
A video encoder or decoder processes portions of video with less delay when its processes are parallelized and avoids delays caused by dependence on the completion of prior processes. In one embodiment, a motion vector predictor from a neighboring block of video is used in a subsequent later block of video before it is finished being refined for use in the neighboring block. In another embodiment, information from a neighboring block is confined to include blocks in the same coding tree unit. In another embodiment, a motion vector predictor is checked to see whether it is already in a list of candidates before adding it to the list to expedite the process.
H04N 19/70 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards
H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
A method and system are provided for processing image content. In one embodiment the method comprises receiving a plurality of captured contents showing same scene as captured by one or more cameras having a different focal length and depth maps and generating a consensus cube by obtaining depth map estimations from said received contents. The visibility of different objects in then analysed to create a soft visibility cube that provides visibility information about each content. A color cube is then generated by using information from the consensus and soft visibility cube. The color cube is then used to combine different received contents and generate a single image for the plurality of contents received.
H04N 13/395 - Volumetric displays, i.e. systems where the image is built up from picture elements distributed through a volume with depth sampling, i.e. the volume being constructed from a stack or sequence of 2D image planes
G06T 5/50 - Image enhancement or restoration by the use of more than one image, e.g. averaging, subtraction
At least one embodiment relates to a method comprising encoding a projection mode associated with an image region of a first depth image, said projection mode indicating if said image region stores either the minimum or the maximum depth value of at least one orthogonally projected 3D point of the point cloud along a same projection direction; and encoding said depth minimum and maximum depth values as a function of said projection mode.
Systems and methods are described for refining first point cloud data using at least second point cloud data and one or more sets of quantizer shifts. An example point cloud decoding method includes obtaining data representing at least a first point cloud and a second point cloud; obtaining information identifying at least a first set of quantizer shifts associated with the first point cloud; and obtaining refined point cloud data based on at least the first point cloud, the first set of quantizer shifts, and the second point cloud. The obtaining of the refined point cloud data may include performing a subtraction based on at least the first set of quantizer shifts. Corresponding encoding systems and methods are also described.
H04N 19/597 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding specially adapted for multi-view video sequence encoding
H04N 19/463 - Embedding additional information in the video signal during the compression process by compressing encoding parameters before transmission
H04N 19/85 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using pre-processing or post-processing specially adapted for video compression
H04N 19/172 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a picture, frame or field
53.
METHODS AND APPARATUS FOR TRANSFORM SELECTION IN VIDEO ENCODING AND DECODING
There are provided methods and apparatus for transform selection in video coding. An apparatus includes a video encoder (300) for encoding at least a block in a picture by selecting a transform (329) to apply to a residue of the block from a set of two or more available transforms (325, 326, 327). The transform is selected based on at least one of an inter prediction mode used to predict at least one reference for the block, one or more values corresponding to a motion vector, a value of a residue of one or more previously encoded blocks, a value of prediction data for the block, one or more transform selections of one or more neighboring reconstructed blocks, and a quantization step applied to transform coefficients for the residue of the block.
H04N 19/61 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding in combination with predictive coding
H04N 19/159 - Prediction type, e.g. intra-frame, inter-frame or bidirectional frame prediction
H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
H04N 19/12 - Selection from among a plurality of transforms or standards, e.g. selection between discrete cosine transform [DCT] and sub-band transform or selection between H.263 and H.264
H04N 19/139 - Analysis of motion vectors, e.g. their magnitude, direction, variance or reliability
H04N 19/18 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being a set of transform coefficients
H04N 19/196 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the adaptation method, adaptation tool or adaptation type used for the adaptive coding being specially adapted for the computation of encoding parameters, e.g. by averaging previously computed encoding parameters
H04N 19/119 - Adaptive subdivision aspects e.g. subdivision of a picture into rectangular or non-rectangular coding blocks
H04N 19/147 - Data rate or code amount at the encoder output according to rate distortion criteria
H04N 19/46 - Embedding additional information in the video signal during the compression process
H04N 19/122 - Selection of transform size, e.g. 8x8 or 2x4x8 DCT; Selection of sub-band transforms of varying structure or type
H04N 19/463 - Embedding additional information in the video signal during the compression process by compressing encoding parameters before transmission
H04N 19/136 - Incoming video signal characteristics or properties
H04N 19/137 - Motion inside a coding unit, e.g. average field, frame or block difference
H04N 19/70 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards
54.
METHOD AND APPARATUS FOR VIDEO ENCODING AND DECODING BASED ON ADAPTIVE COEFFICIENT GROUP
Various implementations for video encoding and decoding are presented involving determining, for a block being encoded or decoded in a picture, a coefficient group mode for coding at least one coefficient of a set of transform coefficients of the image block; encoding or decoding the set of transform coefficients of the image block responsive to the coefficient group mode. The coefficient group mode can be determined from at least one of image block size, number of non-zero transform coefficients inside the image block, position of a transform coefficient inside the image block, decoded syntax element, wherein the coefficient group mode can specify whether a coefficient group significance flag is coded/decoded, indicating that at least one coefficient is non-zero inside a coefficient group and/or specify at least one size of the coefficient group.
H04N 19/18 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being a set of transform coefficients
H04N 19/14 - Coding unit complexity, e.g. amount of activity or edge presence estimation
H04N 19/147 - Data rate or code amount at the encoder output according to rate distortion criteria
H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
H04N 19/70 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards
55.
METHOD AND SYSTEM FOR GAZE-BASED CONTROL OF MIXED REALITY CONTENT
Systems and methods are presented for discovering and positioning content into augmented reality space. A method includes forming a three-dimensional (3D) map of surroundings of a user of an augmented reality (AR) head mounted display (HMD); determining a depth-wise location of a gaze point of a user based on eye gaze direction and eye vergence; determining a visual guidance line pathway in the 3D map; guiding an action of the user along the visual guidance line pathway at one or more identified focal points; and rendering a mixed reality (MR) object along the visual guidance line pathway at a location corresponding to a direction of the user’s gaze.
Coding of the last coded coefficient position is performed by basing the coding of they coordinate of the position of the last coded coefficient on knowledge of the size of the partial transform used to generate a block of coefficients from a block of video pixels. This enables a context adaptive coding of the last coded coefficient parameter to be performed much more efficiently.
H04N 19/18 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being a set of transform coefficients
H04N 19/13 - Adaptive entropy coding, e.g. adaptive variable length coding [AVLC] or context adaptive binary arithmetic coding [CABAC]
H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
H04N 19/60 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding
H04N 19/91 - Entropy coding, e.g. variable length coding [VLC] or arithmetic coding
H04N 19/12 - Selection from among a plurality of transforms or standards, e.g. selection between discrete cosine transform [DCT] and sub-band transform or selection between H.263 and H.264
H04N 19/147 - Data rate or code amount at the encoder output according to rate distortion criteria
H04N 19/132 - Sampling, masking or truncation of coding units, e.g. adaptive resampling, frame skipping, frame interpolation or high-frequency transform coefficient masking
H04N 19/463 - Embedding additional information in the video signal during the compression process by compressing encoding parameters before transmission
57.
METHODS AND APPARATUS FOR ADAPTIVE MOTION VECTOR CANDIDATE ORDERING FOR VIDEO ENCODING AND DECODING
Methods and apparatus are provided for adaptive motion vector candidate ordering for video encoding and decoding. An apparatus includes a video encoder (100) for encoding a block in a picture by selecting an order of motion vector predictor candidates for the block responsive to a characteristic available at both the video encoder and a corresponding decoder. The characteristic excludes a mode in which the block is partitioned.
H04N 19/52 - Processing of motion vectors by encoding by predictive encoding
H04N 19/139 - Analysis of motion vectors, e.g. their magnitude, direction, variance or reliability
H04N 19/567 - Motion estimation based on rate distortion criteria
H04N 19/44 - Decoders specially adapted therefor, e.g. video decoders which are asymmetric with respect to the encoder
H04N 19/70 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards
58.
METHODS AND APPARATUSES FOR ENCODING, DECODING AND RENDERING 6DOF CONTENT FROM 3DOF+ COMPOSED ELEMENTS
A volumetric content is encoded as a set of clusters by an encoder and transmitted to a decoder which retrieves the volumetric content. Clusters common to different viewpoints are obtained and mutualized. Clusters are projected onto 2D images and encoded as independent video streams. Reduction in visual artefacts and reduction of data for storage and streaming are achieved.
Some embodiments of a method may include tracking a distance between a local VR device and a remote VR device; selecting an experience-sharing mode from a plurality of experience-sharing modes based on the tracked distance; and providing a different degree of immersive user experience to a user at the local VR device depending on the selected experience-sharing mode, wherein as the distance between the local VR device and the remote VR device decreases, the experience-sharing mode changes and the degree of immersive user experience increases. Some embodiments of another method may include detecting a gesture made by a user; selecting an immersiveness mode from a plurality of immersiveness modes based on the detected gesture; and providing a degree of immersive user experience to a local VR device user depending on the selected immersiveness mode.
A method for encoding a volumetric video content representative of a 3D scene is disclosed. The method comprises obtaining a reference viewing box and an intermediate viewing box defined within the 3D scene. For the reference viewing bounding box, the volumetric video reference subcontent is encoded as a central image and peripheral patches for parallax. For the intermediate viewing bounding box, the volumetric video intermediate sub-content is encoded as intermediate central patches which are differences between the intermediate central image and the reference central image.
The present principles relate to a method and device method for encoding depth values of orthogonally projected points of a point cloud onto a projection plane. The present principles also relate to a method and device for decoding a point cloud, a computer readable program and a video signal.
H04N 19/597 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding specially adapted for multi-view video sequence encoding
H04N 19/103 - Selection of coding mode or of prediction mode
H04N 19/147 - Data rate or code amount at the encoder output according to rate distortion criteria
H04N 19/17 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object
62.
MULTIFOCAL PLANE BASED METHOD TO PRODUCE STEREOSCOPIC VIEWPOINTS IN A DIBR SYSTEM (MFP-DIBR)
Some embodiments of an example method may include receiving an input image with depth information; mapping the input image to a set of focal plane images; orienting the set of focal plane images using head orientation information to provide stereo disparity between left and right eyes; and displaying the oriented set of focal plane images. Some embodiments of another example method may include: receiving a description of three-dimensional (3D) content; receiving, from a tracker, information indicating motion of a viewer relative to a real-world environment; responsive to receiving the information indicating motion of the viewer, synthesizing motion parallax by altering multi-focal planes of the 3D content; and rendering an image to the multi-focal plane display using the altered multi-focal plane rendering.
H04N 13/366 - Image reproducers using viewer tracking
H04N 13/395 - Volumetric displays, i.e. systems where the image is built up from picture elements distributed through a volume with depth sampling, i.e. the volume being constructed from a stack or sequence of 2D image planes
H04N 13/261 - Image signal generators with monoscopic-to-stereoscopic image conversion
G02B 30/52 - Optical systems or apparatus for producing three-dimensional [3D] effects, e.g. stereoscopic images the image being built up from image elements distributed over a 3D volume, e.g. voxels the 3D volume being constructed from a stack or sequence of 2D planes, e.g. depth sampling systems
G02B 27/00 - Optical systems or apparatus not provided for by any of the groups ,
Methods and apparatus are provided for collaborative partition coding for region based filters. An apparatus includes a video encoder (100) for encoding image data for a plurality of regions in a picture. The video encoder (100) includes multiple filters for filtering the image data based on region partition information for the plurality of regions. The region partition information for the plurality of regions is shared between the multiple filters.
H04N 19/00 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
H04N 19/117 - Filters, e.g. for pre-processing or post-processing
H04N 19/147 - Data rate or code amount at the encoder output according to rate distortion criteria
H04N 19/46 - Embedding additional information in the video signal during the compression process
H04N 19/196 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the adaptation method, adaptation tool or adaptation type used for the adaptive coding being specially adapted for the computation of encoding parameters, e.g. by averaging previously computed encoding parameters
H04N 19/174 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a slice, e.g. a line of blocks or a group of blocks
H04N 19/82 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals - Details of filtering operations specially adapted for video compression, e.g. for pixel interpolation involving filtering within a prediction loop
H04N 19/86 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using pre-processing or post-processing specially adapted for video compression involving reduction of coding artifacts, e.g. of blockiness
A method for decoding or encoding includes obtaining views parameters for a set of views comprising at least one reference view and a current view of a multi-views video content wherein each view comprises a texture layer and a depth layer. For at least one couple of a reference view and the current view of the set of views, an intermediate prediction image applying a forward projection method to pixels of the reference view is generated to project these pixels from a camera coordinates system of the reference view to a camera coordinates system of the current view, the prediction image comprising information allowing reconstructing image data. At least one final prediction image obtained from at least one intermediate prediction image is stored in a buffer of reconstructed images of the current view. A current image of the current view from the images stored in said buffer is reconstructed, said buffer comprising said at least one final prediction image.
H04N 19/597 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding specially adapted for multi-view video sequence encoding
H04N 19/105 - Selection of the reference unit for prediction within a chosen coding or prediction mode, e.g. adaptive choice of position and number of pixels used for prediction
H04N 19/30 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using hierarchical techniques, e.g. scalability
H04N 19/139 - Analysis of motion vectors, e.g. their magnitude, direction, variance or reliability
H04N 19/423 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals - characterised by implementation details or hardware specially adapted for video compression or decompression, e.g. dedicated software implementation characterised by memory arrangements
H04N 19/52 - Processing of motion vectors by encoding by predictive encoding
H04N 19/563 - Motion estimation with padding, i.e. with filling of non-object values in an arbitrarily shaped picture block or region for estimation purposes
H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
H04N 19/70 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards
H04N 19/159 - Prediction type, e.g. intra-frame, inter-frame or bidirectional frame prediction
65.
SYSTEM AND METHOD TO PREVENT SURVEILLANCE AND PRESERVE PRIVACY IN VIRTUAL REALITY
Preserving user privacy and preventing surveillance on behalf of users of a virtual reality world. One or more plans are available when a privacy or surveillance risk to a user is detected. In one plan, configurable scripts execute on behalf of the user to create a confusing array of clone avatars that obfuscate the real user avatar behavior. A malevolent avatar, attempting to surveil the user, may have difficulty distinguishing the clones from the user and may miss out on private insights he might otherwise have learned from the user's behavior. In another exemplary privacy plan, a copy of part of the virtual world is spawned, occupied exclusively by the user's avatar, and then merged into the main world. Privacy plans may be selected manually or automatically in response to perceived privacy threats to strike a balance between privacy and enjoyment within the virtual world.
G06F 21/62 - Protecting access to data via a platform, e.g. using keys or access control rules
G06F 3/04815 - Interaction with a metaphor-based environment or interaction object displayed as three-dimensional, e.g. changing the user viewpoint with respect to the environment or object
A method for encoding a block is disclosed. To this aim, a split mode is determined based on a rate-distortion optimization using a texture-based split prediction set obtained for the block. As an example, the split mode is determined by adapting the texture-based split prediction set according to at least one of a binary or triple split mode non-redundancy constraint or a heuristic-based split mode set pruning. The block is finally encoded using the determined split mode.
H04N 19/119 - Adaptive subdivision aspects e.g. subdivision of a picture into rectangular or non-rectangular coding blocks
H04N 19/102 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
H04N 19/147 - Data rate or code amount at the encoder output according to rate distortion criteria
H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
H04N 19/70 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards
An encoding method is presented. Transform coefficients of a block are first obtained. They are then scanned according to a scan pattern for entropy coding. The scan pattern is determined responsive to a shape of the block.
FIG. 5
H04N 19/159 - Prediction type, e.g. intra-frame, inter-frame or bidirectional frame prediction
H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
H04N 19/129 - Scanning of coding units, e.g. zig-zag scan of transform coefficients or flexible macroblock ordering [FMO]
H04N 19/13 - Adaptive entropy coding, e.g. adaptive variable length coding [AVLC] or context adaptive binary arithmetic coding [CABAC]
H04N 19/147 - Data rate or code amount at the encoder output according to rate distortion criteria
H04N 19/18 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being a set of transform coefficients
68.
METHOD AND DEVICE FOR PREDICTIVE PICTURE ENCODING AND DECODING
A method for encoding a block of a picture is disclosed. The method comprises for at least one sample of the block and for one current component: obtaining a prediction value; determining a mapped residual value from a source value of the sample and from the prediction value responsive to a mapping function; and encoding the mapped residual value into a bitstream; wherein the mapping function is derived to obtain at least one of a reduction of a bit cost of the bitstream for a given quality of reconstruction or an increase of quality of reconstruction for a given bit cost of the bitstream.
H04N 19/132 - Sampling, masking or truncation of coding units, e.g. adaptive resampling, frame skipping, frame interpolation or high-frequency transform coefficient masking
H04N 19/136 - Incoming video signal characteristics or properties
H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
H04N 19/186 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being a colour or a chrominance component
69.
Method and apparatus for processing a medium dynamic range video signal in SL-HDR2 format
Different implementations are described, particularly implementations for processing a medium dynamic range video signal are presented. In a method for processing such a video signal a medium dynamic range video signal and associated metadata are received, wherein the metadata include data representative of a peak luminance value of the medium dynamic range video signal. In addition, data representative of a peak luminance value of a presentation display are received. It is determined whether the peak luminance value of the medium dynamic range video signal is greater or lower than the peak luminance value of the presentation display. A processor is configured based on the determination, wherein the processor has a first mode to reconstruct a high dynamic range video signal based on a received standard dynamic range video signal and associated metadata, and a second mode to optimize a received high dynamic range video signal for the rendering device. The medium dynamic range video signal is processed by the processor in the first mode if the peak luminance value of the medium dynamic range video signal is smaller than the peak luminance value of the presentation display and in the second mode if the peak luminance value of the medium dynamic range video signal is greater than the peak luminance value of the presentation display.
H04N 19/70 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards
H04N 19/186 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being a colour or a chrominance component
Disclosed herein are systems and methods for augmented reality multi-view telepresence. An embodiment takes the form of a method that includes obtaining a session geometry of a multi-location telepresence session that includes a first-location participant at a first location and a second-location participant at a second location, each location having respective pluralities of cameras and display segments. The method includes selecting, according to the session geometry, both a first-to-second-viewpoint second-location camera from the plurality of second-location cameras as well as a first-to-second-viewpoint first-location display segment from the plurality of first-location display segments. The method includes receiving a first-to-second-viewpoint video stream captured by the selected first-to-second-viewpoint second-location camera, and further includes generating a line-of-sight augmented-reality experience for the multi-location telepresence session at least in part by rendering the received first-to-second-viewpoint video stream via the selected first-to-second-viewpoint first-location display segment.
Systems and methods are described for providing a display. In some embodiments, a display device includes a light-emitting layer with an addressable array of light-emitting elements such as OLEDs. A flexible optical layer overlays the light-emitting layer. The flexible optical layer has a plurality of lens regions, where optical powers of the lens regions change in response to changing levels of tensile or compressive force on the flexible optical layer. When no force is applied, the lens regions may have no optical power, and the display may operate as a 2D display. When force is applied (e.g. by bending the display), the lens regions may operate as cylindrical lenses in a lenticular array, and the display may be operated as a 3D multiview display.
G02B 30/29 - Optical systems or apparatus for producing three-dimensional [3D] effects, e.g. stereoscopic images by providing first and second parallax images to an observer’s left and right eyes of the autostereoscopic type involving lenticular arrays characterised by the geometry of the lenticular array, e.g. slanted arrays, irregular arrays or arrays of varying shape or size
72.
METHOD FOR ENCODING AND METHOD FOR DECODING A COLOR TRANSFORM AND CORRESPONDING DEVICES
A method includes encoding a first message including first parameters for a first color transform that transforms a video signal into a first color-mapped video signal; encoding a second message including second parameters for a second color transform that transforms the video signal into a second color-mapped video signal; and encoding the video signal, the first message, and the second message in a video bitstream. A method includes decoding the video bitstream into the video signal, the first message including the first parameters, and the second message including the second parameters, and performing at least one of the following: displaying the video signal; transforming the video signal into the first color-mapped video signal by applying the first color transform and displaying the first color-mapped video signal; and transforming the video signal into the second color-mapped video signal by applying the second color transform and displaying the second color-mapped video signal.
H04N 11/20 - Conversion of the manner in which the individual colour picture signal components are combined, e.g. conversion of colour television standards
H04N 19/70 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards
H04N 19/46 - Embedding additional information in the video signal during the compression process
H04N 19/186 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being a colour or a chrominance component
73.
SYSTEM AND METHOD FOR SPARSE DISTRIBUTED RENDERING
Systems and methods are described for rendering a 3D synthetic scene. A display client receives point cloud samples of the scene, where the point cloud samples include a point location, one or more viewing directions, and color information for each of the viewing directions. The point cloud samples may be generated by a server using ray tracing. The display client combines information from the point cloud samples with a locally generated rendering of at least a portion of the scene to generate a combined rendering of the scene, and the client displays the combined rendering. The number of point cloud samples may be adjusted adaptively based on performance metrics at the client.
A block of video data is split using one or more of several possible partition operations by using the partitioning choices obtained through use of a texture-based image partitioning. In at least one embodiment, the block is split in one or more splitting operations using a convolutional neural network. In another embodiment, inputs to the convolutional neural network come from pixels along the block's causal borders. In another embodiment, boundary information, such as the location of partitions in spatially neighboring blocks, is used by the texture analysis. Methods, apparatus, and signal embodiments are provided for encoding.
H04N 19/119 - Adaptive subdivision aspects e.g. subdivision of a picture into rectangular or non-rectangular coding blocks
H04N 19/14 - Coding unit complexity, e.g. amount of activity or edge presence estimation
H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
75.
METHOD AND SYSTEM FOR MAINTAINING COLOR CALIBRATION USING COMMON OBJECTS
Described herein are systems and methods for maintaining color calibration using common objects. In an exemplary embodiment, an AR system includes a forward-facing camera, an AR display, a processor, and a user interface. The processor is configured to receive image data from the forward-facing camera and identify any known objects depicted in the image data. The processor then determines RGB information at least one test rendering of the identified known object and displays it via the AR display. Input from a user interface, indicating which of the at least one test renderings was a closest match to the real-world object, and a level of satisfaction with the match are received by the processor and used to update an AR display color calibration model. More test renderings may be iteratively provided to improve the accuracy of the calibration.
Methods and apparatus are provided for determining quantization parameter predictors from a plurality of neighboring quantization parameters. An apparatus includes an encoder for encoding image data for at least a portion of a picture using a quantization parameter predictor for a current quantization parameter to be applied to the image data. The quantization parameter predictor is determined using multiple quantization parameters from previously coded neighboring portions. A difference between the current quantization parameter and the quantization parameter predictor is encoded for signaling to a corresponding decoder.
H04N 19/126 - Quantisation - Details of normalisation or weighting functions, e.g. normalisation matrices or variable uniform quantisers
H04N 19/196 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the adaptation method, adaptation tool or adaptation type used for the adaptive coding being specially adapted for the computation of encoding parameters, e.g. by averaging previously computed encoding parameters
H04N 19/61 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding in combination with predictive coding
H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
H04N 19/157 - Assigned coding mode, i.e. the coding mode being predefined or preselected to be further used for selection of another element or parameter
H04N 19/51 - Motion estimation or motion compensation
H04N 19/593 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving spatial prediction techniques
H04N 19/174 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a slice, e.g. a line of blocks or a group of blocks
When dependent scalar quantization is used, the choice of the quantizer depends on the decoding of the preceding transform coefficient, and the entropy decoding of a transform coefficient depends on quantizer choice. To maintain high throughput in hardware implementations for transform coefficient entropy coding, several decision schemes of the scaler quantizer are proposed. In one implementation, the state transition and the context model selection are based on only regular coded bins. For example, the state transition can be based on the sum of the SIG, gt1 and gt2 flags, the exclusive-or function of the SIG, gt1 and gt2 flags, or based on only the gt1 or gt2 flag. When a block of transform coefficients is coded, the regular mode bins can be coded first in one or more scan passes, and the remaining bypass coded bins are grouped together in another one or more scan passes.
H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
H04N 19/18 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being a set of transform coefficients
H04N 19/91 - Entropy coding, e.g. variable length coding [VLC] or arithmetic coding
H04N 19/42 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals - characterised by implementation details or hardware specially adapted for video compression or decompression, e.g. dedicated software implementation
H04N 19/46 - Embedding additional information in the video signal during the compression process
H04N 19/48 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using compressed domain processing techniques other than decoding, e.g. modification of transform coefficients, variable length coding [VLC] data or run-length data
H04N 19/70 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards
78.
Method and device for obtaining a second image from a first image when the dynamic range of the luminance of the first image is greater than the dynamic range of the luminance of the second image
The present disclosure relates to a method and device for obtaining a second image from a first image when the dynamic range of the luminance of the first image is greater than the dynamic range of the luminance of the second image. The disclosure describes deriving at least one component representative of the colors of the second image from the first image, and maximizing at least one derived component according to a maximum value depending on a linear-light luminance component of the first image.
H04N 19/85 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using pre-processing or post-processing specially adapted for video compression
H04N 19/186 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being a colour or a chrominance component
H04N 19/86 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using pre-processing or post-processing specially adapted for video compression involving reduction of coding artifacts, e.g. of blockiness
G06T 5/50 - Image enhancement or restoration by the use of more than one image, e.g. averaging, subtraction
79.
METHOD AND APPARATUS FOR SIGNALING DECODING DATA USING HIGH LEVEL SYNTAX ELEMENTS
Methods (800, 900, 1600, 1700) and apparatuses (2100) for signaling decoding data in a video bitstream, wherein one uses a syntax element indicating whether the decoding data are explicitly coded in the video bitstream or inferred from previous data of the video bitstream. A bitstream, a computer-readable storage medium and a computer program product are also described.
H04N 19/70 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards
H04N 19/46 - Embedding additional information in the video signal during the compression process
H04N 19/136 - Incoming video signal characteristics or properties
H04N 19/44 - Decoders specially adapted therefor, e.g. video decoders which are asymmetric with respect to the encoder
80.
TRANSFORM SELECTION FOR IMPLICIT MULTIPLE TRANSFORM SELECTION
A method and apparatus to improve compression efficiency in a video compression scheme enables use of new tools with multiple transform selection. In one embodiment, transform pair selection is based on a flag indicative of low-frequency non-separable transforms. In another embodiment, transform pair selection is based on a flag indicative of low-frequency non-separable transforms and on a flag indicative of matrix-based intra prediction. In another embodiment, when an implicit multiple transform selection mode is used, transform pair selection is based on a flag indicative of low-frequency non-separable transforms. Bitstream syntax is used to convey the flags.
H04N 19/70 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards
H04N 19/46 - Embedding additional information in the video signal during the compression process
H04N 19/103 - Selection of coding mode or of prediction mode
H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
81.
System and method for hybrid format spatial data distribution and rendering
Systems and methods are described for providing spatial content using a hybrid format. In some embodiments, a client device receives, from a server, surface light field representations of a plurality of scene elements in a 3D scene, including a first scene element. The client device provides to the server an indication of a dynamic behavior of a second scene element different from the first scene element. Further, in response to the indication, the client device receives from the server information defining the first scene element in a 3D asset format. The client device then renders at least the first scene element in the 3D asset format.
G06T 5/50 - Image enhancement or restoration by the use of more than one image, e.g. averaging, subtraction
H04L 67/12 - Protocols specially adapted for proprietary or special-purpose networking environments, e.g. medical networks, sensor networks, networks in vehicles or remote metering networks
82.
Method for improved intra prediction when reference samples are missing
The method for intra frame prediction allows for the recovery of some intra directional modes that are otherwise rendered useless due to the classical padding process. Directional modes are recovered by filling pixel locations in the prediction unit with samples from the portion of the unit that is determined to have samples available for use using a counterpart directional mode to the one that is useless. When it is determined that reference samples are missing for a particular prediction unit or portion thereof given a particular directional mode, a determination is made about whether reference samples along another portion of the prediction unit are available. The other portion is the side of the prediction unit at an opposite end of the particular directional mode. When it is determined that such reference samples are available, it is possible to extend the number of intra directional modes by replacing the particular directional modes that are rendered useless, because of the unavailability of reference samples, with their counterpart directional modes that are oppositely directed.
H04N 19/11 - Selection of coding mode or of prediction mode among a plurality of spatial predictive coding modes
H04N 19/132 - Sampling, masking or truncation of coding units, e.g. adaptive resampling, frame skipping, frame interpolation or high-frequency transform coefficient masking
H04N 19/159 - Prediction type, e.g. intra-frame, inter-frame or bidirectional frame prediction
H04N 19/105 - Selection of the reference unit for prediction within a chosen coding or prediction mode, e.g. adaptive choice of position and number of pixels used for prediction
H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
H04N 19/593 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving spatial prediction techniques
H04N 19/182 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being a pixel
83.
Method and system for the detection and augmentation of tactile interactions in augmented reality
Program actions may be initiated after detection of a predefined gesture by a user with a real-world object. Users may interact with their physical environment in an augmented reality by detecting interactions with real objects using a combination of location and motion detection, material identification using wearable sensors, or both. Based on detected sensor data from user interaction with a real-world object, a predefined gesture may be identified and a program action associated with that target interaction for the real-world object may be executed. In some cases, the user experience may be enhanced by providing haptic feedback in response to tactile gestures and resulting events.
The disclosure discloses methods and devices for transmitting and rendering a 3D scene. The method for transmitting comprises: segmenting a space into m angular sectors, the m angular sectors each corresponding to an angular distance from a viewport, and the space into n depth ranges; obtaining (11) at least one first patch, generated from a first view of the 3D scene, said at least one first patch comprising a texture component and a depth component; obtaining (12) at least one atlas, generated from at least one second view of the 3D scene, said at least one atlas being built by packing together at least one second patch generated for at least one point of one of said second view that is not visible in another view of the 3D scene and that belongs to a same angular sector among the m angular sectors, and a same depth range among the n depth ranges, at least one of morn being greater than or equal to 2, said at least one second patch comprising a texture component and a depth component, wherein each of the at least one first patch and the at least one second patch is based on at least one of a sector and a depth; generating (13) a first subset of streams comprising m′ pairs of streams and a second subset of streams comprising m′×n′ pairs of streams; and transmitting (14) the first and second subsets of streams to the terminal.
Some embodiments of a method may include: streaming a light field lenslet representation of light field video content; and changing resolution of the light field lenslet representation. Some embodiments of a method may include: selecting a lenslet representation from a plurality of lenslet representations of portions of light field content; retrieving a sub-sampled lenslet representation of the selected lenslet representation; and interpolating views from the sub-sampled lenslet representation.
H04N 13/383 - Image reproducers using viewer tracking for tracking with gaze detection, i.e. detecting the lines of sight of the viewer's eyes
H04N 13/307 - Image reproducers for viewing without the aid of special glasses, i.e. using autostereoscopic displays using fly-eye lenses, e.g. arrangements of circular lenses
H04N 13/117 - Transformation of image signals corresponding to virtual viewpoints, e.g. spatial image interpolation the virtual viewpoint locations being selected by the viewers or determined by viewer tracking
86.
METHOD AND APPARATUS FOR VIDEO ENCODING AND DECODING BASED ON NEURAL NETWORK IMPLEMENTATION OF CABAC
Methods and apparatuses for video coding and decoding are provided. The method of video encoding includes accessing a bin of a syntax element associated with a block in a picture of a video, determining a context for the bin of the syntax element associated with the syntax element and entropy encoding the bin of the syntax element based on the determined context wherein either the bin of the syntax element is based on the relevance of a prediction by a neural network of the syntax element or the probability associated to the context is determined by a neural network. A bitstream formatted to include encoded data, a computer-readable storage medium and a computer-readable program product are also described.
H04N 19/13 - Adaptive entropy coding, e.g. adaptive variable length coding [AVLC] or context adaptive binary arithmetic coding [CABAC]
H04N 19/105 - Selection of the reference unit for prediction within a chosen coding or prediction mode, e.g. adaptive choice of position and number of pixels used for prediction
H04N 19/132 - Sampling, masking or truncation of coding units, e.g. adaptive resampling, frame skipping, frame interpolation or high-frequency transform coefficient masking
H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
H04N 19/46 - Embedding additional information in the video signal during the compression process
H04N 19/70 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards
Generating an image from a source image can involve encoding a projection of a part of a three-dimensional scene. Pixels of a source image comprise a depth and a color attribute. Pixels of a source image are de-projected as colored point cloud. A de-projected point in a 3D space has the color attribute of the pixel that it has been de-projected from. Also, a score is attributed to the generated point according to a local depth gradient and/or a local color gradient of the pixel it comes from, the lower the gradient, the higher the score. The generated point cloud is captured by a virtual camera for rendering on a display device. The point cloud is projected onto the viewport image by blending color of points projected on a same pixel, the blending being weighted by the scores of these points.
Methods, devices and data stream are provided for signaling and decoding information representative of restrictions of navigation in a volumetric video. The data stream comprises metadata associated to video data representative of the volumetric video. The metadata comprise data representative of a viewing bounding box, data representative of a curvilinear path in the 3D space of said volumetric video; and data representative of at least one viewing direction range associated with a point on the curvilinear path.
Systems and methods are described for improved color rendering. In some embodiments, video is captured of a real-world scene, and a region of screen content is detected in the captured video. A processor selectively applies a screen-content color transformation on the region of screen content to generate processed video, and the processed video is displayed substantially in real time, for example on a video-see-through head-mounted display. The screen-content color processing may be different for different types of external displays. The screen-content color processing may also be determined based at least in part on illumination conditions. In some embodiments, the color of displayed virtual objects is adjusted based on visual parameters of the screen content.
G06T 19/00 - Manipulating 3D models or images for computer graphics
G09G 3/00 - Control arrangements or circuits, of interest only in connection with visual indicators other than cathode-ray tubes
G06Q 10/04 - Forecasting or optimisation specially adapted for administrative or management purposes, e.g. linear programming or "cutting stock problem"
G06F 17/18 - Complex mathematical operations for evaluating statistical data
G06F 18/2113 - Selection of the most significant subset of features by ranking or filtering the set of features, e.g. using a measure of variance or of feature cross-correlation
G06F 18/21 - Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
90.
Method and apparatus for adaptive illumination compensation in video encoding and decoding
Different implementations are described for determining one or more illumination compensation parameters for a current block being encoded by a video encoder or decoded by a video decoder, based on the selection of one or more neighboring samples. The selection of the one or more neighboring samples is based on information used to reconstruct a plurality of neighboring reconstructed blocks. The selection may be based on the motion information, such as motion vector and reference picture information. In one example, only samples from neighboring reconstructed blocks that have (1) the same reference picture index and/or (2) a motion vector close to the motion vector of the current block is selected. In another example, if the current block derives or inherits some motion information from a top or left neighboring block, then only the top or left neighboring samples are selected for IC parameter calculation.
H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
H04N 19/105 - Selection of the reference unit for prediction within a chosen coding or prediction mode, e.g. adaptive choice of position and number of pixels used for prediction
H04N 19/52 - Processing of motion vectors by encoding by predictive encoding
obtaining at least one color component of the color picture to be decoded at least from said second component and said two chrominance components by determining said at least one color component as a linear combination of the second component and the two chrominance components.
H04N 19/44 - Decoders specially adapted therefor, e.g. video decoders which are asymmetric with respect to the encoder
H04N 19/30 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using hierarchical techniques, e.g. scalability
H04N 19/136 - Incoming video signal characteristics or properties
H04N 19/186 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being a colour or a chrominance component
H04N 19/126 - Quantisation - Details of normalisation or weighting functions, e.g. normalisation matrices or variable uniform quantisers
A method and apparatus for conveying information in a bitstream for deep neural network compression, such as in matrices representing weights, biases and non-linearities, to iteratively compress a pre-trained deep neural network by low displacement rank based approximation of the network layer weight matrices. The low displacement rank approximation allows for replacement of an original layer weight matrices of the pre-trained deep neural network as the sum of small number of structured matrices, allowing compression and low inference complexity. A decoder stage parses a bitstream for inference.
H04N 19/70 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards
A method and apparatus for performing a forward mapping function or its inverse is implemented through information conveyed in a bitstream to perform mapping. The information can be implemented in a uniform or non-uniform lookup table. A decoder can receive the information and derive the inverse function to be used. In one embodiment, a decoder receives information representative of an inverse mapping function and uses the information to implement an inverse mapping lookup table for intra coded picture portions and can derive a forward mapping lookup table to be used with an inverse mapping lookup table for inter coded picture portions.
H04N 19/117 - Filters, e.g. for pre-processing or post-processing
H04N 19/82 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals - Details of filtering operations specially adapted for video compression, e.g. for pixel interpolation involving filtering within a prediction loop
H04N 19/159 - Prediction type, e.g. intra-frame, inter-frame or bidirectional frame prediction
H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
G06F 1/03 - Digital function generators working, at least partly, by table look-up
94.
LOW DISPLACEMENT RANK BASED DEEP NEURAL NETWORK COMPRESSION
A method and an apparatus for performing deep neural network compression use an approximation training set along with information, such as in matrices representing weights, biases and non-linearities, to iteratively compress a pre-trained deep neural network by low displacement rank based approximation of the network layer weight matrices. The low displacement rank approximation allows for replacement of an original layer weight matrices of the pre-trained deep neural network as the sum of a small number of structured matrices, allowing compression and low inference complexity.
At least one embodiment relates to signaling at least one texture patch representing a texture value of at least one in-between 3D sample, a texture patch being a set of 2D samples representing texture values of orthogonally projected 3D samples of a point cloud along projection lines onto a projection plane, and said at least one in-between 3D sample being a 3D sample of the point cloud having a depth value greater than a nearer 3D sample of the point cloud and lower than a farther 3D sample of the point cloud, said at least one in-between 3D sample and said nearer and farther 3D samples being projected along the same projection line
H04N 19/20 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using video object coding
H04N 19/70 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards
H04N 19/184 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being bits, e.g. of the compressed video stream
96.
SYSTEM AND METHOD FOR PRIORITIZING AR INFORMATION BASED ON PERSISTENCE OF REAL-LIFE OBJECTS IN THE USER'S VIEW
Systems and methods described herein in accordance with some embodiments may display augmented reality (AR) information related to an object based on a predicted persistence of the object within view of the user. In some embodiments, display of AR information is prioritized based upon a prediction of how soon an object will disappear from view, perhaps if the predicted disappearance will occur sooner than some threshold time. Systems and methods described herein in accordance with some embodiments extend to mixed reality (MR) systems, and provide for displaying an image of an object that has become obstructed or has disappeared from view and/or for displaying AR information related to an object that has become obstructed or has disappeared from view.
G09G 5/37 - Control arrangements or circuits for visual indicators common to cathode-ray tube indicators and other visual indicators characterised by the display of individual graphic patterns using a bit-mapped memory - Details of the operation on graphic patterns
A decoding method is disclosed. A flag is first decoded. The flag indicates whether a current block of a picture is decoded with a tool of a set of tools. The set of tools comprises tools using reconstructed samples of neighboring blocks decoded prior to the current block or using a block of the picture larger than a hardware unit. The flag is only decoded in the case where the current block has a parent hardware unit. A current block is then decoded responsive to the flag.
H04N 19/503 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
H04N 19/105 - Selection of the reference unit for prediction within a chosen coding or prediction mode, e.g. adaptive choice of position and number of pixels used for prediction
H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
H04N 19/189 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the adaptation method, adaptation tool or adaptation type used for the adaptive coding
98.
SYSTEM AND METHOD FOR MULTIPLEXED RENDERING OF LIGHT FIELDS
An example method in accordance with some embodiments may include: receiving a media manifest file identifying a plurality of representations of a multi-view video, at least a first representation of the plurality of representations comprising a first sub-sampling of views and at least a second representation of the plurality of representations comprising a second sub-sampling of views different from the first sub-sampling of views; selecting a selected representation from the plurality of representations; retrieving the selected representation; and rendering the selected representation.
H04N 21/2343 - Processing of video elementary streams, e.g. splicing of video streams or manipulating MPEG-4 scene graphs involving reformatting operations of video signals for distribution or compliance with end-user requests or end-user device requirements
H04N 19/597 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding specially adapted for multi-view video sequence encoding
H04N 13/282 - Image signal generators for generating image signals corresponding to three or more geometrical viewpoints, e.g. multi-view systems
H04N 13/383 - Image reproducers using viewer tracking for tracking with gaze detection, i.e. detecting the lines of sight of the viewer's eyes
99.
Methods and apparatus for video usability information (VUI) for scalable video coding (SVC)
There are provided methods and apparatus for video usability information (VUI) for scalable video coding (SVC). An apparatus includes an encoder (100) for encoding video signal data into a bitstream. The encoder specifies video user information, excluding hypothetical reference decoder parameters, in the bitstream using a high level syntax element. The video user information corresponds to a set of interoperability points in the bitstream relating to scalable video coding (340, 355).
H04N 7/12 - Systems in which the television signal is transmitted via one channel or a plurality of parallel channels, the bandwidth of each channel being less than the bandwidth of the television signal
H04N 19/30 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using hierarchical techniques, e.g. scalability
H04N 19/70 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards
100.
METHOD AND DEVICE FOR RECONSTRUCTING IMAGE DATA FROM DECODED IMAGE DATA
A method or device for reconstructing image data representative of original image data from decoded image data and parameters obtained from a bitstream is described. The parameters having been processed from the original image data. The method includes checking if parameters are lost, corrupted or not aligned with the decoded image data whose graphics or overlay is added to, when at least one of the parameters is lost, corrupted or not aligned with the decoded image data whose graphics or overlay is added to; selecting a recovery mode according to an information data indicating how the parameters have been processed; and recovering the at least one lost, corrupted or not aligned parameter by applying the selected recovery mode, the reconstruction of image data then taking also into account the recovered parameters.