Vid Scale, Inc.

United States of America

Back to Profile

1-100 of 216 for Vid Scale, Inc. Sort by
Query
Patent
United States - USPTO
Aggregations Reset Report
Date
New (last 4 weeks) 2
2024 April (MTD) 1
2024 March 2
2024 January 1
2023 December 3
See more
IPC Class
H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock 77
H04N 19/105 - Selection of the reference unit for prediction within a chosen coding or prediction mode, e.g. adaptive choice of position and number of pixels used for prediction 58
H04N 19/70 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards 39
H04N 19/597 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding specially adapted for multi-view video sequence encoding 37
H04N 19/186 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being a colour or a chrominance component 33
See more
Status
Pending 70
Registered / In Force 146
Found results for  patents
  1     2     3        Next Page

1.

METRICS AND MESSAGES TO IMPROVE EXPERIENCE FOR 360-DEGREE ADAPTIVE STREAMING

      
Application Number 18543660
Status Pending
Filing Date 2023-12-18
First Publication Date 2024-04-18
Owner VID SCALE, INC. (USA)
Inventor
  • He, Yong
  • Ye, Yan
  • Begen, Ali C.

Abstract

A method for receiving and displaying media content may be provided. The method may include requesting a set of DASH video segments that are associated with various viewports and qualities. The method may include displaying the DASH video segments. The method may include determining a latency metric based on a time difference between the display of a DASH video segment and one of: a device beginning to move, the device ceasing to move, the device determining that the device has begun to move, the device determining that the device has stopped moving, or the display of a different DASH video segment. The different DASH video segment may be associated with one or more of a different quality or a different viewport.

IPC Classes  ?

  • H04N 21/442 - Monitoring of processes or resources, e.g. detecting the failure of a recording device, monitoring the downstream bandwidth, the number of times a movie has been viewed or the storage space available from the internal hard disk
  • H04N 21/2343 - Processing of video elementary streams, e.g. splicing of video streams or manipulating MPEG-4 scene graphs involving reformatting operations of video signals for distribution or compliance with end-user requests or end-user device requirements
  • H04N 21/4728 - End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification or for manipulating displayed content for selecting a ROI [Region Of Interest], e.g. for requesting a higher resolution version of a selected region
  • H04N 21/81 - Monomedia components thereof

2.

AFFINE MOTION MODEL DERIVATION METHOD

      
Application Number 18521810
Status Pending
Filing Date 2023-11-28
First Publication Date 2024-03-28
Owner VID SCALE, INC. (USA)
Inventor
  • Luo, Jiancong
  • He, Yuwen
  • Chen, Wei

Abstract

Systems and methods are described for video coding using affine motion prediction. In an example method, motion vector gradients are determined from respective motion vectors of a plurality of neighboring sub-blocks neighboring a current block. An estimate of at least one affine parameter for the current block is determined based on the motion vector gradients. An affine motion model is determined based at least in part on the estimated affine parameter(s), and a prediction of the current block is generated using the affine motion model. The estimated parameter(s) may be used in the affine motion model itself. Alternatively, the estimated parameter(s) may be used in a prediction of the affine motion model. In some embodiments, only neighboring sub-blocks above and/or to the left of the current block are used in estimating the affine parameter(s).

IPC Classes  ?

  • H04N 19/137 - Motion inside a coding unit, e.g. average field, frame or block difference
  • H04N 19/105 - Selection of the reference unit for prediction within a chosen coding or prediction mode, e.g. adaptive choice of position and number of pixels used for prediction
  • H04N 19/149 - Data rate or code amount at the encoder output by estimating the code amount by means of a model, e.g. mathematical model or statistical model
  • H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock

3.

HISTORY-BASED MOTION VECTOR PREDICTION

      
Application Number 18510196
Status Pending
Filing Date 2023-11-15
First Publication Date 2024-03-14
Owner VID SCALE, INC. (USA)
Inventor
  • Xiu, Xiaoyu
  • He, Yuwen
  • Luo, Jiancong

Abstract

Systems, methods, and instrumentalities are disclosed for processing history-based motion vector prediction (HMVP). A video coding device may generate a history-based motion vector prediction (HMVP) list for a current block. The video coding device derive an HMVP candidate from a previously coded block. The HMVP candidate may include motion information associated with a neighboring block of the current block, one or more reference indices, and a bi-prediction weight index. The video coding device may add the HMVP candidate to the HMVP list for motion compensated prediction of a motion vector associated with the current block. The video coding device use one HMVP selected from the HMVP list to perform motion compensated prediction of the current block. The motion compensated prediction may be performed using the motion information associated with the neighboring block of the current block, the one or more reference indices, and the bi-prediction weight index.

IPC Classes  ?

  • H04N 19/105 - Selection of the reference unit for prediction within a chosen coding or prediction mode, e.g. adaptive choice of position and number of pixels used for prediction
  • H04N 19/159 - Prediction type, e.g. intra-frame, inter-frame or bidirectional frame prediction
  • H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
  • H04N 19/51 - Motion estimation or motion compensation

4.

SYSTEMS AND METHODS FOR SPATIAL PREDICTION

      
Application Number 18370589
Status Pending
Filing Date 2023-09-20
First Publication Date 2024-01-11
Owner VID SCALE, INC. (USA)
Inventor
  • Ye, Yan
  • Chen, Qian
  • Dong, Jie

Abstract

Systems, methods, and instrumentalities are disclosed relating to intra prediction of a video signal based on mode-dependent subsampling. A block of coefficients associated with a first sub block of a video block, one or more blocks of coefficients associated with one or more remaining sub blocks of the video block, and an indication of a prediction mode for the video block may be received. One or more interpolating techniques, a predicted first sub block, and the predicted sub blocks of the one or more remaining sub blocks may be determined. A reconstructed first sub block and one or more reconstructed remaining sub blocks may be generated. A reconstructed video block may be formed based on the prediction mode, the reconstructed first sub block, and the one or more reconstructed remaining sub blocks.

IPC Classes  ?

  • H04N 19/52 - Processing of motion vectors by encoding by predictive encoding
  • H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
  • H04N 19/119 - Adaptive subdivision aspects e.g. subdivision of a picture into rectangular or non-rectangular coding blocks
  • H04N 19/46 - Embedding additional information in the video signal during the compression process
  • H04N 19/593 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving spatial prediction techniques
  • H04N 19/11 - Selection of coding mode or of prediction mode among a plurality of spatial predictive coding modes
  • H04N 19/124 - Quantisation
  • H04N 19/14 - Coding unit complexity, e.g. amount of activity or edge presence estimation
  • H04N 19/80 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals - Details of filtering operations specially adapted for video compression, e.g. for pixel interpolation
  • H04N 19/90 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using coding techniques not provided for in groups , e.g. fractals
  • H04N 19/60 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding
  • H04N 19/59 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving spatial sub-sampling or interpolation, e.g. alteration of picture size or resolution
  • H04N 19/182 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being a pixel

5.

360-DEGREE VIDEO DELIVERY OVER NEXT GENERATION NETWORK

      
Application Number 18237772
Status Pending
Filing Date 2023-08-24
First Publication Date 2023-12-28
Owner VID SCALE, INC. (USA)
Inventor
  • He, Yong
  • Ye, Yan

Abstract

Systems, methods, and instrumentalities are disclosed for a 360-degree video streaming. A video streaming device may receive a 360-degree video stream from a network node. The video streaming device may determine a viewport associated with the video streaming device and/or the 360-degree video stream. The video streaming device may determine (e.g., based on the viewport) to request in advance a first segment and a second segment of the 360-degree video stream. The video streaming device may determine a relative priority order for the first segment and the second segment. The video streaming device may generate an anticipated requests message. The anticipated requests message may indicate the determined relative priority order, for example, by listing the first segment and the second segment in decreasing relative priority based on the determined relative priority order. The video streaming device may send the anticipated requests message to the network node.

IPC Classes  ?

  • H04N 21/81 - Monomedia components thereof
  • H04L 65/60 - Network streaming of media packets
  • H04L 65/80 - Responding to QoS
  • H04L 67/02 - Protocols based on web technology, e.g. hypertext transfer protocol [HTTP]
  • H04N 21/218 - Source of audio or video content, e.g. local disk arrays
  • H04N 21/44 - Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to MPEG-4 scene graphs
  • H04N 21/472 - End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification or for manipulating displayed content
  • H04N 21/6587 - Control parameters, e.g. trick play commands or viewpoint selection
  • H04N 21/845 - Structuring of content, e.g. decomposing content into time segments

6.

Geometry Conversion for 360-degree Video Coding

      
Application Number 18240174
Status Pending
Filing Date 2023-08-30
First Publication Date 2023-12-21
Owner VID SCALE, INC. (USA)
Inventor
  • He, Yuwen
  • Ye, Yan
  • Hamza, Ahmed

Abstract

Processing a 360-degree video content for video coding may include receiving the video content in a first geometry. The video content may include unaligned chroma and luma components associated with a first chroma sampling scheme. The unaligned chroma and luma components may be aligned to a sampling grid associated with a second chroma sampling scheme that has aligned chroma and luma components. A geometric conversion to the video content may be performed. The video content, that may comprise the aligned chroma and luma components, in the first geometry may be converted to a second geometry. The first geometry may be a stitched geometry, and the second geometry may be a coding geometry. The converted video content in the second geometry may include the chroma and luma components aligned to the sampling grid associated with the second chroma sampling scheme.

IPC Classes  ?

  • H04N 19/597 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding specially adapted for multi-view video sequence encoding
  • H04N 19/82 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals - Details of filtering operations specially adapted for video compression, e.g. for pixel interpolation involving filtering within a prediction loop
  • H04N 19/59 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving spatial sub-sampling or interpolation, e.g. alteration of picture size or resolution
  • H04N 19/186 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being a colour or a chrominance component

7.

LEARNED VIDEO COMPRESSION FRAMEWORK FOR MULTIPLE MACHINE TASKS

      
Application Number 18033693
Status Pending
Filing Date 2021-11-03
First Publication Date 2023-12-07
Owner VID SCALE, INC. (USA)
Inventor
  • Racape, Fabien
  • Hewa Gamage, Lahiru Dulanjana
  • Pushparaja, Akshay
  • Begaint, Jean

Abstract

Processing of a compressed representation of a video signal is optimized for multiple tasks, such as object detection, viewing of displayed video, or other machine tasks. In one embodiment, multiple analysis stages and a single synthesis is performed as part of a coding/decoding operation with training of an encoder side analysis and, optionally, a corresponding machine task. In another embodiment, multiple synthesis operations are performed on the decoding side, so that respective analysis, synthesis, and task stages are optimized. Other embodiments comprise feeding decoded feature maps to tasks, predictive coding, and using hyperprior-based models.

IPC Classes  ?

  • H04N 19/597 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding specially adapted for multi-view video sequence encoding
  • H04N 19/105 - Selection of the reference unit for prediction within a chosen coding or prediction mode, e.g. adaptive choice of position and number of pixels used for prediction
  • H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
  • H04N 19/136 - Incoming video signal characteristics or properties
  • H04N 19/42 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals - characterised by implementation details or hardware specially adapted for video compression or decompression, e.g. dedicated software implementation

8.

METHODS FOR SIMPLIFYING ADAPTIVE LOOP FILTER IN VIDEO CODING

      
Application Number 18140806
Status Pending
Filing Date 2023-04-28
First Publication Date 2023-11-30
Owner VID SCALE, INC. (USA)
Inventor
  • Vanam, Rahul
  • He, Yuwen
  • Ye, Yan

Abstract

Systems, methods and instrumentalities are disclosed for adaptively selecting an adaptive loop filter (ALF) procedure for a frame based on which temporal layer the frame is in. ALF procedures may vary in computational complexity. One or more frames including the current frame may be in a temporal layer of a coding scheme. The decoder may determine the current frame's temporal layer level within the coding scheme. The decoder may select an ALF procedure based on the current frame's temporal layer level. If the current frame's temporal layer level is higher within the coding scheme than some other temporal layer levels, an ALF procedure that is less computationally complex may be selected for the current frame. Then the decoder may perform the selected ALF procedure on the current frame.

IPC Classes  ?

  • H04N 19/82 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals - Details of filtering operations specially adapted for video compression, e.g. for pixel interpolation involving filtering within a prediction loop
  • H04N 19/105 - Selection of the reference unit for prediction within a chosen coding or prediction mode, e.g. adaptive choice of position and number of pixels used for prediction
  • H04N 19/117 - Filters, e.g. for pre-processing or post-processing
  • H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
  • G06F 18/211 - Selection of the most significant subset of features
  • G06F 18/241 - Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
  • G06F 18/2431 - Multiple classes

9.

LEARNED VIDEO COMPRESSION AND CONNECTORS FOR MULTIPLE MACHINE TASKS

      
Application Number 18033697
Status Pending
Filing Date 2021-11-03
First Publication Date 2023-11-16
Owner VID SCALE, INC. (USA)
Inventor
  • Racape, Fabien
  • Hewa Gamage, Lahiru Dulanjana
  • Begaint, Jean
  • Feltman, Simon

Abstract

A processing module, or connector, adapts an output of a codec, or a decoded output, to a form suitable for an alternate task. In one embodiment, the output of a codec is used for a machine task and the connector adapts this output to a form suitable for a video display. In another embodiment, metadata accompanies the codec output, which can instruct the connector how to adapt the codec output for an alternate task. In other embodiments, the processing module performs averaging over a N×M window, or convolution.

IPC Classes  ?

  • H04N 19/186 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being a colour or a chrominance component
  • H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
  • H04N 19/60 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding

10.

SYSTEMS, APPARATUS AND METHODS FOR INTER PREDICTION REFINEMENT WITH OPTICAL FLOW

      
Application Number 18199238
Status Pending
Filing Date 2023-05-18
First Publication Date 2023-09-14
Owner VID SCALE, INC. (USA)
Inventor
  • Luo, Jiancong
  • He, Yuwen

Abstract

Method, apparatus and systems are disclosed. In one embodiment, a method of decoding includes obtaining a sub-block based motion prediction signal for a current block of the video; obtaining one or more spatial gradients of the sub-block based motion prediction signal or one or more motion vector difference values; obtaining a refinement signal for the current block based on the one or more obtained spatial gradients or the one or more obtained motion vector difference values; obtaining a refined motion prediction signal for the current block based on the sub-block based motion prediction signal and the refinement signal; and decoding the current block based on the refined motion prediction signal.

IPC Classes  ?

  • H04N 19/52 - Processing of motion vectors by encoding by predictive encoding
  • H04N 19/105 - Selection of the reference unit for prediction within a chosen coding or prediction mode, e.g. adaptive choice of position and number of pixels used for prediction
  • H04N 19/132 - Sampling, masking or truncation of coding units, e.g. adaptive resampling, frame skipping, frame interpolation or high-frequency transform coefficient masking
  • H04N 19/139 - Analysis of motion vectors, e.g. their magnitude, direction, variance or reliability
  • H04N 19/167 - Position within a video image, e.g. region of interest [ROI]
  • H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
  • H04N 19/55 - Motion estimation with spatial constraints, e.g. at image or region borders

11.

TILE TRACKS FOR GEOMETRY-BASED POINT CLOUD DATA

      
Application Number 18017989
Status Pending
Filing Date 2021-08-06
First Publication Date 2023-09-07
Owner VID SCALE, INC. (USA)
Inventor
  • Gudumasu, Srinivas
  • Hamza, Ahmed

Abstract

A method and apparatus include receiving a timed-metadata track identifying point cloud tiles corresponding to one or more spatial regions within a point cloud scene. A decoding device determines one or more point cloud tiles to be used for rendering an image. One or more geometry tile tracks are retrieved, via a communications network, corresponding to the determined one or more point cloud tiles. Each geometry tile track comprises point cloud geometry data for a respective tile. The retrieved geometry tile tracks are processed

IPC Classes  ?

  • G06T 17/00 - 3D modelling for computer graphics
  • G06V 10/25 - Determination of region of interest [ROI] or a volume of interest [VOI]
  • H04N 19/597 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding specially adapted for multi-view video sequence encoding

12.

METHODS AND APPARATUS TO REDUCE LATENCY FOR 360-DEGREE VIEWPORT ADAPTIVE STREAMING

      
Application Number 18197323
Status Pending
Filing Date 2023-05-15
First Publication Date 2023-09-07
Owner Vid Scale, Inc. (USA)
Inventor
  • He, Yong
  • Ye, Yan
  • Gudumasu, Srinivas
  • Asbun, Eduardo
  • Hamza, Ahmed
  • Begen, Ali C.

Abstract

A client device adaptively streams a 360-degree video. A first segment is displayed based on a first viewing direction at a first time, where the first viewing direction is associated with a first viewport. The client requests a first base buffer segment based on the first viewport. The first base buffer segment has a presentation time after the first segment. At a second time, the viewing direction changes to a second viewing direction associated with a second viewport. The client requests, prior to the presentation time, a first viewport buffer segment based on the second viewport, with the same presentation time. The client device displays a second segment at the presentation time, wherein the second segment is either the first viewport buffer segment or the first base buffer segment. The client provides reports on viewport switching latency and on the most-requested segments.

IPC Classes  ?

  • H04L 65/80 - Responding to QoS
  • H04N 21/6587 - Control parameters, e.g. trick play commands or viewpoint selection
  • H04N 21/81 - Monomedia components thereof
  • H04N 21/2343 - Processing of video elementary streams, e.g. splicing of video streams or manipulating MPEG-4 scene graphs involving reformatting operations of video signals for distribution or compliance with end-user requests or end-user device requirements
  • H04N 21/442 - Monitoring of processes or resources, e.g. detecting the failure of a recording device, monitoring the downstream bandwidth, the number of times a movie has been viewed or the storage space available from the internal hard disk
  • H04N 21/647 - Control signaling between network components and server or clients; Network processes for video distribution between server and clients, e.g. controlling the quality of the video stream, by dropping packets, protecting content from unauthorised alteration within the network, monitoring of network load or bridging bet
  • H04N 21/845 - Structuring of content, e.g. decomposing content into time segments
  • H04N 21/218 - Source of audio or video content, e.g. local disk arrays
  • H04L 65/65 - Network streaming protocols, e.g. real-time transport protocol [RTP] or real-time control protocol [RTCP]
  • H04L 65/70 - Media network packetisation
  • H04L 65/75 - Media network packet handling
  • H04L 65/612 - Network streaming of media packets for supporting one-way streaming services, e.g. Internet radio for unicast
  • H04N 13/111 - Transformation of image signals corresponding to virtual viewpoints, e.g. spatial image interpolation

13.

SYSTEMS AND METHODS FOR GENERALIZED HTTP HEADERS IN DYNAMIC ADAPTIVE STREAMING OVER HTTP (DASH)

      
Application Number 18111363
Status Pending
Filing Date 2023-02-17
First Publication Date 2023-06-29
Owner Vid Scale, Inc. (USA)
Inventor Giladi, Alexander

Abstract

Systems and methods for streaming content are disclosed. A Media Presentation Description (MPD) may be associated with streaming content, for example, in Dynamic Adaptive Streaming Over HTTP (DASH). An MPD array comprise an element, e.g., a header element or URL, query element, which provides a name value functionality. An element may allow a request that results in insertion of custom headers. Requests may be used for retrieval of a subsegment, segment, or MPD, re-referencing of remote elements, or triggered by an event. The event may be embedded either in an MPD or in segments.

IPC Classes  ?

  • H04N 21/235 - Processing of additional data, e.g. scrambling of additional data or processing content descriptors
  • H04L 65/80 - Responding to QoS
  • H04L 65/65 - Network streaming protocols, e.g. real-time transport protocol [RTP] or real-time control protocol [RTCP]
  • H04L 65/70 - Media network packetisation
  • H04L 65/612 - Network streaming of media packets for supporting one-way streaming services, e.g. Internet radio for unicast
  • H04L 65/613 - Network streaming of media packets for supporting one-way streaming services, e.g. Internet radio for the control of the source by the destination
  • H04L 65/75 - Media network packet handling
  • H04L 67/1001 - Protocols in which an application is distributed across nodes in the network for accessing one among a plurality of replicated servers
  • H04N 21/2381 - Adapting the multiplex stream to a specific network, e.g. an IP [Internet Protocol] network

14.

HYBRID CUBEMAP PROJECTION FOR 360-DEGREE VIDEO CODING

      
Application Number 18112234
Status Pending
Filing Date 2023-02-21
First Publication Date 2023-06-22
Owner VID SCALE, INC. (USA)
Inventor
  • Xiu, Xiaoyu
  • He, Yuwen
  • Ye, Yan

Abstract

A system, method, and/or instrumentality may be provided for coding a 360-degree video. A picture of the 360-degree video may be received. The picture may include one or more faces associated with one or more projection formats. A first projection format indication may be received that indicates a first projection format may be associated with a first face. A second projection format indication may be received that indicates a second projection format may be associated with a second face. Based on the first projection format, a first transform function associated with the first face may be determined. Based on the second projection format, a second transform function associated with the second face may be determined. At least one decoding process may be performed on the first face using the first transform function and/or at least one decoding process may be performed on the second face using the second transform function.

IPC Classes  ?

  • H04N 19/597 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding specially adapted for multi-view video sequence encoding
  • H04N 19/12 - Selection from among a plurality of transforms or standards, e.g. selection between discrete cosine transform [DCT] and sub-band transform or selection between H.263 and H.264
  • H04N 19/61 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding in combination with predictive coding
  • G06T 3/00 - Geometric image transformation in the plane of the image

15.

SECONDARY CONTENT INSERTION IN 360-DEGREE VIDEO

      
Application Number 18106128
Status Pending
Filing Date 2023-02-06
First Publication Date 2023-06-22
Owner VID SCALE, INC. (USA)
Inventor
  • Asbun, Eduardo
  • He, Yong
  • Gudumasu, Srinivas
  • Hamza, Ahmed
  • Ye, Yan

Abstract

A secondary content such as an advertisement may be inserted based on users’ interests in 360 degree video streaming. Users may have different interests and may watch different areas within a 360 degree video. The information about area(s) of 360 degree scenes that users watch the most may be used to select an ad(s) relevant to their interests. One or more secondary content viewports may be defined within a 360 degree video frame. Secondary content viewport parameter(s) may be tracked. For example, statistics of the user’s head orientation for some time leading to the presentation of the ad(s) may be collected. Secondary content may be determined based on the tracked secondary content viewport parameter(s).

IPC Classes  ?

  • H04N 21/2668 - Creating a channel for a dedicated end-user group, e.g. by inserting targeted commercials into a video stream based on end-user profiles
  • H04N 21/414 - Specialised client platforms, e.g. receiver in car or embedded in a mobile appliance
  • H04N 21/81 - Monomedia components thereof
  • H04N 21/258 - Client or end-user data management, e.g. managing client capabilities, user preferences or demographics or processing of multiple end-users preferences to derive collaborative data
  • G06T 19/00 - Manipulating 3D models or images for computer graphics
  • H04N 21/845 - Structuring of content, e.g. decomposing content into time segments
  • H04N 21/482 - End-user interface for program selection
  • H04N 21/431 - Generation of visual interfaces; Content or additional data rendering
  • H04N 21/435 - Processing of additional data, e.g. decrypting of additional data or reconstructing software from modules extracted from the transport stream
  • H04N 21/442 - Monitoring of processes or resources, e.g. detecting the failure of a recording device, monitoring the downstream bandwidth, the number of times a movie has been viewed or the storage space available from the internal hard disk
  • H04N 21/6587 - Control parameters, e.g. trick play commands or viewpoint selection

16.

Sample Derivation For 360-degree Video Coding

      
Application Number 18105290
Status Pending
Filing Date 2023-02-03
First Publication Date 2023-06-15
Owner Vid Scale, Inc. (USA)
Inventor
  • Hanhart, Philippe
  • He, Yuwen
  • Ye, Yan

Abstract

Systems, methods, and instrumentalities are disclosed for performing horizontal geometry padding on a current sample based on receiving a wraparound enabled indication that indicates whether a horizontal wraparound motion compensation is enabled. If the horizontal wraparound motion compensation is enabled based on the wraparound enabled indication, a video coding device may determine a reference sample wraparound offset of a current sample in a picture. The reference sample wraparound offset may indicate a face width of the picture. The video coding device may determine a reference sample location for the current sample based on the reference sample wraparound offset, a picture width of the picture, and a current sample location. The video coding device may predict the current sample based on the reference sample location in a horizontal direction. Repetitive padding or clipping may be used in the vertical direction.

IPC Classes  ?

  • H04N 19/597 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding specially adapted for multi-view video sequence encoding
  • H04N 19/105 - Selection of the reference unit for prediction within a chosen coding or prediction mode, e.g. adaptive choice of position and number of pixels used for prediction
  • H04N 19/132 - Sampling, masking or truncation of coding units, e.g. adaptive resampling, frame skipping, frame interpolation or high-frequency transform coefficient masking
  • H04N 19/172 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a picture, frame or field
  • H04N 19/184 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being bits, e.g. of the compressed video stream
  • H04N 19/51 - Motion estimation or motion compensation

17.

BI-DIRECTIONAL OPTICAL FLOW METHOD WITH SIMPLIFIED GRADIENT DERIVATION

      
Application Number 18106181
Status Pending
Filing Date 2023-02-06
First Publication Date 2023-06-15
Owner VID SCALE, INC. (USA)
Inventor
  • He, Yuwen
  • Xiu, Xiaoyu
  • Ye, Yan

Abstract

A video coding device may be configured to perform directional Bi-directional optical flow (BDOF) refinement on a coding unit (CU). The device may determine the direction in which to perform directional BDOF refinement. The device may calculate the vertical direction gradient difference and the horizontal direction gradient difference for the CU. The vertical direction gradient difference may indicate the difference between the vertical gradients for a first reference picture and the vertical gradients for a second reference picture. The horizontal direction gradient difference may indicate the difference between the horizontal gradients for the first reference picture and the horizontal gradients for the second reference picture. The video coding device may determine the direction in which to perform directional BDOF refinement based on the vertical direction gradient difference and the horizontal direction gradient difference. The video coding device may perform directional BDOF refinement in the determined direction.

IPC Classes  ?

  • H04N 19/577 - Motion compensation with bidirectional frame interpolation, i.e. using B-pictures
  • H04N 19/103 - Selection of coding mode or of prediction mode
  • H04N 19/139 - Analysis of motion vectors, e.g. their magnitude, direction, variance or reliability
  • H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock

18.

Bl-PREDICTION FOR VIDEO CODING

      
Application Number 18094515
Status Pending
Filing Date 2023-01-09
First Publication Date 2023-06-08
Owner VID SCALE, INC. (USA)
Inventor
  • He, Yuwen
  • Luo, Jiancong
  • Xiu, Xiaoyu
  • Ye, Yan

Abstract

Systems, methods, and instrumentalities may be provided for determining whether to bypass bi-directional optical flow (BDOF) if BDOF is used in combination with bi-prediction with coding unit (CU) weights (e.g., generalized bi-prediction (GBi)). A coding system may combine coding modes, coding techniques, and/or coding tools. The coding system may include a wireless transmit/receive unit (WTRU). For example, the coding system may combine BDOF and bi-prediction with CU weights (BCW). BDOF may include refining a motion vector associated with a current CU based at least in part on gradients associated with a location in the current CU. The coding system may determine that BDOF is enabled, and/or that bi-prediction with CU weights is enabled for the current CU. The coding system’s determination that bi-prediction with CU weights is enabled and/or that BDOF is enabled may be based on one or more indications.

IPC Classes  ?

  • H04N 19/577 - Motion compensation with bidirectional frame interpolation, i.e. using B-pictures

19.

SUB-BLOCK MOTION DERIVATION AND DECODER-SIDE MOTION VECTOR REFINEMENT FOR MERGE MODE

      
Application Number 18097390
Status Pending
Filing Date 2023-01-16
First Publication Date 2023-05-18
Owner VID SCALE, INC. (USA)
Inventor
  • Xiu, Xiaoyu
  • He, Yuwen
  • Ye, Yan

Abstract

Systems, methods, and instrumentalities for sub-block motion derivation and motion vector refinement for merge mode may be disclosed herein. Video data may be coded (e.g., encoded and/or decoded). A collocated picture for a current slice of the video data may be identified. The current slice may include one or more coding units (CUs). One or more neighboring CUs may be identified for a current CU. A neighboring CU (e.g., each neighboring CU) may correspond to a reference picture. A (e.g., one) neighboring CU may be selected to be a candidate neighboring CU based on the reference pictures and the collocated picture. A motion vector (MV) (e.g., collocated MV) may be identified from the collocated picture based on an MV (e.g., a reference MV) of the candidate neighboring CU. The current CU may be coded (e.g., encoded and/or decoded) using the collocated MV.

IPC Classes  ?

  • H04N 19/51 - Motion estimation or motion compensation
  • H04N 19/105 - Selection of the reference unit for prediction within a chosen coding or prediction mode, e.g. adaptive choice of position and number of pixels used for prediction
  • H04N 19/174 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a slice, e.g. a line of blocks or a group of blocks
  • H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock

20.

ADAPTIVE MOTION VECTOR PRECISION FOR AFFINE MOTION MODEL BASED VIDEO CODING

      
Application Number 18089027
Status Pending
Filing Date 2022-12-27
First Publication Date 2023-05-04
Owner Vid Scale, Inc. (USA)
Inventor
  • He, Yuwen
  • Xiu, Xiaoyu
  • Ye, Yan
  • Luo, Jiancong

Abstract

Systems and methods are described for video coding using affine motion models with adaptive precision. In an example, a block of video is encoded in a bitstream using an affine motion model, where the affine motion model is characterized by at least two motion vectors. A precision is selected for each of the motion vectors, and the selected precisions are signaled in the bitstream. In some embodiments, the precisions are signaled by including in the bitstream information that identifies one of a plurality of elements in a selected predetermined precision set. The identified element indicates the precision of each of the motion vectors that characterize the affine motion model. In some embodiments, the precision set to be used is signaled expressly in the bitstream; in other embodiments, the precision set may be inferred, e.g., from the block size, block shape or temporal layer.

IPC Classes  ?

  • H04N 19/137 - Motion inside a coding unit, e.g. average field, frame or block difference
  • H04N 19/159 - Prediction type, e.g. intra-frame, inter-frame or bidirectional frame prediction
  • H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
  • H04N 19/192 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the adaptation method, adaptation tool or adaptation type used for the adaptive coding the adaptation method, adaptation tool or adaptation type being iterative or recursive

21.

SYMMETRIC MOTION VECTOR DIFFERENCE CODING

      
Application Number 18082525
Status Pending
Filing Date 2022-12-15
First Publication Date 2023-04-27
Owner VID SCALE, INC. (USA)
Inventor
  • Luo, Jiancong
  • Xiu, Xiaoyu
  • He, Yuwen
  • Yang, Hua

Abstract

Bi-directional optical flow (BDOF) may be bypassed, for a current coding block, based on whether symmetric motion vector difference (SMVD) is used in motion vector coding for the current coding block. A coding device (e.g., an encoder or a decoder) may determine whether to bypass BDOF for the current coding block based at least in part on an SMVD indication for the current coding block. The coding device may obtain the SMVD indication that indicates whether SMVD is used in motion vector coding for the current coding block. If SMVD indication indicates that SMVD is used in the motion vector coding for the current coding block, the coding device may bypass BDOF for the current coding block. The coding device may reconstruct the current coding block without performing BDOF if it determines to bypass BDOF for the current coding block.

IPC Classes  ?

  • H04N 19/137 - Motion inside a coding unit, e.g. average field, frame or block difference
  • H04N 19/105 - Selection of the reference unit for prediction within a chosen coding or prediction mode, e.g. adaptive choice of position and number of pixels used for prediction
  • H04N 19/159 - Prediction type, e.g. intra-frame, inter-frame or bidirectional frame prediction
  • H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
  • H04N 19/46 - Embedding additional information in the video signal during the compression process

22.

METHODS AND APPARATUS FOR REDUCING THE CODING LATENCY OF DECODER-SIDE MOTION REFINEMENT

      
Application Number 18075169
Status Pending
Filing Date 2022-12-05
First Publication Date 2023-04-20
Owner VID SCALE, INC. (USA)
Inventor
  • Xiu, Xiaoyu
  • He, Yuwen
  • Ye, Yan

Abstract

Embodiments of video coding systems and methods are described for reducing coding latency introduced by decoder-side motion vector refinement (DMVR). In one example, two non-refined motion vectors are identified for coding of a first block of samples (e.g. a first coding unit) using bi-prediction. One or both of the non-refined motion vectors are used to predict motion information for a second block of samples (e.g. a second coding unit). The two non-refined motion vectors are refined using DMVR, and the refined motion vectors are used to generate a prediction signal of the first block of samples. Such embodiments allow the second block of samples to be coded substantially in parallel with the first block without waiting for completion of DMVR on the first block. In additional embodiments, optical-flow-based techniques are described for motion vector refinement.

IPC Classes  ?

  • H04N 19/513 - Processing of motion vectors
  • H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
  • H04N 19/577 - Motion compensation with bidirectional frame interpolation, i.e. using B-pictures
  • H04N 19/86 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using pre-processing or post-processing specially adapted for video compression involving reduction of coding artifacts, e.g. of blockiness

23.

METHODS AND APPARATUS FOR POINT CLOUD COMPRESSION BITSTREAM FORMAT

      
Application Number 18081050
Status Pending
Filing Date 2022-12-14
First Publication Date 2023-04-20
Owner VID SCALE, INC. (USA)
Inventor
  • Hamza, Ahmed
  • He, Yong

Abstract

Methods, apparatus, systems, architectures and interfaces for encoding and/or decoding point cloud bitstreams including coded point cloud sequences are provided. Included among such methods, apparatuses, systems, architectures, and interfaces is an apparatus that may include a processor and memory. A method may include any of: mapping components of the point cloud bitstream into tracks; generating information identifying any of geometry streams or texture streams according to the mapping of the components; generating information associated with layers corresponding to respective geometry component streams; and generating information indicating operation points associated with the point cloud bitstream.

IPC Classes  ?

  • G06T 9/00 - Image coding
  • H04N 19/44 - Decoders specially adapted therefor, e.g. video decoders which are asymmetric with respect to the encoder
  • H04N 19/70 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards

24.

COMPLEXITY REDUCTION OF OVERLAPPED BLOCK MOTION COMPENSATION

      
Application Number 17993470
Status Pending
Filing Date 2022-11-23
First Publication Date 2023-04-06
Owner VID SCALE, INC. (USA)
Inventor
  • Zhang, Yan
  • Xiu, Xiaoyu
  • He, Yuwen
  • Ye, Yan

Abstract

Overlapped block motion compensation (OBMC) may be performed for a current video block based on motion information associated with the current video block and motion information associated with one or more neighboring blocks of the current video block. Under certain conditions, some or all of these neighboring blocks may be omitted from the OBMC operation of the current block. For instance, a neighboring block may be skipped during the OBMC operation if the current video block and the neighboring block are both uni-directionally or bi-directionally predicted, if the motion vectors associated with the current block and the neighboring block refer to a same reference picture, and if a sum of absolute differences between those motion vectors is smaller than a threshold value. Further, OBMC may be conducted in conjunction with regular motion compensation and may use simplified filters than traditionally allowed.

IPC Classes  ?

  • H04N 19/583 - Motion compensation with overlapping blocks
  • H04N 19/139 - Analysis of motion vectors, e.g. their magnitude, direction, variance or reliability
  • H04N 19/189 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the adaptation method, adaptation tool or adaptation type used for the adaptive coding

25.

DETECTING MAN-IN-THE-MIDDLE ATTACKS IN ADAPTIVE STREAMING

      
Application Number 17990084
Status Pending
Filing Date 2022-11-18
First Publication Date 2023-03-23
Owner Vid Scale, Inc. (USA)
Inventor Giladi, Alexander

Abstract

Systems and methods for adaptively streaming video content to a wireless transmit/receive unit (WTRU) or wired transmit/receive unit may comprise obtaining a media presentation description that comprises a content authenticity, requesting a key for a hash-based message authentication code; receiving the key for the hash-based message authentication code, determining a determined hash for a segment of the media presentation description, requesting a reference hash for the segment from a server, receiving the reference hash for the segment from the server, and comparing the reference hash to the determined hash to determine whether the requested hash matches the determined hash.

IPC Classes  ?

  • H04L 9/40 - Network security protocols
  • H04L 67/02 - Protocols based on web technology, e.g. hypertext transfer protocol [HTTP]
  • H04N 21/61 - Network physical structure; Signal processing
  • H04N 21/845 - Structuring of content, e.g. decomposing content into time segments
  • H04W 12/10 - Integrity
  • H04N 21/854 - Content authoring
  • H04N 21/835 - Generation of protective data, e.g. certificates
  • H04L 65/65 - Network streaming protocols, e.g. real-time transport protocol [RTP] or real-time control protocol [RTCP]
  • H04L 65/75 - Media network packet handling

26.

Power Aware Video Decoding and Streaming

      
Application Number 18070255
Status Pending
Filing Date 2022-11-28
First Publication Date 2023-03-23
Owner VID SCALE, Inc. (USA)
Inventor
  • He, Yuwen
  • Ye, Yan
  • He, Yong
  • Mcclellan, George W.
  • Ryu, Eun Seok

Abstract

Methods and systems are disclosed for a mobile device to decode video based on available power and/or energy. For example, the mobile device may receive a media description file (MDF) from for a video stream from a video server. The MDF may include complexity information associated with a plurality of video segments. The complexity information may be related to the amount of processing power to be utilized for decoding the segment at the mobile device. The mobile device may determine at least one power metric for the mobile device. The mobile device may determine a first complexity level to be requested for a first video segment based on the complexity information from the MDF and the power metric. The mobile device may dynamically alter the decoding process to save energy based on the detected power/energy level.

IPC Classes  ?

  • H04N 19/187 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being a scalable video layer
  • G06F 1/3212 - Monitoring battery levels, e.g. power saving mode being initiated when battery voltage goes below a certain level
  • G06F 1/329 - Power saving characterised by the action undertaken by task scheduling
  • H04N 19/117 - Filters, e.g. for pre-processing or post-processing
  • H04N 19/127 - Prioritisation of hardware or computational resources
  • H04N 19/136 - Incoming video signal characteristics or properties
  • H04N 19/156 - Availability of hardware or computational resources, e.g. encoding based on power-saving criteria
  • H04N 19/44 - Decoders specially adapted therefor, e.g. video decoders which are asymmetric with respect to the encoder
  • H04N 21/2343 - Processing of video elementary streams, e.g. splicing of video streams or manipulating MPEG-4 scene graphs involving reformatting operations of video signals for distribution or compliance with end-user requests or end-user device requirements
  • H04N 21/414 - Specialised client platforms, e.g. receiver in car or embedded in a mobile appliance
  • H04N 21/442 - Monitoring of processes or resources, e.g. detecting the failure of a recording device, monitoring the downstream bandwidth, the number of times a movie has been viewed or the storage space available from the internal hard disk
  • H04N 21/845 - Structuring of content, e.g. decomposing content into time segments
  • H04N 21/854 - Content authoring
  • H04L 65/612 - Network streaming of media packets for supporting one-way streaming services, e.g. Internet radio for unicast
  • H04L 65/613 - Network streaming of media packets for supporting one-way streaming services, e.g. Internet radio for the control of the source by the destination

27.

ADAPTIVE CONTROL POINT SELECTION FOR AFFINE MOTION MODEL BASED VIDEO CODING

      
Application Number 17985652
Status Pending
Filing Date 2022-11-11
First Publication Date 2023-03-09
Owner Vid Scale, Inc. (USA)
Inventor
  • He, Yuwen
  • Xiu, Xiaoyu
  • Ye, Yan
  • Hanhart, Philippe

Abstract

Systems, methods, and instrumentalities are disclosed for motion vector clipping when affine motion mode is enabled for a video block. A video coding device may determine that an affine mode for a video block is enabled. The video coding device may determine a plurality of control point affine motion vectors associated with the video block. The video coding device may store the plurality of clipped control point affine motion vectors for motion vector prediction of a neighboring control point affine motion vector. The video coding device may derive a sub-block motion vector associated with a sub-block of the video block, clip the derived sub-block motion vector, and store it for spatial motion vector prediction or temporal motion vector prediction. For example, the video coding device may clip the derived sub-block motion vector based on a motion field range that may be based on a bit depth value.

IPC Classes  ?

  • H04N 19/52 - Processing of motion vectors by encoding by predictive encoding
  • H04N 19/54 - Motion estimation other than block-based using feature points or meshes

28.

ADAPTIVE FRAME PACKING FOR 360-DEGREE VIDEO CODING

      
Application Number 17892911
Status Pending
Filing Date 2022-08-22
First Publication Date 2023-03-09
Owner VID SCALE, INC. (USA)
Inventor
  • Hanhart, Philippe
  • He, Yuwen
  • Ye, Yan

Abstract

A video coding device may be configured to periodically select the frame packing configuration (e.g., face layout and/or face rotations parameters) associated with a RAS. The device may receive a plurality of pictures, which may each comprise a plurality of faces. The pictures may be grouped into a plurality of RASs. The device may select a frame packing configuration with the lowest cost for a first RAS. For example, the cost of a frame packing configuration may be determined based on the first picture of the first RAS. The device may select a frame packing configuration for a second RAS. The frame packing configuration for the first RAS may be different than the frame packing configuration for the second RAS. The frame packing configuration for the first RAS and the frame packing configuration for the second RAS may be signaled in the video bitstream.

IPC Classes  ?

  • H04N 13/161 - Encoding, multiplexing or demultiplexing different image signal components
  • H04N 19/172 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a picture, frame or field
  • H04N 19/186 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being a colour or a chrominance component

29.

COMPLEXITY REDUCTION AND BIT-WIDTH CONTROL FOR BI-DIRECTIONAL OPTICAL FLOW

      
Application Number 17962928
Status Pending
Filing Date 2022-10-10
First Publication Date 2023-02-16
Owner Vid Scale, Inc. (Germany)
Inventor
  • Xiu, Xiaoyu
  • He, Yuwen
  • Ye, Yan
  • Luo, Jiancong

Abstract

Systems and methods are described for reducing the complexity of using bi-directional optical flow (BIO) in video coding. In some embodiments, bit-width reduction steps are introduced in the BIO motion refinement process to reduce the maximum bit-width used for BIO calculations. In some embodiments, simplified interpolation filters are used to generate predicted samples in an extended region around a current coding unit. In some embodiments, different interpolation filters are used for vertical versus horizontal interpolation. In some embodiments, BIO is disabled for coding units with small heights and/or for coding units that are predicted using a sub-block level inter prediction technique, such as advanced temporal motion vector prediction (ATMVP) or affine prediction.

IPC Classes  ?

  • H04N 19/105 - Selection of the reference unit for prediction within a chosen coding or prediction mode, e.g. adaptive choice of position and number of pixels used for prediction
  • H04N 19/132 - Sampling, masking or truncation of coding units, e.g. adaptive resampling, frame skipping, frame interpolation or high-frequency transform coefficient masking
  • H04N 19/146 - Data rate or code amount at the encoder output
  • H04N 19/159 - Prediction type, e.g. intra-frame, inter-frame or bidirectional frame prediction
  • H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
  • H04N 19/513 - Processing of motion vectors

30.

COMPLEXITY REDUCTION AND BIT-WIDTH CONTROL FOR BI-DIRECTIONAL OPTICAL FLOW

      
Application Number 17962918
Status Pending
Filing Date 2022-10-10
First Publication Date 2023-02-16
Owner Vid Scale, Inc. (USA)
Inventor
  • Xiu, Xiaoyu
  • He, Yuwen
  • Ye, Yan
  • Luo, Jiancong

Abstract

Systems and methods are described for reducing the complexity of using bi-directional optical flow (BIO) in video coding. In some embodiments, bit-width reduction steps are introduced in the BIO motion refinement process to reduce the maximum bit-width used for BIO calculations. In some embodiments, simplified interpolation filters are used to generate predicted samples in an extended region around a current coding unit. In some embodiments, different interpolation filters are used for vertical versus horizontal interpolation. In some embodiments, BIO is disabled for coding units with small heights and/or for coding units that are predicted using a sub-block level inter prediction technique, such as advanced temporal motion vector prediction (ATMVP) or affine prediction.

IPC Classes  ?

  • H04N 19/105 - Selection of the reference unit for prediction within a chosen coding or prediction mode, e.g. adaptive choice of position and number of pixels used for prediction
  • H04N 19/132 - Sampling, masking or truncation of coding units, e.g. adaptive resampling, frame skipping, frame interpolation or high-frequency transform coefficient masking
  • H04N 19/146 - Data rate or code amount at the encoder output
  • H04N 19/159 - Prediction type, e.g. intra-frame, inter-frame or bidirectional frame prediction
  • H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
  • H04N 19/513 - Processing of motion vectors

31.

Affine motion estimation for affine model-based video coding

      
Application Number 17278202
Grant Number 11943467
Status In Force
Filing Date 2019-09-20
First Publication Date 2023-02-02
Grant Date 2024-03-26
Owner VID Scale, Inc. (USA)
Inventor
  • He, Yuwen
  • Xiu, Xiaoyu
  • Ye, Yan
  • Luo, Jiancong

Abstract

Systems, methods, and instrumentalities for affine motion estimation for affine model-based video coding may be disclosed herein. A first motion vector (MV) set including one or more MVs may be derived for a first coding block. The MVs may be control point MVs (CPMVs) and the MVs may be derived by performing affine motion estimation (ME) associated with the first coding block. The first MV set may be added to a recently-estimated MV list. A head of the recently-estimated MV list may be set to the first MV set. The recently-estimated MV list may be empty or may contain one or more previously-added MV sets.

IPC Classes  ?

  • H04N 19/52 - Processing of motion vectors by encoding by predictive encoding
  • H04N 19/147 - Data rate or code amount at the encoder output according to rate distortion criteria
  • H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock

32.

METHOD AND APPARATUS FOR SIGNALING AND CONSTRUCTION OF VIDEO CODING REFERENCE PICTURE LISTS

      
Application Number 17945714
Status Pending
Filing Date 2022-09-15
First Publication Date 2023-01-26
Owner VID SCALE, INC. (USA)
Inventor
  • Ye, Yan
  • He, Yong

Abstract

Improved method and apparatus for signaling of reference pictures used for temporal prediction. The signaling schemes and construction process for different reference picture lists in HEVC Working Draft 5 (WD5) are improved.

IPC Classes  ?

  • H04N 19/105 - Selection of the reference unit for prediction within a chosen coding or prediction mode, e.g. adaptive choice of position and number of pixels used for prediction
  • H04N 19/70 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards
  • H04N 19/103 - Selection of coding mode or of prediction mode
  • H04N 19/577 - Motion compensation with bidirectional frame interpolation, i.e. using B-pictures
  • H04N 19/503 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
  • H04N 19/117 - Filters, e.g. for pre-processing or post-processing
  • H04N 19/124 - Quantisation
  • H04N 19/13 - Adaptive entropy coding, e.g. adaptive variable length coding [AVLC] or context adaptive binary arithmetic coding [CABAC]
  • H04N 19/139 - Analysis of motion vectors, e.g. their magnitude, direction, variance or reliability
  • H04N 19/15 - Data rate or code amount at the encoder output by monitoring actual compressed data size at the memory before deciding storage at the transmission buffer
  • H04N 19/61 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding in combination with predictive coding

33.

METHODS AND APPARATUS FOR PERFORMING REAL-TIME VVC DECODING

      
Application Number 17779212
Status Pending
Filing Date 2020-11-24
First Publication Date 2023-01-05
Owner VID SCALE, INC. (USA)
Inventor
  • Gudumasu, Srinivas
  • Bandyopadhyay, Saurav
  • He, Yuwen
  • He, Yong
  • Srivastava, Asit

Abstract

Apparatus and methods for implementing a real-time Versatile Video Coding (VVC) decoder use multiple threads to address the limitation with existing parallelization techniques and fully utilizes the available CPU computation resource without compromising on the coding efficiency. The proposed Multi-threaded (MT) framework uses CTU level parallel processing techniques without compromising on the memory bandwidth. Picture level parallel processing separates the sequence into temporal levels by considering the picture's referencing hierarchy. Embodiments are provided using various optimization techniques to achieve real-time VVC decoding on heterogenous platforms with multi-core CPUs, for those bitstreams generated using a VVC reference encoder with a default configuration.

IPC Classes  ?

  • H04N 19/436 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals - characterised by implementation details or hardware specially adapted for video compression or decompression, e.g. dedicated software implementation using parallelised computational arrangements
  • H04N 19/124 - Quantisation
  • H04N 19/52 - Processing of motion vectors by encoding by predictive encoding
  • H04N 19/593 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving spatial prediction techniques
  • H04N 19/82 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals - Details of filtering operations specially adapted for video compression, e.g. for pixel interpolation involving filtering within a prediction loop
  • H04N 19/96 - Tree coding, e.g. quad-tree coding
  • H04N 19/139 - Analysis of motion vectors, e.g. their magnitude, direction, variance or reliability
  • H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
  • H04N 19/159 - Prediction type, e.g. intra-frame, inter-frame or bidirectional frame prediction

34.

Systems and methods for spatial prediction

      
Application Number 17896350
Grant Number 11785249
Status In Force
Filing Date 2022-08-26
First Publication Date 2022-12-22
Grant Date 2023-10-10
Owner VID Scale, Inc. (USA)
Inventor
  • Ye, Yan
  • Chen, Qian
  • Dong, Jie

Abstract

Systems, methods, and instrumentalities are disclosed relating to intra prediction of a video signal based on mode-dependent subsampling. A block of coefficients associated with a first sub block of a video block, one or more blocks of coefficients associated with one or more remaining sub blocks of the video block, and an indication of a prediction mode for the video block may be received. One or more interpolating techniques, a predicted first sub block, and the predicted sub blocks of the one or more remaining sub blocks may be determined. A reconstructed first sub block and one or more reconstructed remaining sub blocks may be generated. A reconstructed video block may be formed based on the prediction mode, the reconstructed first sub block, and the one or more reconstructed remaining sub blocks.

IPC Classes  ?

  • H04N 19/52 - Processing of motion vectors by encoding by predictive encoding
  • H04N 19/119 - Adaptive subdivision aspects e.g. subdivision of a picture into rectangular or non-rectangular coding blocks
  • H04N 19/11 - Selection of coding mode or of prediction mode among a plurality of spatial predictive coding modes
  • H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
  • H04N 19/46 - Embedding additional information in the video signal during the compression process
  • H04N 19/593 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving spatial prediction techniques
  • H04N 19/124 - Quantisation
  • H04N 19/14 - Coding unit complexity, e.g. amount of activity or edge presence estimation
  • H04N 19/80 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals - Details of filtering operations specially adapted for video compression, e.g. for pixel interpolation
  • H04N 19/90 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using coding techniques not provided for in groups , e.g. fractals
  • H04N 19/60 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding
  • H04N 19/59 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving spatial sub-sampling or interpolation, e.g. alteration of picture size or resolution
  • H04N 19/182 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being a pixel

35.

REFERENCE PICTURE SET (RPS) SIGNALING FOR SCALABLE HIGH EFFICIENCY VIDEO CODING (HEVC)

      
Application Number 17891352
Status Pending
Filing Date 2022-08-19
First Publication Date 2022-12-15
Owner VID SCALE, INC. (USA)
Inventor
  • He, Yong
  • Ye, Yan
  • Mcclellan, George W.
  • Dong, Jie

Abstract

Embodiments contemplate coding video data by generating a video encoded bitstream that may include reference picture set (RPS) extensions for inter-layer reference pictures, and the extensions may include inter-layer delta Picture Order Counts (POCs). Embodiments may also include signaling that lower layer reference pictures may be available in a lower layer decoder picture buffer (DPB), and/or an aggregate DPB, that may be added to the RPS set of a higher layer. The bitstream may include a signal indicating whether the higher layer RPS may be specified by a lower layer RPS, and the lower layer RPS may be temporal, inter-layer prediction (ILP), or both.

IPC Classes  ?

  • H04N 19/105 - Selection of the reference unit for prediction within a chosen coding or prediction mode, e.g. adaptive choice of position and number of pixels used for prediction
  • H04N 19/187 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being a scalable video layer
  • H04N 19/70 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards
  • H04N 19/463 - Embedding additional information in the video signal during the compression process by compressing encoding parameters before transmission
  • H04N 19/33 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using hierarchical techniques, e.g. scalability in the spatial domain
  • H04N 19/31 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using hierarchical techniques, e.g. scalability in the temporal domain

36.

DECODING COMPLEXITY FOR MOBILE MULTIMEDIA STREAMING

      
Application Number 17842490
Status Pending
Filing Date 2022-06-16
First Publication Date 2022-12-08
Owner Vid Scale, Inc. (USA)
Inventor
  • Lotfallah, Osama
  • Asbun, Eduardo
  • Liu, Hang
  • Reznik, Yuriy

Abstract

A decoding complexity may be used to predict power consumption for receiving, decoding, and/or displaying multimedia content at a wireless transmit/receive unit (WTRU). The decoding complexity may be based on decoding complexity feedback received from a reference device, such as another WTRU. The decoding complexity feedback may be based on measurements performed at the reference device for receiving decoding, and/or displaying the multimedia content. A content providing device may indicate the decoding complexity of requested media content to a WTRU, or another network entity. The decoding complexity may be indicated in a streaming protocol or file associated with the media content. The WTRU, or other network entity, may use the decoding complexity determine its preferences regarding transmission of the media content. The content providing device may determine whether to transmit the media content based on the decoding complexity and/or the preferences of the WTRU or other network entity.

IPC Classes  ?

  • H04L 65/70 - Media network packetisation
  • H04L 65/80 - Responding to QoS
  • H04W 52/22 - TPC being performed according to specific parameters taking into account previous information or commands
  • H04N 21/462 - Content or additional data management e.g. creating a master electronic program guide from data received from the Internet and a Head-end or controlling the complexity of a video stream by scaling the resolution or bit-rate based on the client capabi
  • H04N 21/2662 - Controlling the complexity of the video stream, e.g. by scaling the resolution or bitrate of the video stream based on the client capabilities
  • H04L 65/612 - Network streaming of media packets for supporting one-way streaming services, e.g. Internet radio for unicast
  • H04L 65/613 - Network streaming of media packets for supporting one-way streaming services, e.g. Internet radio for the control of the source by the destination
  • H04L 65/75 - Media network packet handling
  • H04W 52/02 - Power saving arrangements

37.

ADAPTIVE INTERPOLATION FILTER FOR MOTION COMPENSATION

      
Application Number 17761254
Status Pending
Filing Date 2020-09-18
First Publication Date 2022-12-01
Owner VID SCALE, INC. (USA)
Inventor
  • Chen, Wei
  • He, Yuwen
  • Yang, Hua

Abstract

A video processing apparatus may comprise one or more processors that are configured to determine an interpolation filter length for an interpolation filter associated with a coding unit (CU) based on a size of the CU. The one or more processor may be configured to determine an interpolated reference sample based on the determined interpolation filter length for the interpolation filter and a reference sample for the CU. The one or more processor may be configured to predict the CU based on the interpolated reference sample. For example, if a first CU has a size that is greater than the size of a second CU, the one or more processors may be configured to use a shorter interpolation filter for the first CU than for the second CU.

IPC Classes  ?

  • H04N 19/117 - Filters, e.g. for pre-processing or post-processing
  • H04N 19/80 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals - Details of filtering operations specially adapted for video compression, e.g. for pixel interpolation
  • H04N 19/105 - Selection of the reference unit for prediction within a chosen coding or prediction mode, e.g. adaptive choice of position and number of pixels used for prediction
  • H04N 19/132 - Sampling, masking or truncation of coding units, e.g. adaptive resampling, frame skipping, frame interpolation or high-frequency transform coefficient masking
  • H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock

38.

FACE DISCONTINUITY FILTERING FOR 360-DEGREE VIDEO CODING

      
Application Number 17884399
Status Pending
Filing Date 2022-08-09
First Publication Date 2022-12-01
Owner VID SCALE, INC. (USA)
Inventor
  • Hanhart, Philippe
  • Ye, Yan
  • He, Yuwen

Abstract

Systems, methods, and instrumentalities are disclosed for discontinuous face boundary filtering for 360-degree video coding. A face discontinuity may be filtered (e.g., to reduce seam artifacts) in whole or in part, for example, using coded samples or padded samples on either side of the face discontinuity. Filtering may be applied, for example, as an in-loop filter or a post-processing step. 2D positional information related to two sides of the face discontinuity may be signaled in a video bitstream so that filtering may be applied independent of projection formats and/or frame packing techniques.

IPC Classes  ?

  • H04N 19/597 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding specially adapted for multi-view video sequence encoding
  • H04N 19/117 - Filters, e.g. for pre-processing or post-processing
  • H04N 19/85 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using pre-processing or post-processing specially adapted for video compression

39.

JOINT COMPONENT VIDEO FRAME FILTERING

      
Application Number 17762991
Status Pending
Filing Date 2020-09-23
First Publication Date 2022-11-24
Owner VID SCALE, INC. (USA)
Inventor
  • Yang, Hua
  • He, Yuwen
  • Chen, Wei
  • Li, Hongyu

Abstract

A filter may be applied to a subset of components associated with a sample in a coding block. The output of the filter may be used to modify values for other component(s). For example, a filter may be applied to a selected (for example, dominant) component(s). The output of the filter may be used to modify a value for one of the other components (for example, non-dominant components). The output of the filter may be used, for example, after a weighting factor is applied to the filter output, to modify a value for another one of the other components. A joint refinement signal may be obtained, for example, as the filtered output signal minus the filter input signal of the selected component(s). A properly weighted version of the joint refinement signal may be applied to modify the other components.

IPC Classes  ?

  • H04N 19/117 - Filters, e.g. for pre-processing or post-processing
  • H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
  • H04N 19/186 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being a colour or a chrominance component

40.

SWITCHING LOGIC FOR BI-DIRECTIONAL OPTICAL FLOW

      
Application Number 17763024
Status Pending
Filing Date 2020-09-23
First Publication Date 2022-11-24
Owner VID SCALE, INC. (USA)
Inventor
  • Chen, Wei
  • He, Yuwen
  • Yang, Hua

Abstract

An apparatus may be configured to determine a reference picture listed in a first reference picture list and a reference picture listed in a second reference picture list, for a coding block. The apparatus may be configured to determine whether to perform bi-directional optical flow (BDOF) for the coding block based at least in part on whether a distance between a picture associated with the coding block and the reference picture listed in the first reference picture list differs from a distance between the picture associated with the coding block and the reference picture listed in the second reference picture list. The apparatus may be configured to decode the coding block based on the determination of whether to perform BDOF for the coding block.

IPC Classes  ?

  • H04N 19/105 - Selection of the reference unit for prediction within a chosen coding or prediction mode, e.g. adaptive choice of position and number of pixels used for prediction
  • H04N 19/137 - Motion inside a coding unit, e.g. average field, frame or block difference
  • H04N 19/117 - Filters, e.g. for pre-processing or post-processing
  • H04N 19/132 - Sampling, masking or truncation of coding units, e.g. adaptive resampling, frame skipping, frame interpolation or high-frequency transform coefficient masking
  • H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock

41.

HANDLING FACE DISCONTINUITIES IN 360-DEGREE VIDEO CODING

      
Application Number 17876708
Status Pending
Filing Date 2022-07-29
First Publication Date 2022-11-24
Owner Vid Scale, Inc. (USA)
Inventor
  • Hanhart, Philippe
  • He, Yuwen
  • Ye, Yan

Abstract

Systems, methods, and instrumentalities may be provided for discounting reconstructed samples and/or coding information from spatial neighbors across face discontinuities. Whether a current block is located at a face discontinuity may be determined. The face discontinuity may be a face boundary between two or more adjoining blocks that are not spherical neighbors. The coding availability of a neighboring block of the current block may be determined, e.g., based on whether the neighboring block is on the same side of the face discontinuity as the current block. For example, the neighboring block may be determined to be available for decoding the current block if it is on the same side of the face discontinuity as the current block, and unavailable if it is not on the same side of the face discontinuity. The neighboring block may be a spatial neighboring block or a temporal neighboring block.

IPC Classes  ?

  • H04N 19/82 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals - Details of filtering operations specially adapted for video compression, e.g. for pixel interpolation involving filtering within a prediction loop
  • H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
  • H04N 19/184 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being bits, e.g. of the compressed video stream
  • H04N 19/597 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding specially adapted for multi-view video sequence encoding

42.

MOTION-COMPENSATION PREDICTION BASED ON BI-DIRECTIONAL OPTICAL FLOW

      
Application Number 17749991
Status Pending
Filing Date 2022-05-20
First Publication Date 2022-11-17
Owner VID SCALE, INC. (USA)
Inventor
  • Ye, Yan
  • Xiu, Xiaoyu
  • He, Yuwen

Abstract

A device may determine whether to enable or disable bi-directional optical flow (BIO) for a current coding unit (CU) (e.g., block and/or sub-block). Prediction information for the CU may be identified and may include prediction signals associated with a first reference block and a second reference block (e.g., or a first reference sub-block and a second reference sub-block). A prediction difference may be calculated and may be used to determine the similarity between the two prediction signals. The CU may be reconstructed based on the similarity. For example, whether to reconstruct the CU with BIO enabled or BIO disabled may be based on whether the two prediction signals are similar. It may be determined to enable BIO for the CU when the two prediction signals are determined to be dissimilar. For example, the CU may be reconstructed with BIO disabled when the two prediction signals are determined to be similar.

IPC Classes  ?

  • H04N 19/577 - Motion compensation with bidirectional frame interpolation, i.e. using B-pictures
  • H04N 19/103 - Selection of coding mode or of prediction mode
  • H04N 19/139 - Analysis of motion vectors, e.g. their magnitude, direction, variance or reliability
  • H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock

43.

360-DEGREE VIDEO CODING USING GEOMETRY PROJECTION

      
Application Number 17860629
Status Pending
Filing Date 2022-07-08
First Publication Date 2022-11-17
Owner VID SCALE, INC. (USA)
Inventor
  • He, Yuwen
  • Ye, Yan
  • Hanhart, Philippe
  • Xiu, Xiaoyu

Abstract

Processing video data may include capturing the video data with multiple cameras and stitching the video data together to obtain a 360-degree video. A frame-packed picture may be provided based on the captured and stitched video data. A current sample location may be identified in the frame-packed picture. Whether a neighboring sample location is located outside of a content boundary of the frame-packed picture may be determined. When the neighboring sample location is located outside of the content boundary, a padding sample location may be derived based on at least one circular characteristic of the 360-degree video content and the projection geometry. The 360-degree video content may be processed based on the padding sample location.

IPC Classes  ?

  • H04N 19/597 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding specially adapted for multi-view video sequence encoding
  • H04N 13/194 - Transmission of image signals
  • H04N 13/383 - Image reproducers using viewer tracking for tracking with gaze detection, i.e. detecting the lines of sight of the viewer's eyes
  • G06T 17/30 - Surface description, e.g. polynomial surface description
  • H04N 13/117 - Transformation of image signals corresponding to virtual viewpoints, e.g. spatial image interpolation the virtual viewpoint locations being selected by the viewers or determined by viewer tracking
  • G06T 17/10 - Volume description, e.g. cylinders, cubes or using CSG [Constructive Solid Geometry]
  • H04N 13/161 - Encoding, multiplexing or demultiplexing different image signal components
  • H04N 13/344 - Displays for viewing with the aid of special glasses or head-mounted displays [HMD] with head-mounted left-right displays
  • H04N 19/563 - Motion estimation with padding, i.e. with filling of non-object values in an arbitrarily shaped picture block or region for estimation purposes
  • H04N 19/593 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving spatial prediction techniques
  • H04N 19/105 - Selection of the reference unit for prediction within a chosen coding or prediction mode, e.g. adaptive choice of position and number of pixels used for prediction
  • H04N 19/132 - Sampling, masking or truncation of coding units, e.g. adaptive resampling, frame skipping, frame interpolation or high-frequency transform coefficient masking
  • H04N 19/172 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a picture, frame or field

44.

METHODS AND APPARATUS FOR KERNEL TENSOR AND TREE PARTITION BASED NEURAL NETWORK COMPRESSION FRAMEWORK

      
Application Number 17621476
Status Pending
Filing Date 2020-07-02
First Publication Date 2022-11-10
Owner VID SCALE, INC. (USA)
Inventor
  • Yang, Hua
  • Li, Duanshun
  • Tian, Dong
  • He, Yuwen

Abstract

A method of encoding or decoding a video comprising a current picture, a first reference picture, and a weight tensor associated with a trained neural network (NN) model are provided. The method includes generating any number of kernel tensors, input channels and output channels associated with the weight tensor, each kernel tensor being associated with any of: a layer type, an input signal type, and a tree partition type, and each kernel tensor including weight coefficients, generating, for each of the any number of kernel tensors, tree partitions for any of a coding tree unit (CTU), a coding unit (CU), a prediction unit (PU), and a transform unit (TU) according to respective tree partition types associated with each of the any number of kernel tensors, and generating a compressed representation of the trained NN model by compressing and coding the any number of kernel tensors.

IPC Classes  ?

  • H04N 19/119 - Adaptive subdivision aspects e.g. subdivision of a picture into rectangular or non-rectangular coding blocks
  • H04N 19/96 - Tree coding, e.g. quad-tree coding
  • H04N 19/105 - Selection of the reference unit for prediction within a chosen coding or prediction mode, e.g. adaptive choice of position and number of pixels used for prediction
  • H04N 19/172 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a picture, frame or field

45.

Dynamic adaptation of volumetric content component sub-bitstreams in streaming services

      
Application Number 17623999
Grant Number 11917177
Status In Force
Filing Date 2020-07-02
First Publication Date 2022-11-03
Grant Date 2024-02-27
Owner VID Scale, Inc. (USA)
Inventor Hamza, Ahmed

Abstract

A media content processing device may decode visual volumetric content based on one or more messages, which may indicate which attribute sub-bitstream of one or more attribute sub-bitstreams indicated in a parameter set is active, The parameter set may include a visual volumetric video-based parameter set. The message indicating one or more active attribute sub-bitstreams may be received by the decoder, A decoder may perform decoding, such as determining which attribute sub-bitstream to use for decoding visual media content, based on the one or more messages, The one or more messages may be generated and sent to a decoder, for example, to indicate the deactivation of the one or more attribute sub-bitstreams. The decoder may determine an inactive attribute sub-bitstream and skip the inactive attribute sub-bitstream for decoding the visual media content based on the one or more messages.

IPC Classes  ?

  • H04N 19/44 - Decoders specially adapted therefor, e.g. video decoders which are asymmetric with respect to the encoder
  • H04N 19/46 - Embedding additional information in the video signal during the compression process
  • H04N 19/184 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being bits, e.g. of the compressed video stream

46.

INTRA SUB-PARTITIONS RELATED INFRA CODING

      
Application Number 17760985
Status Pending
Filing Date 2020-09-17
First Publication Date 2022-10-27
Owner VID SCALE, INC. (USA)
Inventor
  • Yang, Hua
  • He, Yuwen
  • Chen, Wei

Abstract

Systems, methods, and instrumentalities may be used for decoding and/or encoding a coding unit (CD), An intra-prediction mode for a CD may be determined. A split mode may be determined based on the intra-prediction mode, to generate a plurality of sub-partitions in the CU. A prediction for a first sub-partition of the plurality of sub-partitions in the CU may be based on a reference sample in a second sub-partition of the plurality of sub-partitions in the CU. The CU may be decoded and/or encoded, for example, based on the determined split mode.

IPC Classes  ?

  • H04N 19/119 - Adaptive subdivision aspects e.g. subdivision of a picture into rectangular or non-rectangular coding blocks
  • H04N 19/105 - Selection of the reference unit for prediction within a chosen coding or prediction mode, e.g. adaptive choice of position and number of pixels used for prediction
  • H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
  • H04N 19/159 - Prediction type, e.g. intra-frame, inter-frame or bidirectional frame prediction

47.

SYSTEMS AND METHODS FOR GENERALIZED MULTI-HYPOTHESIS PREDICTION FOR VIDEO CODING

      
Application Number 17838729
Status Pending
Filing Date 2022-06-13
First Publication Date 2022-09-29
Owner VID SCALE, INC. (USA)
Inventor
  • Chen, Chun-Chi
  • Xiu, Xiaoyu
  • He, Yuwen
  • Ye, Yan

Abstract

Systems and methods are described for video coding using generalized bi-prediction. In an exemplary embodiment, to code a current block of a video in a bitstream, a first reference block is selected from a first reference picture and a second reference block is selected from a second reference picture. Each reference block is associated with a weight, where the weight may be an arbitrary weight ranging, e.g., between 0 and 1. The current block is predicted using a weighted sum of the reference blocks. The weights may be selected from among a plurality of candidate weights. Candidate weights may be signaled in the bitstream or may be derived implicitly based on a template. Candidate weights may be pruned to avoid out-of-range or substantially duplicate candidate weights. Generalized bi-prediction may additionally be used in frame rate up conversion.

IPC Classes  ?

  • H04N 19/105 - Selection of the reference unit for prediction within a chosen coding or prediction mode, e.g. adaptive choice of position and number of pixels used for prediction
  • H04N 19/96 - Tree coding, e.g. quad-tree coding
  • H04N 19/463 - Embedding additional information in the video signal during the compression process by compressing encoding parameters before transmission
  • H04N 19/139 - Analysis of motion vectors, e.g. their magnitude, direction, variance or reliability
  • H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
  • H04N 19/573 - Motion compensation with multiple frame prediction using two or more reference frames in a given prediction direction

48.

PRECISION REFINEMENT FOR MOTION COMPENSATION WITH OPTICAL FLOW

      
Application Number 17619192
Status Pending
Filing Date 2020-06-18
First Publication Date 2022-09-08
Owner Vid Scale, Inc. (USA)
Inventor
  • Chen, Wei
  • He, Yuwen

Abstract

Systems and methods are described for refining motion compensated predictions in block-based video coding. In an example embodiment, motion-compensated prediction is used to generate predicted sample values in a current block of samples. A precision difference value and a motion vector refinement for the current block are signaled in the bitstream. For each sample in the current block, a spatial gradient is calculated at the sample, and a scalar product is calculated between the spatial gradient and the motion vector refinement. The scalar product is scaled (e.g. bit-shifted) by an amount indicated by the precision difference value to generate a sample difference value, and the sample difference value is added to the predicted sample value to generate a refined sample value.

IPC Classes  ?

  • H04N 19/137 - Motion inside a coding unit, e.g. average field, frame or block difference
  • H04N 19/132 - Sampling, masking or truncation of coding units, e.g. adaptive resampling, frame skipping, frame interpolation or high-frequency transform coefficient masking
  • H04N 19/33 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using hierarchical techniques, e.g. scalability in the spatial domain
  • H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock

49.

Motion compensated bi-prediction based on local illumination compensation

      
Application Number 17732466
Grant Number 11962759
Status In Force
Filing Date 2022-04-28
First Publication Date 2022-08-25
Grant Date 2024-04-16
Owner VID SCALE, Inc. (USA)
Inventor
  • Xiu, Xiaoyu
  • He, Yuwen
  • Ye, Yan
  • Bandyopadhyay, Saurav

Abstract

Systems, methods, and instrumentalities are described herein for calculating local illumination compensation (LIC) parameters for bi-predicted coding unit (CU). The LIC parameters may be used to generate adjusted samples for the current CU and to address local illumination changes that may exist among temporal neighboring pictures. LIC parameters may be calculated based on bi-predicted reference template samples and template samples for a current CU. Bi-predicted reference template samples may be generated based on reference template samples neighboring temporal reference CUs. For example, the bi-predicted reference template samples may be generated based on averaging the reference template samples. The reference template samples may correspond to template samples for the current CU. A CU may be or may include a coding block and/or a sub-block that may be derived by dividing the coding block.

IPC Classes  ?

  • H04N 19/105 - Selection of the reference unit for prediction within a chosen coding or prediction mode, e.g. adaptive choice of position and number of pixels used for prediction
  • G06F 17/18 - Complex mathematical operations for evaluating statistical data
  • H04N 19/132 - Sampling, masking or truncation of coding units, e.g. adaptive resampling, frame skipping, frame interpolation or high-frequency transform coefficient masking
  • H04N 19/159 - Prediction type, e.g. intra-frame, inter-frame or bidirectional frame prediction
  • H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
  • H04N 19/186 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being a colour or a chrominance component
  • H04N 19/30 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using hierarchical techniques, e.g. scalability

50.

CLUSTERING-BASED QUANTIZATION FOR NEURAL NETWORK COMPRESSION

      
Application Number 17622954
Status Pending
Filing Date 2020-07-01
First Publication Date 2022-08-18
Owner VID SCALE, INC. (USA)
Inventor
  • Li, Duanshun
  • Tian, Dong
  • Yang, Hua
  • He, Yuwen

Abstract

Systems, methods, and instrumentalities are disclosed for clustering-based quantization for neural network (NN) compression. A distribution of weights in weight tensors in NN layers may be analyzed to identify cluster outliers. Cluster inliers may be coded from cluster outliers, for example, using scalar and/or vector quantization. Weight-rearrangement may rearrange weights for higher dimensional weight tensors into lower dimensional matrices. For example, weight rearrangement may flatten a convolutional kernel into a vector. Correlation between kernels may be preserved, for example, by treating a filter or kernels across a channel as a point. A tensor may be split into multiple subspaces, for example, along an input and/or an output channel. Predictive coding may be performed for a current block of weights or weight matrix based on a reshaped or previously coded block or matrix. Arrangement, inlier, outlier, and/or prediction information may be signaled to a decoder for reconstruction of a compressed NN.

IPC Classes  ?

  • G06N 3/04 - Architecture, e.g. interconnection topology

51.

VIDEO-BASED POINT CLOUD STREAMS

      
Application Number 17613634
Status Pending
Filing Date 2020-05-21
First Publication Date 2022-07-28
Owner VID SCALE, INC. (USA)
Inventor
  • Ahmed, Hamza
  • He, Yong

Abstract

Systems, methods, and instrumentalities are disclosed herein that related to video-based point cloud streams in one or more ISO Base Media File Format (ISOBMFF) container files, A container format for point cloud data is provided and the container format indicates at least a relationship between a 3D region of the point cloud and one or more video-based point cloud compression (V-PCC) tracks. The V-PCC tracks may be grouped together and linked to the 3D region to allow spatial access to the 3D region.

IPC Classes  ?

  • H04N 19/597 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding specially adapted for multi-view video sequence encoding
  • H04N 19/132 - Sampling, masking or truncation of coding units, e.g. adaptive resampling, frame skipping, frame interpolation or high-frequency transform coefficient masking
  • H04N 19/169 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
  • H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock

52.

BLOCK BOUNDARY PREDICTION REFINEMENT WITH OPTICAL FLOW

      
Application Number 17616449
Status Pending
Filing Date 2020-06-02
First Publication Date 2022-07-28
Owner VID SCALE, INC. (USA)
Inventor
  • Chen, Wei
  • Luo, Jiancong
  • He, Yuwen

Abstract

Systems, methods, and instrumentalities are disclosed for sub-block/block refinement, including sub-block/block boundary refinement, such as block boundary prediction refinement with optical flow (BBPROF). A block comprising a current sub-block may be decoded based on a sample value for a first pixel that is obtained based on, for example, an MV for a current sub-block, an MV for a sub-block adjacent the current sub-block, and a sample value for a second pixel adjacent the first pixel. BBPROF may include determining spatial gradients at pixel(s)/sample location(s). An MV difference may be calculated between a current sub-block and one or more neighboring sub-blocks. An MV offset may be determined at pixel(s)/sample location(s) based on the MV difference. A sample value offset for the pixel in a current sub-block may be determined. The prediction for a reference picture list may be refined by adding the calculated sample value offset to the sub-block prediction.

IPC Classes  ?

  • H04N 19/137 - Motion inside a coding unit, e.g. average field, frame or block difference
  • H04N 19/132 - Sampling, masking or truncation of coding units, e.g. adaptive resampling, frame skipping, frame interpolation or high-frequency transform coefficient masking
  • H04N 19/182 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being a pixel
  • H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock

53.

Systems, apparatus and methods for inter prediction refinement with optical flow

      
Application Number 17426976
Grant Number 11695950
Status In Force
Filing Date 2020-02-04
First Publication Date 2022-06-30
Grant Date 2023-07-04
Owner VID SCALE, INC. (USA)
Inventor
  • Luo, Jiancong
  • He, Yuwen

Abstract

Method, apparatus and systems are disclosed. In one embodiment, a method of decoding includes obtaining a sub-block based motion prediction signal for a current block of the video; obtaining one or more spatial gradients of the sub-block based motion prediction signal or one or more motion vector difference values; obtaining a refinement signal for the current block based on the one or more obtained spatial gradients or the one or more obtained motion vector difference values; obtaining a refined motion prediction signal for the current block based on the sub-block based motion prediction signal and the refinement signal; and decoding the current block based on the refined motion prediction signal.

IPC Classes  ?

  • H04N 19/52 - Processing of motion vectors by encoding by predictive encoding
  • H04N 19/55 - Motion estimation with spatial constraints, e.g. at image or region borders
  • H04N 19/105 - Selection of the reference unit for prediction within a chosen coding or prediction mode, e.g. adaptive choice of position and number of pixels used for prediction
  • H04N 19/132 - Sampling, masking or truncation of coding units, e.g. adaptive resampling, frame skipping, frame interpolation or high-frequency transform coefficient masking
  • H04N 19/139 - Analysis of motion vectors, e.g. their magnitude, direction, variance or reliability
  • H04N 19/167 - Position within a video image, e.g. region of interest [ROI]
  • H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock

54.

Predictive coding for 360-degree video based on geometry padding

      
Application Number 17690996
Grant Number 11825116
Status In Force
Filing Date 2022-03-09
First Publication Date 2022-06-23
Grant Date 2023-11-21
Owner VID Scale, Inc. (USA)
Inventor
  • Xiu, Xiaoyu
  • He, Yuwen
  • Ye, Yan

Abstract

A video coding system (e.g., an encoder and/or a decoder) may perform face-based sub-block motion compensation for 360-degree video to predict samples (e.g., of a sub-block). The video coding system may receive a 360-degree video content. The 360-degree video content may include a current block. The current block may include a plurality of sub-blocks. The system may determine whether a sub-block mode is used for the current block. The system may predict a sample in the current block based on the sub-block level face association. For a first sub-block in the current block, the system may identify a first location of the first sub-block. The system may associate the first sub-block with a first face based on the identified first location of the first sub-block. The system may predict a first sample in the first sub-block based on the first face that is associated with the first sub-block.

IPC Classes  ?

  • H04N 19/563 - Motion estimation with padding, i.e. with filling of non-object values in an arbitrarily shaped picture block or region for estimation purposes
  • H04N 19/105 - Selection of the reference unit for prediction within a chosen coding or prediction mode, e.g. adaptive choice of position and number of pixels used for prediction
  • H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
  • H04N 19/597 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding specially adapted for multi-view video sequence encoding

55.

TEMPLATE-BASED INTER PREDICTION TECHNIQUES BASED ON ENCODING AND DECODING LATENCY REDUCTION

      
Application Number 17694168
Status Pending
Filing Date 2022-03-14
First Publication Date 2022-06-23
Owner VID SCALE, INC. (USA)
Inventor
  • Xiu, Xiaoyu
  • He, Yuwen
  • Ye, Yan

Abstract

Video coding methods are described for reducing latency in template-based inter coding. In some embodiments, a method is provided for coding a video that includes a current picture and at least one reference picture. For at least a current block in the current picture, a respective predicted value is generated (e.g. using motion compensated prediction) for each sample in a template region adjacent to the current block. Once the predicted values are generated for each sample in the template region, a process is invoked to determine a template-based inter prediction parameter by using predicted values in the template region and sample values the reference picture. This process can be invoked without waiting for reconstructed sample values in the template region. Template-based inter prediction of the current block is then performed using the determined template-based inter prediction parameter.

IPC Classes  ?

  • H04N 19/105 - Selection of the reference unit for prediction within a chosen coding or prediction mode, e.g. adaptive choice of position and number of pixels used for prediction
  • H04N 19/132 - Sampling, masking or truncation of coding units, e.g. adaptive resampling, frame skipping, frame interpolation or high-frequency transform coefficient masking
  • H04N 19/159 - Prediction type, e.g. intra-frame, inter-frame or bidirectional frame prediction
  • H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
  • H04N 19/184 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being bits, e.g. of the compressed video stream
  • H04N 19/46 - Embedding additional information in the video signal during the compression process
  • H04N 19/583 - Motion compensation with overlapping blocks
  • H04N 19/625 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding using discrete cosine transform [DCT]
  • H04N 19/64 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding using sub-band based transform, e.g. wavelets characterised by ordering of coefficients or of bits for transmission

56.

METHODS AND APPARATUS FOR SUB-PICTURE ADAPTIVE RESOLUTION CHANGE

      
Application Number 17438300
Status Pending
Filing Date 2020-03-11
First Publication Date 2022-06-16
Owner VID SCALE, INC. (USA)
Inventor
  • He, Yong
  • Neff, Ralph
  • He, Yuwen

Abstract

Methods and apparatus relate to picture and video coding in communication systems are provided. Included therein is a method comprising determining one or more layers associated with a parameter set, generating a syntax element including an indication indicating whether the one or more layers associated with the parameter set are independently coded, and generating a message including the syntax element.

IPC Classes  ?

  • H04N 19/59 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving spatial sub-sampling or interpolation, e.g. alteration of picture size or resolution
  • H04N 19/70 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards
  • H04N 19/33 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using hierarchical techniques, e.g. scalability in the spatial domain
  • H04N 19/172 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a picture, frame or field

57.

METHODS AND APPARATUS FOR PREDICTION REFINEMENT FOR DECODER SIDE MOTION VECTOR REFINEMENT WITH OPTICAL FLOW

      
Application Number 17442536
Status Pending
Filing Date 2020-03-26
First Publication Date 2022-06-16
Owner VID SCALE, INC. (USA)
Inventor
  • Chen, Wei
  • He, Yuwen
  • Luo, Jiancong

Abstract

Methods, devices, apparatus, systems, architectures and interfaces to improve motion vector (MV) refinement based sub-block (SB) level motion compensated prediction are provided. A decoding method includes receiving a bitstream of encoded video data, the bitstream including at least one block of video data including a plurality of SBs; performing a MV derivation, including a decoder based MV (DMVR) process, for at least one SB in the block to generate a refined MV for each SB; performing SB based motion compensation on the at least one sub-block to generate a SB based prediction within each SB; obtaining a spatial gradient for the prediction within each SB; determining a MV offset for each pixel in each SB; obtaining an intensity change in each SB based on the spatial gradients and MV offsets via an optical flow equation; and refining the prediction within each SB based on the obtained intensity changes.

IPC Classes  ?

  • H04N 19/137 - Motion inside a coding unit, e.g. average field, frame or block difference
  • H04N 19/513 - Processing of motion vectors
  • H04N 19/105 - Selection of the reference unit for prediction within a chosen coding or prediction mode, e.g. adaptive choice of position and number of pixels used for prediction
  • H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock

58.

METHODS AND SYSTEMS FOR POST-RECONSTRUCTION FILTERING

      
Application Number 17437377
Status Pending
Filing Date 2020-03-10
First Publication Date 2022-06-09
Owner Vid Scale, Inc. (USA)
Inventor
  • Wu, Shilin
  • Hanhart, Philippe
  • He, Yuwen

Abstract

Systems and methods are described for video coding using adaptive Hadamard filtering of reconstructed blocks, such as coding units. In some embodiments, where Hadamard filtering might otherwise encompass samples outside the current coding unit, extrapolated samples are generated for use in the filtering. Reconstructed samples from neighboring blocks may be used in the filtering where available (e.g. in a line buffer). In some embodiments, different filter strengths are applied to different spectrum components in the transform domain. In some embodiments, filter strength is based on position of filtered samples within the block. In some embodiments, filter strength is based on the prediction mode used to code the current block.

IPC Classes  ?

  • H04N 19/132 - Sampling, masking or truncation of coding units, e.g. adaptive resampling, frame skipping, frame interpolation or high-frequency transform coefficient masking
  • H04N 19/60 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding
  • H04N 19/80 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals - Details of filtering operations specially adapted for video compression, e.g. for pixel interpolation
  • H04N 19/159 - Prediction type, e.g. intra-frame, inter-frame or bidirectional frame prediction
  • H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock

59.

SYMMETRIC MERGE MODE MOTION VECTOR CODING

      
Application Number 17438119
Status Pending
Filing Date 2020-03-11
First Publication Date 2022-06-09
Owner VID SCALE, INC. (USA)
Inventor
  • Yang, Hua
  • He, Yuwen

Abstract

Systems, devices, and methods are described herein for symmetric merge mode motion vector coding. Symmetric bi-prediction (bi-pred) motion vectors (MVs) may be constructed from available candidates in a merge candidate list for regular inter prediction merge mode and/or affine prediction merge mode. Available MV merge candidates may be symmetrically extended or mapped in either direction (e.g., between reference pictures before and after a current picture), for example, when coding a picture that allows bi-directional motion compensation prediction (MCP). A symmetric bi-pred merge candidate may be selected among merge candidates for predicting the motion information of a current prediction unit (PU). The symmetric mapping construction may be repeated by a decoder (e.g., based on a coded index of the MV merge candidate list), for example, to obtain the same merge candidates and coded MV at an encoder.

IPC Classes  ?

  • H04N 19/52 - Processing of motion vectors by encoding by predictive encoding
  • H04N 19/30 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using hierarchical techniques, e.g. scalability
  • H04N 19/105 - Selection of the reference unit for prediction within a chosen coding or prediction mode, e.g. adaptive choice of position and number of pixels used for prediction
  • H04N 19/139 - Analysis of motion vectors, e.g. their magnitude, direction, variance or reliability
  • H04N 19/159 - Prediction type, e.g. intra-frame, inter-frame or bidirectional frame prediction
  • H04N 19/172 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a picture, frame or field
  • H04N 19/577 - Motion compensation with bidirectional frame interpolation, i.e. using B-pictures

60.

Method and system for signaling of 360-degree video information

      
Application Number 17671494
Grant Number 11876981
Status In Force
Filing Date 2022-02-14
First Publication Date 2022-06-02
Grant Date 2024-01-16
Owner VID SCALE, INC. (USA)
Inventor
  • Hanhart, Philippe
  • He, Yuwen
  • Ye, Yan

Abstract

Coding techniques for 360-degree video are described. An encoder selects a projection format and maps the 360-degree video to a 2D planar video using the selected projection format. The encoder encodes the 2D planar video in a bitstream and further signals, in the bitstream, parameters identifying the projection format. The parameters identifying the projection format may be signaled in a video parameter set, sequence parameter set, and/or picture parameter set of the bitstream. Different projection formats that may be signaled include formats using geometries such as equirectangular, cubemap, equal-area, octahedron, icosahedron, cylinder, and user-specified polygon. Other parameters that may be signaled include different arrangements of geometric faces or different encoding quality for different faces. Corresponding decoders are also described. In some embodiments, projection parameters may further include relative geometry rotation parameters that define an orientation of the projection geometry.

IPC Classes  ?

  • H04N 19/167 - Position within a video image, e.g. region of interest [ROI]
  • H04N 19/85 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using pre-processing or post-processing specially adapted for video compression
  • H04N 19/184 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being bits, e.g. of the compressed video stream
  • H04N 19/597 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding specially adapted for multi-view video sequence encoding

61.

METHODS AND APPARATUS FOR ADAPTIVE STREAMING OF POINT CLOUDS

      
Application Number 17436891
Status Pending
Filing Date 2020-03-06
First Publication Date 2022-05-26
Owner VID SCALE, INC. (USA)
Inventor
  • Hamza, Ahmed
  • He, Yong

Abstract

Methods, apparatus, and systems directed to adaptive streaming of V-PCC (Video-based Point Cloud Compression) data using an adaptive HTTP streaming protocol, such as MPEG DASH. A method includes signaling the point cloud data of the point cloud in a DASH MPD including: a main AdaptationSet for the point cloud, including at least (1) a @codecs attribute that is set to a unique value signifying that the corresponding AdaptationSet corresponds to V-PCC data and (2) an initialization segment containing at least one V-PCC sequence parameter set for a representation of the point cloud; and a plurality of component AdaptationSets, each corresponding to one of the V-PCC components and including at least (1) a VPCCComponent descriptor identifying a type of the corresponding V-PCC component and (2) at least one property of the V-PCC component; and transmitting the DASH bitstream over the network.

IPC Classes  ?

  • H04L 65/61 - Network streaming of media packets for supporting one-way streaming services, e.g. Internet radio
  • H04N 19/70 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards
  • H04L 67/02 - Protocols based on web technology, e.g. hypertext transfer protocol [HTTP]

62.

Symmetric motion vector difference coding

      
Application Number 17416205
Grant Number 11546604
Status In Force
Filing Date 2019-12-19
First Publication Date 2022-05-12
Grant Date 2023-01-03
Owner VID SCALE, Inc. (USA)
Inventor
  • Luo, Jiancong
  • Xiu, Xiaoyu
  • He, Yuwen
  • Yang, Hua

Abstract

Bi-directional optical flow (BDOF) may be bypassed, for a current coding block, based on whether symmetric motion vector difference (8MVD) is used in motion vector coding for the current coding block, A coding device (e.g., an encoder or a decoder) may determine whether to bypass BDOF for the current coding block based at least in part on an SMVD indication for the current coding block, The coding device may obtain the SMVD indication that indicates whether SMVD is used in motion vector coding for the current coding block. If SMVD Indication indicates that SMVD is used in the motion vector coding for the current coding block, the coding device may bypass BDOF for the current coding block. The coding device may reconstruct, the current coding block without performing BDOF if it determines to bypass BDOF for the current coding block.

IPC Classes  ?

  • H04N 19/137 - Motion inside a coding unit, e.g. average field, frame or block difference
  • H04N 19/105 - Selection of the reference unit for prediction within a chosen coding or prediction mode, e.g. adaptive choice of position and number of pixels used for prediction
  • H04N 19/159 - Prediction type, e.g. intra-frame, inter-frame or bidirectional frame prediction
  • H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
  • H04N 19/46 - Embedding additional information in the video signal during the compression process

63.

Affine motion model derivation method

      
Application Number 17434974
Grant Number 11870997
Status In Force
Filing Date 2020-02-28
First Publication Date 2022-05-12
Grant Date 2024-01-09
Owner VID SCALE, INC. (USA)
Inventor
  • Luo, Jiancong
  • He, Yuwen
  • Chen, Wei

Abstract

Systems and methods are described for video coding using affine motion prediction. In an example method, motion vector gradients are determined from respective motion vectors of a plurality of neighboring sub-blocks neighboring a current block. An estimate of at least one affine parameter for the current block is determined based on the motion vector gradients. An affine motion model is determined based at least in part on the estimated affine parameter(s), and a prediction of the current block is generated using the affine motion model. The estimated parameter(s) may be used in the affine motion model itself. Alternatively, the estimated parameter(s) may be used in a prediction of the affine motion model. In some embodiments, only neighboring sub-blocks above and/or to the left of the current block are used in estimating the affine parameter(s).

IPC Classes  ?

  • H04N 19/137 - Motion inside a coding unit, e.g. average field, frame or block difference
  • H04N 19/105 - Selection of the reference unit for prediction within a chosen coding or prediction mode, e.g. adaptive choice of position and number of pixels used for prediction
  • H04N 19/149 - Data rate or code amount at the encoder output by estimating the code amount by means of a model, e.g. mathematical model or statistical model
  • H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock

64.

INTRA SUB-PARTITIONS IN VIDEO CODING

      
Application Number 17438042
Status Pending
Filing Date 2020-03-11
First Publication Date 2022-05-12
Owner VID SCALE, INC. (USA)
Inventor
  • Yang, Hua
  • Vanam, Rahul
  • He, Yuwen

Abstract

Intra sub-partitions (ISP) may be enabled for a current block, for example, based on an ISP indication. The block may be partitioned into multiple sub-partitions, and a sub-partition may belong to a prediction unit (PU). A sub-partition width for the current block and a minimum prediction block width may be obtained. A PU corresponding to a current sub-partition may be determined based on the sub-partition width and the minimum prediction block width. For example, when the sub-partition width is less than the minimum prediction block width, the PU may include multiple sub-partitions. In examples, the minimum prediction block width may be four samples. Reference samples may be determined, and the PU may be predicted using the reference samples.

IPC Classes  ?

  • H04N 19/119 - Adaptive subdivision aspects e.g. subdivision of a picture into rectangular or non-rectangular coding blocks
  • H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
  • H04N 19/593 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving spatial prediction techniques
  • H04N 19/61 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding in combination with predictive coding
  • H04N 19/105 - Selection of the reference unit for prediction within a chosen coding or prediction mode, e.g. adaptive choice of position and number of pixels used for prediction

65.

SUB-PICTURE BITSTREAM EXTRACTION AND REPOSITION

      
Application Number 17435669
Status Pending
Filing Date 2020-03-11
First Publication Date 2022-05-05
Owner VID SCALE, INC. (USA)
Inventor He, Yong

Abstract

Systems and methods described herein employ a high-level syntax design that supports a sub-picture extraction and reposition process. An input video may be encoded into multiple representations, each representation may be represented as a layer. A layer picture may be partitioned into multiple sub-pictures. Each sub-picture may have its own tile partitioning, resolution, color format and bit depth. Each sub-picture is encoded independently from other sub-pictures of the same layer, but it may be inter-predicted from the corresponding sub-pictures from its dependent layers. Each sub-picture may refer to a sub-picture parameter set where the sub-picture properties such as resolution and coordinate is signaled. Each sub-picture parameter set may refer to a PPS where the resolution of the entire picture is signaled.

IPC Classes  ?

  • H04N 19/597 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding specially adapted for multi-view video sequence encoding
  • H04N 19/70 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards
  • H04N 19/172 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a picture, frame or field
  • H04N 19/30 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using hierarchical techniques, e.g. scalability

66.

INTER PREDICTION BANDWIDTH REDUCTION METHOD WITH OPTICAL FLOW COMPENSATION

      
Application Number 17439760
Status Pending
Filing Date 2020-03-16
First Publication Date 2022-04-28
Owner Vid Scale, Inc. (USA)
Inventor
  • Luo, Jiancong
  • He, Yuwen
  • Chen, Wei

Abstract

Systems and methods are described for video coding. In some embodiments, inter prediction of a sample in a current block is performed by rounding an initial motion vector and determining a rounding error vector caused by the rounding. An unrefined prediction of the sample is generated using the rounded motion vector. Unrefined predictions are similarly generated for other samples in the current block. Based on the unrefined predictions, a spatial gradient is determined for each sample position in the block. A refined prediction is generated for each sample position by adding, to the unrefined prediction, a scalar product between the spatial gradient and the rounding error vector at the sample position. Example methods can reduce the number of reference pixels used to predict a current block and thus may reduce memory access bandwidth.

IPC Classes  ?

  • H04N 19/137 - Motion inside a coding unit, e.g. average field, frame or block difference
  • H04N 19/105 - Selection of the reference unit for prediction within a chosen coding or prediction mode, e.g. adaptive choice of position and number of pixels used for prediction
  • H04N 19/132 - Sampling, masking or truncation of coding units, e.g. adaptive resampling, frame skipping, frame interpolation or high-frequency transform coefficient masking
  • H04N 19/117 - Filters, e.g. for pre-processing or post-processing
  • H04N 19/82 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals - Details of filtering operations specially adapted for video compression, e.g. for pixel interpolation involving filtering within a prediction loop
  • H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock

67.

CONTENT ADAPTIVE TRANSFORM PRECISION FOR VIDEO CODING

      
Application Number 17442699
Status Pending
Filing Date 2020-03-25
First Publication Date 2022-04-28
Owner VID SCALE, INC. (USA)
Inventor
  • Kerofsky, Louis
  • He, Yuwen
  • Hanhart, Philippe

Abstract

Systems, methods, and instrumentalities are disclosed for obtaining coded video data comprising quantized transform coefficients for a plurality of blocks, obtaining a first precision factor associated with a first block for performing at least one decoding function on the first block, obtaining a second precision factor associated with a second block for performing the at least one decoding function on the second block, and performing the at least one decoding function on the quantized transform coefficients for the first block using the first precision factor and on the quantized transform coefficients for the second block using the second precision factor.

IPC Classes  ?

  • H04N 19/12 - Selection from among a plurality of transforms or standards, e.g. selection between discrete cosine transform [DCT] and sub-band transform or selection between H.263 and H.264
  • H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
  • H04N 19/18 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being a set of transform coefficients
  • H04N 19/126 - Quantisation - Details of normalisation or weighting functions, e.g. normalisation matrices or variable uniform quantisers
  • H04N 19/14 - Coding unit complexity, e.g. amount of activity or edge presence estimation
  • H04N 19/60 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding

68.

IMPROVED INTRA PLANAR PREDICTION USING MERGE MODE MOTION VECTOR CANDIDATES

      
Application Number 17421973
Status Pending
Filing Date 2020-01-10
First Publication Date 2022-04-14
Owner VID SCALE, INC. (USA)
Inventor
  • Vanam, Rahul
  • He, Yuwen

Abstract

Methods, procedures, architectures, apparatuses, systems, devices, interfaces, and computer program products for encoding/decoding data (e.g. a data stream) are provided. A video coding method for predicting a current block includes identifying a first block adjacent to the current block, the first block having motion information, performing motion compensation using the motion information to generate a set of reference samples adjacent to the current block, identifying a first line of reference samples from the set of generated reference samples to be used for intra prediction of the current block, and performing intra prediction of the current block using at least the first line of reference samples.

IPC Classes  ?

  • H04N 19/593 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving spatial prediction techniques
  • H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
  • H04N 19/137 - Motion inside a coding unit, e.g. average field, frame or block difference
  • H04N 19/132 - Sampling, masking or truncation of coding units, e.g. adaptive resampling, frame skipping, frame interpolation or high-frequency transform coefficient masking
  • H04N 19/105 - Selection of the reference unit for prediction within a chosen coding or prediction mode, e.g. adaptive choice of position and number of pixels used for prediction
  • H04N 19/11 - Selection of coding mode or of prediction mode among a plurality of spatial predictive coding modes

69.

Methods, architectures, apparatuses and systems directed to improved linear model estimation for template based video coding

      
Application Number 17415176
Grant Number 11909977
Status In Force
Filing Date 2019-12-20
First Publication Date 2022-03-24
Grant Date 2024-02-20
Owner Vid Scale, Inc. (USA)
Inventor
  • Bandyopadhyay, Saurav
  • Xiu, Xiaoyu
  • He, Yuwen

Abstract

Procedures, methods, architectures, apparatuses, systems, devices, and computer program products directed to improved linear model estimation for template-based video coding are provided. Included therein is a method comprising determining minimum and maximum (“min/max”) values of luma and chroma samples neighboring a coding block, wherein the min/max chroma values correspond to the min/max luma values; determining a first linear model parameter of a template-based video coding technique (i) based on a single look-up table and the min/max chroma values and (ii) at a precision no greater than 16 bits; determining a second linear model parameter of the template-based video coding technique (i) based on the first linear model parameter and the minimum chroma and luma values and (ii) at a precision no greater than 16 bits; and predicting chroma samples of the coding block based on reconstructed luma samples of the coding block and the first and second linear model parameters.

IPC Classes  ?

  • H04N 19/132 - Sampling, masking or truncation of coding units, e.g. adaptive resampling, frame skipping, frame interpolation or high-frequency transform coefficient masking
  • H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
  • H04N 19/186 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being a colour or a chrominance component
  • H04N 19/105 - Selection of the reference unit for prediction within a chosen coding or prediction mode, e.g. adaptive choice of position and number of pixels used for prediction

70.

COMBINED INTER AND INTRA PREDICTION

      
Application Number 17420080
Status Pending
Filing Date 2019-12-20
First Publication Date 2022-03-03
Owner VID SCALE, INC. (USA)
Inventor
  • Xiu, Xiaoyu
  • He, Yuwen
  • Luo, Jiancong

Abstract

Systems, methods, and instrumentalities are disclosed for a combined inter and intra prediction, A video coding device may receive a motion vector difference (MMVD) mode indication that indicates whether MMVD mode is used to generate inter prediction of a coding unit (CU). The video coding device may receive a combined inter merge/intra prediction (CUP) indication, for example, when the MMVD mode indication indicates that MMVD mode is not used to generate the inter prediction of the CU, The video coding device may determine whether to use triangle merge mode for the CU, for example, based on the MMVD mode indication and/or the CUP indication. On a condition that the CUP indication indicates that CUP is applied for the CU or the MMVD mode indication indicates that MMVD mode is used to generate the inter prediction, the video coding device may disable the triangle merge mode for the CU.

IPC Classes  ?

  • H04N 19/105 - Selection of the reference unit for prediction within a chosen coding or prediction mode, e.g. adaptive choice of position and number of pixels used for prediction
  • H04N 19/159 - Prediction type, e.g. intra-frame, inter-frame or bidirectional frame prediction
  • H04N 19/46 - Embedding additional information in the video signal during the compression process
  • H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock

71.

Methods and apparatus for point cloud compression bitstream format

      
Application Number 17277064
Grant Number 11568573
Status In Force
Filing Date 2019-09-11
First Publication Date 2022-02-17
Grant Date 2023-01-31
Owner Vid Scale, Inc. (USA)
Inventor
  • Hamza, Ahmed
  • He, Yong

Abstract

Methods, apparatus, systems, architectures and interfaces for encoding and/or decoding point cloud bitstreams including coded point cloud sequences are provided. Included among such methods, apparatuses, systems, architectures, and interfaces is an apparatus that may include a processor and memory. A method may include any of: mapping components of the point cloud bitstream into tracks; generating information identifying any of geometry streams or texture streams according to the mapping of the components; generating information associated with layers corresponding to respective geometry component streams; and generating information indicating operation points associated with the point cloud bitstream.

IPC Classes  ?

  • G06T 9/00 - Image coding
  • H04N 19/44 - Decoders specially adapted therefor, e.g. video decoders which are asymmetric with respect to the encoder
  • H04N 19/70 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards
  • G06T 15/00 - 3D [Three Dimensional] image rendering
  • G06T 15/08 - Volume rendering
  • G06T 15/10 - Geometric effects

72.

Metrics and messages to improve experience for 360-degree adaptive streaming

      
Application Number 17505982
Grant Number 11849178
Status In Force
Filing Date 2021-10-20
First Publication Date 2022-02-10
Grant Date 2023-12-19
Owner VID SCALE, Inc. (USA)
Inventor
  • He, Yong
  • Ye, Yan
  • Begen, Ali C.

Abstract

A method for receiving and displaying media content may be provided. The method may include requesting a set of DASH video segments that are associated with various viewports and qualities. The method may include displaying the DASH video segments. The method may include determining a latency metric based on a time difference between the display of a DASH video segment and one of: a device beginning to move, the device ceasing to move, the device determining that the device has begun to move, the device determining that the device has stopped moving, or the display of a different DASH video segment. The different DASH video segment may be associated with one or more of a different quality or a different viewport.

IPC Classes  ?

  • H04N 21/81 - Monomedia components thereof
  • H04N 21/442 - Monitoring of processes or resources, e.g. detecting the failure of a recording device, monitoring the downstream bandwidth, the number of times a movie has been viewed or the storage space available from the internal hard disk
  • H04N 21/2343 - Processing of video elementary streams, e.g. splicing of video streams or manipulating MPEG-4 scene graphs involving reformatting operations of video signals for distribution or compliance with end-user requests or end-user device requirements
  • H04N 21/4728 - End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification or for manipulating displayed content for selecting a ROI [Region Of Interest], e.g. for requesting a higher resolution version of a selected region

73.

METHODS AND APPARATUS FOR FLEXIBLE GRID REGIONS

      
Application Number 17275249
Status Pending
Filing Date 2019-09-13
First Publication Date 2022-02-03
Owner VID SCALE, INC. (USA)
Inventor
  • He, Yong
  • Ye, Yan
  • Hamza, Ahmed

Abstract

Methods and apparatus for using flexible grid regions in picture or video frames are disclosed. In one embodiment, a method includes receiving a set of first parameters that defines a plurality of first grid regions comprising a frame. For each first grid region, the method includes receiving a set of second parameters that defines a plurality of second grid regions, and the plurality of second grid regions partitions the respective first grid region. The method further includes partitioning the frame into the plurality of first grid regions based on the set of first parameters, and partitioning each first grid region into the plurality of second grid regions based on the respective set of second parameters.

IPC Classes  ?

  • H04N 19/563 - Motion estimation with padding, i.e. with filling of non-object values in an arbitrarily shaped picture block or region for estimation purposes
  • H04N 19/119 - Adaptive subdivision aspects e.g. subdivision of a picture into rectangular or non-rectangular coding blocks
  • H04N 19/174 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a slice, e.g. a line of blocks or a group of blocks
  • H04N 19/117 - Filters, e.g. for pre-processing or post-processing

74.

GENERALIZED BI-PREDICTION FOR VIDEO CODING WITH REDUCED CODING COMPLEXITY

      
Application Number 17504150
Status Pending
Filing Date 2021-10-18
First Publication Date 2022-02-03
Owner Vid Scale, Inc. (USA)
Inventor
  • Zhang, Yan
  • Xiu, Xiaoyu
  • He, Yuwen
  • Ye, Yan

Abstract

Exemplary embodiments include systems and methods for coding a video comprising a plurality of pictures including a current picture, a first reference picture, and a second reference picture, where each picture includes a plurality of blocks. In one method, for at least a current block in the current picture, a number of available bi-prediction weights is determined based at least in part on a temporal layer and/or a quantization parameter of the current picture. From among available bi-prediction weights a pair of weights are identified. Using the identified weights, the current block is then predicted as a weighted sum of a first reference block in the first reference picture and a second reference block in the second reference picture. Encoding techniques are also described for efficient searching and selection of a pair of bi-prediction weights to use for prediction of a block.

IPC Classes  ?

  • H04N 19/577 - Motion compensation with bidirectional frame interpolation, i.e. using B-pictures
  • H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
  • H04N 19/31 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using hierarchical techniques, e.g. scalability in the temporal domain
  • H04N 19/573 - Motion compensation with multiple frame prediction using two or more reference frames in a given prediction direction

75.

Complexity reduction and bit-width control for bi-directional optical flow

      
Application Number 17276448
Grant Number 11470308
Status In Force
Filing Date 2019-09-17
First Publication Date 2022-02-03
Grant Date 2022-10-11
Owner VID SCALE, INC. (USA)
Inventor
  • Xiu, Xiaoyu
  • He, Yuwen
  • Ye, Yan
  • Luo, Jiancong

Abstract

Systems and methods are described for reducing the complexity of using bi-directional optical flow (BIO) in video coding. In some embodiments, bit-width reduction steps are introduced in the BIO motion refinement process to reduce the maximum bit-width used for BIO calculations. In some embodiments, simplified interpolation filters are used to generate predicted samples in an extended region around a current coding unit. In some embodiments, different interpolation filters are used for vertical versus horizontal interpolation. In some embodiments, BIO is disabled for coding units with small heights and/or for coding units that are predicted using a sub-block level inter prediction technique, such as advanced temporal motion vector prediction (ATMVP) or affine prediction.

IPC Classes  ?

  • H04N 7/12 - Systems in which the television signal is transmitted via one channel or a plurality of parallel channels, the bandwidth of each channel being less than the bandwidth of the television signal
  • H04N 19/105 - Selection of the reference unit for prediction within a chosen coding or prediction mode, e.g. adaptive choice of position and number of pixels used for prediction
  • H04N 19/132 - Sampling, masking or truncation of coding units, e.g. adaptive resampling, frame skipping, frame interpolation or high-frequency transform coefficient masking
  • H04N 19/146 - Data rate or code amount at the encoder output
  • H04N 19/159 - Prediction type, e.g. intra-frame, inter-frame or bidirectional frame prediction
  • H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
  • H04N 19/513 - Processing of motion vectors

76.

Sample derivation for 360-degree video coding

      
Application Number 17280183
Grant Number 11601676
Status In Force
Filing Date 2019-09-25
First Publication Date 2022-01-06
Grant Date 2023-03-07
Owner VID Scale, Inc. (USA)
Inventor
  • Hanhart, Philippe
  • He, Yuwen
  • Ye, Yan

Abstract

Systems, methods, and instrumentalities are disclosed for performing horizontal geometry padding on a current sample based on receiving a wraparound enabled indication that indicates whether a horizontal wraparound motion compensation is enabled. If the horizontal wraparound motion compensation is enabled based on the wraparound enabled indication, a video coding device may determine a reference sample wraparound offset of a current sample in a picture. The reference sample wraparound offset may indicate a face width of the picture. The video coding device may determine a reference sample location for the current sample based on the reference sample wraparound offset, a picture width of the picture, and a current sample location. The video coding device may predict the current sample based on the reference sample location in a horizontal direction. Repetitive padding or clipping may be used in the vertical direction.

IPC Classes  ?

  • H04N 19/597 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding specially adapted for multi-view video sequence encoding
  • H04N 19/172 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a picture, frame or field
  • H04N 19/184 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being bits, e.g. of the compressed video stream
  • H04N 19/105 - Selection of the reference unit for prediction within a chosen coding or prediction mode, e.g. adaptive choice of position and number of pixels used for prediction
  • H04N 19/132 - Sampling, masking or truncation of coding units, e.g. adaptive resampling, frame skipping, frame interpolation or high-frequency transform coefficient masking
  • H04N 19/51 - Motion estimation or motion compensation

77.

Bl-prediction for video coding

      
Application Number 17280175
Grant Number 11553206
Status In Force
Filing Date 2019-09-26
First Publication Date 2022-01-06
Grant Date 2023-01-10
Owner VID SCALE, Inc. (USA)
Inventor
  • He, Yuwen
  • Luo, Jiancong
  • Xiu, Xiaoyu
  • Ye, Yan

Abstract

Systems, methods, and instrumentalities may be provided for determining whether to bypass bi-directional optical flow (BDOF) if BDOF is used in combination with bi-prediction with coding unit (CU) weights (e.g., generalized bi-prediction (GBi)). A coding system may combine coding modes, coding techniques, and/or coding tools. The coding system may include a wireless transmit/receive unit (WTRU). For example, the coding system may combine BDOF and bi-prediction with GU weights (BCW). BDOF may include refining a motion vector associated with a current CU based at least in part on gradients associated with a location in the current CU. The coding system may determine that BDOF is enabled, and/or that bi-prediction with CU weights is enabled for the current CU. The coding system's determination that bi-prediction with CU weights is enabled and/or that BDOF is enabled may be based on one or more indications.

IPC Classes  ?

  • H04N 19/577 - Motion compensation with bidirectional frame interpolation, i.e. using B-pictures

78.

TILE GROUP PARTITIONING

      
Application Number 17311200
Status Pending
Filing Date 2019-12-04
First Publication Date 2022-01-06
Owner VID SCALE, INC. (USA)
Inventor He, Yong

Abstract

A system may identify a defined rectangular picture area and render video corresponding to the defined rectangular picture area. The system may receive a video bitstream comprising a picture having a header and may receive data specifying a structure of the picture. The system may parse the data specifying the structure of the picture for an identifier corresponding to a defined rectangular area in the first picture and for a tile index of a top left tile in the defined rectangular area. The system may determine one or more tiles comprised in the defined rectangular area based on the identifier corresponding to the defined rectangular area and the tile index of the top left tile. The system may reconstruct the picture including a sub-picture that comprises the defined rectangular area based upon the identifier corresponding to the defined rectangular area. The computing system may render the sub-picture in the defined rectangular area.

IPC Classes  ?

  • H04N 19/119 - Adaptive subdivision aspects e.g. subdivision of a picture into rectangular or non-rectangular coding blocks
  • H04N 19/174 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a slice, e.g. a line of blocks or a group of blocks
  • H04N 19/105 - Selection of the reference unit for prediction within a chosen coding or prediction mode, e.g. adaptive choice of position and number of pixels used for prediction
  • H04N 19/167 - Position within a video image, e.g. region of interest [ROI]

79.

Early packet loss detection and feedback

      
Application Number 17395342
Grant Number 11824664
Status In Force
Filing Date 2021-08-05
First Publication Date 2021-12-02
Grant Date 2023-11-21
Owner VID Scale, Inc. (USA)
Inventor
  • Liu, Weimin
  • Vanam, Rahul
  • Ma, Liangping
  • Reznik, Yuriy
  • Sternberg, Gregory S.
  • Chen, Wei
  • Veer, Dharm

Abstract

A video encoding device (e.g., a wireless transmit/receive unit (WTRU)) may transmit an encoded frame with a frame sequence number using a transmission protocol. The video encoding device, an application on the video encoding device, and/or a protocol layer on the encoding device may detect a packet loss by receiving an error notification. The packet loss may be detected at the MAC layer. The packet loss may be signaled using spoofed packets, such as a spoofed NAM packet, a spoofed XR packet, or a spoofed ACK packet. A lost packet may be retransmitted at the MAC layer (e.g., by the encoding device or another device on the wireless path). Packet loss detection may be performed in uplink operations and/or downlink operations, and/or may be performed in video gaining applications via the cloud. The video encoding device may generate and send a second encoded frame based on the error notification.

IPC Classes  ?

  • H04L 1/1867 - Arrangements specially adapted for the transmitter end
  • H04W 28/04 - Error control
  • H04L 43/0823 - Errors, e.g. transmission errors
  • H04N 19/164 - Feedback from the receiver or from the transmission channel
  • H04N 21/6437 - RTP [Real-time Transport Protocol]
  • H04L 1/18 - Automatic repetition systems, e.g. Van Duuren systems
  • H04N 19/107 - Selection of coding mode or of prediction mode between spatial and temporal predictive coding, e.g. picture refresh
  • H04N 21/647 - Control signaling between network components and server or clients; Network processes for video distribution between server and clients, e.g. controlling the quality of the video stream, by dropping packets, protecting content from unauthorised alteration within the network, monitoring of network load or bridging bet
  • H04N 19/105 - Selection of the reference unit for prediction within a chosen coding or prediction mode, e.g. adaptive choice of position and number of pixels used for prediction
  • H04N 19/172 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a picture, frame or field
  • H04L 47/74 - Admission control; Resource allocation measures in reaction to resource unavailability
  • H04L 47/50 - Queue scheduling
  • H04W 8/04 - Registration at HLR or HSS [Home Subscriber Server]

80.

ADAPTIVE QUANTIZATION METHOD FOR 360-DEGREE VIDEO CODING

      
Application Number 16625144
Status Pending
Filing Date 2018-06-21
First Publication Date 2021-10-28
Owner VID SCALE, INC. (USA)
Inventor
  • Xiu, Xiaoyu
  • He, Yuwen
  • Ye, Yan

Abstract

Systems, procedures, and instrumentalities may be provided for adaptively adjusting quantization parameters (QPs) for 360-degree video coding. For example, a first luma QP for a first region may be identified. Based on the first luma QP, a first chroma QP for the first region may be determined. A QP offset for a second region may be identified. A second luma QP for the second region may be determined based on the first luma QP and/or the QP offset for the second region. A second chroma QP of the second region may be determined based on the first chroma QP and/or the QP offset for the second region. An inverse quantization may be performed for the second region based on the second luma QP for the second region and/or the second chroma QP for the second region. The QP offset may be adapted based on a spherical sampling density.

IPC Classes  ?

  • H04N 19/124 - Quantisation
  • H04N 19/186 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being a colour or a chrominance component
  • H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
  • H04N 19/167 - Position within a video image, e.g. region of interest [ROI]

81.

360-degree video delivery over next generation network

      
Application Number 17245788
Grant Number 11770594
Status In Force
Filing Date 2021-04-30
First Publication Date 2021-10-14
Grant Date 2023-09-26
Owner VID SCALE, Inc. (USA)
Inventor
  • He, Yong
  • Ye, Yan

Abstract

Systems, methods, and instrumentalities are disclosed for a 360-degree video streaming. A video streaming device may receive a 360-degree video stream from a network node. The video streaming device may determine a viewport associated with the video streaming device and/or the 360-degree video stream. The video streaming device may determine (e.g., based on the viewport) to request in advance a first segment and a second segment of the 360-degree video stream. The video streaming device may determine a relative priority order for the first segment and the second segment. The video streaming device may generate an anticipated requests message. The anticipated requests message may indicate the determined relative priority order, for example, by listing the first segment and the second segment in decreasing relative priority based on the determined relative priority order. The video streaming device may send the anticipated requests message to the network node.

IPC Classes  ?

  • H04N 21/81 - Monomedia components thereof
  • H04L 65/60 - Network streaming of media packets
  • H04L 65/80 - Responding to QoS
  • H04L 67/02 - Protocols based on web technology, e.g. hypertext transfer protocol [HTTP]
  • H04N 21/218 - Source of audio or video content, e.g. local disk arrays
  • H04N 21/44 - Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to MPEG-4 scene graphs
  • H04N 21/472 - End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification or for manipulating displayed content
  • H04N 21/6587 - Control parameters, e.g. trick play commands or viewpoint selection
  • H04N 21/845 - Structuring of content, e.g. decomposing content into time segments

82.

Methods and apparatus for reducing the coding latency of decoder-side motion refinement

      
Application Number 17256155
Grant Number 11546628
Status In Force
Filing Date 2019-06-20
First Publication Date 2021-09-02
Grant Date 2023-01-03
Owner Vid Scale, Inc. (USA)
Inventor
  • Xiu, Xiaoyu
  • He, Yuwen
  • Ye, Yan

Abstract

Embodiments of video coding systems and methods are described for reducing coding latency introduced by decoder-side motion vector refinement (DMVR). In one example, two non-refined motion vectors are identified for coding of a first block of samples (e.g. a first coding unit) using bi-prediction. One or both of the non-refined motion vectors are used to predict motion information for a second block of samples (e.g. a second coding unit). The two non-refined motion vectors are refined using DMVR, and the refined motion vectors are used to generate a prediction signal of the first block of samples. Such embodiments allow the second block of samples to be coded substantially in parallel with the first block without waiting for completion of DMVR on the first block. In additional embodiments, optical-flow-based techniques are described for motion vector refinement.

IPC Classes  ?

  • H04N 19/577 - Motion compensation with bidirectional frame interpolation, i.e. using B-pictures
  • H04N 19/513 - Processing of motion vectors
  • H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
  • H04N 19/86 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using pre-processing or post-processing specially adapted for video compression involving reduction of coding artifacts, e.g. of blockiness

83.

Power aware video decoding and streaming

      
Application Number 17323139
Grant Number 11516485
Status In Force
Filing Date 2021-05-18
First Publication Date 2021-09-02
Grant Date 2022-11-29
Owner VID SCALE, Inc. (USA)
Inventor
  • He, Yuwen
  • Ye, Yan
  • He, Yong
  • Mcclellan, George W.
  • Ryu, Eun Seok

Abstract

Methods and systems are disclosed for a mobile device to decode video based on available power and/or energy. For example, the mobile device may receive a media description file (MDF) from for a video stream from a video server. The MDF may include complexity information associated with a plurality of video segments. The complexity information may be related to the amount of processing power to be utilized for decoding the segment at the mobile device. The mobile device may determine at least one power metric for the mobile device. The mobile device may determine a first complexity level to be requested for a first video segment based on the complexity information from the MDF and the power metric. The mobile device may dynamically alter the decoding process to save energy based on the detected power/energy level.

IPC Classes  ?

  • H04N 7/12 - Systems in which the television signal is transmitted via one channel or a plurality of parallel channels, the bandwidth of each channel being less than the bandwidth of the television signal
  • H04N 21/442 - Monitoring of processes or resources, e.g. detecting the failure of a recording device, monitoring the downstream bandwidth, the number of times a movie has been viewed or the storage space available from the internal hard disk
  • H04N 19/187 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being a scalable video layer
  • G06F 1/3212 - Monitoring battery levels, e.g. power saving mode being initiated when battery voltage goes below a certain level
  • G06F 1/329 - Power saving characterised by the action undertaken by task scheduling
  • H04N 19/117 - Filters, e.g. for pre-processing or post-processing
  • H04N 19/127 - Prioritisation of hardware or computational resources
  • H04N 19/136 - Incoming video signal characteristics or properties
  • H04N 19/156 - Availability of hardware or computational resources, e.g. encoding based on power-saving criteria
  • H04N 19/44 - Decoders specially adapted therefor, e.g. video decoders which are asymmetric with respect to the encoder
  • H04N 21/2343 - Processing of video elementary streams, e.g. splicing of video streams or manipulating MPEG-4 scene graphs involving reformatting operations of video signals for distribution or compliance with end-user requests or end-user device requirements
  • H04N 21/414 - Specialised client platforms, e.g. receiver in car or embedded in a mobile appliance
  • H04N 21/845 - Structuring of content, e.g. decomposing content into time segments
  • H04N 21/854 - Content authoring
  • H04L 65/612 - Network streaming of media packets for supporting one-way streaming services, e.g. Internet radio for unicast
  • H04L 65/613 - Network streaming of media packets for supporting one-way streaming services, e.g. Internet radio for the control of the source by the destination

84.

Adaptive control point selection for affine motion model based video coding

      
Application Number 17256769
Grant Number 11503328
Status In Force
Filing Date 2019-06-27
First Publication Date 2021-09-02
Grant Date 2022-11-15
Owner VID SCALE, Inc. (USA)
Inventor
  • He, Yuwen
  • Xiu, Xiaoyu
  • Ye, Yan
  • Hanhart, Philippe

Abstract

Systems, methods, and instrumentalities are disclosed for motion vector clipping when affine motion mode is enabled for a video block. A video coding device may determine that an affine mode for a video block is enabled. The video coding device may determine a plurality of control point affine motion vectors associated with the video block. The video coding device may store the plurality of clipped control point affine motion vectors for motion vector prediction of a neighboring control point affine motion vector. The video coding device may derive a sub-block motion vector associated with a sub-block of the video block, clip the derived sub-block motion vector, and store it for spatial motion vector prediction or temporal motion vector prediction. For example, the video coding device may clip the derived sub-block motion vector based on a motion field range that may be based on a bit depth value.

IPC Classes  ?

  • H04N 19/00 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
  • H04N 19/52 - Processing of motion vectors by encoding by predictive encoding
  • H04N 19/54 - Motion estimation other than block-based using feature points or meshes

85.

SYSTEMS AND METHODS FOR RGB VIDEO CODING ENHANCEMENT

      
Application Number 17211498
Status Pending
Filing Date 2021-03-24
First Publication Date 2021-09-02
Owner VID SCALE, INC. (USA)
Inventor
  • Xiu, Xiaoyu
  • He, Yuwen
  • Tsai, Chia-Ming
  • Ye, Yan

Abstract

Systems, methods, and devices are disclosed for performing adaptive residue color space conversion. A video bitstream may be received and a first flag may be determined based on the video bitstream. A residual may also be generated based on the video bitstream. The residual may be converted from a first color space to a second color space in response to the first flag.

IPC Classes  ?

  • H04N 19/44 - Decoders specially adapted therefor, e.g. video decoders which are asymmetric with respect to the encoder
  • H04N 19/46 - Embedding additional information in the video signal during the compression process
  • H04N 19/174 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a slice, e.g. a line of blocks or a group of blocks
  • H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
  • H04N 19/12 - Selection from among a plurality of transforms or standards, e.g. selection between discrete cosine transform [DCT] and sub-band transform or selection between H.263 and H.264

86.

INTRA BLOCK COPY MODE FOR SCREEN CONTENT CODING

      
Application Number 17240473
Status Pending
Filing Date 2021-04-26
First Publication Date 2021-08-12
Owner VID SCALE, INC. (USA)
Inventor
  • Xiu, Xiaoyu
  • Ye, Yan
  • He, Yuwen

Abstract

A video block of a current picture may be coded in an intra block copy (IBC) mode. Weighted prediction may be disabled for the IBC-coded screen content video block. Fractional block vectors may be used for the chroma components of the IBC-coded video block. An interpolation filter may be utilized to generate chroma prediction samples for the video block. A decoded version of the current reference picture may be added to both reference picture list L0 and reference picture list L1 that are associated with the IBC-coded video block. When constrained intra prediction is applied, reference samples that may be used to predict an intra-coded video block may be limited to those in intra-coded neighboring blocks. The range of IBC searches may be restricted by imposing a maximum absolute value for block vectors.

IPC Classes  ?

  • H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
  • H04N 19/593 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving spatial prediction techniques
  • H04N 19/184 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being bits, e.g. of the compressed video stream
  • H04N 19/186 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being a colour or a chrominance component
  • H04N 19/126 - Quantisation - Details of normalisation or weighting functions, e.g. normalisation matrices or variable uniform quantisers

87.

Codec architecture for multiple layer video coding

      
Application Number 17246025
Grant Number 11627340
Status In Force
Filing Date 2021-04-30
First Publication Date 2021-08-12
Grant Date 2023-04-11
Owner VID SCALE, Inc. (USA)
Inventor
  • Ye, Yan
  • Mcclellan, George W.
  • He, Yong
  • Xiu, Xiaoyu
  • He, Yuwen
  • Dong, Jie
  • Bal, Can
  • Ryu, Eun Seok

Abstract

Systems, methods, and instrumentalities are provided to implement video coding system (VCS). The VCS may be configured to receive a video signal, which may include one or more layers (e.g., a base layer (BL) and/or one or more enhancement layers (ELs)). The VCS may be configured to process a BL picture into an inter-layer reference (ILR) picture, e.g., using picture level inter-layer prediction process. The VCS may be configured to select one or both of the processed ILR picture or an enhancement layer (EL) reference picture. The selected reference picture(s) may comprise one of the EL reference picture, or the ILR picture. The VCS may be configured to predict a current EL picture using one or more of the selected ILR picture or the EL reference picture. The VCS may be configured to store the processed ILR picture in an EL decoded picture buffer (DPB).

IPC Classes  ?

  • H04N 19/30 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using hierarchical techniques, e.g. scalability
  • H04N 19/117 - Filters, e.g. for pre-processing or post-processing
  • H04N 19/157 - Assigned coding mode, i.e. the coding mode being predefined or preselected to be further used for selection of another element or parameter
  • H04N 19/33 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using hierarchical techniques, e.g. scalability in the spatial domain
  • H04N 19/70 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards
  • H04N 19/31 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using hierarchical techniques, e.g. scalability in the temporal domain
  • H04N 19/36 - Scalability techniques involving formatting the layers as a function of picture distortion after decoding, e.g. signal-to-noise [SNR] scalability

88.

Methods and systems for video delivery supporting adaptation to viewing conditions

      
Application Number 17225191
Grant Number 11849153
Status In Force
Filing Date 2021-04-08
First Publication Date 2021-07-29
Grant Date 2023-12-19
Owner VID SCALE, Inc. (USA)
Inventor
  • Reznik, Yuriy
  • Asbun, Eduardo
  • Chen, Zhifeng
  • Ye, Yan
  • Zeira, Eldad M.
  • Zeira, Ariela
  • Soni, Naresh
  • Liu, Hang

Abstract

Described herein are methods and systems associated with viewing condition adaption of multimedia content. A method for receiving multimedia content with a device from a network may include determining a viewing parameter, transmitting a request for the multimedia content to the network, whereby the request may be based on the viewing parameter, and receiving the multimedia content from the network, whereby the multimedia content may be processed at a rate according to the viewing parameter. The viewing parameter may include at least one of: a user viewing parameter, a device viewing parameter, or a content viewing parameter. The method may further include receiving a multimedia presentation description (MPD) file from the network. The MPD file may include information relating to the rate of the multimedia content and information relating to the rate may include a descriptor relating to the viewing parameter, whereby the descriptor may be required or optional.

IPC Classes  ?

  • H04N 19/85 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using pre-processing or post-processing specially adapted for video compression
  • H04N 21/6587 - Control parameters, e.g. trick play commands or viewpoint selection
  • H04N 21/442 - Monitoring of processes or resources, e.g. detecting the failure of a recording device, monitoring the downstream bandwidth, the number of times a movie has been viewed or the storage space available from the internal hard disk
  • H04N 21/2343 - Processing of video elementary streams, e.g. splicing of video streams or manipulating MPEG-4 scene graphs involving reformatting operations of video signals for distribution or compliance with end-user requests or end-user device requirements
  • H04L 67/306 - User profiles
  • H04L 67/303 - Terminal profiles
  • H04N 21/81 - Monomedia components thereof
  • H04L 65/80 - Responding to QoS
  • H04L 65/612 - Network streaming of media packets for supporting one-way streaming services, e.g. Internet radio for unicast
  • H04L 65/75 - Media network packet handling

89.

Adaptive motion vector precision for affine motion model based video coding

      
Application Number 17269937
Grant Number 11582458
Status In Force
Filing Date 2019-08-28
First Publication Date 2021-07-01
Grant Date 2023-02-14
Owner Vid Scale, Inc. (USA)
Inventor
  • He, Yuwen
  • Xiu, Xiaoyu
  • Ye, Yan
  • Luo, Jiancong

Abstract

Systems and methods are described for video coding using affine motion models with adaptive precision. In an example, a block of video is encoded in a bitstream using an affine motion model, where the affine motion model is characterized by at least two motion vectors. A precision is selected for each of the motion vectors, and the selected precisions are signaled in the bitstream. In some embodiments, the precisions are signaled by including in the bitstream information that identifies one of a plurality of elements in a selected predetermined precision set. The identified element indicates the precision of each of the motion vectors that characterize the affine motion model. In some embodiments, the precision set to be used is signaled expressly in the bitstream; in other embodiments, the precision set may be inferred, e.g., from the block size, block shape or temporal layer.

IPC Classes  ?

  • H04N 19/137 - Motion inside a coding unit, e.g. average field, frame or block difference
  • H04N 19/159 - Prediction type, e.g. intra-frame, inter-frame or bidirectional frame prediction
  • H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
  • H04N 19/192 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the adaptation method, adaptation tool or adaptation type used for the adaptive coding the adaptation method, adaptation tool or adaptation type being iterative or recursive

90.

Overlapped block motion compensation

      
Application Number 16760905
Grant Number 11425418
Status In Force
Filing Date 2018-11-01
First Publication Date 2021-06-17
Grant Date 2022-08-23
Owner VID SCALE, Inc. (USA)
Inventor
  • Xiu, Xiaoyu
  • He, Yuwen
  • Ye, Yan
  • Zhang, Yan

Abstract

External overlapped block motion compensation (OBMC) may be performed for samples of a coding unit (CU) located along an inter-CU boundary of the CU while internal OBMC may be performed separately for samples located along inter-sub-block boundaries inside the CU. External OBMC may be applied based on substantially similar motion information associated with multiple external blocks neighboring the CU. The external blocks may be treated as a group to provide OBMC for multiple boundary samples together in an external OBMC operation. Internal OBMC may be applied using the same sub-block size used for sub-block level motion derivation. Internal OBMC may be disabled for the CU, for example, if the CU is coded in a spatial-temporal motion vector prediction (STMVP) mode.

IPC Classes  ?

  • H04N 19/583 - Motion compensation with overlapping blocks
  • H04N 19/184 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being bits, e.g. of the compressed video stream
  • H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
  • H04N 19/55 - Motion estimation with spatial constraints, e.g. at image or region borders
  • H04N 19/517 - Processing of motion vectors by encoding

91.

Sub-block motion derivation and decoder-side motion vector refinement for merge mode

      
Application Number 16761039
Grant Number 11558633
Status In Force
Filing Date 2018-11-01
First Publication Date 2021-06-17
Grant Date 2023-01-17
Owner VID SCALE, Inc. (USA)
Inventor
  • Xiu, Xiaoyu
  • He, Yuwen
  • Ye, Yan

Abstract

Systems, methods, and instrumentalities for sub-block motion derivation and motion vector refinement for merge mode may be disclosed herein. Video data may be coded (e.g., encoded and/or decoded). A collocated picture for a current slice of the video data may be identified. The current slice may include one or more coding units (CUs). One or more neighboring CUs may be identified for a current CU. A neighboring CU (e.g., each neighboring CU) may correspond to a reference picture. A (e.g., one) neighboring CU may be selected to be a candidate neighboring CU based on the reference pictures and the collocated picture. A motion vector (MV) (e.g., collocated MV) may be identified from the collocated picture based on an MV (e.g., a reference MV) of the candidate neighboring CU. The current CU may be coded (e.g., encoded and/or decoded) using the collocated MV.

IPC Classes  ?

  • H04N 19/51 - Motion estimation or motion compensation
  • H04N 19/105 - Selection of the reference unit for prediction within a chosen coding or prediction mode, e.g. adaptive choice of position and number of pixels used for prediction
  • H04N 19/174 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a slice, e.g. a line of blocks or a group of blocks
  • H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
  • H04N 19/70 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards

92.

Hybrid cubemap projection for 360-degree video coding

      
Application Number 17131902
Grant Number 11616981
Status In Force
Filing Date 2020-12-23
First Publication Date 2021-06-10
Grant Date 2023-03-28
Owner VID Scale, Inc. (USA)
Inventor
  • Xiu, Xiaoyu
  • He, Yuwen
  • Ye, Yan

Abstract

A system, method, and/or instrumentality may be provided for coding a 360-degree video. A picture of the 360-degree video may be received. The picture may include one or more faces associated with one or more projection formats. A first projection format indication may be received that indicates a first projection format may be associated with a first face. A second projection format indication may be received that indicates a second projection format may be associated with a second face. Based on the first projection format, a first transform function associated with the first face may be determined. Based on the second projection format, a second transform function associated with the second face may be determined. At least one decoding process may be performed on the first face using the first transform function and/or at least one decoding process may be performed on the second face using the second transform function.

IPC Classes  ?

  • H04N 19/597 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding specially adapted for multi-view video sequence encoding
  • H04N 19/12 - Selection from among a plurality of transforms or standards, e.g. selection between discrete cosine transform [DCT] and sub-band transform or selection between H.263 and H.264
  • H04N 19/61 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding in combination with predictive coding
  • G06T 3/00 - Geometric image transformation in the plane of the image

93.

360-degree video coding using geometry projection

      
Application Number 17111683
Grant Number 11388438
Status In Force
Filing Date 2020-12-04
First Publication Date 2021-05-06
Grant Date 2022-07-12
Owner VID SCALE, Inc. (USA)
Inventor
  • He, Yuwen
  • Ye, Yan
  • Hanhart, Philippe
  • Xiu, Xiaoyu

Abstract

Processing video data may include capturing the video data with multiple cameras and stitching the video data together to obtain a 360-degree video. A frame-packed picture may be provided based on the captured and stitched video data. A current sample location may be identified in the frame-packed picture. Whether a neighboring sample location is located outside of a content boundary of the frame-packed picture may be determined. When the neighboring sample location is located outside of the content boundary, a padding sample location may be derived based on at least one circular characteristic of the 360-degree video content and the projection geometry. The 360-degree video content may be processed based on the padding sample location.

IPC Classes  ?

  • H04N 19/597 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding specially adapted for multi-view video sequence encoding
  • H04N 13/194 - Transmission of image signals
  • H04N 19/105 - Selection of the reference unit for prediction within a chosen coding or prediction mode, e.g. adaptive choice of position and number of pixels used for prediction
  • H04N 13/383 - Image reproducers using viewer tracking for tracking with gaze detection, i.e. detecting the lines of sight of the viewer's eyes
  • H04N 13/117 - Transformation of image signals corresponding to virtual viewpoints, e.g. spatial image interpolation the virtual viewpoint locations being selected by the viewers or determined by viewer tracking
  • H04N 13/344 - Displays for viewing with the aid of special glasses or head-mounted displays [HMD] with head-mounted left-right displays
  • H04N 19/132 - Sampling, masking or truncation of coding units, e.g. adaptive resampling, frame skipping, frame interpolation or high-frequency transform coefficient masking
  • H04N 19/172 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a picture, frame or field
  • G06T 17/30 - Surface description, e.g. polynomial surface description
  • G06T 17/10 - Volume description, e.g. cylinders, cubes or using CSG [Constructive Solid Geometry]
  • H04N 13/161 - Encoding, multiplexing or demultiplexing different image signal components
  • H04N 19/563 - Motion estimation with padding, i.e. with filling of non-object values in an arbitrarily shaped picture block or region for estimation purposes
  • H04N 19/593 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving spatial prediction techniques
  • H04N 5/232 - Devices for controlling television cameras, e.g. remote control

94.

Method and apparatus of motion vector prediction for scalable video coding

      
Application Number 17132789
Grant Number 11343519
Status In Force
Filing Date 2020-12-23
First Publication Date 2021-04-22
Grant Date 2022-05-24
Owner VID SCALE. Inc. (USA)
Inventor
  • Xiu, Xiaoyu
  • Ye, Yan
  • He, Yong
  • He, Yuwen

Abstract

Inter-layer motion mapping information may be used to enable temporal motion vector prediction (TMVP) of an enhancement layer of a bitstream. For example, a reference picture and a motion vector (MV) of an inter-layer video block may be determined. The reference picture may be determined based on a collocated base layer video block. For example, the reference picture may be a collocated inter-layer reference picture of the reference picture of the collocated base layer video block. The MV may be determined based on a MV of the collocated base layer video block. For example, the MV may be determined by determining the MV of the collocated base layer video block and scaling the MV of the collocated base layer video block according to a spatial ratio between the base layer and the enhancement layer. TMVP may be performed on the enhancement layer picture using the MV of the inter-layer video block.

IPC Classes  ?

  • H04N 19/30 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using hierarchical techniques, e.g. scalability
  • H04N 19/52 - Processing of motion vectors by encoding by predictive encoding
  • H04N 19/70 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards
  • H04N 19/51 - Motion estimation or motion compensation
  • H04N 19/33 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using hierarchical techniques, e.g. scalability in the spatial domain
  • H04N 19/31 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using hierarchical techniques, e.g. scalability in the temporal domain
  • H04N 19/46 - Embedding additional information in the video signal during the compression process
  • H04N 19/587 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal sub-sampling or interpolation, e.g. decimation or subsequent interpolation of pictures in a video sequence
  • H04N 19/59 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving spatial sub-sampling or interpolation, e.g. alteration of picture size or resolution

95.

Frame-rate up conversion with low complexity

      
Application Number 16965571
Grant Number 11470344
Status In Force
Filing Date 2019-01-28
First Publication Date 2021-02-18
Grant Date 2022-10-11
Owner VID SCALE, INC. (USA)
Inventor
  • Xiu, Xiaoyu
  • He, Yuwen
  • Ye, Yan

Abstract

Systems and methods are described for selecting a motion vector (MV) to use in frame-rate up conversion (FRUC) coding of a block of video. In one embodiment, a first set of motion vector candidates is identified for FRUC prediction of the block. A search center is defined based on the first set of motion vector candidates, and a search window is determined, the search window having a selected width and being centered on the search center. A search for a selected MV is performed within the search window. In some embodiments, an initial set of MVs is processed with a clustering algorithm to generate a smaller number of MVs that are used as the first set. The selected MV may be subject to a motion refinement search, which may also be performed over a constrained search range. In additional embodiments, search iterations are constrained to limit complexity.

IPC Classes  ?

  • H04N 19/56 - Motion estimation with initialisation of the vector search, e.g. estimating a good candidate to initiate a search
  • H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
  • H04N 19/57 - Motion estimation characterised by a search window with variable size or shape

96.

Generalized bi-prediction for video coding with reduced coding complexity

      
Application Number 16964144
Grant Number 11153602
Status In Force
Filing Date 2019-01-23
First Publication Date 2021-02-04
Grant Date 2021-10-19
Owner Vid Scale, Inc. (USA)
Inventor
  • Zhang, Yan
  • Xiu, Xiaoyu
  • He, Yuwen
  • Ye, Yan

Abstract

Exemplary embodiments include systems and methods for coding a video comprising a plurality of pictures including a current picture, a first reference picture, and a second reference picture, where each picture includes a plurality of blocks. In one method, for at least a current block in the current picture, a number of available bi-prediction weights is determined based at least in part on a temporal layer and/or a quantization parameter of the current picture. From among available bi-prediction weights a pair of weights are identified. Using the identified weights, the current block is then predicted as a weighted sum of a first reference block in the first reference picture and a second reference block in the second reference picture. Encoding techniques are also described for efficient searching and selection of a pair of bi-prediction weights to use for prediction of a block.

IPC Classes  ?

  • H04N 19/577 - Motion compensation with bidirectional frame interpolation, i.e. using B-pictures
  • H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
  • H04N 19/31 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using hierarchical techniques, e.g. scalability in the temporal domain
  • H04N 19/573 - Motion compensation with multiple frame prediction using two or more reference frames in a given prediction direction

97.

Bi-directional optical flow method with simplified gradient derivation

      
Application Number 17045683
Grant Number 11575933
Status In Force
Filing Date 2019-04-05
First Publication Date 2021-01-28
Grant Date 2023-02-07
Owner VID Scale, Inc. (USA)
Inventor
  • He, Yuwen
  • Xiu, Xiaoyu
  • Ye, Yan

Abstract

A video coding device may be configured to perform directional Bi-directional optical flow (BDOF) refinement on a coding unit (CU). The device may determine the direction in which to perform directional BDOF refinement. The device may calculate the vertical direction gradient difference and the horizontal direction gradient difference for the CU. The vertical direction gradient difference may indicate the difference between the vertical gradients for a first reference picture and the vertical gradients for a second reference picture. The horizontal direction gradient difference may indicate the difference between the horizontal gradients for the first reference picture and the horizontal gradients for the second reference picture. The video coding device may determine the direction in which to perform directional BDOF refinement based on the vertical direction gradient difference and the horizontal direction gradient difference. The video coding device may perform directional BDOF refinement in the determined direction.

IPC Classes  ?

  • H04N 19/00 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
  • H04N 19/577 - Motion compensation with bidirectional frame interpolation, i.e. using B-pictures
  • H04N 19/103 - Selection of coding mode or of prediction mode
  • H04N 19/139 - Analysis of motion vectors, e.g. their magnitude, direction, variance or reliability
  • H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock

98.

METHODS AND APPARATUS OF VIEWPORT ADAPTIVE 360 DEGREE VIDEO DELIVERY

      
Application Number 17036578
Status Pending
Filing Date 2020-09-29
First Publication Date 2021-01-14
Owner Vid Scale, Inc. (USA)
Inventor
  • He, Yong
  • Ye, Yan
  • Gudumasu, Srinivas
  • Asbun, Eduardo
  • Hamza, Ahmed

Abstract

Systems, methods, and instrumentalities are disclosed for client centric service quality control. A first viewport of a 360 degree video may be determined. The 360 degree video may comprise one or more of an equirectangular, a cube-map, a cylindrical, a pyramidal, and/or a spherical projection mapping. The first viewport may be associated with a spatial region of the 360 degree video. An adjacent area that extends around the spatial region may be determined. A second viewport of the 360 degree video may be determined. A bitstream associated with the 360 degree video may be received. One or more enhanced regions may be included in the bitstream. The one or more enhanced regions may correspond to the first and/or second viewport. A high coding bitrate may be associated with the first viewport and/or the second viewport.

IPC Classes  ?

  • H04N 13/282 - Image signal generators for generating image signals corresponding to three or more geometrical viewpoints, e.g. multi-view systems
  • H04N 19/597 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding specially adapted for multi-view video sequence encoding
  • H04N 21/8543 - Content authoring using a description language, e.g. MHEG [Multimedia and Hypermedia information coding Expert Group] or XML [eXtensible Markup Language]
  • H04N 21/2343 - Processing of video elementary streams, e.g. splicing of video streams or manipulating MPEG-4 scene graphs involving reformatting operations of video signals for distribution or compliance with end-user requests or end-user device requirements
  • H04N 19/70 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards
  • H04N 21/845 - Structuring of content, e.g. decomposing content into time segments
  • H04N 21/4728 - End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification or for manipulating displayed content for selecting a ROI [Region Of Interest], e.g. for requesting a higher resolution version of a selected region
  • H04N 21/81 - Monomedia components thereof
  • H04N 13/117 - Transformation of image signals corresponding to virtual viewpoints, e.g. spatial image interpolation the virtual viewpoint locations being selected by the viewers or determined by viewer tracking
  • H04N 13/354 - Multi-view displays for displaying three or more geometrical viewpoints without viewer tracking for displaying sequentially

99.

Template-based inter prediction techniques based on encoding and decoding latency reduction

      
Application Number 16969190
Grant Number 11317085
Status In Force
Filing Date 2019-03-22
First Publication Date 2020-11-26
Grant Date 2022-04-26
Owner VID SCALE, INC. (USA)
Inventor
  • Xiu, Xiaoyu
  • He, Yuwen
  • Ye, Yan

Abstract

Video coding methods are described for reducing latency in template-based inter coding. In some embodiments, a method is provided for coding a video that includes a current picture and at least one reference picture. For at least a current block in the current picture, a respective predicted value is generated (e.g. using motion compensated prediction) for each sample in a template region adjacent to the current block. Once the predicted values are generated for each sample in the template region, a process is invoked to determine a template-based inter prediction parameter by using predicted values in the template region and sample values the reference picture. This process can be invoked without waiting for reconstructed sample values in the template region. Template-based inter prediction of the current block is then performed using the determined template-based inter prediction parameter.

IPC Classes  ?

  • H04N 19/132 - Sampling, masking or truncation of coding units, e.g. adaptive resampling, frame skipping, frame interpolation or high-frequency transform coefficient masking
  • H04N 19/176 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
  • H04N 19/46 - Embedding additional information in the video signal during the compression process
  • H04N 19/583 - Motion compensation with overlapping blocks
  • H04N 19/105 - Selection of the reference unit for prediction within a chosen coding or prediction mode, e.g. adaptive choice of position and number of pixels used for prediction
  • H04N 19/159 - Prediction type, e.g. intra-frame, inter-frame or bidirectional frame prediction
  • H04N 19/184 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being bits, e.g. of the compressed video stream
  • H04N 19/625 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding using discrete cosine transform [DCT]
  • H04N 19/64 - Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding using sub-band based transform, e.g. wavelets characterised by ordering of coefficients or of bits for transmission

100.

TRACKED VIDEO ZOOMING

      
Application Number 16642628
Status Pending
Filing Date 2018-08-23
First Publication Date 2020-11-05
Owner Vid Scale, Inc. (USA)
Inventor
  • Kerofsky, Louis
  • Asbun, Eduardo

Abstract

Systems, methods, and instrumentalities are disclosed for dynamic picture-in-picture (PIP) by a client. The client may reside on any device. The client may receive video content from a server, and identify an object within the video content using at least one of object recognition or metadata. The metadata may include information that indicates a location of an object within a frame of the video content. The client may receive a selection of the object by a user, and determine positional data of the object across frames of the video content using at least one of object recognition or metadata. The client may display an enlarged and time-delayed version of the object within a PIP window across the frames of the video content. Alternatively or additionally, the location of the PIP window within each frame may be fixed or may be based on the location of the object within each frame.

IPC Classes  ?

  • H04N 21/431 - Generation of visual interfaces; Content or additional data rendering
  • H04N 5/45 - Picture in picture
  • H04N 21/4728 - End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification or for manipulating displayed content for selecting a ROI [Region Of Interest], e.g. for requesting a higher resolution version of a selected region
  • H04N 21/44 - Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to MPEG-4 scene graphs
  1     2     3        Next Page