Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

Method and system for processing multiview videos for view synthesis using motion vector predictor list

A motion vector prediction, multi-view video technology, applied in the field of multi-view video encoding and decoding, can solve the problem of not being able to encode temporally correlated and spatially correlated multi-view videos, etc.

Active Publication Date: 2015-03-18
MITSUBISHI ELECTRIC CORP
View PDF4 Cites 12 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

However, this method is also not capable of encoding multi-view videos where both temporal and spatial correlations between views are used, since there is no efficient method for predicting views considering temporal correlations

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Method and system for processing multiview videos for view synthesis using motion vector predictor list
  • Method and system for processing multiview videos for view synthesis using motion vector predictor list
  • Method and system for processing multiview videos for view synthesis using motion vector predictor list

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0048] One embodiment of the present invention provides a joint temporal / inter-view processing method for encoding and decoding frames of multi-view video. A multi-view video is a video of a scene acquired by multiple cameras with different poses. Defines a pose camera as its 3D (x,y,z) position and its orientation. Each pose corresponds to a "view" of the scene.

[0049] The method uses temporal correlation between frames in the same video acquired for a specific camera pose, and spatial correlation between synchronized frames in different videos acquired from multiple camera views. Additionally, "composite" frames can be associated, as described below.

[0050] In one embodiment, the temporal correlation uses motion compensated temporal filtering (MCTF), while the spatial correlation uses disparity compensated inter-view filtering (DCVF).

[0051] In another embodiment of the invention, the spatial correlation uses predictions from one view of a synthesized frame genera...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

Multiview videos are acquired by overlapping cameras. Side information is used to synthesize multiview videos. A reference picture list is maintained for current frames of the multiview videos, the reference picture indexes temporal reference pictures and spatial reference pictures of the acquired multiview videos and the synthesized reference pictures of the synthesized multiview video. Each current frame of the multiview videos is predicted according to reference pictures indexed by the associated reference picture list with a skip mode and a direct mode, whereby the side information is inferred from the synthesized reference picture. In addition, the skip and merge modes for single view video coding are modified to support multiview video coding by generating a motion vector prediction list by also considering neighboring blocks that are associated with synthesized reference pictures.

Description

technical field [0001] The present invention relates generally to encoding and decoding multi-view video, and more particularly to synthesizing multi-view video. Background technique [0002] Multiview video encoding and decoding is important for applications such as three-dimensional television (3DTV), free viewpoint television (FTV) and multi-camera surveillance. Multi-view video encoding and decoding is also known as dynamic light field compression. [0003] figure 1 A prior art "simulcast" system 100 for multi-view video coding is shown. Cameras 1-4 capture a series of frames or video 101-104 of scene 5. Each camera has a different view of the scene. Each video is independently encoded 111-114 into a corresponding encoded video 121-124. The system uses conventional 2D video coding techniques. Therefore, the system does not establish a relationship between different videos captured by the camera from different viewpoints while predicting frames of the encoded video....

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
IPC IPC(8): H04N19/597H04N19/52H04N19/423H04N19/70
CPCH04N19/423H04N19/597H04N19/52H04N13/0048H04N13/161
Inventor 田栋邹峰安东尼·韦特罗
Owner MITSUBISHI ELECTRIC CORP
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products