Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

Video prediction method based on space-time propagation level codec

A codec and prediction method technology, applied in the field of computer vision, can solve problems such as inaccurate location, fuzzy prediction results, underutilization, etc., to achieve high prediction quality and improve performance

Active Publication Date: 2021-09-21
HANGZHOU DIANZI UNIV
View PDF7 Cites 3 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0004] The shortcomings of the above methods are mainly manifested in the following aspects: First, the method of stacking multiple ConvLSTMs does not integrate the learned low-level visual features and high-level semantic features well, and does not make full use of the learned features, resulting in prediction results There are still ambiguities; second, the video prediction method based on the dual-stream architecture does not effectively propagate the low-level visual features of the video in time sequence, resulting in inaccurate prediction of the position of the object outline in the video; third, the self-recursive method passes the The video frames generated by the network are sent to the network again to realize the video prediction of multiple video frames, but there are errors in the video frames generated by the network, and they will continue to accumulate due to this form, resulting in blurring in the late stage of multi-video frame prediction
Therefore, in order to alleviate the problems of insufficient fusion of different levels of features, inaccurate contour positions of video objects, and blurring in the later stage of prediction, there is an urgent need for a method that can fuse different levels of features, provide more accurate contour position information, and alleviate error accumulation. thereby improving the accuracy of video predictions

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Video prediction method based on space-time propagation level codec

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0039] The present invention will be further described below in conjunction with accompanying drawing.

[0040] The video prediction method based on the spatio-temporal propagation hierarchical codec first samples the given video and inputs it into the low-level visual memory encoder to obtain the low-level visual coding features and low-level memory state features; then uses the spatio-temporal propagation module to extract from the low-level visual coding features Spatio-temporal encoding features; then use a high-level semantic encoder consisting of a two-dimensional convolutional layer and a convolutional long-term short-term memory module to extract high-level semantic encoding features; finally, the obtained low-level visual encoding features, low-level memory state features and high-level semantics The encoded features are fused through a hierarchical feature decoder to obtain predicted video frames. This method uses a hierarchical codec to fuse the low-level visual and...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention discloses a video prediction method based on a space-time propagation level codec. The method comprises the following steps: firstly, sampling a given original video to obtain a frame sequence, and inputting the frame sequence into a low-layer visual memory coder to obtain low-layer visual coding features and low-layer memory state features; then, extracting space-time coding features from low-layer visual coding representation through a space-time propagation module, and extracting high-layer semantic features through a high-layer semantic coder; and carrying out information fusion on the obtained low-layer visual coding features, low-layer memory state features and high-layer semantic coding features through a hierarchical feature decoder, and outputting a predicted video frame. According to the method, the low-level visual memory features and the high-level semantic features can be fused, the low-level visual information is propagated in the time sequence direction through the space-time propagation module, the problem of video frame blurring can be solved to a certain extent by utilizing the priori knowledge of the first frame of the video, and the definition and visual quality of the predicted video frame are improved as a whole.

Description

technical field [0001] The invention belongs to the field of computer vision, in particular to the technical field of video prediction in video perception, and relates to a video prediction method based on a time-space propagation layer codec. Background technique [0002] In the era of Internet +, thousands of videos are generated on various terminal devices every day. Video perception has attracted extensive attention from academia and industry, and video prediction is one of the challenges and has high application value. visual task. This task aims to generate video segments at subsequent moments given a partial video segment. Video prediction has a wide range of applications in practical scenarios such as radar weather map prediction, traffic flow prediction, robot object interaction prediction, and driverless driving. For example, in an unmanned driving environment, traditional radar ranging can only judge close-range vehicle interaction, while video prediction can us...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
IPC IPC(8): H04N19/107H04N19/30H04N19/61G06K9/62G06N3/04G06N3/08
CPCH04N19/107H04N19/30H04N19/61G06N3/08G06N3/044G06F18/253Y02A90/10
Inventor 李平张陈翰王然徐向华
Owner HANGZHOU DIANZI UNIV
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products