Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

Video prediction method based on spatiotemporal propagation hierarchical codec

A codec and prediction method technology, applied in the field of computer vision, can solve problems such as inaccurate location, blurred prediction results, blurred predictions in the later stage, etc., to achieve high prediction quality and improve performance

Active Publication Date: 2022-05-31
HANGZHOU DIANZI UNIV
View PDF0 Cites 0 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0004] The shortcomings of the above methods are mainly manifested in the following aspects: First, the method of stacking multiple ConvLSTMs does not integrate the learned low-level visual features and high-level semantic features well, and does not make full use of the learned features, resulting in prediction results There are still ambiguities; second, the video prediction method based on the dual-stream architecture does not effectively propagate the low-level visual features of the video in time sequence, resulting in inaccurate prediction of the position of the object outline in the video; third, the self-recursive method passes the The video frames generated by the network are sent to the network again to realize the video prediction of multiple video frames, but there are errors in the video frames generated by the network, and they will continue to accumulate due to this form, resulting in blurring in the late stage of multi-video frame prediction
Therefore, in order to alleviate the problems of insufficient fusion of different levels of features, inaccurate contour positions of video objects, and blurring in the later stage of prediction, there is an urgent need for a method that can fuse different levels of features, provide more accurate contour position information, and alleviate error accumulation. thereby improving the accuracy of video predictions

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Video prediction method based on spatiotemporal propagation hierarchical codec

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0039] The present invention will be further described below in conjunction with the accompanying drawings.

[0040] The video prediction method based on the spatio-temporal propagation hierarchical codec, firstly sample the given video and input

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention discloses a video prediction method based on a time-space propagation hierarchical codec. The method of the present invention first samples the given original video to obtain the frame sequence and inputs it to the low-level visual memory encoder to obtain the low-level visual coding features and low-level memory state features; features, and extract high-level semantic features through a high-level semantic encoder; then fuse the obtained low-level visual coding features, low-level memory state features, and high-level semantic coding features through a hierarchical feature decoder to output predicted video frames. The method of the present invention can not only integrate the low-level visual memory features and high-level semantic features, but also make the low-level visual information spread along the time sequence direction through the spatio-temporal propagation module, and can also use the prior knowledge of the first frame of the video to solve the video frame blur problem to a certain extent, Improve overall clarity and visual quality of predicted video frames.

Description

Video prediction method based on spatiotemporal propagation hierarchical codec technical field The invention belongs to the field of computer vision, especially the video prediction technology field in video perception, relate to a kind of A video prediction method based on the spatiotemporal propagation hierarchical codec. Background technique [0002] In the Internet+ era, thousands of videos are generated on various terminal devices every day, and video perception is affected by It has attracted widespread attention from academia and industry, and video prediction is one of the most challenging and highly valuable applications. sense task. This task aims to generate video segments of subsequent moments given a partial video segment. Video Prediction on Radar It is widely used in practical scenarios such as weather map prediction, traffic flow prediction, robot-object interaction prediction, and unmanned driving. example For example, in an unmanned environment, ...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Patents(China)
IPC IPC(8): H04N19/107H04N19/30H04N19/61G06K9/62G06V10/40G06N3/04G06N3/08
CPCH04N19/107H04N19/30H04N19/61G06N3/08G06N3/044G06F18/253Y02A90/10
Inventor 李平张陈翰王然徐向华
Owner HANGZHOU DIANZI UNIV
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products