Human Behavior Recognition Method Based on Quaternion Space-Time Convolutional Neural Network

A convolutional neural network and recognition method technology, applied in neural learning methods, biological neural network models, character and pattern recognition, etc., can solve the problem of destroying color features, affecting the robustness of recognition methods, not considering three-channel correlation and Integrity and other issues, to achieve the effect of rich feature information and high recognition rate

Active Publication Date: 2020-07-14
NORTHEAST DIANLI UNIVERSITY
View PDF7 Cites 0 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

However, the current convolutional neural network model only processes grayscale image sequences or RGB three-channels separately, without considering the correlation and integrity of the three channels, which destroys the color characteristics of the real environment and affects the recognition method in the actual environment. robustness

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Human Behavior Recognition Method Based on Quaternion Space-Time Convolutional Neural Network
  • Human Behavior Recognition Method Based on Quaternion Space-Time Convolutional Neural Network
  • Human Behavior Recognition Method Based on Quaternion Space-Time Convolutional Neural Network

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0076] In order to make the object, technical solution and advantages of the present invention clearer, the present invention will be further described in detail below in conjunction with the accompanying drawings. It should be understood that the specific embodiments described here are only used to explain the present invention, not to limit the present invention.

[0077] On the contrary, the invention covers any alternatives, modifications, equivalent methods and schemes within the spirit and scope of the invention as defined by the claims. Further, in order to make the public have a better understanding of the present invention, some specific details are described in detail in the detailed description of the present invention below. The present invention can be fully understood by those skilled in the art without the description of these detailed parts.

[0078] refer to figure 1 ,

[0079] A human behavior recognition method based on a quaternion spatiotemporal convolu...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

PUM

No PUM Login to view more

Abstract

The invention belongs to the field of computer vision, and in particular relates to a human behavior recognition method based on quaternion spatiotemporal convolutional nerves. The specific implementation steps of the present invention are as follows: (1) inputting the action video set to be recognized; (2) image preprocessing to extract images of key regions of human motion; (3) constructing a quaternion spatiotemporal convolutional neural network; (4) adopting The BP algorithm trains the network and outputs the training results; (5) Input the video test set and output the test results. The present invention uses the codebook model to extract the image of the human body movement area, and can detect the movement of the human body under the condition of complex background. The quaternion spatiotemporal convolutional neural network of the present invention directly takes the color image as the input, solves the problem of missing image features in the process of converting the color image into a grayscale image or dividing the channel by the traditional convolutional neural network, and improves the network feature extraction. performance, making the recognition of human behavior more accurate.

Description

technical field [0001] The invention belongs to the field of computer vision, and further relates to a human behavior recognition method based on a quaternion space-time convolutional neural network in target recognition. The invention can be used in applications such as human-computer interaction and intelligent monitoring. Background technique [0002] Human behavior recognition is an important research direction of computer vision, pattern recognition, image processing and artificial intelligence. It has great application value and theoretical significance in the fields of human-computer interaction, intelligent monitoring and medical treatment. It mainly analyzes and processes moving image sequences containing people, extracts features, and classifies motions to realize the recognition and understanding of individual human actions, interactions between people and between people and the external environment. [0003] Compared with object classification based on still imag...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

Application Information

Patent Timeline
no application Login to view more
Patent Type & Authority Patents(China)
IPC IPC(8): G06K9/00G06K9/62G06N3/08
CPCG06N3/084G06V40/23G06V20/46G06F18/2163G06F18/24G06F18/214
Inventor 孟勃刘雪君王晓霖
Owner NORTHEAST DIANLI UNIVERSITY
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products