Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

Multi-scale convolution kernel method based on text-image generative adversarial network model

An image generation and network model technology, applied in the field of deep learning neural network, can solve the problem of slow learning features and achieve the effect of improving efficiency

Active Publication Date: 2018-04-06
SOUTH CHINA UNIV OF TECH
View PDF4 Cites 16 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0003] In the traditional adversarial network model, the discriminator and the generator use the convolution method after receiving the picture. Originally, only one convolution kernel is used for the single-layer image channel, which makes the network learn features slowly during the training process. For Each layer of image channels can only learn one feature

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Multi-scale convolution kernel method based on text-image generative adversarial network model
  • Multi-scale convolution kernel method based on text-image generative adversarial network model
  • Multi-scale convolution kernel method based on text-image generative adversarial network model

Examples

Experimental program
Comparison scheme
Effect test

Embodiment

[0031] This embodiment discloses a multi-scale convolution kernel method based on a text-image generation confrontation network model, which specifically includes the following steps:

[0032] Step S1, constructing a text-image generation confrontation network model, the generator generates images and inputs them to the discriminator for network training.

[0033] Step S2, using a deep convolutional neural network to function as a generator and a discriminator;

[0034] Different convolution kernels are reflected in different matrix values ​​and different numbers of rows and columns.

[0035] Construct multiple convolution kernels. In the process of processing images, different convolution kernels mean that different features of generated images can be learned during network training.

[0036] In the network model involved in the present invention, compared with the traditional generative confrontation network model, there are more encoding operations for text content, so tha...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention discloses a multi-scale convolution kernel method based on a text-image generative adversarial network model. The method comprises the following steps that: S1: constructing the text-image generative adversarial network model; S2: utilizing a deep convolutional neural network to serve as the functions of a generator and a discriminator; S3: after a text is coded, combining with random noise, and inputting the combined text and random noise into the generator; S4: in the text-image generative adversarial network model, utilizing multi-scale convolution to carry out a convolution operation on an image; and S5: inputting a loss function obtained by the multi-scale convolution operation into the generator for subsequent training. By use of the text-image generative adversarial network model constructed by the method, a convolution way generated after the generator and the discriminator receive pictures is changed through the multi-scale convolution, an original operation thatonly one convolution kernel is used by aiming at a single-layer image channel is changed into a situation that a plurality of convolution kernels are simultaneously adopted, so that the whole networkcan learn more characteristics when the single-layer image channel is convoluted, and network training efficiency is improved.

Description

technical field [0001] The invention relates to the technical field of deep learning neural networks, in particular to a multi-scale convolution kernel method based on a text-image generation confrontation network model. Background technique [0002] Generative Adversarial Network (GAN for short) is a deep learning framework proposed by Goodfellow in 2014. It is based on the idea of ​​"game theory" and constructs two models, the generator and the discriminator. The former The image is generated by inputting (0, 1) uniform noise or Gaussian random noise, which discriminates the input image to determine whether it is an image from the dataset or an image produced by the generator. [0003] In the traditional adversarial network model, the discriminator and the generator use the convolution method after receiving the picture. Originally, only one convolution kernel is used for the single-layer image channel, which makes the network learn features slowly during the training proc...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Applications(China)
IPC IPC(8): G06N3/08G06K9/62
CPCG06N3/08G06F18/214
Inventor 周智恒李立军黄俊楚
Owner SOUTH CHINA UNIV OF TECH
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products