Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

Pre-training method and device, equipment and storage medium

A pre-training and character technology, applied in the computer field, can solve the problem of spending a lot of data computing resources and computing time text sentences, and achieve the effect of reducing data computing resources and computing time

Active Publication Date: 2022-03-15
BEIJING SANKUAI ONLINE TECH CO LTD
View PDF11 Cites 0 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0005] The embodiment of the present application provides a pre-training method, which can solve the problem in the prior art that it takes a lot of data computing resources and computing time to get the vector of the text sentence

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Pre-training method and device, equipment and storage medium
  • Pre-training method and device, equipment and storage medium
  • Pre-training method and device, equipment and storage medium

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0063] In order to make the purpose, technical solution and advantages of the present application clearer, the implementation manners of the present application will be further described in detail below in conjunction with the accompanying drawings.

[0064] The embodiment of the present application provides a pre-training method, which can be implemented by a server. The server may be a single server or a server cluster composed of multiple servers.

[0065] The server may include a processor, a memory, and a communication component, and the processor is respectively connected to the memory and the communication component.

[0066] The processor may be a CPU (Central Processing Unit, central processing unit). The processor can be used to read instructions and process data, for example, perform character masking on the initial text sentence, determine the target text sentence, determine the mask matrix corresponding to the target text sentence, train the feature extraction mo...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention discloses a pre-training method and device, equipment and a storage medium, and belongs to the technical field of computers. The method comprises the following steps: obtaining an initial text sentence after character masking processing; obtaining a target text sentence based on the initial text sentence after the character masking processing and the additional characters before the sentence; a mask matrix corresponding to the target text sentence is determined, the mask matrix comprises a plurality of elements, each element is used for indicating the operation association degree of two characters corresponding to the element in the target text sentence in the feature extraction process to a to-be-trained feature extraction model, and the element corresponding to the additional character in front of the sentence is not 0; and training a to-be-trained feature extraction model based on the initial text sentence, the target text sentence and the mask matrix. By adopting the method and the device, not only can the feature vector corresponding to each character in the target text sentence be obtained, but also the feature vector corresponding to the target text sentence can be obtained, other training does not need to be carried out additionally, and data operation resources and operation time are reduced.

Description

technical field [0001] The present application relates to the field of computer technology, in particular to a pre-training method, device, equipment and storage medium. Background technique [0002] With the continuous development of machine learning technology, pre-training models are widely emerging. The pre-trained model is a pre-trained model using a large number of unlabeled training sets. After the pre-training is completed, according to different needs, a small number of labeled training sets can be used for targeted training to obtain a trained model. For example, if you want to get a model for predicting the emotion of a text sentence (for example, predicting whether a text sentence is a happy mood or a sad mood, etc.), you can first use a BERT (Bidirectional Encoder Representation from Transformers, based on the two-way encoding of the converter Representation) model is pre-trained to obtain the pre-trained BERT model. Then, add a classifier after the BERT model...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
IPC IPC(8): G06F16/332G06F16/33G06K9/62
CPCG06F16/3329G06F16/3344G06F18/2411G06F18/214
Inventor 李如寐王思睿张富峥武威
Owner BEIJING SANKUAI ONLINE TECH CO LTD
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products