Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

An OCR method based on depth learning

A deep learning and algorithmic technology, applied in the direction of instruments, character and pattern recognition, character recognition, etc., can solve the problems that the actual effect is not always satisfactory, achieve the effect of fast retrieval, fast processing speed, and improve work efficiency

Active Publication Date: 2019-02-22
SUNYARD SYST ENG CO LTD +1
View PDF25 Cites 42 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

However, considering the quality of the scanning effect, the quality of the paper document itself (such as printing quality, font clarity, font standardization, etc.), and the content layout (the arrangement of the text, compared with ordinary text and table text and bills), The actual effect of OCR is not always satisfactory
The requirements for recognition accuracy of different paper documents are different. For example, the recognition of bills has very high requirements for accuracy, because if a digital recognition error may lead to fatal consequences, traditional OCR recognition cannot meet the requirements. Such high-precision identification requirements

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • An OCR method based on depth learning

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0035] The present invention will be described in detail below according to the accompanying drawings and preferred embodiments, and the purpose and effect of the present invention will become clearer. The present invention will be further described in detail below in conjunction with the accompanying drawings and embodiments. It should be understood that the specific embodiments described here are only used to explain the present invention, not to limit the present invention.

[0036] Such as figure 1 Shown, the OCR method based on deep learning of the present invention, it comprises the steps:

[0037] S1: Acquiring the image to be recognized;

[0038] S2: Scaling the image to be recognized, and then preprocessing the zoomed image, the preprocessing is any one or more of sharpening, grayscale, binarization, tilt correction, noise reduction, and official seal removal item;

[0039] Image preprocessing to remove the official seal adopts the following method:

[0040] S2.1:...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention discloses an OCR method based on depth learning, comprising the following steps of obtaining an image to be recognized; scaling the image to be recognized, and then preprocessing the scaled image, wherein the preprocessing comprises sharpening, graying, binarization, correcting tilt, reducing noise, and removing any one or more items in the official seal; putting the preprocessed image into the depth learning neural network to obtain a plurality of candidate text regions of the scaled image; putting a plurality of candidate text regions of the scaled image into a depth-based learning neural network, and converting the contents of the candidate text regions into computer-readable text; putting the computer-readable text into the deep learning language model to correct the textrecognition results. The method of the invention can realize the intelligent extraction of the text content of the image, provides the service of fast searching and fast obtaining information for thepersonnel of various trades, improves the working efficiency, has high identification field accuracy rate and improves the processing speed.

Description

technical field [0001] The invention relates to the technical field of image recognition in computer software, in particular to an OCR method based on deep learning. Background technique [0002] The text area detection and positioning recognition technology based on optical character recognition (hereinafter referred to as OCR) refers to the use of computers and other equipment to automatically extract and recognize valid information in paper materials and perform corresponding processing. One of the key technologies of paper-based computer automatic processing. The traditional image text recognition is OCR, which is based on scanning the paper documents to be recognized into electronic images. However, considering the quality of the scanning effect, the quality of the paper document itself (such as printing quality, font clarity, font standardization, etc.), and the content layout (the arrangement of the text, compared with ordinary text and table text and bills), The ac...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
IPC IPC(8): G06K9/00G06K9/32G06K9/62
CPCG06V30/40G06V20/62G06V30/10G06F18/214
Inventor 王慜骊林路桂晓雷安通鉴林康陈立强
Owner SUNYARD SYST ENG CO LTD
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products