Movatterモバイル変換


[0]ホーム

URL:


CN114419626B - High-precision bill identification method and system based on OCR technology - Google Patents

High-precision bill identification method and system based on OCR technology
Download PDF

Info

Publication number
CN114419626B
CN114419626BCN202111640612.3ACN202111640612ACN114419626BCN 114419626 BCN114419626 BCN 114419626BCN 202111640612 ACN202111640612 ACN 202111640612ACN 114419626 BCN114419626 BCN 114419626B
Authority
CN
China
Prior art keywords
sentences
group
text
handwriting
bill
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202111640612.3A
Other languages
Chinese (zh)
Other versions
CN114419626A (en
Inventor
丁云波
陈伟
谢勇
赖显成
唐文华
董双翼
柳青
潘朝晖
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Sichuan Huaxi Jicai E Commerce Co ltd
Original Assignee
Sichuan Huaxi Jicai E Commerce Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Sichuan Huaxi Jicai E Commerce Co ltdfiledCriticalSichuan Huaxi Jicai E Commerce Co ltd
Priority to CN202111640612.3ApriorityCriticalpatent/CN114419626B/en
Publication of CN114419626ApublicationCriticalpatent/CN114419626A/en
Application grantedgrantedCritical
Publication of CN114419626BpublicationCriticalpatent/CN114419626B/en
Activelegal-statusCriticalCurrent
Anticipated expirationlegal-statusCritical

Links

Landscapes

Abstract

The application relates to a high-precision bill identification method and a high-precision bill identification system based on an OCR technology, which are used for converting characters on a bill into plane image information based on an optical scanning technology; scanning the character track on the front side of the bill and the character concave depth of the handwritten characters; based on the same principle, acquiring the handwriting protrusion height of the handwritten characters on the back of the bill; for handwritten characters with unclear handwriting overlapping, sequentially listing characters possibly represented by the overlapped handwriting based on the depth of the handwriting concave trace and the height of the handwriting convex trace; acquiring text information from the plane image information, and inserting the text information one by one from the listed text to obtain a first group of text groups; based on logic judgment, correct text groups and sentences are selected to finish document identification, and the application can effectively improve the identification precision of handwritten documents, especially for the scenes of correction of wrong words, nonstandard writing and the like.

Description

High-precision bill identification method and system based on OCR technology
Technical Field
The application relates to the technical field of data sharing, in particular to a high-precision bill identification method and system based on an OCR technology.
Background
OCR is an abbreviation (OpticalCharacterRecognition) for optical character recognition, which is a computer input technology that converts characters of various notes, newspapers, books, manuscripts, and other printed matters into image information by means of optical input methods such as scanning, and then converts the image information into usable ones by means of the character recognition technology. The method can be applied to the fields of inputting and processing bank notes, a large amount of text data, file files and texts. The automatic scanning identification method is suitable for automatic scanning identification and long-term storage of a large number of bill forms in the industries of banks, tax and the like. Compared with a general text, 4 aspects of final recognition rate, recognition speed, layout understanding accuracy and layout reduction satisfaction are generally used as evaluation basis of OCR technology; in contrast, the recognition rate, the overall pass rate, and the recognition speed are generally practical standards for measuring OCR technology with respect to forms and notes. The OCR recognition technology not only has the advantages that various general printed forms can be automatically judged, split, recognized and restored, satisfactory practical results are made on form understanding, the layout of the manuscript can be automatically analyzed, corresponding attributes such as titles, bars, images and forms are automatically divided, recognition sequences are judged, and recognition results can be restored to new texts consistent with the layout of scanned manuscript. The automatic form input technology can automatically recognize the printing of specific form or print Chinese character, letter and number, recognize handwriting Chinese character, handwriting letter, number and various handwriting symbols, and output according to form format. The form input efficiency is improved, and a large amount of manpower can be saved. Meanwhile, the form identification is supported to be directly restored into PTF, PDF, HTML and other format documents; and the automatic typesetting surface analysis can be carried out on the embedded horizontal text, the vertical text and the form text of the image. At present, the OCR is utilized to directly extract important data such as money, account numbers and the like from the certificate image, replaces manual input of people, is tightly combined with bar code identification/running water identification, and realizes the work of establishing an after-the-fact copy account and completing after-the-fact supervision.
The defects of the prior art are that: the OCR technology is limited by the definition and the integrity of the bill, and compared with manual identification, the identification accuracy of the OCR technology still has larger error, and especially has the defects of low intelligence and larger error for the handwritten bill.
Disclosure of Invention
The application aims to overcome the defects of the prior art, and provides a high-precision receipt identification method and a high-precision receipt identification system based on an OCR technology, which can effectively improve the identification precision of handwriting receipts, and particularly can be used for the scenes of correction of wrong characters, nonstandard writing and the like.
The aim of the application is realized by the following technical scheme:
The first aspect of the application provides a high-precision bill identification method based on OCR technology, which comprises the following steps:
s101: based on an optical scanning technology, converting characters on a bill into planar image information;
s102: scanning the character track on the front side of the bill and the character concave depth of the handwritten characters; based on the same principle, acquiring the handwriting protrusion height of the handwritten characters on the back of the bill;
S103: for handwritten characters with unclear handwriting overlapping, sequentially listing characters possibly represented by the overlapped handwriting based on the depth of the handwriting concave trace and the height of the handwriting convex trace;
S104: acquiring text information from the plane image information, and inserting the text information one by one from the listed text to obtain a first group of text groups;
s105: based on logic judgment, correct text groups and sentences are selected, and bill identification is completed.
The application is different from the traditional optical scanning technology, has better recognition effect based on correction in the handwritten characters, detects based on the characteristic that writing is harder for protruding display effect during correction, and judges correct characters based on concave and protruding stroke marks, thereby improving recognition accuracy of the characters.
Further, the handwritten character detection for unclear overlapping of handwriting also includes a recognition of the thickness of the handwritten character trace, which includes:
Extracting handwriting of the handwriting overlapping area, and decomposing according to strokes to obtain each stroke;
sequencing the strokes from thick to thin, and then combining the characters according to the strokes from thick to thin to obtain the possibly represented characters of the overlapped handwriting;
And inserting the characters into the character information one by one to obtain a second group of character groups and sentences.
Further, the method also comprises a mutual verification step, and the content is as follows:
Comparing the first group of text group sentences with the second group of text group sentences, selecting a first logic group sentence in the first group of text group sentences or the second group of text group sentences, judging whether the first logic group sentence appears in the other text group sentences, if so, judging that the first logic group sentence is a correct text group sentence;
The first logical group sentence refers to the best option in the first group text group sentence and the second group text group sentence under the corresponding logical judgment.
Further, when the first logical group of sentences of the first group of sentences and the second group of sentences are the same, mutual verification is not required.
Further, when the first logical group of sentences of the first group of sentences and the second group of sentences are different, the correct text group sentence is determined before ranking and examination of the first logical group of sentences in the other group of sentences.
Optionally, the application also adds the characteristic of thicker strokes during correction, combines the characteristics with the writing strength of the strokes, and further improves the character recognition precision by comprehensively considering the characteristics, especially the scene of the correction of the strokes.
Further, when the first logical group sentence in the first group of text group sentences or the second group of text group sentences is not in the other logical group sentence, the review is performed manually.
Further, the logic judgment comprises semantic logic, word order logic and context logic in an application scene.
The second aspect of the present application also provides a high-precision document identification system based on OCR technology, for implementing the method according to the first aspect, the system comprising:
The optical scanning module scans the bill and converts characters on the bill into plane image information;
The surface trace identification module is used for scanning the front and back of the bill to obtain the writing concave depth of the front and the protruding height of the back of the handwritten character;
the image processing module converts the plane image information into text information;
the word processing module is used for analyzing the concave depth of the handwriting and the convex height of the back surface to obtain a first group of word groups and obtaining a second group of word groups and sentences based on the thickness of strokes in the word information;
The logic judging module is used for obtaining correct text groups and sentences from the first text groups and sentences of the second text groups based on logic judgment;
A server storing a computer program which, when run, configures the modules as described in the first aspect.
Further, the system also comprises an alarm module, and when the first logical group sentence in the first group of text group sentences or the second group of text group sentences is not in the other logical group sentence, an alarm is sent.
Further, the surface trace recognition module is a detection device based on the principle of optical technology or acoustic wave technology.
The beneficial effects of the application are as follows: the application utilizes the characteristics of large stroke force and thick stroke during correction to identify the handwritten characters, and particularly has higher identification precision aiming at the corrected handwritten characters, thereby improving the accuracy of bill identification.
Drawings
FIG. 1 is a schematic flow chart of the present application;
FIG. 2 is a system block diagram of an embodiment of the present application;
Fig. 3 is a schematic structural diagram of a server according to the present application.
Detailed Description
The technical scheme of the present application is described in further detail below in connection with specific embodiments, but the scope of the present application is not limited to the following.
As shown in fig. 1, a first aspect of the present embodiment provides a high-precision document identification method based on OCR technology, where the method includes:
s101: based on an optical scanning technology, converting characters on a bill into planar image information;
s102: scanning the character track on the front side of the bill and the character concave depth of the handwritten characters; based on the same principle, acquiring the handwriting protrusion height of the handwritten characters on the back of the bill;
s103: for handwritten characters with unclear handwriting overlapping, the characters possibly represented by the overlapped characters are listed in sequence based on the depth of the concave trace of the handwriting and the height of the convex trace of the handwriting;
S104: acquiring text information from the plane image information, and inserting the text information one by one from the listed text to obtain a first group of text groups;
s105: based on logic judgment, correct text groups and sentences are selected, and bill identification is completed.
Optionally, in some embodiments, the detecting handwriting that is not clear of the handwriting overlap further includes identifying a handwriting trace thickness, including: extracting handwriting of the handwriting overlapping area, and decomposing according to strokes to obtain each stroke; sequencing the strokes from thick to thin, and then combining the characters according to the strokes from thick to thin to obtain the possibly represented characters of the overlapped handwriting; inserting the characters into the character information one by one to obtain a second group of character groups. Further, the method also comprises a mutual verification step, and the content of the mutual verification step is as follows: comparing the first set of text set sentences with the second set of text set sentences, selecting a first logic set sentence in the first set of text set sentences or the second set of text set sentences, judging whether the first logic set sentence appears in the other text set sentences, and if so, judging that the first logic set sentence is a correct text set sentence. The first logical group sentence refers to the best option in the first group word group sentence and the second group word group sentence under the corresponding logical judgment. In some specific applications, the specific steps are that a first logic set of sentences in a first set of text sets is selected, whether identical text sets exist or not is searched in a second set of text sets, and if so, the first logic set of sentences is the correct set of sentences. And simultaneously, selecting a first logic group sentence in the second group of text group sentences, searching whether identical text group sentences exist in the first group of text group sentences, and if so, judging that the first logic group sentence is a correct group sentence. At this time, mutual check is not effective if the first logical group sentences are identical twice, that is, mutual check is not required when the first logical group sentences of the first group text group sentences and the second group text group sentences are identical. When one logical set of sentences of the two results is different, further checking is performed, in some embodiments, when the first logical set of sentences of the first set of sentences and the second set of sentences are different, determining the correct sentence set by the ranking of the first logical set of sentences in the other set of sentences. For example, a first logical group of sentences in the first set of sentences in the second logical order of sentences in the second set of sentences, and the first logical group of sentences in the second group of text group sentences arranges a third logical sequence in the first group of text group sentences, and the first logical group of sentences in the first group of text group sentences is selected as correct words.
Alternatively, in some embodiments, the review is performed manually when the first logical set of sentences in the first set of sentences or the second set of sentences is not in the other logical set of sentences. The logic judgment in the application comprises semantic logic, word order logic and context logic in an application scene.
As shown in fig. 2, the second aspect of the present application further provides a high-precision document identification system based on OCR technology, for implementing the method as in the first aspect, the system comprising:
The optical scanning module scans the bill and converts characters on the bill into plane image information;
The surface trace identification module is used for scanning the front and back of the bill to obtain the writing concave depth of the front and the protruding height of the back of the handwritten character;
The image processing module converts the plane image information into text information;
The word processing module is used for analyzing the concave depth of the handwriting and the convex height of the back surface to obtain a first group of word groups and sentences, and obtaining a second group of word groups and sentences based on the thickness of strokes in the word information;
the logic judging module is used for obtaining correct text groups and sentences from the first text groups and sentences of the second text groups based on logic judgment;
A server storing a computer program which, when run, configures the modules as in the method of the first aspect.
Optionally, in some embodiments, an alarm module is further included to issue an alarm when a first logical group sentence in the first group of text group sentences or the second group of text group sentences is not in the other logical group sentence. The alarm module preferably adopts a voice alarm device to carry out language reminding.
Alternatively, in some embodiments, the surface trace recognition module is a detection device based on optical or sonic principles.
FIG. 3 is a schematic diagram of a server in the embodiment of FIG. 2 of the present application. As shown in fig. 3, the server of this embodiment includes: a processor, a memory, and a computer program stored in the memory and executable on the processor, such as a push message program. The steps in the above-described embodiments of the feature point extraction method under each dynamic scenario, such as steps S101 to S105 shown in fig. 1, are implemented when the processor executes the computer program. Or the processor, when executing the computer program, performs the functions of the modules/units in the above-described device embodiments.
For example, a computer program may be split into one or more modules/units, which are stored in a memory and executed by a processor to perform the present application. One or more of the modules/units may be a series of computer program instruction segments capable of performing specific functions to describe the execution of the computer program in a server. For example, the computer program may be divided into an acquisition module, an analysis module, a search module, and a push module, where each module specifically functions as follows:
The server may be a computing device such as a desktop computer, notebook, palm computer, etc. The server may include, but is not limited to, a processor, memory. It will be appreciated by those skilled in the art that fig. 3 is merely an example of a server and is not limiting of the server, and may include more or fewer components than shown, or may combine certain components, or different components, e.g., the server may also include input and output devices, network access devices, buses, etc.
The Processor may be a central processing unit (Central Processing Unit, CPU), other general purpose Processor, digital signal Processor (DIGITAL SIGNAL Processor, DSP), application SPECIFIC INTEGRATED Circuit (ASIC), off-the-shelf Programmable gate array (Field-Programmable GATE ARRAY, FPGA) or other Programmable logic device, discrete gate or transistor logic device, discrete hardware components, or the like. A general purpose processor may be a microprocessor or the processor may be any conventional processor or the like.
The memory may be an internal storage unit of the server, such as a hard disk or a memory of the server. The memory may also be an external storage device of the server, such as a plug-in hard disk provided on the server, a smart memory card (SMART MEDIA CARD, SMC), a Secure Digital (SD) card, a flash memory card (FLASH CARD), or the like. Further, the memory may also include both internal storage units of the server and external storage devices. The memory is used to store computer programs and other programs and data required by the server. The memory may also be used to temporarily store data that has been output or is to be output.
It will be apparent to those skilled in the art that, for convenience and brevity of description, only the above-described division of the functional units and modules is illustrated, and in practical application, the above-described functional distribution may be performed by different functional units and modules according to needs, i.e. the internal structure of the apparatus is divided into different functional units or modules to perform all or part of the above-described functions. The functional units and modules in the embodiment may be integrated in one processing unit, or each unit may exist alone physically, or two or more units may be integrated in one unit, where the integrated units may be implemented in a form of hardware or a form of a software functional unit. In addition, the specific names of the functional units and modules are only for distinguishing from each other, and are not used for limiting the protection scope of the present application. The specific working process of the units and modules in the above system may refer to the corresponding process in the foregoing method embodiment, which is not described herein again.
In the foregoing embodiments, the descriptions of the embodiments are emphasized, and in part, not described or illustrated in any particular embodiment, reference is made to the related descriptions of other embodiments.
Those of ordinary skill in the art will appreciate that the various illustrative elements and algorithm steps described in connection with the embodiments disclosed herein may be implemented as electronic hardware, or combinations of computer software and electronic hardware. Whether such functionality is implemented as hardware or software depends upon the particular application and design constraints imposed on the solution. Skilled artisans may implement the described functionality in varying ways for each particular application, but such implementation decisions should not be interpreted as causing a departure from the scope of the present application.
In the embodiments provided in the present application, it should be understood that the disclosed server and method may be implemented in other manners. For example, the above-described server embodiments are merely illustrative, and the division of the modules or units, for example, is merely a logical functional division, and there may be additional divisions when actually implemented, for example, multiple units or components may be combined or integrated into another system, or some features may be omitted, or not performed. Alternatively, the coupling or direct coupling or communication connection shown or discussed may be an indirect coupling or communication connection via interfaces, devices or units, which may be in electrical, mechanical or other forms.
The units described as separate units may or may not be physically separate, and units shown as units may or may not be physical units, may be located in one place, or may be distributed on a plurality of network units. Some or all of the units may be selected according to actual needs to achieve the purpose of the solution of this embodiment.
In addition, each functional unit in the embodiments of the present application may be integrated in one processing unit, or each unit may exist alone physically, or two or more units may be integrated in one unit. The integrated units may be implemented in hardware or in software functional units.
The integrated modules/units, if implemented in the form of software functional units and sold or used as stand-alone products, may be stored in a computer readable storage medium. Based on such understanding, the present application may implement all or part of the flow of the method of the above embodiment, or may be implemented by a computer program to instruct related hardware, where the computer program may be stored in a computer readable storage medium, and when the computer program is executed by a processor, the computer program may implement the steps of each of the method embodiments described above. Wherein the computer program comprises computer program code which may be in source code form, object code form, executable file or some intermediate form etc. The computer readable medium may include: any entity or device capable of carrying the computer program code, a recording medium, a U disk, a removable hard disk, a magnetic disk, an optical disk, a computer Memory, a Read-Only Memory (ROM), a random access Memory (RAM, random Access Memory), an electrical carrier signal, a telecommunications signal, a software distribution medium, and so forth. It should be noted that the computer readable medium contains content that can be appropriately scaled according to the requirements of jurisdictions in which such content is subject to legislation and patent practice, such as in certain jurisdictions in which such content is subject to legislation and patent practice, the computer readable medium does not include electrical carrier signals and telecommunication signals.
The foregoing is merely a preferred embodiment of the application, and it is to be understood that the application is not limited to the form disclosed herein but is not to be construed as excluding other embodiments, but is capable of numerous other combinations, modifications and environments and is capable of modifications within the scope of the inventive concept, either as taught or as a matter of routine skill or knowledge in the relevant art. And that modifications and variations which do not depart from the spirit and scope of the application are intended to be within the scope of the appended claims.

Claims (10)

CN202111640612.3A2021-12-292021-12-29High-precision bill identification method and system based on OCR technologyActiveCN114419626B (en)

Priority Applications (1)

Application NumberPriority DateFiling DateTitle
CN202111640612.3ACN114419626B (en)2021-12-292021-12-29High-precision bill identification method and system based on OCR technology

Applications Claiming Priority (1)

Application NumberPriority DateFiling DateTitle
CN202111640612.3ACN114419626B (en)2021-12-292021-12-29High-precision bill identification method and system based on OCR technology

Publications (2)

Publication NumberPublication Date
CN114419626A CN114419626A (en)2022-04-29
CN114419626Btrue CN114419626B (en)2024-07-26

Family

ID=81268833

Family Applications (1)

Application NumberTitlePriority DateFiling Date
CN202111640612.3AActiveCN114419626B (en)2021-12-292021-12-29High-precision bill identification method and system based on OCR technology

Country Status (1)

CountryLink
CN (1)CN114419626B (en)

Citations (2)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
CN111291761A (en)*2020-02-172020-06-16北京百度网讯科技有限公司 Method and apparatus for recognizing text
CN111695539A (en)*2020-06-172020-09-22北京一起教育信息咨询有限责任公司Evaluation method and device for handwritten Chinese characters and electronic equipment

Family Cites Families (3)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
JP5043093B2 (en)*2009-12-112012-10-10新日本製鐵株式会社 Character information recognition method and character information recognition apparatus
US8363947B2 (en)*2010-07-312013-01-29International Business Machines CorporationHandwritten character recognition based on frequency variations in characters
US10521654B2 (en)*2018-03-292019-12-31Fmr LlcRecognition of handwritten characters in digital images using context-based machine learning

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
CN111291761A (en)*2020-02-172020-06-16北京百度网讯科技有限公司 Method and apparatus for recognizing text
CN111695539A (en)*2020-06-172020-09-22北京一起教育信息咨询有限责任公司Evaluation method and device for handwritten Chinese characters and electronic equipment

Also Published As

Publication numberPublication date
CN114419626A (en)2022-04-29

Similar Documents

PublicationPublication DateTitle
US8401293B2 (en)Word recognition of text undergoing an OCR process
CN110135411A (en) Business card recognition method and device
CN112508011A (en)OCR (optical character recognition) method and device based on neural network
EP2144188A2 (en)Word detection method and system
CN109816118A (en) A method and terminal for creating structured documents based on a deep learning model
CN109635805B (en)Image text positioning method and device and image text identification method and device
CN104966097A (en)Complex character recognition method based on deep learning
US20120039536A1 (en)Optical character recognition with two-pass zoning
Engin et al.Offline signature verification on real-world documents
CN111950557A (en) Error handling method, image forming apparatus, and electronic device
CN111462388A (en)Bill inspection method and device, terminal equipment and storage medium
CN112149401A (en)Document comparison identification method and system based on ocr
CN112766255A (en)Optical character recognition method, device, equipment and storage medium
CN117612182A (en)Document classification method, device, electronic equipment and medium
CN115083024B (en) Signature recognition method, device, medium and equipment based on regional division
CN114612919B (en)Bill information processing system, method and device
CN102968610A (en)Method and device for processing receipt images
CN114419626B (en)High-precision bill identification method and system based on OCR technology
JosephAdvanced digital image processing technique based optical character recognition of scanned document
Almohri et al.A real-time DSP-based optical character recognition system for isolated Arabic characters using the TI TMS320C6416T
CN112632926A (en)Data processing method and device for bill, electronic equipment and storage medium
CN111881880A (en)Bill text recognition method based on novel network
Kumar et al.Line based robust script identification for indianlanguages
CN116030469A (en)Processing method, processing device, processing equipment and computer readable storage medium
CN115661839A (en)Bill filing method, device, terminal and storage medium

Legal Events

DateCodeTitleDescription
PB01Publication
PB01Publication
SE01Entry into force of request for substantive examination
SE01Entry into force of request for substantive examination
GR01Patent grant
GR01Patent grant

[8]ページ先頭

©2009-2025 Movatter.jp