JPH03260888A - Standard pattern preparing method - Google Patents
Standard pattern preparing methodInfo
- Publication number
- JPH03260888A JPH03260888A JP2057993A JP5799390A JPH03260888A JP H03260888 A JPH03260888 A JP H03260888A JP 2057993 A JP2057993 A JP 2057993A JP 5799390 A JP5799390 A JP 5799390A JP H03260888 A JPH03260888 A JP H03260888A
- Authority
- JP
- Japan
- Prior art keywords
- pattern
- standard pattern
- feature
- function
- density
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Landscapes
- Character Discrimination (AREA)
Abstract
Description
【発明の詳細な説明】
〔産業上の利用分野〕
本発明は、パターンマツチングによる文字認識処理に用
いる標準パターンの作成方法に関する。DETAILED DESCRIPTION OF THE INVENTION [Field of Industrial Application] The present invention relates to a method for creating a standard pattern used in character recognition processing by pattern matching.
パターンマツチングを用いた一般的な文字認識処理手順
を第2図に示す。同図において、ステップS1は文字画
像入力、ステップS2は特徴抽出、ステップS3は字種
判別、を示す。FIG. 2 shows a general character recognition processing procedure using pattern matching. In the figure, step S1 shows character image input, step S2 shows feature extraction, and step S3 shows character type discrimination.
ここで、字種判別部(ステップ33)では、認識対象と
なる文字ごとにあらかじめ用意された標準パターンと入
力画像から抽出された特徴パターンとの間で類似度が算
出され、入力文字の字種が判別される。Here, the character type discrimination unit (step 33) calculates the degree of similarity between a standard pattern prepared in advance for each character to be recognized and the feature pattern extracted from the input image, and calculates the degree of similarity between the character type of the input character. is determined.
類似度Gは、標準パターンと人力画像の特徴パターンそ
れぞれをn次元のベクトルS (n)、 P (n)と
して、以下の式から求められる。The degree of similarity G is obtained from the following equation, where the standard pattern and the characteristic pattern of the human image are respectively n-dimensional vectors S (n) and P (n).
ここで、(s、p)=Σ= S (i) P (i)で
あり、ベクトルの内積をあられす。(S、S)、(P、
P)はそれぞれ、標準パターン、特徴パターンの自己相
関値と呼ばれる。Here, (s, p) = Σ = S (i) P (i), and the inner product of the vectors is calculated. (S, S), (P,
P) are called autocorrelation values of the standard pattern and the characteristic pattern, respectively.
このようにして使用される標準パターンの一般的な作成
手順を第3図に示す。A general procedure for creating a standard pattern used in this manner is shown in FIG.
同図に見られるように、ステップSlで得た複数のサン
プルからの特徴抽出をステップS2で行い、その結果を
ステップS3で各要素ごとに足し込み、ステップS4で
得られる総和パターンに対し、必要に応じた濃度変換を
ステップS5で施して標準パターンを得る。ここで言う
濃度とは特徴要素の大きさをさす。以後同様とする。As seen in the figure, features are extracted from the multiple samples obtained in step Sl in step S2, the results are added for each element in step S3, and the necessary A standard pattern is obtained by performing density conversion according to step S5. The density here refers to the size of the feature element. The same shall apply thereafter.
一般に濃度変換としては、それぞれの標準パターンの中
の最大要素を所定の値Fに揃えるように、他の要素もそ
れに比例して一律的に大きさを変換する処理が行われる
。複数ある字種の中のj番目字種の総和パターンJ(n
)の最大要素をM、とすれば、標準パターンS、(n)
のi番目の要素は、5J(i)=Hj(i)・F/MJ
・・・・・・(1)として得られる。Generally, density conversion is performed by uniformly converting the size of other elements in proportion to the largest element in each standard pattern so as to equalize it to a predetermined value F. Total pattern J(n
), the standard pattern S, (n)
The i-th element of is 5J(i)=Hj(i)・F/MJ
......obtained as (1).
これは認識処理を行う装置の有効桁数や、標準パターン
の格納領域の制約から、限られたビット幅を有効に使う
ためである。または、入力画像の特徴パターンと標準パ
ターンの濃度のオーダを一致させるため、総和パターン
の各要素をサンプル数で割る処理が行われる。This is to effectively use the limited bit width due to the number of effective digits of the device that performs the recognition process and the storage area of the standard pattern. Alternatively, in order to match the density order of the characteristic pattern of the input image and the standard pattern, a process is performed in which each element of the total pattern is divided by the number of samples.
所で、このようにして作成された標準パターンが、結果
的に、少数の突出した値の特徴要素をもつことがあり、
その場合、これらの突出した要素が字種の判別に大きな
影響を与えることが判明した。すなわち、文字の変動に
弱くなり、認識率が低くなる。However, the standard pattern created in this way may end up having a small number of characteristic elements with outstanding values.
In that case, it was found that these prominent elements had a significant influence on character type discrimination. In other words, it becomes vulnerable to character fluctuations and the recognition rate becomes low.
第4図は、従来手法で作成された標準パターンにおける
濃度分布を示したグラフであるが、横軸に沿って濃度2
55.208及び207の所にそれぞれ、1個の画素が
存在し、少数の突出した値の特徴要素になっていること
が分かる。Figure 4 is a graph showing the density distribution in a standard pattern created using the conventional method.
It can be seen that there is one pixel at each of 55.208 and 207, which is a feature element with a small number of outstanding values.
本発明の目的は、かかる問題点を改善し、少数の突出し
た値の特徴要素をもつ標準パターンの場合には、それら
の特徴要素をうまく処理することにより、認識率の低下
を招かないようにした標準パターンの作成方法を提供す
ることにある。The purpose of the present invention is to improve this problem and, in the case of a standard pattern having feature elements with a small number of outstanding values, to prevent the recognition rate from decreasing by effectively processing those feature elements. The objective is to provide a method for creating standard patterns.
標準パターン作成時、各サンプルの特徴抽出結果を足し
込んだ総和パターンが、少数の突出した特徴要素をもつ
とき、該総和パターンの各特徴要素に対し、a乗根(a
>1.0)をとる演算や対数に類似した関数f (x)
を用いた演算をほどこすことにより、第6図に示すよう
な濃度の非線形変換を行い、この後、必要に応じた濃度
の線形変換を行う。When creating a standard pattern, if the summation pattern obtained by adding up the feature extraction results of each sample has a small number of prominent feature elements, the a-th root (a
>1.0) and functions similar to logarithms f (x)
By performing calculations using , non-linear conversion of density as shown in FIG. 6 is performed, and then linear conversion of density is performed as required.
f (x)は、任意のX(≧0)に対して以下の条件を
満たすような凸関数である。f (x) is a convex function that satisfies the following conditions for any X (≧0).
f(x) −f(x−711x)≧f(x+、dx)−
f(x)≧0・・・・・・(2)
さらに、以下の条件を満たす。f(x) −f(x−711x)≧f(x+,dx)−
f(x)≧0 (2) Furthermore, the following conditions are satisfied.
ここで、Axは任意の正の実数であり、関数の増分を表
す。Here, Ax is any positive real number and represents the increment of the function.
上記の如き非線形変換を施すことにより、突出した特徴
要素を、そうでない特徴要素よりも一段と押え込むこと
が可能になるから、これにより、文字の変動に強い安定
した標準パターンを得ることができ、従来と同様の処理
を用いた認識処理において認識率を向上させることがで
きる。By applying the non-linear transformation as described above, it is possible to suppress the prominent feature elements to a greater extent than the non-protruding feature elements, so it is possible to obtain a stable standard pattern that is resistant to character fluctuations. The recognition rate can be improved in recognition processing using processing similar to conventional methods.
〔実施例]
第1図は本発明の一実施例としての標準パターン作成方
法を示すフローチャートである。同図に見られるように
、各サンプル画像に対し順次特徴抽出処理が施され足し
込まれる。この総和パターンに対して、ステップS5に
見られる如く、各特徴要素のa乗根(aは1.0より大
きな実数)をとり、その結果で新たな総和パターンを得
る。すなわち、a乗根をとる演算は、下記に示すような
非線形関数f (x)を用いた演算により、非線形な濃
度変換を行うことを意味する。[Embodiment] FIG. 1 is a flowchart showing a standard pattern creation method as an embodiment of the present invention. As seen in the figure, feature extraction processing is sequentially performed and added to each sample image. As shown in step S5, the a-th root of each characteristic element (a is a real number greater than 1.0) is taken for this summation pattern, and a new summation pattern is obtained as a result. That is, the operation of taking the a-th root means performing nonlinear concentration conversion by the operation using the nonlinear function f (x) as shown below.
f (x)=χI/11 ・・・・
・・(4)その後、更に、従来も行われていた濃度変換
(第1図のステップS6或いは第3図のステップS5)
として、前記(1)式を用いた濃度変換が行われる。こ
こでは、各特徴要素のビット長を8ピントに圧縮するた
めF=255として一律的な濃度変換を行っている。従
来手法と異なるのは各特徴要素の8乗根をとって非線形
変換を行う点である。f (x)=χI/11...
...(4) After that, density conversion, which has been conventionally performed (step S6 in FIG. 1 or step S5 in FIG. 3)
Then, density conversion is performed using equation (1) above. Here, in order to compress the bit length of each feature element to 8 points, uniform density conversion is performed with F=255. The difference from the conventional method is that nonlinear transformation is performed by taking the 8th root of each feature element.
従来手法と本手法(a=4)を用いた場合の標準パター
ンの例を、“あ”という文字について、学習サンプル数
80個で、特徴抽出手法として、(16x16)メツシ
ュに正規化された画像の輪郭画素を白画素を挟んだ輪郭
画素間の相対距離で複数の特徴面(例えば15面)に割
り付ける手法(特願平11年第71757号)を用いて
作成し比較してみると、後者の方が、突出した特徴要素
を低く抑え、その後、全体をカサ上げするように濃度変
換しているので、全体に濃度が濃くなっているのが分か
る。An example of a standard pattern when using the conventional method and the present method (a = 4) is an image normalized to a (16x16) mesh for the character "a" with 80 learning samples and a feature extraction method. When creating and comparing contour pixels using a method (Patent Application No. 71757 of 1999) in which the contour pixels of In this case, the prominent feature elements are kept low, and then the density is converted to increase the bulk of the whole image, so you can see that the overall density is higher.
本発明による手法で標準パターンをひらがな75文字に
対して作成し、判別式として類似度を用いた従来と同様
の認識処理を学習サンプル80文字、ひらがな75字種
に対して行った場合、従来手法で作成された標準パター
ンを用いた場合の認識率90.5%に対して、95.5
%に認識率が向上した。変数a (8乗根のa)と認識
率の関係を第5図に示したので参照されたい。When a standard pattern is created for 75 hiragana characters using the method according to the present invention, and the same recognition processing as before using similarity as a discriminant is performed on 80 learning samples and 75 hiragana character types, the conventional method The recognition rate was 95.5% compared to 90.5% when using the standard pattern created by
% of the recognition rate. Please refer to FIG. 5, which shows the relationship between the variable a (a of the 8th root) and the recognition rate.
本発明は、前記特徴抽出手法に依存するものではなく、
伝播停止処理を用いた背景特徴(特開昭59−792号
公報)でも、その有効性が確認されている。The present invention does not depend on the feature extraction method described above,
The effectiveness of background features using propagation stop processing (Japanese Patent Laid-Open No. 59-792) has also been confirmed.
ここでは前記(4)式において、a=4としたが、aの
最適値は認識対象とするフィールドや特徴抽出手法、判
別式の種類によって異なると考えられる。また、総和パ
ターンの各特徴要素の対数をとる非線形変換も、8乗根
をとる非線形変換と同様に有効である。すなわち、f
(x)として下記に示す関数を用いる。Here, in equation (4), a=4, but the optimal value of a is considered to vary depending on the field to be recognized, the feature extraction method, and the type of discriminant. Furthermore, nonlinear transformation that takes the logarithm of each characteristic element of the summation pattern is also effective, as is nonlinear transformation that takes the 8th root. That is, f
The function shown below is used as (x).
f (x)=log+o(x + 1) −
−(5)ただし、前記(4〉式におけるaを大きくして
いくと特徴パターンは2値画像に近づき、認識率は低下
する。そこで、サンプル画像から抽出された特徴パター
ンの各要素をa乗して足し込み、前記総和パターンを得
ることで、総和パターンにおける濃度分布をある程度保
存したまま、突出した特徴要素の値を低く押さえること
ができる。f(x)=log+o(x+1)−
- (5) However, as a in equation (4) is increased, the feature pattern approaches a binary image and the recognition rate decreases.Therefore, each element of the feature pattern extracted from the sample image is raised to the a power. By adding them together to obtain the summation pattern, it is possible to keep the values of prominent feature elements low while preserving the density distribution in the summation pattern to some extent.
本発明によれば、標準パターン作成時に、各サンプル文
字画像から抽出した特徴要素の足し込み総和パターンに
おいて、生しることのある突出した特徴要素を押え込む
ような非線形変換を、各特徴要素に施して標準パターン
を作成することにより、認識率が向上するという利点が
ある。非線形変換として4乗根を用いる場合、認識率は
90.5%から95.5%に向上し、認識率にして53
%の改善がみられた。According to the present invention, when creating a standard pattern, a nonlinear transformation is applied to each feature element to suppress any prominent feature elements that may occur in the summation pattern of feature elements extracted from each sample character image. By applying this method to create a standard pattern, there is an advantage that the recognition rate is improved. When using the fourth root as a nonlinear transformation, the recognition rate improves from 90.5% to 95.5%, and the recognition rate is 53%.
% improvement was seen.
第1図は本発明の一実施例を示すフローチャート、第2
図はパターンマツチングを用いる一般的な文字認識手順
を示すフローチャート、第3図は従来の標準パターン作
成方法を示すフローチャート、第4図は従来手法により
得られる標準パターンの濃度分布を示すグラフ、第5図
は変数aと認識率の関係を示す特性図、第6図は本発明
による非線形変換の特性例を示す特性図、である。
符号の説明
S1〜S6・・・ステップ
第 1 図
第
2
図
第
図
第
5
図
第
図FIG. 1 is a flowchart showing one embodiment of the present invention, and FIG.
The figure is a flowchart showing a general character recognition procedure using pattern matching. FIG. 5 is a characteristic diagram showing the relationship between the variable a and the recognition rate, and FIG. 6 is a characteristic diagram showing an example of the characteristics of the nonlinear transformation according to the present invention. Explanation of symbols S1 to S6...Step 1 Figure 2 Figure 5 Figure 5
Claims (1)
いる標準パターンを複数のサンプル文字画像から作成す
る標準パターン作成方法において、各サンプル文字画像
からの特徴抽出結果を各特徴要素毎に足し込んでその総
和パターンを得た後、得られた該総和パターンを構成す
る各特徴要素に対し、その大きさのa乗根(但し、a>
1.0)をとる演算又は対数をとる演算の如き、非線形
変換を行う演算を施して、その演算結果から成る総和パ
ターンを得る段階を含むことを特徴とする標準パターン
作成方法。1) In a standard pattern creation method in which a standard pattern used for character recognition processing by pattern matching is created from multiple sample character images, the feature extraction results from each sample character image are added for each feature element to create a summation pattern. After obtaining the summation pattern, the a-th root of its size (where a>
1.0) or a logarithm, which performs an operation that performs nonlinear transformation, and obtains a summation pattern consisting of the result of the operation.
Priority Applications (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| JP2057993A JP2749692B2 (en) | 1990-03-12 | 1990-03-12 | Standard pattern creation method |
Applications Claiming Priority (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| JP2057993A JP2749692B2 (en) | 1990-03-12 | 1990-03-12 | Standard pattern creation method |
Publications (2)
| Publication Number | Publication Date |
|---|---|
| JPH03260888A true JPH03260888A (en) | 1991-11-20 |
| JP2749692B2 JP2749692B2 (en) | 1998-05-13 |
Family
ID=13071535
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| JP2057993A Expired - Lifetime JP2749692B2 (en) | 1990-03-12 | 1990-03-12 | Standard pattern creation method |
Country Status (1)
| Country | Link |
|---|---|
| JP (1) | JP2749692B2 (en) |
Citations (3)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| JPS60160488A (en) * | 1984-02-01 | 1985-08-22 | Nec Corp | Framing system of standard pattern in pattern recognition |
| JPS63313283A (en) * | 1987-06-17 | 1988-12-21 | Agency Of Ind Science & Technol | Image normalizing method |
| JPH01116892A (en) * | 1987-10-30 | 1989-05-09 | Agency Of Ind Science & Technol | Non-linear normalizing system |
-
1990
- 1990-03-12 JP JP2057993A patent/JP2749692B2/en not_active Expired - Lifetime
Patent Citations (3)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| JPS60160488A (en) * | 1984-02-01 | 1985-08-22 | Nec Corp | Framing system of standard pattern in pattern recognition |
| JPS63313283A (en) * | 1987-06-17 | 1988-12-21 | Agency Of Ind Science & Technol | Image normalizing method |
| JPH01116892A (en) * | 1987-10-30 | 1989-05-09 | Agency Of Ind Science & Technol | Non-linear normalizing system |
Also Published As
| Publication number | Publication date |
|---|---|
| JP2749692B2 (en) | 1998-05-13 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| CN113011581B (en) | Neural network model compression method and device, electronic equipment and readable storage medium | |
| Cai et al. | Deep learning with low precision by half-wave gaussian quantization | |
| CN110874636B (en) | A neural network model compression method, device and computer equipment | |
| CN109800865B (en) | Neural network generation and image processing method and device, platform and electronic equipment | |
| CN109284761B (en) | Image feature extraction method, device and equipment and readable storage medium | |
| US20230143985A1 (en) | Data feature extraction method and related apparatus | |
| CN110909874A (en) | Convolution operation optimization method and device for neural network model | |
| JP7020331B2 (en) | Clustering equipment, methods, and programs | |
| CN117997619B (en) | A Web attack end-to-end detection method, system and medium based on deep learning | |
| KR102476695B1 (en) | Method and server for generating adaptive convolution for text classification | |
| KR20240113154A (en) | Method and apparatus for classifying text using text classifying model | |
| Hirata et al. | Iterative design of morphological binary image operators | |
| KR102726377B1 (en) | Apparatus and method for generating binary neural network | |
| JPH03260888A (en) | Standard pattern preparing method | |
| US11853387B2 (en) | Method, electronic device and storage medium for data projection | |
| Prono et al. | A non-conventional sum-and-max based neural network layer for low power classification | |
| Sattiraju et al. | Adaptive BPSO based feature selection and skin detection based background removal for enhanced face recognition | |
| CN111382764B (en) | Neural network model building method and device for face recognition or gesture recognition and computer readable storage medium | |
| KR102834846B1 (en) | Method, computing device and computer program for calculating softmax with divided processing of large input value | |
| Jeng et al. | Multidimensional Quantum Convolution with Arbitrary Filtering and Unity Stride | |
| Tao et al. | Image segmentation based on PCNN model | |
| CN110889413A (en) | Terminal equipment determining method and device for realizing optical character recognition | |
| US20240378436A1 (en) | Partial Quantization To Achieve Full Quantized Model On Edge Device | |
| Barrah et al. | New technique for face recognition based on singular value decomposition (SVD) | |
| KR102707241B1 (en) | Method and Apparatus for Image Data Processing in Image Signal Processor Adapted Neural Algorithm |