JPH0352058A - Document processor for voice input - Google Patents
Document processor for voice inputInfo
- Publication number
- JPH0352058A JPH0352058A JP1188126A JP18812689A JPH0352058A JP H0352058 A JPH0352058 A JP H0352058A JP 1188126 A JP1188126 A JP 1188126A JP 18812689 A JP18812689 A JP 18812689A JP H0352058 A JPH0352058 A JP H0352058A
- Authority
- JP
- Japan
- Prior art keywords
- voice
- section
- speech
- character
- displayed
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
Landscapes
- Document Processing Apparatus (AREA)
Abstract
Description
【発明の詳細な説明】
[概要]
音声入力により文書等を作成する音声入力の文書処理装
置に係り、特に文字表示される際の文字の濃さ、大きざ
、字体等を音声の大きさ、音声の高さ、音声の長さによ
り指示する音声入力の文書処理装置に関し、
表示される文字の属性を音声の特徴により判定して決定
し、操作性の優れた文書処理装置を提供することを目的
とし、
入力音声を電気信号に変換する信号変換手段と、変換さ
れた電気信号を分析する音声分析手段と、分析結果にし
たがって音声区間を切り出す音声区間切り出し手段と、
切り出された音声区間の音声を認識する!!識手段と、
!!tA結果を文字コードに変換する文字コード発生手
段と、変換された文字コードより文書を編集する編集手
段と、変換された電気信号から入力音声の特徴を抽出す
る音声特徴抽出手段と、抽出された音声の特徴を判定し
て文字コードで示される文字の属性を決定する判定手段
と、を備えて構成されている.
[産業上の利用分野]
本発明は音声入力により文書等を作成する音声入力の文
書処理装置に間するものである.[従来の技術]
従来の音声入力の文書処理装置は、第6図に示す様に発
声された音声は電気信号変換部30(例えばマイク)に
より音声電気信号に変換される.電気信号は音声分析部
32においーで十数〜数十ms毎にその音声のスペクト
ルと音声のパワー(音量)が分析され、ベクトル量のス
ペクトルとスカラ量パワーの時系列が出力される.
音声区間切り出し部34では、特に分析されたパワー時
系列を用いて、音声区間の切り出しが行われる.すなわ
ち、音声パワー時系列上で、音声の始端と終端が求めら
れる.この始端と終端はスペクトル時系列上の音声の始
端と終端と一致することになる.
認識部36では音声区間切り出し部34で求まった始端
と終端の間のスペクトルを用いて認識を行う.認識方式
については各種方式が間発されており、たとえば、動的
計画法(DP法)を用いた方式、ヒドゥンマルコフモデ
ル(HMM)を用いた方式、判別分析を用いた方式、ニ
ューラルネットワークを用いた方式がある.又認識単位
としては、単音節、文字、単語、文節及び文章がある.
何れの認識単位であっても、言!!(日本語や英語など
)には同音異義語があるために仮名漢字変換の候補選択
を行う必要がある.
よって認識部36としては音素の記号列(子音と母音の
列であって、日本語の場合はほぼローマ字表記)を出力
するにとどまるものが多い.仮名漢字変換の候補選択は
文字コード発生部38および第一の指示部40によって
行われる.すなわち、まず認識部36より送られてきた
音素記号列を漢字列に変換し、その第一候補を編集部4
2を通して表示部44に表示する.
次に使用者は、表示された漢字候補が望ましくない場合
は第一の指示部40(具体的には次候補キー)を押す.
文字コード発生部38では、第一の指示部40が押され
た場合には、漢字候補の次の候補を表示する.
ここで使用者は正しい漢字が表示された場合には、次の
発声を行う.その時点で前の漢字候補は確定する.
編集g42では入力されたく表示された)文章の編集を
、第二の指示部46(スイッチ等で構成される)からの
指示により行う.
編集の具体的内容は、文字の太き・さの変更、字体(明
朝などやゴシック体など)の変更、文字の濃さの変更、
誤り入力の削除などがある.表示部44は文字列(文章
など)を例えばCRTディスプレイに表示し、縞集する
.
そして印刷部48は&W集された文字列(文章など)を
用紙に印刷する.
[発明が解決しようとする譚題]
しかし、このような従来の装置であると、例えば文字の
大きざを変更するためには第二の指示部46からの指示
が必要となり、このため音声入力を中断してスイッチ操
作を行う必要があり、極めて使い勝手が悪いものであっ
た.
本発明は上記従来の課題に対処するために創案されたも
のであり、表示される文字の属性を音声の特徴により判
定して決定できる操作性の優れた音声入力の文書処理装
置を提供することを目的とする.
[課題を解決するための手段]
上記目的を達成するために、本発明にかかる装置は第1
t!!Iのように構成ざれている.同図において、音声
は信号変換手段10によって電気信号に変換される.
変換された電気信号は音声分析手段l2によって分析さ
れる.
分析結果にしたがって音声区間切り出し手段14により
音声区間が切り出される.
切り出された音声区間の音声は認識手段16によりL!
識され、その認識結果は文字コード発生手段18により
文字コードに変換される.さらに、変換された文字コー
ドから編集手段20により文書が編集される.
ここで、音声特徴抽出手段22により電算信号から入力
音声の特徴が抽出され、判定手段24により抽出された
音声の特徴が判定ざれ、文字コードで示される文字の属
性が決定される.[作用]
本発明では、音声特徴抽出手段2・2は音声区間内の音
声パワー、音声ピッチ、また音声区間長などの音声の特
徴を抽出する.
判定手段24は音声特徴抽出手段22で得られた音声の
各特徴を判定し、表示文字の大きさ、濶さ、字体などを
決定して編集手段20に伝える.すなわち、判定手段2
4は音声の情報から、従来装置における第二の指示部4
6の代わりを行なう.
よって使用者は従来のようにスイッチ等により指示を与
える必要はなく、たとえば、大きな声で喋れば潔い文字
で表示され、ゆっくり喋れば倍角文字で表示される等の
操作が可能となる.[実施例]
以下、図面に基づいて本発明にかかる装置の好適な実施
例を説明する.
第2E!Iには第一実施例の全体構成が示されており、
同図において従来例と同一符号のブロックは同一の機能
を有する.
ここで、ピッチ計算部50は音声区間切り出し部34で
得られた音声区間(始端から終端)内の音声の平均ビッ
チFを計算する.
パワー計算部52は音声区間切り出し部34で得られた
音声区間(始端から終端〉内の音声の平均パワーPを計
算する.
さらに区間長計算部54は音声区間切り出し部34で得
られた始端と終端から音声区間の長さ(時間長)Lを計
算する.
そして音素数計算部56は認識部36で得られた認識候
補のうち、一位の候補(最も允もらしい候補)の音素数
Cをカウントする.
判定部58ではピッチ計算部50、パワー計算部52、
区間長計算部54、音素数計算部56で得られたfli
F.P.L.およびCに基づいて総合的な判断をくだし
、その結果をI!集部42に通知する.編集部42では
これら結果に基づいて決定された文字の属性で表示部4
4に表示し、また印刷84Bでは印字する.
ここで、判定の基準としては、たとえば第4図に示す基
準が用いられる.
すなわち、第4図から理解されるようにピッチ計算部5
0、パワー計算部52で得られた値F.Pが、判定[5
8であらかじめ決められたilll[Fth,pthよ
り大なるときは濃い文字で、またイタリック文字で表示
するよう判定される.次に区間長計算部54、音素数計
算[5Bで得られた値L. CにおいてそのL/Cなる
罐が判定部58であらかじめ決められた閾[T}{より
大なるときは倍角文字で表示するよう判定される.次に
、第3図に本発明の第二実施例を示す.第一実施旌で使
用される判定基準(第41!I参wi)の閾値Pth.
Fth.THは不特定多数を前提にして決定されたもの
であり、現実に音声入力する操作者の微妙な音声の違い
まで考慮されたものではない.
そこで本実施例では、現実に音声入力する操作者の音声
特徴を記憶し、当該音声を判定基準の閾値として使用す
るものである.
そして、本実施例では例えば第5図に示す判定基準が使
用され、パワー計算部52で得られた値Pが前回特徴記
憶部60で記憶された一発声前の罐pbより50%以上
大なるときには濃い文字で表示するように判定される.
また、ピッチ計算部50で得られた値Fが前回特徴記憶
部60で記憶された一発声前の[Fbより50%以上大
なるときはイタリック文字で表示するように判定される
.
さらに、区間長計算部54、音素計算部56で得られた
ML.CにおいてL/Cの値が前回特徴記憶部60で記
憶ざれた一発声前のfllL b /C bより50%
以上大なるときは倍角文字で表示するよう判定される.
このように本実施例によれば、現実に音声入力を行う操
作者の微妙な音声の違いまで把握することができ、本装
置の操作性をより向上させることができる.
[発明の効果コ
以上説明したように、本発明によれば、表示される文字
の大きさ、濃さおよび字体・等の変換を、音声入力する
際の音声の大きさやピッチあるいは発声の速さを変える
だけで行うことができ、操作性の優れた音声入力の文書
処理装置を提供できることになる,[Detailed Description of the Invention] [Summary] This invention relates to a voice input document processing device that creates documents, etc. by voice input, and in particular, when characters are displayed, the density, size, font, etc. of the characters can be adjusted according to the size of the voice. An object of the present invention is to provide a document processing device with excellent operability, which determines and determines the attributes of displayed characters based on the characteristics of the voice, regarding a document processing device that inputs instructions based on the pitch and length of the voice. A signal conversion means for converting input speech into an electrical signal, a speech analysis means for analyzing the converted electrical signal, and a speech section cutting means for cutting out a speech section according to the analysis result.
Recognize the audio of the extracted audio section! ! means of knowledge,
! ! A character code generation means for converting the tA result into a character code, an editing means for editing a document from the converted character code, an audio feature extraction means for extracting the characteristics of the input voice from the converted electric signal, and an extracted character code. A determination means for determining the attributes of a character indicated by a character code by determining the characteristics of the voice. [Industrial Field of Application] The present invention relates to a voice input document processing device that creates documents, etc. by voice input. [Prior Art] In a conventional voice input document processing device, as shown in FIG. 6, uttered voice is converted into an audio electrical signal by an electrical signal converter 30 (for example, a microphone). The electrical signal is analyzed by the audio analysis section 32 for its audio spectrum and audio power (volume) every 10 to several tens of milliseconds, and a time series of vector quantity spectrum and scalar quantity power is output. In the speech segment extraction unit 34, speech segments are extracted using the analyzed power time series. In other words, the beginning and end of the voice can be found on the voice power time series. These start and end points coincide with the start and end points of the voice on the spectral time series. The recognition unit 36 performs recognition using the spectrum between the start and end determined by the speech segment extraction unit 34. Various recognition methods have been developed, including methods using dynamic programming (DP method), methods using hidden Markov models (HMM), methods using discriminant analysis, and methods using neural networks. There is a method. Recognition units include monosyllables, characters, words, phrases, and sentences.
No matter what unit of recognition it is, it is a word! ! (Japanese, English, etc.) have homonyms, so it is necessary to select candidates for kana-kanji conversion. Therefore, in many cases, the recognition unit 36 only outputs a symbol string of phonemes (a string of consonants and vowels, which in the case of Japanese is almost always written in Roman letters). Candidate selection for kana-kanji conversion is performed by the character code generation section 38 and first instruction section 40. That is, first, the phoneme symbol string sent from the recognition section 36 is converted into a kanji string, and the first candidate is sent to the editing section 4.
2 on the display section 44. Next, if the displayed kanji candidate is not desirable, the user presses the first instruction section 40 (specifically, the next candidate key).
In the character code generation section 38, when the first instruction section 40 is pressed, the next kanji candidate is displayed. Here, if the correct kanji is displayed, the user makes the following utterance. At that point, the previous kanji candidate is confirmed. In editing g42, the text (displayed to be input) is edited according to instructions from the second instruction section 46 (consisting of switches, etc.). The specific contents of editing include changing the thickness and width of the characters, changing the font (Mincho, etc., Gothic, etc.), changing the darkness of the characters,
This includes deleting incorrect entries. The display unit 44 displays character strings (sentences, etc.) on, for example, a CRT display, and displays them as stripes. Then, the printing unit 48 prints the &W collected character strings (sentences, etc.) on paper. [Problem to be solved by the invention] However, with such a conventional device, an instruction from the second instruction unit 46 is required to change the size of characters, for example, and therefore voice input is not required. It was extremely inconvenient to use, as it was necessary to interrupt the process to operate the switch. The present invention has been devised to address the above-mentioned conventional problems, and provides a voice input document processing device with excellent operability that can determine the attributes of displayed characters by determining them based on the characteristics of voice. With the goal. [Means for Solving the Problems] In order to achieve the above object, the device according to the present invention has the following features:
T! ! It is structured like I. In the figure, audio is converted into an electrical signal by signal conversion means 10. The converted electrical signal is analyzed by the audio analysis means l2. According to the analysis result, a speech section is cut out by the speech section cutting means 14. The voice of the cut out voice section is recognized by the recognition means 16 as L!
The recognition result is converted into a character code by the character code generating means 18. Furthermore, the document is edited by the editing means 20 using the converted character code. Here, the voice feature extracting means 22 extracts the characteristics of the input voice from the computer signal, and the determining means 24 determines the extracted voice characteristics and determines the attributes of the character indicated by the character code. [Operation] In the present invention, the speech feature extraction means 2 extracts speech characteristics such as speech power, speech pitch, and speech section length within a speech section. The determining means 24 determines each feature of the voice obtained by the voice feature extracting means 22, determines the size, sharpness, font, etc. of displayed characters, and transmits them to the editing means 20. That is, determining means 2
4 is the second instruction unit 4 in the conventional device based on the audio information.
Substitute 6. Therefore, the user does not need to give instructions using switches or the like as in the past; for example, if the user speaks loudly, the text will be displayed in plain text, or if he speaks slowly, the text will be displayed in double-width text. [Embodiments] Hereinafter, preferred embodiments of the apparatus according to the present invention will be described based on the drawings. 2nd E! I shows the overall configuration of the first embodiment,
In the figure, blocks with the same symbols as in the conventional example have the same functions. Here, the pitch calculation section 50 calculates the average pitch F of the speech within the speech section (from the beginning to the end) obtained by the speech section cutting section 34. The power calculation section 52 calculates the average power P of the speech within the speech section (from the start to the end) obtained by the speech section extraction section 34. Furthermore, the section length calculation section 54 calculates the average power P of the speech within the speech section (starting end to end) obtained by the speech section extraction section 34. The length of the speech interval (time length) L is calculated from the end. Then, the phoneme number calculation unit 56 calculates the phoneme number C of the first candidate (the most plausible candidate) among the recognition candidates obtained by the recognition unit 36. The determination unit 58 includes a pitch calculation unit 50, a power calculation unit 52,
fli obtained by the section length calculation unit 54 and the phoneme number calculation unit 56
F. P. L. Make a comprehensive judgment based on C and I! Notify the collection department 42. The editing section 42 displays the character attributes determined based on these results on the display section 4.
4 and printed in print 84B. Here, as a criterion for judgment, for example, the criterion shown in FIG. 4 is used. That is, as understood from FIG.
0, the value F. obtained by the power calculation unit 52. P makes a judgment [5
When the value is larger than the predetermined illll[Fth, pth in step 8, it is determined that the text is displayed in dark or italic text. Next, the section length calculation unit 54 calculates the number of phonemes [the value L obtained in step 5B]. When the can L/C in C is larger than a predetermined threshold [T}{, the determination unit 58 determines to display it in double-width characters. Next, FIG. 3 shows a second embodiment of the present invention. Threshold value Pth of the judgment criteria used in the first implementation (see 41st!I)
Fth. The TH was determined based on the assumption that the number of people would be unspecified, and it did not take into account the subtle differences in the voices of the operators who actually input voice. Therefore, in this embodiment, the voice characteristics of the operator who actually inputs voice are stored, and the voice is used as a threshold value for the judgment criterion. In this embodiment, for example, the determination criteria shown in FIG. Sometimes it is determined to be displayed in dark letters. Further, when the value F obtained by the pitch calculation section 50 is 50% or more larger than [Fb of the previous utterance stored in the previous feature storage section 60, it is determined that the value is displayed in italics. Furthermore, the ML. At C, the value of L/C is 50% from the value of fllL b /C b stored in the previous feature storage unit 60 before the first utterance.
If the value is larger than that, it will be determined to display double-width characters. As described above, according to this embodiment, it is possible to grasp even subtle differences in the voices of operators who actually perform voice input, thereby further improving the operability of the present device. [Effects of the Invention] As explained above, according to the present invention, the size, density, font, etc. of displayed characters can be converted by changing the volume, pitch, or speed of voice input when inputting voice. This can be done by simply changing the
第l図は本発明の原理説明図、第21!Iは本発明の第
1実施例構成図、第3図は本発明の第2実施例構成図、
第4図は3111実jv例で用いられる判定基準の具体
例、第5図は第2実施例で用いられる判定基準の具体例
、第6v;!iは従来例の構成説明図である.
10・・・信号変換手段、
12・・・音声分析手段、
14・・・音声区間切り出し手段、
I6・・・認識手段、
18・・・文字コード発生手段、
20・・・編集手段、
22・・・音声特徴抽出手段、
24・・・判定手段、
本発明の原理説明図
第 1rEJ
第
2
図
p > p th
F>Fth
L/C>TH
ならば
ならば
ならば
濃い文字
イタリック体文字
倍角文字
ここで、
pth,
Fth,
THは予め決められたa値
P>1.5XPb ならば 濃い文字(ボー
ルド文字)F>1.5XFb ならば イタ
リック体文字L/C> 1.5XLb/Cbならば
倍角文字ここで、Pb,Fb,Lb,Cbは一発声前
の各々の値である.第
6
図Figure 1 is a diagram explaining the principle of the present invention, and Figure 21! I is a block diagram of the first embodiment of the present invention, FIG. 3 is a block diagram of the second embodiment of the present invention,
Fig. 4 is a specific example of the judgment criteria used in the 3111 actual JV example, Fig. 5 is a specific example of the judgment criteria used in the second example, and 6th v;! i is an explanatory diagram of the configuration of a conventional example. 10...Signal conversion means, 12...Speech analysis means, 14...Speech section extraction means, I6...Recognition means, 18...Character code generation means, 20...Editing means, 22. ...Voice feature extraction means, 24...Judgment means, Principle explanatory diagram of the present invention No. 1rEJ Fig. 2 p>p th F>Fth L/C>TH If then then dark italic characters double-width characters Here, pth, Fth, TH are predetermined a values If P>1.5XPb If dark text (bold text) F>1.5XFb If italic text L/C> 1.5XLb/Cb
Double-width characters Here, Pb, Fb, Lb, and Cb are each value before one utterance. Figure 6
Claims (1)
、 変換された電気信号を分析する音声分析手段(12)と
、 分析結果にしたがって音声区間を切り出す音声区間切り
出し手段(14)と、 切り出された音声区間の音声を認識する認識手段(16
)と、 認識結果を文字コードに変換する文字コード発生手段(
18)と、 変換された文字コードより文書を編集する編集手段(2
0)と、 変換された電気信号から入力音声の特徴を抽出する音声
特徴抽出手段(22)と、 抽出された音声の特徴を判定して文字コードで示される
文字の属性を決定する判定手段(24)と、 を備えたことを特徴とする音声入力の文書処理装置。[Scope of Claims] Signal converting means (10) for converting input speech into an electrical signal, speech analysis means (12) for analyzing the converted electrical signal, and speech section cutting means for cutting out a speech section according to the analysis result. (14), and a recognition means (16) for recognizing the voice of the cut out voice section.
) and a character code generation means (
18) and editing means (2) for editing the document using the converted character code.
0), a voice feature extracting means (22) for extracting the characteristics of the input voice from the converted electrical signal, and a determining means (22) for determining the attributes of the character indicated by the character code by determining the characteristics of the extracted voice. 24) A voice input document processing device comprising:
Priority Applications (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| JP1188126A JPH0352058A (en) | 1989-07-20 | 1989-07-20 | Document processor for voice input |
Applications Claiming Priority (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| JP1188126A JPH0352058A (en) | 1989-07-20 | 1989-07-20 | Document processor for voice input |
Publications (1)
| Publication Number | Publication Date |
|---|---|
| JPH0352058A true JPH0352058A (en) | 1991-03-06 |
Family
ID=16218173
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| JP1188126A Pending JPH0352058A (en) | 1989-07-20 | 1989-07-20 | Document processor for voice input |
Country Status (1)
| Country | Link |
|---|---|
| JP (1) | JPH0352058A (en) |
Cited By (15)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US5426488A (en) * | 1992-10-19 | 1995-06-20 | Sharp Kabushiki Kaisha | Method of charging a built-in electrophotographic charge member |
| US5933681A (en) * | 1997-08-29 | 1999-08-03 | Canon Kabushiki Kaisha | Image forming apparatus |
| WO1999063425A1 (en) * | 1998-06-02 | 1999-12-09 | Sony Corporation | Method and apparatus for information processing, and medium for provision of information |
| US6292640B1 (en) | 1999-07-29 | 2001-09-18 | Canon Kabushiki Kaisha | Image forming apparatus for replenishing magnetic brush charging device with magnetic particles |
| KR20010106689A (en) * | 2000-05-22 | 2001-12-07 | 이성희 | Spray Gun for Painting |
| US6330411B1 (en) | 1999-07-29 | 2001-12-11 | Canon Kabushiki Kaisha | Replenishing container for replenishing charging particles and toner |
| US6549223B2 (en) | 2000-12-12 | 2003-04-15 | Canon Kabushiki Kaisha | Electrophotographic apparatus, process cartridge, and electrophotographic photosensitive member |
| US6999690B2 (en) | 2002-07-18 | 2006-02-14 | Canon Kabushiki Kaisha | Image forming apparatus |
| JP2008170505A (en) * | 2007-01-09 | 2008-07-24 | Yamaha Corp | Speech processing apparatus and program |
| US7463849B2 (en) | 2005-01-25 | 2008-12-09 | Ricoh Printing Systems, Ltd. | Image forming apparatus including static pre-eliminator |
| US7603062B2 (en) | 2005-09-16 | 2009-10-13 | Ricoh Company, Ltd. | Conductive member, and charging roller, process cartridge and image forming apparatus using same |
| US7693458B2 (en) | 2006-06-20 | 2010-04-06 | Ricoh Company, Ltd. | Charging device, process cartridge, and image forming apparatus |
| US7865114B2 (en) | 2006-11-21 | 2011-01-04 | Ricoh Company Limited | Image forming apparatus, image forming method and process cartridge |
| US7962055B2 (en) | 2006-04-19 | 2011-06-14 | Ricoh Company, Ltd. | Image forming apparatus |
| US8032050B2 (en) | 2006-10-27 | 2011-10-04 | Ricoh Company, Ltd. | Charge assembly and image formation apparatus including the same |
Citations (4)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| JPS5665242A (en) * | 1979-10-31 | 1981-06-02 | Toshiba Corp | Japanese language information input device |
| JPS56153395A (en) * | 1980-04-30 | 1981-11-27 | Nippon Telegraph & Telephone | Voice form kanji input system |
| JPS62299898A (en) * | 1986-06-18 | 1987-12-26 | 日本電気株式会社 | Character generator |
| JPS63111568A (en) * | 1986-10-29 | 1988-05-16 | Nec Corp | Kana/kanji converter with voice input |
-
1989
- 1989-07-20 JP JP1188126A patent/JPH0352058A/en active Pending
Patent Citations (4)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| JPS5665242A (en) * | 1979-10-31 | 1981-06-02 | Toshiba Corp | Japanese language information input device |
| JPS56153395A (en) * | 1980-04-30 | 1981-11-27 | Nippon Telegraph & Telephone | Voice form kanji input system |
| JPS62299898A (en) * | 1986-06-18 | 1987-12-26 | 日本電気株式会社 | Character generator |
| JPS63111568A (en) * | 1986-10-29 | 1988-05-16 | Nec Corp | Kana/kanji converter with voice input |
Cited By (16)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US5426488A (en) * | 1992-10-19 | 1995-06-20 | Sharp Kabushiki Kaisha | Method of charging a built-in electrophotographic charge member |
| US5933681A (en) * | 1997-08-29 | 1999-08-03 | Canon Kabushiki Kaisha | Image forming apparatus |
| KR100638198B1 (en) * | 1998-06-02 | 2006-10-26 | 소니 가부시끼 가이샤 | Information processing apparatus and method and information providing medium |
| WO1999063425A1 (en) * | 1998-06-02 | 1999-12-09 | Sony Corporation | Method and apparatus for information processing, and medium for provision of information |
| US6292640B1 (en) | 1999-07-29 | 2001-09-18 | Canon Kabushiki Kaisha | Image forming apparatus for replenishing magnetic brush charging device with magnetic particles |
| US6330411B1 (en) | 1999-07-29 | 2001-12-11 | Canon Kabushiki Kaisha | Replenishing container for replenishing charging particles and toner |
| KR20010106689A (en) * | 2000-05-22 | 2001-12-07 | 이성희 | Spray Gun for Painting |
| US6549223B2 (en) | 2000-12-12 | 2003-04-15 | Canon Kabushiki Kaisha | Electrophotographic apparatus, process cartridge, and electrophotographic photosensitive member |
| US6999690B2 (en) | 2002-07-18 | 2006-02-14 | Canon Kabushiki Kaisha | Image forming apparatus |
| US7463849B2 (en) | 2005-01-25 | 2008-12-09 | Ricoh Printing Systems, Ltd. | Image forming apparatus including static pre-eliminator |
| US7603062B2 (en) | 2005-09-16 | 2009-10-13 | Ricoh Company, Ltd. | Conductive member, and charging roller, process cartridge and image forming apparatus using same |
| US7962055B2 (en) | 2006-04-19 | 2011-06-14 | Ricoh Company, Ltd. | Image forming apparatus |
| US7693458B2 (en) | 2006-06-20 | 2010-04-06 | Ricoh Company, Ltd. | Charging device, process cartridge, and image forming apparatus |
| US8032050B2 (en) | 2006-10-27 | 2011-10-04 | Ricoh Company, Ltd. | Charge assembly and image formation apparatus including the same |
| US7865114B2 (en) | 2006-11-21 | 2011-01-04 | Ricoh Company Limited | Image forming apparatus, image forming method and process cartridge |
| JP2008170505A (en) * | 2007-01-09 | 2008-07-24 | Yamaha Corp | Speech processing apparatus and program |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| US6490563B2 (en) | Proofreading with text to speech feedback | |
| JP2815579B2 (en) | Word candidate reduction device in speech recognition | |
| JPH0352058A (en) | Document processor for voice input | |
| JPS61107430A (en) | Editing unit for voice information | |
| US4468756A (en) | Method and apparatus for processing languages | |
| JP2004046807A (en) | Multi-modal input of ideographic language | |
| JPH0535783A (en) | Document translation device with typographical information | |
| JP5611270B2 (en) | Word dividing device and word dividing method | |
| JPS5971536A (en) | Japanese sentence creation device | |
| JPS634206B2 (en) | ||
| JPS607514A (en) | Word processor | |
| JPS6211731B2 (en) | ||
| JPH01290032A (en) | Voice recognition/display device | |
| JP2615643B2 (en) | Word speech recognition device | |
| JPH10143503A (en) | Speech word processor | |
| JP2755792B2 (en) | Voice recognition device | |
| JP2615649B2 (en) | Word speech recognition device | |
| JP2000010579A (en) | Speech synthesizer and computer readable recording medium | |
| JPS59176835A (en) | Method and device for producing sound input sentence | |
| JPH08272780A (en) | Chinese input processing apparatus, Chinese input processing method, language processing apparatus and language processing method | |
| JPS62224859A (en) | Japanese language processing system | |
| JPH06103457B2 (en) | Voice typewriter | |
| JPS61223971A (en) | Sentence generating device | |
| JPH0664571B2 (en) | Character processing method | |
| JPH0128977B2 (en) |