[go: up one dir, main page]

WO2018179426A1 - Système, procédé et programme de correction de reconnaissance vocale - Google Patents

Système, procédé et programme de correction de reconnaissance vocale Download PDF

Info

Publication number
WO2018179426A1
WO2018179426A1 PCT/JP2017/013826 JP2017013826W WO2018179426A1 WO 2018179426 A1 WO2018179426 A1 WO 2018179426A1 JP 2017013826 W JP2017013826 W JP 2017013826W WO 2018179426 A1 WO2018179426 A1 WO 2018179426A1
Authority
WO
WIPO (PCT)
Prior art keywords
information
voice
user
speech
position information
Prior art date
Application number
PCT/JP2017/013826
Other languages
English (en)
Japanese (ja)
Inventor
俊二 菅谷
Original Assignee
株式会社オプティム
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 株式会社オプティム filed Critical 株式会社オプティム
Priority to JP2018516873A priority Critical patent/JP6457154B1/ja
Priority to PCT/JP2017/013826 priority patent/WO2018179426A1/fr
Publication of WO2018179426A1 publication Critical patent/WO2018179426A1/fr

Links

Images

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue

Definitions

  • the present invention relates to a speech recognition correction system, method, and program.
  • the sound collecting device of the system may not be able to collect the voice due to ambient noise or the like. In this case, it is required to provide a system that can correct the content of the voice that the user is trying to recognize and correctly recognize the voice.
  • the present invention has been made in view of such a demand, and even if the sound collecting device of the system cannot collect sound due to ambient noise or the like, the sound that the user is trying to recognize is intended. It is an object of the present invention to provide a system capable of estimating the content of the voice and correctly recognizing the voice.
  • the present invention provides the following solutions.
  • the invention according to the first feature is Position information acquisition means for acquiring position information of a place visited by a user before a specific time; Speech recognition means for recognizing speech uttered by the user; Correction means for correcting the speech-recognized content based on the acquired position information;
  • a speech recognition correction system comprising:
  • the location information acquisition unit acquires location information of a place visited by the user before a specific time
  • the correction unit is based on the location information acquired by the location information acquisition unit. Correct the content of the voice recognition.
  • the sound collecting device of the system cannot pick up the sound due to ambient noise or the like, the user tried to make the user recognize from the location information of the place where the user visited before a specific time. It is possible to provide a system capable of estimating the content of a voice and correctly recognizing the voice.
  • the invention according to the second feature is the invention according to the first feature,
  • the position information acquisition means provides a voice recognition correction system that acquires position information of a place visited by the user before a specific time from the user's portable terminal.
  • the position information of the place visited by the user before a specific time is acquired from the mobile terminal owned by the user himself, and the voice that the user tries to recognize from the position information. Guess the contents. Therefore, it is possible to provide a system that can further improve the accuracy of recognizing the voice.
  • the invention according to the third feature is the invention according to the first or second feature,
  • the correction means provides a voice recognition correction system that corrects the voice-recognized content with reference to Web content related to the acquired position information.
  • the content recognized by the voice is corrected by referring to the Web content related to the position information. Therefore, it is possible to provide a system that can further increase the accuracy of recognizing voice.
  • the invention according to the fourth feature is the invention according to any one of the first to third features,
  • the correction means provides a voice recognition correction system that specifies weather information in the acquired position information and corrects the voice-recognized content.
  • the weather information related to the location information is specified, and the speech-recognized content is corrected. Therefore, it is possible to provide a system that can further improve the accuracy of recognizing voices regarding the weather of a place visited by a user before a specific time.
  • the invention according to the fifth feature is the invention according to any one of the first to fourth features,
  • the correction means provides a voice recognition correction system that specifies time information in the acquired position information and corrects the voice-recognized content.
  • the time information related to the position information is specified, and the speech-recognized content is corrected. For this reason, it is possible to provide a system that can further improve the accuracy of recognizing the voice with respect to the time when the user visited the predetermined place.
  • the invention according to a sixth feature is the invention according to any one of the first to fifth features, Further comprising state information acquisition means for acquiring state information indicating the state of the user from the portable terminal of the user;
  • the correction means provides a voice recognition correction system that specifies state information in the acquired position information and corrects the voice-recognized content.
  • the state information in the position information is specified, and the speech-recognized content is corrected. Therefore, it is possible to provide a system that can further improve the accuracy of recognizing voices regarding the state of the user at the place where the user has visited.
  • the invention according to a seventh feature is the invention according to any one of the first to sixth features, Payment information acquisition means for acquiring payment information settled by the user;
  • the correction means provides a voice recognition correction system that specifies payment information in the acquired position information and corrects the voice-recognized content.
  • the settlement information in the position information is specified, and the speech-recognized content is corrected. Therefore, it is possible to provide a system that can further improve the accuracy of recognizing voices with respect to matters related to the settlement status at the place where the user has visited.
  • the invention according to the eighth feature is the invention according to any one of the first to seventh features, Comprising a plurality of portable terminals and a management computer connected to the plurality of portable terminals via a network;
  • the plurality of portable terminals include the position information acquisition unit and voice information acquisition unit that acquires voice information related to a voice uttered by the user,
  • the management computer is configured to receive the location information and the audio information acquired by the plurality of mobile terminals,
  • the management computer includes a determination unit that determines whether the portable terminal that has transmitted the position information and the portable terminal that has transmitted the audio information are the same portable terminal, and the correction unit.
  • the correction unit provides a voice recognition correction system that corrects the voice-recognized content based on the acquired position information when the determination unit determines that the mobile terminals are the same portable terminal. .
  • the voice recognition correction system is a plurality of portable terminals and a network type system connected to the plurality of portable terminals via a network.
  • the voice recognition correction system is a plurality of portable terminals and a network type system connected to the plurality of portable terminals via a network.
  • the invention according to a ninth feature is the invention according to any one of the first to eighth features, A repeater that repeats the corrected content; There is provided a voice recognition system further comprising recording means for recording the corrected content when there is no problem as a result of the repetition.
  • the ninth aspect of the invention since the corrected content is repeated, even if the user is moving, the content corrected by the correcting means can be confirmed without paying attention to the screen display. it can.
  • the recording means records the corrected content when there is no problem as a result of the repetition. Therefore, according to the ninth aspect of the invention, when there is an error in the corrected content, it is possible to prevent the corrected content from being recorded, and as a result, the accuracy of recognizing the voice is further improved. A system that can be enhanced can be provided.
  • the user can recognize the position information from a location visited by the user before a specific time. It is possible to provide a system capable of estimating the content of such a voice and correctly recognizing the voice.
  • FIG. 1 is a block diagram showing a hardware configuration and software functions of a speech recognition correction system 1 according to the first embodiment of the present invention.
  • FIG. 2 is a flowchart showing the speech recognition correction method according to this embodiment.
  • FIG. 3 is an example of the position information database 31 in the present embodiment.
  • FIG. 4 is an example of the stay time measurement area 32 in the present embodiment.
  • FIG. 5 is an example for explaining the collected sound contents.
  • FIG. 6 is an example of the voice database 34 in the present embodiment.
  • FIG. 7 is an example of the dictionary database 35 in the present embodiment.
  • FIG. 8 is an example of the classification database 36 in the present embodiment.
  • FIG. 9 is an example of display content and audio output content in the speech recognition correction system 1 according to the present embodiment.
  • FIG. 10 is an example of the voice database 34 after being overwritten and saved in the present embodiment.
  • FIG. 11 is a block diagram showing a hardware configuration and software functions of the speech recognition correction system 1 according to the first embodiment of the present invention.
  • the voice recognition correction system may be a stand-alone type system that is provided integrally with a mobile terminal such as a smartphone, smart glass, or smart watch, or is connected to the mobile terminal via the network.
  • a network type system including a management computer may be used.
  • the speech recognition correction system is a stand-alone system.
  • the speech recognition correction system will be described as a network type system.
  • FIG. 1 is a block diagram for explaining the hardware configuration and software functions of a speech recognition correction system 1 according to this embodiment.
  • the speech recognition correction system 1 includes a control unit 10 that controls data, a communication unit 20 that communicates with other devices, a storage unit 30 that stores data, an input unit 40 that receives user operations, and user voices.
  • a sound collection unit 50 that collects sound, a position detection unit 60 that detects a position where the speech recognition correction system 1 exists, a timer 70 that measures a staying time at a certain place, and data controlled by the control unit 10
  • an image display unit 80 for outputting and displaying an image.
  • the control unit 10 includes a CPU (Central Processing Unit), a RAM (Random Access Memory), a ROM (Read Only Memory), and the like.
  • a CPU Central Processing Unit
  • RAM Random Access Memory
  • ROM Read Only Memory
  • the communication unit 20 includes a device for enabling communication with other devices, for example, a Wi-Fi (Wireless Fidelity) compatible device compliant with IEEE 802.11.
  • Wi-Fi Wireless Fidelity
  • the control unit 10 reads a predetermined program and cooperates with the communication unit 20 as necessary, so that the position information position information acquisition module 11, the state information etc. acquisition module 12, the voice recognition module 13, and the correction module 14, a repeat module 15, and a recording module 16 are realized.
  • the storage unit 30 is a device that stores data and files, and includes a data storage unit such as a hard disk, a semiconductor memory, a recording medium, and a memory card.
  • the storage unit 30 stores a history information database 31, a map database 32, a stay time measurement area 33, a voice database 34, a dictionary database 35, and a classification database 36, which will be described later.
  • the storage unit 30 also stores image data to be displayed on the image display unit 80.
  • the type of the input unit 40 is not particularly limited. Examples of the input unit 40 include a keyboard, a mouse, and a touch panel.
  • the type of the sound collecting unit 50 is not particularly limited. Examples of the sound collecting unit 50 include a microphone.
  • the position detection unit 60 is not particularly limited as long as it is a device that can detect the latitude and longitude where the voice recognition correction system 1 is located.
  • Examples of the position detection unit 60 include a GPS (Global Positioning System).
  • the type of the timer 70 is not particularly limited as long as the staying time at a certain place can be measured.
  • the type of the image display unit 80 is not particularly limited. Examples of the image display unit 80 include a monitor and a touch panel.
  • FIG. 2 is a flowchart showing a voice recognition correction method using the voice recognition correction system 1. The processing executed by each hardware and the software module described above will be described.
  • Step S10 Acquisition of Position Information
  • the control unit 10 of the voice recognition correction system 1 executes the position information acquisition module 11 and acquires position information of a place visited by the user before a specific time (step S10).
  • the position detection unit 60 of the voice recognition correction system 1 detects the latitude and longitude where the voice recognition correction system 1 is located at any time. Then, the control unit 10 refers to the map database 32 and searches for a place corresponding to the latitude and longitude detected by the position detection unit 60. Then, the control unit 10 records the searched place in the history information database 31.
  • FIG. 3 shows an example of the history information database 31.
  • the history information database 31 information on the date and time when the position detection unit 60 detects the position information and the location corresponding to the position detected by the position detection unit 60 is recorded in association with the identification number.
  • the date can be recorded by referring to a calendar function (not shown) built in the audio content correction system 1.
  • the time can be recorded by referring to a clock function (not shown) built in the audio content correction system 1.
  • the control unit 10 can acquire position information of a place visited by the user before a specific time by referring to the history information database 31.
  • Step S11 Acquisition of status information and the like
  • the control unit 10 executes the status information acquisition module 12, and acquires status information indicating the user status, current weather information, payment information regarding a credit card and electronic payment, and the like (step S11).
  • the timer 70 of the voice recognition correction system 1 measures the time during which the voice recognition correction system 1 stays at a certain place and records it in the stay time measurement area 32.
  • FIG. 4 is an example of the stay time measurement area 32.
  • the stay time measurement area 32 information on the stay location, stay start date and time and stay end date and time of the speech recognition correction system 1 is recorded.
  • the control unit 10 determines that the user is staying at a certain place and records the history information database.
  • the item of “state” in 31 is updated to “staying”.
  • control unit 10 accesses an external weather forecast providing website via the communication unit 20. And the control part 10 reads the information of the weather in the spot corresponded to the latitude and the longitude which the position detection part 60 detected from the said weather forecast provision Web site. Then, the control unit 10 records the read weather information in the history information database 31.
  • control unit 10 records the payment information regarding the credit card or the electronic payment in the history information database 31.
  • the history information database 31 shown in FIG. 3 includes not only date, time and place information when the position detection unit 60 detects position information, but also state information indicating the user state, current weather information, credit card Also, payment information relating to electronic payment is recorded in association with an identification number.
  • the control unit 10 can acquire the state information, weather information, settlement information, and the like by referring to the history information database 31.
  • Step S12 Sound collection
  • the control unit 10 When the sound collection unit 50 collects the user's voice, the control unit 10 performs A / D conversion on the voice collected by the sound collection unit 50 and sets the A / D converted information in a predetermined area of the storage unit 30. To do.
  • the sound collection unit 50 of the speech recognition correction system 1 collects the sound. Then, the control unit 10 A / D converts the sound collected by the sound collection unit 50 and sets the A / D converted information in a predetermined area of the storage unit 30.
  • step S12 determines whether the determination in step S12 is YES or not. If the determination in step S12 is YES, the process proceeds to step S13. On the other hand, when the determination in step S12 is NO, the process returns to step S10.
  • Step S13 Speech recognition
  • the control unit 10 refers to the voice database 34 shown in FIG. 6 and transcribes the voice collected by the sound collection unit 50 from the sound wave waveform included in the A / D converted information.
  • the information that has been A / D converted is “Kyoha ??? Nidekaketa / Harete Yokatta / ??? Nyotte Brand ??????
  • “???” is a place where the sound collection unit 50 of the sound content correction system 1 cannot collect sound due to ambient noise or the like.
  • control unit 10 refers to the dictionary database 35 shown in FIG. 7, replaces the transcribed information with a language, and creates a sentence.
  • the information that has been A / D converted is “I went out to today. It was good to be sunny.
  • the documented information is set in a predetermined area of the storage unit 30 in association with the A / D converted information.
  • Step S14 Correction of Recognized Content
  • the control unit 10 executes the correction module 14 and corrects the content recognized in the process of step S13 based on the position information acquired in the process of step S10, the state information acquired in the process of step S11, and the like. (Step S14).
  • the control unit 10 refers to the classification database 36.
  • FIG. 8 is an example of the classification database 36.
  • the classification database 36 records in advance the relationship between words and the like included in the documented contents and items listed in the history information database 31.
  • items such as “date”, “time”, “location”, “state”, “weather”, “payment information” are listed in the history information database 31 (FIG. 3).
  • word groups related to these items are recorded.
  • the control unit 10 refers to the classification database 36, associates “today” included in this information with the item “date”, and associates “going out” with the item “location”.
  • “good” is associated with the item “weather”, and “stop” is associated with the item “location”.
  • “clothes” is associated with the item “settlement information”, and “purchase” is associated with the item “settlement information”.
  • control unit 10 refers to the history information database 31.
  • the control unit 10 refers to the item “date” in the history information database 31 and extracts an item related to “today” included in the speech-recognized content.
  • it can be grasped
  • control unit 10 refers to the item “place” in the history information database 31 and extracts items relating to “going out” and “stopping” included in the speech-recognized content.
  • control unit 10 determines the “going out” location, “stopping” from the content recorded in the history information database 31. It can be inferred that the location is “Yurakucho”, “Yurakucho Station”, “A Department Store”, “Department Store”, “Ginza”, or “A Department Store Ginza Store”.
  • control unit 10 refers to the voice database 34 (FIG. 6), and the voice corresponding to “Yurakucho”, “Yurakucho Station”, “A Department Store”, “Department Store”, “Ginza”, “A Department Store Ginza Store”. Synthesize data (waveform data). Subsequently, the control unit 10 compares the synthesized voice data with the voice data that has been A / D converted in the process of step S ⁇ b> 13, and went to “???”. The voice closest to the voice data corresponding to “???” is extracted.
  • control unit 10 refers to the item “payment information” in the history information database 31 and extracts items related to “clothes” and “purchase” included in the speech-recognized content.
  • control unit 10 purchases “clothes” from the content recorded in the history information database 31. It can be inferred that it is one of “brand X”, “shirt”, “7560 yen”, “credit card”, or “card payment”.
  • control unit 10 refers to the voice database 34 (FIG. 6), and the voice data (waveform data) corresponding to “brand X”, “shirt”, “7560 yen”, “credit card”, “card payment”. Is synthesized. Subsequently, the control unit 10 compares the synthesized voice data with the voice data that has been A / D converted in the process of step S ⁇ b> 13 and sets “???” to “???”. The voice closest to the corresponding voice data is extracted.
  • control unit 10 can presume that “???” in “?????? purchases” is “brand X”.
  • the control unit 10 acquires position information of a place visited by the user before a specific time in the process of step S10, and the process of step S10 in the process of step S14. Based on the acquired position information, the content of voice recognition is corrected. As a result, even if the sound collecting device of the system cannot pick up the sound due to ambient noise or the like, the user tried to make the user recognize from the location information of the place where the user visited before a specific time. It is possible to provide the audio content correction system 1 that can estimate the audio content and correctly recognize the audio.
  • the control unit 10 identifies the weather information, the time information, the state information indicating the user state, and the payment information settled by the user in the position information acquired in the process of step S11. It is possible to correct the content recognized by the voice processing. According to the invention described in the present embodiment, in addition to the position information of a place visited by a user before a specific time point, it is possible to specify various information related to the position information and correct the content recognized by voice. To do. Therefore, it is possible to provide the audio content correction system 1 that can further increase the accuracy of recognizing the audio.
  • control unit 10 refers to the Web content related to the position information acquired in the process of step S10 and corrects the content recognized in the process of step S13. By doing so, it is possible to provide the audio content correction system 1 that can further increase the accuracy of recognizing the audio.
  • Step S15 Recurrence of correction contents
  • FIG. 9 shows an example of the state of the audio content correction system 1 at that time.
  • the corrected content is repeated not only as a screen display on the image display unit 80 but also as a sound from a speaker.
  • the content corrected in the process of step S14 can be confirmed without paying attention to the above.
  • Step S16 Recording Correction Contents
  • the contents of the part whose contents are unknown only by the process of step S13 are “Ginza”, “A department store”, and “Brand X”. There was found.
  • the control unit 10 changes the voice data A / D converted in the process of step S13 to “Ginza”, “A”, “Department Store”, “A Department Store”, “Brand”, “X”, “Brand X”.
  • the waveform at the corresponding location is extracted and overwritten and saved in the voice database 34 that was originally stored.
  • FIG. 10 shows an example of the voice database 34 after being overwritten. Audio data of “Ginza”, “A”, “Department Store”, “A Department Store”, “Brand”, “X”, and “Brand X” is newly added to the audio database 34.
  • the corrected content is recorded when there is no problem as a result of the repetition in step S15. Therefore, when there is an error in the content corrected in the process of step S14, it is possible to prevent the incorrect content from being recorded, and as a result, the speech recognition that can further improve the accuracy of recognizing the speech.
  • a correction system 1 can be provided.
  • the voice recognition correction system is described as a stand-alone type system.
  • the second embodiment is different in that the voice recognition correction system is a network type system, and the rest is the same.
  • FIG. 11 is a block diagram for explaining the hardware configuration and software functions of the speech recognition correction system 100 according to this embodiment.
  • the voice recognition correction system 100 includes a plurality of portable terminals 200 and a management computer 300 connected to the plurality of portable terminals 200 via a network.
  • the mobile terminal 200 includes a control unit 210, a communication unit 220, a storage unit 230, an input unit 240, a sound collection unit 250, a position detection unit 260, and an image display unit 280, respectively.
  • the control unit 210 includes a position information acquisition module 211, a state information acquisition module 212, and a repetition module 215.
  • the sound collection unit 250 functions as a voice information acquisition unit that acquires voice information related to the voice uttered by the user.
  • the functions of the communication unit 220, the storage unit 230, the input unit 240, the position detection unit 260, and the image display unit 280 are the same as those of the communication unit 20, the storage unit 30, the input unit 40, and the position detection unit 60 in the first embodiment. , And the function of the image display unit 80.
  • the functions of the position information acquisition module 211, the state information acquisition module 212, and the repetition module 215 are the same as the functions of the position information acquisition module 11, the state information acquisition module 12, and the repetition module 15 in the first embodiment. It is.
  • the management computer 300 includes a control unit 310, a communication unit 320, a storage unit 330, an input unit 340, and an image display unit 380.
  • the control unit 310 includes a voice recognition module 313, a correction module 314, and a recording module 316.
  • the communication unit 320 is configured to be able to receive position information and audio information acquired by the plurality of mobile terminals 200.
  • the storage unit 330 stores a history information database 331, a map database 332, a stay time measurement area 333, a voice database 334, a dictionary database 335, and a classification database 336.
  • the control part 310 discriminate
  • the correction module 314 of the control unit 310 when the portable terminal that transmitted the position information and the portable terminal that transmitted the audio information are the same portable terminal, based on the position information acquired by the portable terminal, The speech-recognized content of the sound collected by the sound collection unit 250 of the portable terminal is corrected.
  • the speech recognition correction system 1 is a network type system including a plurality of portable terminals 200 and a management computer 300 connected to the plurality of portable terminals 200 via a network. it can. Therefore, it is possible to provide the network type speech recognition correction system 1 that can further increase the accuracy of speech recognition.
  • the functions of the input unit 340 and the image display unit 380 are the same as the functions of the input unit 40 and the image display unit 80 in the first embodiment.
  • the functions of the speech recognition module 313, the correction module 314, and the recording module 316 are basically the same as the functions of the speech recognition module 13, the correction module 14, and the recording module 16 in the first embodiment.
  • the history information database 331, the map database 332, the stay time measurement area 333, the voice database 334, the dictionary database 335, and the classification database 336 are configured in the history information database 31, the map database 32, the stay time in the first embodiment.
  • the configuration is the same as that of the measurement area 33, the voice database 34, the dictionary database 35, and the classification database 36.
  • the means and functions described above are realized by a computer (including a CPU, an information processing apparatus, and various terminals) reading and executing a predetermined program.
  • the program is provided in a form recorded on a computer-readable recording medium such as a flexible disk, CD (CD-ROM, etc.), DVD (DVD-ROM, DVD-RAM, etc.).
  • the computer reads the program from the recording medium, transfers it to the internal storage device or the external storage device, stores it, and executes it.
  • the program may be recorded in advance in a storage device (recording medium) such as a magnetic disk, an optical disk, or a magneto-optical disk, and provided from the storage device to a computer via a communication line.

Landscapes

  • Engineering & Computer Science (AREA)
  • Computational Linguistics (AREA)
  • Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • Acoustics & Sound (AREA)
  • Multimedia (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)

Abstract

Le problème décrit par la présente invention est de fournir un système susceptible de reconnaître correctement la parole qu'un utilisateur souhaite reconnaître, en déduisant les détails de la parole même lorsqu'un dispositif de capture de son ne parvient pas à collecter complètement la parole en raison du bruit et autres de la zone environnante. À cet effet, dans un système de correction de reconnaissance vocale (1) selon la présente invention, une unité de commande (10) exécute un module d'acquisition d'informations de position (11) pour acquérir des informations de position concernant un emplacement qu'un utilisateur a visité avant un point temporel spécifique. En outre, l'unité de commande (10) exécute un module de reconnaissance vocale (13) pour effectuer une reconnaissance vocale de la parole prononcée par l'utilisateur. L'unité de commande (10) exécute un module de correction (14) pour corriger, sur la base des informations de position acquises par le module d'acquisition d'informations de position (11), les détails de la reconnaissance vocale effectuée par l'exécution du module de reconnaissance vocale (13).
PCT/JP2017/013826 2017-03-31 2017-03-31 Système, procédé et programme de correction de reconnaissance vocale WO2018179426A1 (fr)

Priority Applications (2)

Application Number Priority Date Filing Date Title
JP2018516873A JP6457154B1 (ja) 2017-03-31 2017-03-31 音声認識補正システム、方法及びプログラム
PCT/JP2017/013826 WO2018179426A1 (fr) 2017-03-31 2017-03-31 Système, procédé et programme de correction de reconnaissance vocale

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/JP2017/013826 WO2018179426A1 (fr) 2017-03-31 2017-03-31 Système, procédé et programme de correction de reconnaissance vocale

Publications (1)

Publication Number Publication Date
WO2018179426A1 true WO2018179426A1 (fr) 2018-10-04

Family

ID=63674781

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2017/013826 WO2018179426A1 (fr) 2017-03-31 2017-03-31 Système, procédé et programme de correction de reconnaissance vocale

Country Status (2)

Country Link
JP (1) JP6457154B1 (fr)
WO (1) WO2018179426A1 (fr)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110534112A (zh) * 2019-08-23 2019-12-03 王晓佳 基于位置与时间的分布式语音识别纠错装置及方法

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2004163265A (ja) * 2002-11-13 2004-06-10 Nissan Motor Co Ltd ナビゲーション装置
JP2006349427A (ja) * 2005-06-14 2006-12-28 Toyota Motor Corp 車載音声認識装置
JP2012093508A (ja) * 2010-10-26 2012-05-17 Nec Corp 音声認識支援システム、音声認識支援装置、利用者端末、方法およびプログラム

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP3948441B2 (ja) * 2003-07-09 2007-07-25 松下電器産業株式会社 音声認識方法及び、車載装置

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2004163265A (ja) * 2002-11-13 2004-06-10 Nissan Motor Co Ltd ナビゲーション装置
JP2006349427A (ja) * 2005-06-14 2006-12-28 Toyota Motor Corp 車載音声認識装置
JP2012093508A (ja) * 2010-10-26 2012-05-17 Nec Corp 音声認識支援システム、音声認識支援装置、利用者端末、方法およびプログラム

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
KUMIKO OMORI ET AL.: "A Spoken Dialogue Interface through Natural and Efficient Responses", JOURNAL OF NATURAL LANGUAGE PROCESSING, vol. 10, no. 5, 10 October 2003 (2003-10-10), pages 23 - 40, ISSN: 1340-7619 *

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110534112A (zh) * 2019-08-23 2019-12-03 王晓佳 基于位置与时间的分布式语音识别纠错装置及方法

Also Published As

Publication number Publication date
JP6457154B1 (ja) 2019-01-23
JPWO2018179426A1 (ja) 2019-04-04

Similar Documents

Publication Publication Date Title
CN112214418B (zh) 一种应用程序的合规检测方法、装置和电子设备
JP6107409B2 (ja) 位置特定処理装置及び位置特定処理プログラム
US9188456B2 (en) System and method of fixing mistakes by going back in an electronic device
US8918320B2 (en) Methods, apparatuses and computer program products for joint use of speech and text-based features for sentiment detection
US8521681B2 (en) Apparatus and method for recognizing a context of an object
US10127907B2 (en) Control device and message output control system
WO2011093025A1 (fr) Système, procédé et programme d'aide à l'entrée de données
US20120224707A1 (en) Method and apparatus for identifying mobile devices in similar sound environment
US20140324428A1 (en) System and method of improving speech recognition using context
US20130065611A1 (en) Method and apparatus for providing information based on a location
US10515634B2 (en) Method and apparatus for searching for geographic information using interactive voice recognition
CN110998719A (zh) 信息处理设备和信息处理方法
US11495245B2 (en) Urgency level estimation apparatus, urgency level estimation method, and program
WO2019205398A1 (fr) Procédé et dispositif d'incitation de comportement d'utilisateur, appareil informatique et support de stockage
CN103828400A (zh) 信息处理装置、信息提供方法和程序
JP5929393B2 (ja) 位置推定方法、装置及びプログラム
US9224388B2 (en) Sound recognition method and system
CN112951274A (zh) 语音相似度确定方法及设备、程序产品
JP6457154B1 (ja) 音声認識補正システム、方法及びプログラム
JP7314975B2 (ja) 音声操作装置及びその制御方法
KR20150037104A (ko) 클라우드 소싱 기반 관심 지점 정보 업데이트 방법, 이를 위한 시스템 및 장치
CN112863496B (zh) 一种语音端点检测方法以及装置
CN110263135B (zh) 一种数据交换匹配方法、装置、介质和电子设备
JP4408665B2 (ja) 音声認識用発話データ収集装置、音声認識用発話データ収集方法、及びコンピュータプログラム
CN113453135A (zh) 智能音箱优化方法及测试方法、装置、设备和存储介质

Legal Events

Date Code Title Description
ENP Entry into the national phase

Ref document number: 2018516873

Country of ref document: JP

Kind code of ref document: A

121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 17902628

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 17902628

Country of ref document: EP

Kind code of ref document: A1