WO2001039178A1 - Referencing web pages by categories for voice navigation - Google Patents
Referencing web pages by categories for voice navigation Download PDFInfo
- Publication number
- WO2001039178A1 WO2001039178A1 PCT/EP2000/011299 EP0011299W WO0139178A1 WO 2001039178 A1 WO2001039178 A1 WO 2001039178A1 EP 0011299 W EP0011299 W EP 0011299W WO 0139178 A1 WO0139178 A1 WO 0139178A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- search
- internet
- search criterion
- assigned
- phoneme sequence
- Prior art date
Links
- 238000000034 method Methods 0.000 claims abstract description 54
- 238000004590 computer program Methods 0.000 claims description 21
- 238000010187 selection method Methods 0.000 claims description 7
- 230000003213 activating effect Effects 0.000 claims description 4
- 230000000694 effects Effects 0.000 claims 1
- 235000013550 pizza Nutrition 0.000 description 15
- 230000005236 sound signal Effects 0.000 description 5
- 230000001419 dependent effect Effects 0.000 description 2
- 238000010586 diagram Methods 0.000 description 2
- 238000009434 installation Methods 0.000 description 1
- 230000000699 topical effect Effects 0.000 description 1
Classifications
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/22—Procedures used during a speech recognition process, e.g. man-machine dialogue
Definitions
- the invention relates to a speech recognition method of selecting Internet addresses of Internet pages assigned to a search criterion, to a computer program product in accordance with the introductory part of claim 8 and to a selection method of selecting Internet addresses of Internet pages assigned to a search criterion.
- Such a speech recognition method such a selection method and such a computer program product is known from the speech recognition software "Free Speech Browser", which has been marketed by Philips since mid-November 1999.
- the known speech recognition software is loaded into an internal memory of the computer and is processed by the computer, the known speech recognition method is executed.
- this computer further processes the software of the so-called Internet browser (for example, "Microsoft Explorer” by Microsoft) and the computer is connected to the Internet, a user of the computer can select by commands spoken into a microphone Internet pages which are then displayed on a monitor of the computer.
- the user of the computer speaks an Internet address into the microphone, after which the text information of the Internet address recognized via the speech recognition method is delivered to the Internet browser. Then the Internet browser scans the Internet page of the respective computer server connected to the Internet featured by this Internet address and displays this Internet page on the monitor.
- a user liked to search for Internet pages containing information about a certain search criterion - such as, for example, books of motorcars - the user is to proceed in accordance with the known speech recognition method of selecting Internet pages as follows.
- the user speaks the Internet address of a so-called search engine - such as, for example, YAHOO or ALTAVISTA - into the microphone and waits for the search engine to be displayed on the monitor.
- the user speaks the search criterion into the microphone, which is recognized in accordance with the speech recognition method and is inserted into an entry field of the Internet page of the search engine.
- the user receives a survey Internet page with hyperlinks to Internet pages on the monitor, which pages contain information about the entered search criterion.
- a user can, for example, speak the command "search" and then a search criterion into the microphone and a search phoneme sequence will be determined for this search criterion in accordance with the speech recognition method.
- the search phoneme sequence is compared to a number of stored search criterion phoneme sequences of search criterions for which at least one Internet address of an Internet page is assignedly stored which page contains the information corresponding to the search criterion. If one of the stored search criterion phoneme sequences sufficiently corresponds to the determined search criterion phoneme sequence, the Internet page of the Internet address assigned to this search criterion phoneme sequence is displayed on the monitor.
- the speech recognition method as claimed in claim 1 it has proved to be advantageous to provide the measures as claimed in claim 3.
- the database stored in a computer server connected to the Internet and continuously updated containing Internet addresses and search criterions is retrieved after long or short time distances by the computer which executes the speech recognition method and stored by the computer. This advantageously achieves, on the one hand, that the result of a search is displayed on the monitor very fast and, on the other hand, always topical data are processed.
- a user himself can then enter a search criterion as search criterion text information and one or more Internet addresses with information about the search criterion via a keyboard of the computer. Subsequently, a search criterion phoneme sequence corresponding to the entered search criterion text information is detected and stored assigned to the entered Internet address(es) in the database.
- the thus updated database can be stored with the computer, but also be conveyed to the computer server over the Internet to update the database stored in the computer server.
- a search phoneme sequence associated to a search criterion is delivered to the Internet server, which stores the database with Internet addresses and search criterion phoneme sequences.
- the Internet server compares the received search phoneme sequence with stored search criterion phoneme sequences and the result of the search is then delivered to the user's computer via the Internet and displayed on the monitor.
- Fig. 1 diagrammatically shows in the form of a block diagram a computer connected to the Internet by which a speech recognition method is executed for selecting Internet addresses from Internet pages assigned to a search criterion
- Fig. 2 shows text information and picture information of an Internet page displayed on a monitor
- Fig. 3 shows a command table stored in a command determining stage of the computer shown in Fig. 1,
- Fig. 4 shows a first search table stored in search storage means of the computer shown in Fig. 1
- Fig. 5 shows a second search table stored in the search storage means of the computer shown in Fig. 1
- Fig. 6 shows an Internet index page displayed on the monitor
- Fig. 7 shows a second search table shown in Fig. 5 with a further search criterion and an associated Internet address entered in the second search table.
- Fig. 1 diagrammatically shows in the form of a block diagram a computer 1, a first computer server 2 and a second computer server 3, which are all connected to the Internet NET.
- a monitor 5 To the monitor port 4 of the computer 1 is connected a monitor 5 by which picture information BI and text information TI of an Internet page can be displayed.
- a monitor signal MS containing picture information BI and text information TI can be delivered through the monitor port 4 of the computer 1 to the monitor 5.
- the first computer program product contains software code sections and may be formed, for example, by the known computer software "Microsoft Explorer” of Microsoft or, for example, by the known computer software "Netscape Navigator” of Netscape.
- An Internet address URL can be applied to the Internet browser 6 via a keyboard 7 and the Internet browser 6 is then arranged for searching for the computer server 2 or 3 connected to the Internet NET and featured by the Internet address URL.
- the Internet browser 6 scans and receives the Internet page featured by the Internet address URL and stored with a computer server 2 or 3, which page usually contains text information TI and picture information BI and is HTML coded.
- the Internet browser 6, after receiving the text information TI and picture information BI of an Internet page, delivers this monitor signal MS containing this information to the monitor port 4.
- Fig. 2 is shown an Internet page 8, which can be displayed or reproduced respectively by the monitor 5.
- the Internet page 8 contains text information Til and TI2 as well as picture information BI 1 and B 12.
- the Internet page 8 contains further text information TI3, TI4, TI5, TI6 and TI7, which are shown underlined and form hypertexts HT(HL) of so-called hyperlinks HL.
- Each hyperlink contains both a hypertext HT(HL) and an Internet address URL(HL) of the hyperlink HL assigned to the hypertext HT(HL) which, however, is not displayed by the monitor 5.
- the Internet browser 6 loads text information TI and picture information BI of the Internet page featured by the Internet address URL(HL) of the activated hyperlink HL, as this was described above.
- a second computer program product which, when it runs on the computer 1, forms a speech recognition device 9, as the result of which the computer 1 executes a speech recognition method.
- the speech recognition device 9 is arranged for controlling the Internet browser 6.
- the speech recognition device 9 is arranged for delivering the Internet address URL of an Internet page selected by a user of the computer 1 via a spoken command.
- the speech recognition device 9 has hyperlink identification means 10.
- the hyperlink identification means 10 can be applied text information TI of the Internet page 8 displayed by the monitor 5 and detected by the Internet browser 6.
- the hyperlink identification means 10 are arranged for detecting the text information TI3 to TI7 of hypertexts HT(HL) of the hyperlink HL from the text information TI of the received Internet page 8.
- This text information TI3 to TI7 can be delivered as hypertexts HT(HL) of the hyperlink HL of the Internet page 8 by the hyperlink identification means 10.
- the hyperlink identification means 10 are further arranged for detecting the
- the speech recognition device 9 further includes correlation means 11 for determining first phoneme sequences PI1[HT(HL)] corresponding to these hypertexts HT(HL).
- the correlation means 11 include a correlation stage 12 and a word memory 13. In the word memory 13 are stored 64,000 words in English as a so-called background lexicon. Stored in the word memory 13 and assigned to each of these words is a phoneme sequence PI, which corresponds to the acoustic pronunciation of this word.
- the correlation stage 12 is arranged for determining a first phoneme sequence PI1[HT(HL)] for each hypertext HT(HL) of a hyperlink HL delivered to the correlation stage 12 by the hyperlink identification means 10.
- the correlation stage 12 is then arranged for comparing text portions of the hypertext HT(HL) of a hyperlink HL with words stored in the word memory 13.
- the phoneme sequence PI assigned to this word and stored in the word memory 13 is incorporated in the first phoneme sequence PI1[HT(HL)] of this hypertext HT(HL).
- the speech recognition device 9 further includes a command determining stage 14 by which a command table 15 shown in Fig. 3 is stored.
- a command table 15 shown in Fig. 3 is stored.
- the first phoneme sequences PI1[HT(HL)] determined by the correlation stage 12 and delivered to the command determining stage 14.
- the command table 15 of the command determining stage 14 are stored first phoneme sequences PI1[HT(HL)] and Internet addresses URL(HL) for each hyperlink HL of the Internet page 8 displayed by the monitor 5.
- the hypertexts HT(HL) of the phoneme sequences PI1[HT(HL)] stored in the command determining stage 14 form the spoken commands that can be recognized by the speech recognition device 9 when the Internet page 8 is displayed by the monitor 5.
- the speech recognition device 9 also recognizes search commands and subsequent search criterions in spoken commands, which will be further discussed hereinbelow.
- the computer 1 has an audio port 16 to which a microphone 17 can be connected to the computer 1.
- a user of the computer 1 can speak a command into the microphone 17, after which an audio signal AS corresponding to the command is delivered to the audio port 16 by the microphone 17.
- the user can speak a part of or also the whole text information TI3, TI4, TI5, TI6 or TI7 of a hypertext HT(HL) of a hyperlink HL into the microphone 17 as a command.
- the user of the computer 1 can also speak a command into the microphone 17, which command represents one of the search commands that can be recognized by the speech recognition device 9 and represents a search criterion to find Internet pages whose contents correspond to the search criterion. This will be further explained below with reference to the following examples of application.
- the speech recognition device further includes receiving means 18 for receiving an audio signal AS of a user-uttered command applied to the audio port 16.
- the receiving means 18 include an input amplifier for amplifying the audio signal AS and an analog-to-digital converter for digitizing the analog audio signal AS.
- the receiving means 18 can produce digital audio data AD representing the command uttered by the user.
- the speech recognition device 9 further includes speech recognition means 19 for detecting a phoneme sequence P corresponding to the spoken command and for detecting the hyperlink HL selected by the user by comparing the determined phoneme sequence P with the phoneme sequences P[HT(HL)] stored in the command word determining stage 14.
- the speech recognition means 19 include a speech recognition stage 20 and the command word determining stage 14.
- the speech recognition stage 20 can be supplied with digital audio data AD which can be delivered by the receiving means 18.
- the speech recognition stage 20 is arranged for detecting the phoneme sequence P corresponding to the digital audio data AD of the command spoken by the user, as this has already been known for a long time with speech recognition devices.
- a phoneme sequence P detected by the speech recognition stage 20 can be delivered by this stage to the command word determining stage 14.
- the speech recognition stage 20 is further arranged for comparing the detected phoneme sequence P with phoneme sequences of recognizable search commands stored in the speech recognition stage 20.
- a command spoken by the user represents a search command known to the search recognition stage 20 or contains same
- the phoneme sequence detected by the speech recognition stage 20 for the search criterion following the search command in the spoken command can be delivered as a search phoneme sequence SP to the command determining stage 14.
- the command determining stage 14 After receiving a phoneme sequence P from the speech recognition stage 20, the command determining stage 14 is arranged for comparing the received phoneme sequence P with phoneme sequences P[HT(HL)] stored in the command table 15. The command determining stage 14 is further arranged for delivering the hyperlink HL Internet address URL(HL) stored in the command table 15, of which the hyperlink phoneme sequence P[HT(HL)] of the hypertext HT(HL) corresponds best to the phoneme sequence P delivered to the command determining stage 14.
- the speech recognition device 9 further includes control means 21 for controlling the Internet browser 6 to enable reception of text information TI and picture information BI of the Internet page featured by the hyperlink HL selected by the user.
- the hyperlink HL Internet address URL(HL) determined by the command determining stage 14 can be applied to the control means 21.
- the control means 21 form an interface to the Internet browser 6 and deliver the Internet address URL(HL) of the selected hyperlink HL applied to the control means 21 to the Internet browser 6 in a data format so that the first computer program product can immediately process the Internet address URL.
- the speech recognition device 9 henceforth has search storage means 22 which store a first search table 23 shown in Fig. 4 and a second search table 24 shown in Fig. 5.
- search storage means 22 which store a first search table 23 shown in Fig. 4 and a second search table 24 shown in Fig. 5.
- search criterion phoneme sequences KP(KT) of search criterions KT In a first column of the search tables 23 and 24 are stored search criterion phoneme sequences KP(KT) of search criterions KT.
- a second column of the search tables 23 and 24 are stored one or more Internet addresses URL[KT(HL)] of Internet pages stored and assigned to the search criterion phoneme sequences KP(KT) contained in the first column, the contents of these Internet pages containing information that was assigned to the search criterion KT.
- URL[BOOKS(l)] http://www.amazone.com features an
- a third column of the search tables 23 and 24 contains hypertexts
- the Internet page 8 is displayed by the monitor 5 and that the user of the computer 1 would like to receive information about books from the Internet NET.
- the user speaks the command "SEARCH BOOKS" into the microphone 17, after which a respective audio signal AS is delivered to the receiving means 18 and corresponding audio data AD are delivered to the speech recognition stage 20 by the receiving means 18.
- the speech recognition stage 20 detects phoneme sequences P corresponding to the words "SEARCH” and "BOOKS" and compares the phoneme sequence P of the first word "SEARCH” to phoneme sequences of search commands that can be recognized and are stored in the speech recognition stage 20.
- the speech recognition stage 20 recognizes the first word of the spoken command as a search command and delivers the phoneme sequence P of the second word "BOOKS" as a search phoneme sequence SP to the command determining stage 14.
- the command determining stage 14 is then arranged for comparing the search phoneme sequences SP delivered thereto with search criterion phoneme sequences KP(KT) contained in the first column of the first search table 23.
- This Internet address URL[BOOKS(l)] http://www.amazon.com is then delivered to the control means 21 by the command determining stage 14.
- a further advantage of the speech recognition method is provided in that the provider of the first search table 23 forming a database may ask for a registration fee from persons or businesses whose Internet address URL[KT(HL)] is to be registered in the first search table 23 with a certain search criterion KT, as a result of which an economically interesting method of doing business on the Internet is obtained.
- the provider of the first search table 23 uses a third computer server 25 which is connected to the Internet NET and by which the first search table 23 respectively updated by the provider is stored.
- the control means 21 deliver an Internet address URL featuring the third computer server 25 to the Internet browser 6, after which the updated search table 23 is received by the Internet browser 6 and stored in the search storage means 22.
- This updating of the first search table 23 may advantageously be effected automatically and without the user of the computer 1 being involved. This offers the advantage that the first search table 23 is continuously updated in all the computers running the second computer program product.
- the provider of the first search table 23 may also demand different registration fees for entering the Internet address URL[KT(HL)] from a person or business for different time ranges in the first search table 23.
- the Internet page 8 is displayed by the monitor 5 and the user of the computer 1 would like to have information about baby clothing from the Internet NET.
- the user speaks the command "SEARCH BABY CLOTHES" into the microphone 17, after which the speech recognition stage 20 - as described above - applies a phoneme sequence P corresponding to the search criterion "BABY CLOTHES" as a search phoneme sequence SP to the command determining stage 14.
- the command determining stage 14 determines that the search criterion phoneme sequence KP(BABY CLOTHES) entered on the third line of the first search table 23 has a sufficiently large correspondence to the search phoneme sequence SP. Subsequently, the command determining stage 14 reads the four Internet addresses URL[B ABY CLOTHES(l)] to URL[BABY CLOTHES(4)] and the four associated hypertexts HT[BABY CLOTHES(l)] to HT[BABY CLOTHES(4)] stored and assigned to the search criterion phoneme sequence KP(BABY CLOTHES) in the first search table 23 from the search storage means 22 and delivers this information to the control means 21.
- the control means 21 then generate text information TI of an Internet index page 26 shown in Fig. 6 and apply this to the Intemet browser 6 to be displayed on the monitor 5.
- Text information TI8 indicates the search criterion KT entered as a search command by the user.
- Text information TI9 to Til 2 forms hyperlinks HL to the Internet pages featured by the Internet addresses URL[BABY CLOTHES(l)] to URL[BABY CLOTHES(4)] with information about the search criterion "BABY CLOTHES", which hyperlinks can be activated by the user .
- This information entered with the keyboard 7 is received by the correlation stage 12.
- the search table 24 shown in Fig. 7 is then stored in the search storage means 22.
- the user of the computer 1 can also assign search criterions KT to Internet addresses URL of interest to himself and store them in the second search table 24.
- the first search table 23 is then provided by the provider of the database and updated, in contrast to which the second search table 24 can be provided and updated continuously by the user of the computer 1 himself.
- the user can also enter a plurality of Internet addresses URL[KT(HL)] and hypertexts HT(HL) of hyperlinks HL to the search criterion text information KT by using the keyboard 7. They are then also entered in the search table 24, as this is represented in the third row of the search table 24 under the search criterion "VIDEO".
- the user would like to order a pizza from his pizza service. For this purpose the user speaks the command "MY PIZZA SERVICE" into the microphone 17. The speech recognition stage 20 then compares the detected phoneme sequence P with phoneme sequences of recognizable speech commands and recognizes the speech command "MY".
- a search phoneme sequence SP of the search criterion "PHOTOS" detected by the speech recognition stage 20 is compared by the command determining stage 14 to search criterion phoneme sequences KP[KT(HL)] stored in the first search table 23.
- the command determining stage 14 determines that none of the stored search criterion phoneme sequences KP[KT(HL)] has a sufficiently great correspondence to the search phoneme sequence SP.
- Search text information ST "PHOTOS" detected by the speech recognition stage 20 is delivered by it to the command determining stage 14 and delivered by this stage to the control means 21.
- the first search table 23 is not stored, as explained with reference to the first example of application, in the second computer program product, thus in the search storage means 22 of the computer 1, and continuously updated.
- the Internet address URL of the third computer server 25 stored in the control means of the second computer program product and, subsequently, the detected search phoneme sequence SP of the search criterion KT is delivered to the Internet browser 6 which delivers the search phoneme sequence SP to the third computer server 25 via the Internet NET.
- the third computer server 25 comprises means corresponding to the command determining stage 14 and the search storage means 22 of the computer 1.
- the third computer server 25 executes a selection method of selecting Internet addresses URL[KT(HL)] of Internet pages assigned to a search criterion
- a search phoneme sequence SP sent via the Internet NET to the third computer server 25 is received by the third computer server 25.
- the third computer server 25 detects from a search table stored in the third computer server 25 a search criterion phoneme sequence KP(KT) that has a sufficiently large correspondence to the received search phoneme sequence SP and detects at least one Internet address URL[(KT(HL)] stored and assigned to this search criterion phoneme sequence KP(KT). These one or more Internet addresses URL[KT(HL)] detected by the third computer server 25 via the search criterion KT are then delivered to the computer by the third computer sever 25, by which computer the search phoneme sequence SP is received.
- the speech recognition method is subdivided - as described with reference to the second example of embodiment - into a part to be processed by the computer (client) of the user and a part to be processed by the computer server, in which phonemes or information corresponding to the phonemes are transmitted from the client to the server, there are two essential advantages.
- the speaker-dependent processing operations of the speech recognition method are processed at the client's, so that the server advantageously need not process speaker-dependent information.
- all processing operations that cost much memory space are uniformly processed by the server, so that the computers (clients) of the user advantageously need not have much memory capacity.
- the second computer program product can be loaded from a CD-ROM or a floppy disc into the internal memory of the computer 1 and thus, advantageously, can be installed in the computer 1 in an extremely simple manner.
- the speech recognition method and the second computer program product can be implemented and processed respectively by any product that can be connected to the Internet NET.
- Such products may be, for example, a personal digital assistant, a set top box or a mobile telephone, which can set up a connection to the Internet.
- the hypertext HT(HL) of a hyperlink HL of an Internet page displayed by the monitor 5 contains the same text information as a search command that can be detected by the speech recognition stage 20 together with the subsequent search criterion KT, the hyperlink HL of the Internet page can be activated, for example, by speaking a command "CLICK" before the hypertext HT(HL) is spoken.
- the second computer program product may also form part of the first computer program product - thus part of the Internet browser.
Landscapes
- Engineering & Computer Science (AREA)
- Computational Linguistics (AREA)
- Health & Medical Sciences (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Human Computer Interaction (AREA)
- Physics & Mathematics (AREA)
- Acoustics & Sound (AREA)
- Multimedia (AREA)
- Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
Abstract
Description
Claims
Priority Applications (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2001540761A JP2003515832A (en) | 1999-11-25 | 2000-11-10 | Browse Web Pages by Category for Voice Navigation |
EP00977543A EP1157373A1 (en) | 1999-11-25 | 2000-11-10 | Referencing web pages by categories for voice navigation |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
EP99890371.0 | 1999-11-25 | ||
EP99890371 | 1999-11-25 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2001039178A1 true WO2001039178A1 (en) | 2001-05-31 |
Family
ID=8244028
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/EP2000/011299 WO2001039178A1 (en) | 1999-11-25 | 2000-11-10 | Referencing web pages by categories for voice navigation |
Country Status (3)
Country | Link |
---|---|
EP (1) | EP1157373A1 (en) |
JP (1) | JP2003515832A (en) |
WO (1) | WO2001039178A1 (en) |
Cited By (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7162424B2 (en) | 2001-04-26 | 2007-01-09 | Siemens Aktiengesellschaft | Method and system for defining a sequence of sound modules for synthesis of a speech signal in a tonal language |
US7289960B2 (en) | 2001-10-24 | 2007-10-30 | Agiletv Corporation | System and method for speech activated internet browsing using open vocabulary enhancement |
US7324947B2 (en) | 2001-10-03 | 2008-01-29 | Promptu Systems Corporation | Global speech user interface |
US7428273B2 (en) | 2003-09-18 | 2008-09-23 | Promptu Systems Corporation | Method and apparatus for efficient preamble detection in digital data receivers |
US7519534B2 (en) | 2002-10-31 | 2009-04-14 | Agiletv Corporation | Speech controlled access to content on a presentation medium |
US7729910B2 (en) | 2003-06-26 | 2010-06-01 | Agiletv Corporation | Zero-search, zero-memory vector quantization |
US8321427B2 (en) | 2002-10-31 | 2012-11-27 | Promptu Systems Corporation | Method and apparatus for generation and augmentation of search terms from external and internal sources |
USRE44326E1 (en) | 2000-06-08 | 2013-06-25 | Promptu Systems Corporation | System and method of voice recognition near a wireline node of a network supporting cable television and/or video delivery |
WO2016058425A1 (en) * | 2014-10-17 | 2016-04-21 | 百度在线网络技术(北京)有限公司 | Voice search method, apparatus and device, and computer storage medium |
Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5960399A (en) * | 1996-12-24 | 1999-09-28 | Gte Internetworking Incorporated | Client/server speech processor/recognizer |
WO1999050830A1 (en) * | 1998-03-30 | 1999-10-07 | Microsoft Corporation | Information retrieval and speech recognition based on language models |
-
2000
- 2000-11-10 WO PCT/EP2000/011299 patent/WO2001039178A1/en not_active Application Discontinuation
- 2000-11-10 EP EP00977543A patent/EP1157373A1/en not_active Ceased
- 2000-11-10 JP JP2001540761A patent/JP2003515832A/en active Pending
Patent Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5960399A (en) * | 1996-12-24 | 1999-09-28 | Gte Internetworking Incorporated | Client/server speech processor/recognizer |
WO1999050830A1 (en) * | 1998-03-30 | 1999-10-07 | Microsoft Corporation | Information retrieval and speech recognition based on language models |
Non-Patent Citations (3)
Title |
---|
DATABASE INSPEC [online] INSTITUTE OF ELECTRICAL ENGINEERS, STEVENAGE, GB; KATSUURA M ET AL: "The WWW browser system with spoken keyword recognition", XP002160022, Database accession no. 6264744 * |
LAU R ET AL: "WebGALAXY: beyond point and click -- a conversational interface to a browser", COMPUTER NETWORKS AND ISDN SYSTEMS,NL,NORTH HOLLAND PUBLISHING. AMSTERDAM, vol. 29, no. 8-13, 1 September 1997 (1997-09-01), pages 1385 - 1393, XP004095333, ISSN: 0169-7552 * |
TRANSACTIONS OF THE INFORMATION PROCESSING SOCIETY OF JAPAN, FEB. 1999, INF. PROCESS. SOC. JAPAN, JAPAN, vol. 40, no. 2, pages 443 - 452, ISSN: 0387-5806 * |
Cited By (27)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
USRE44326E1 (en) | 2000-06-08 | 2013-06-25 | Promptu Systems Corporation | System and method of voice recognition near a wireline node of a network supporting cable television and/or video delivery |
US7162424B2 (en) | 2001-04-26 | 2007-01-09 | Siemens Aktiengesellschaft | Method and system for defining a sequence of sound modules for synthesis of a speech signal in a tonal language |
US9848243B2 (en) | 2001-10-03 | 2017-12-19 | Promptu Systems Corporation | Global speech user interface |
US11070882B2 (en) | 2001-10-03 | 2021-07-20 | Promptu Systems Corporation | Global speech user interface |
US11172260B2 (en) | 2001-10-03 | 2021-11-09 | Promptu Systems Corporation | Speech interface |
US10932005B2 (en) | 2001-10-03 | 2021-02-23 | Promptu Systems Corporation | Speech interface |
US8005679B2 (en) | 2001-10-03 | 2011-08-23 | Promptu Systems Corporation | Global speech user interface |
US10257576B2 (en) | 2001-10-03 | 2019-04-09 | Promptu Systems Corporation | Global speech user interface |
US8983838B2 (en) | 2001-10-03 | 2015-03-17 | Promptu Systems Corporation | Global speech user interface |
US8407056B2 (en) | 2001-10-03 | 2013-03-26 | Promptu Systems Corporation | Global speech user interface |
US7324947B2 (en) | 2001-10-03 | 2008-01-29 | Promptu Systems Corporation | Global speech user interface |
US8818804B2 (en) | 2001-10-03 | 2014-08-26 | Promptu Systems Corporation | Global speech user interface |
US7289960B2 (en) | 2001-10-24 | 2007-10-30 | Agiletv Corporation | System and method for speech activated internet browsing using open vocabulary enhancement |
US9305549B2 (en) | 2002-10-31 | 2016-04-05 | Promptu Systems Corporation | Method and apparatus for generation and augmentation of search terms from external and internal sources |
US8959019B2 (en) | 2002-10-31 | 2015-02-17 | Promptu Systems Corporation | Efficient empirical determination, computation, and use of acoustic confusability measures |
US12067979B2 (en) | 2002-10-31 | 2024-08-20 | Promptu Systems Corporation | Efficient empirical determination, computation, and use of acoustic confusability measures |
US11587558B2 (en) | 2002-10-31 | 2023-02-21 | Promptu Systems Corporation | Efficient empirical determination, computation, and use of acoustic confusability measures |
US8321427B2 (en) | 2002-10-31 | 2012-11-27 | Promptu Systems Corporation | Method and apparatus for generation and augmentation of search terms from external and internal sources |
US8862596B2 (en) | 2002-10-31 | 2014-10-14 | Promptu Systems Corporation | Method and apparatus for generation and augmentation of search terms from external and internal sources |
US9626965B2 (en) | 2002-10-31 | 2017-04-18 | Promptu Systems Corporation | Efficient empirical computation and utilization of acoustic confusability |
US10121469B2 (en) | 2002-10-31 | 2018-11-06 | Promptu Systems Corporation | Efficient empirical determination, computation, and use of acoustic confusability measures |
US10748527B2 (en) | 2002-10-31 | 2020-08-18 | Promptu Systems Corporation | Efficient empirical determination, computation, and use of acoustic confusability measures |
US7519534B2 (en) | 2002-10-31 | 2009-04-14 | Agiletv Corporation | Speech controlled access to content on a presentation medium |
US8185390B2 (en) | 2003-06-26 | 2012-05-22 | Promptu Systems Corporation | Zero-search, zero-memory vector quantization |
US7729910B2 (en) | 2003-06-26 | 2010-06-01 | Agiletv Corporation | Zero-search, zero-memory vector quantization |
US7428273B2 (en) | 2003-09-18 | 2008-09-23 | Promptu Systems Corporation | Method and apparatus for efficient preamble detection in digital data receivers |
WO2016058425A1 (en) * | 2014-10-17 | 2016-04-21 | 百度在线网络技术(北京)有限公司 | Voice search method, apparatus and device, and computer storage medium |
Also Published As
Publication number | Publication date |
---|---|
JP2003515832A (en) | 2003-05-07 |
EP1157373A1 (en) | 2001-11-28 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US6604076B1 (en) | Speech recognition method for activating a hyperlink of an internet page | |
US9202247B2 (en) | System and method utilizing voice search to locate a product in stores from a phone | |
US6941273B1 (en) | Telephony-data application interface apparatus and method for multi-modal access to data applications | |
US7228327B2 (en) | Method and apparatus for delivering content via information retrieval devices | |
US6400806B1 (en) | System and method for providing and using universally accessible voice and speech data files | |
US7522711B1 (en) | Delivery of audio driving directions via a telephone interface | |
US20090304161A1 (en) | system and method utilizing voice search to locate a product in stores from a phone | |
JP2002536755A (en) | Intercommunication system and method between user and system | |
US20120253800A1 (en) | System and Method for Modifying and Updating a Speech Recognition Program | |
US20070050191A1 (en) | Mobile systems and methods of supporting natural language human-machine interactions | |
US20100114571A1 (en) | Information retrieval system, information retrieval method, and information retrieval program | |
JPH06208389A (en) | Method and device for information processing | |
EP1157373A1 (en) | Referencing web pages by categories for voice navigation | |
JP2002539481A (en) | Method using multiple speech recognizers | |
US7346651B2 (en) | Method of searching information site by item keyword and action keyword | |
US20050102147A1 (en) | Method of speech-based navigation in a communications network and of implementing a speech input possibility in private information units | |
JP2003005778A (en) | Voice recognition portal system | |
KR20010044834A (en) | System and method for processing speech-order | |
CN1316076A (en) | User-profile-driven mapping of hyperlinks onto URLS | |
JPH10164249A (en) | Information processor | |
KR100381605B1 (en) | Ars voice web hosting service system and the method thereof | |
JP2003319085A (en) | Voice information search device and voice information search method | |
US20080133240A1 (en) | Spoken dialog system, terminal device, speech information management device and recording medium with program recorded thereon | |
JP2003140682A (en) | Voice recognition device and voice dictionary generation method | |
JP2001222494A (en) | Communication address search device, system and method |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AK | Designated states |
Kind code of ref document: A1 Designated state(s): JP |
|
AL | Designated countries for regional patents |
Kind code of ref document: A1 Designated state(s): AT BE CH CY DE DK ES FI FR GB GR IE IT LU MC NL PT SE |
|
ENP | Entry into the national phase |
Ref country code: JP Ref document number: 2001 540761 Kind code of ref document: A Format of ref document f/p: F |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application | ||
WWE | Wipo information: entry into national phase |
Ref document number: 2000977543 Country of ref document: EP |
|
WWP | Wipo information: published in national office |
Ref document number: 2000977543 Country of ref document: EP |
|
WWR | Wipo information: refused in national office |
Ref document number: 2000977543 Country of ref document: EP |
|
WWW | Wipo information: withdrawn in national office |
Ref document number: 2000977543 Country of ref document: EP |