Method and apparatus for presenting proximal feedback in...

Data processing: speech signal processing – linguistics – language – Speech signal processing – Application

Reexamination Certificate

Rate now

  [ 0.00 ] – not rated yet Voters 0   Comments 0

Details

C345S215000

Reexamination Certificate

active

06233560

ABSTRACT:

BACKGROUND OF THE INVENTION
1. Field of the Invention
This invention relates to voice command computer systems and, more particularly, to such systems which present command feedback to the end-user.
2. Background and Related Art
As the computer field has matured, vast improvements have been made in easing the manner in which end-users may interface with the systems. What was originally a crude command line interface requiring keyboard input of the end-user has now evolved to a rich panoply of techniques and devices for facilitating a more natural human interface with the computer. Representative examples of this include various pointing devices such as mice, track balls, touchscreens and voice activated interfaces made possible by advances in computerized speech recognition.
Such speech recognition falls into two major categories. First, systems have been developed for voice data input, examples of which include speech dictation or “speech-to-text” systems such as those marketed by the IBM Corporation and Dragon Systems. Yet a second category of computerized speech recognition technology is represented by command and control systems wherein end-user speech patterns are detected and analyzed. A specific form of this command and control technology is represented by a “desktop navigator”, wherein a user may navigate through a computer system's graphical user interface (GUI) by merely speaking menu command words, labels on buttons, directional commands such as “up” and “left” to move the familiar mouse cursor, and the like.
Due to inherent voice recognition errors such speech recognition systems are prone to—partly because of the state of the technology and partly due to the variability of the user's speech patterns and memory of the correct inputs to utter, such systems typically reserve a dedicated, predetermined area of the GUI to display the system's interpretation of the user's utterances. The user, by inspecting this area, is looking for confirmation that the utterance he or she just made was in fact interpreted or acted upon correctly.
Current speech recognition systems display this “best guess” in a control window at a fixed position on the screen of the GUI in a control window. Although the spatial positioning of this control window may in some implementations be altered by the end-user as desired or is placed in a fixed location such as the title bar of the current application being executed, numerous problems nevertheless remain associated with this technology. These methods of displaying these interpreted commands attempt to place the confirmation area in a position so as to not block the visibility of other important portions of the application. However, in doing so, this may necessitate distracting and disrupting eye movement from the area of the screen where the end-user is focusing to the confirmation area of the voice recognition, whereupon the user must then recall and reposition his or her eyes at the area of the screen which had the user's focus prior to the interruption caused by display of the confirmation area. Moreover, current systems display the results of the speech perception system without taking into account interpretation of the command, as it is simply reported.
It will be readily appreciated that in sessions of any length with such a voice recognition computer system these drawbacks can become extremely tiresome both physically and mentally and can severely impact the productivity of the end-user in a manner as to almost effectively negate the aforementioned benefits to voice recognition or navigation systems. Current methods and technology have simply been unable to eliminate these serious problems of maintaining visibility of important areas of the display and the focusing and refocusing just described.
Accordingly, such a voice command and control system was highly desired which could avoid the distractions to end-users and the degradation of visibility of applications presented on GUI user interfaces.
SUMMARY OF THE INVENTION
In a voice actuated computer system, voice input from an end-user, such as command and/or control utterances relevant to or prompted from a discrete location on a display screen, is analyzed by the system. In a response to such analysis, a confirmation area is thereby displayed on the display screen in a location functionally related to the analyzed contents of the voice input or the location the utterance was prompted from. Within the confirmation area the computer system's interpretation of the utterance is displayed for a preselected time and then gradually dissolved also at a preselected rate, such persistence and dissolution being selectively adjustable by the end-user. In this manner, display of the interpretation of the command or control utterance will be placed in a confirmation area which is in turn spatially positioned on the display screen at variable locations functionally related to the content of the command or control utterance so as to be proximal to the location on the display screen which may have had the user's eye focus and which gave rise to the utterance. In this manner, distractions are thereby avoided associated with a fixed location confirmation in the prior art (1) which obscures content of interest on the display screen and/or (2) destroys end-user focus by requiring the eyes to shift from positions of interest on the display screen to a different location wherein the confirmation is displayed. An important factor in the invention is the interpretation of the utterance and therefore the true target of the command. Feedback will be preferably withheld, momentarily in most cases, until such correct interpretation to application functionality can be determined.
In a more general case, the invention is not intended to be limited to the prior art programmer's graphical model comprised of interaction and association with cursor targets, menu items, pointing devices, and toolbars. Rather, the invention contemplates a visual mechanism for display of recognized speech commands in the form of natural language and resulting feedback actions proximal to the area of focus in a graphical user interface which gave rise to and prompted the speech command. Such feedback actions may include text over text, visuals, color change, animation, and gradual reduction of an image or message fading into the target area automatically so as to provide feedback to the end-user that the correct spoken action has been taken in the desired focus or target area.


REFERENCES:
patent: 4451895 (1984-05-01), Sliwkowski
patent: 4506142 (1985-03-01), Takano et al.
patent: 4726065 (1988-02-01), Froessl
patent: 4766529 (1988-08-01), Nakano et al.
patent: 4783759 (1988-11-01), Borgendale et al.
patent: 5027406 (1991-06-01), Roberts et al.
patent: 5068900 (1991-11-01), Searcy et al.
patent: 5133011 (1992-07-01), McKiel, Jr.
patent: 5157384 (1992-10-01), Greanias et al.
patent: 5222146 (1993-06-01), Bahl et al.
patent: 5231670 (1993-07-01), Goldhor et al.
patent: 5251130 (1993-10-01), Andrews et al.
patent: 5305244 (1994-04-01), Newman et al.
patent: 5386494 (1995-01-01), White
patent: 5408582 (1995-04-01), Colier
patent: 5428707 (1995-06-01), Gould et al.
patent: 5452397 (1995-09-01), Ittycheriah et al.
patent: 5465317 (1995-11-01), Epstein
patent: 5500920 (1996-03-01), Kupiec
patent: 5526407 (1996-06-01), Russell et al.
patent: 5553121 (1996-09-01), Martin et al.
patent: 5602963 (1997-02-01), Bissonnette et al.
patent: 5604840 (1997-02-01), Asai et al.
patent: 5632002 (1997-05-01), Hashimoto et al.
patent: 5638486 (1997-06-01), Wang et al.
patent: 5664061 (1997-09-01), Andreshak et al.
patent: 5671328 (1997-09-01), Fitzpatrick et al.
patent: 5698834 (1997-12-01), Worthington et al.
patent: 5706399 (1998-01-01), Bareis
patent: 5729659 (1998-03-01), Potter
patent: 5864815 (1999-01-01), Rozak et al.
patent: 6018711 (2000-01-01), French-St. George et al.

LandOfFree

Say what you really think

Search LandOfFree.com for the USA inventors and patents. Rate them and share your experience with other people.

Rating

Method and apparatus for presenting proximal feedback in... does not yet have a rating. At this time, there are no reviews or comments for this patent.

If you have personal experience with Method and apparatus for presenting proximal feedback in..., we encourage you to share that experience with our LandOfFree.com community. Your opinion is very important and Method and apparatus for presenting proximal feedback in... will most certainly appreciate the feedback.

Rate now

     

Profile ID: LFUS-PAI-O-2546596

  Search
All data on this website is collected from public sources. Our data reflects the most accurate information available at the time of publication.