Links

Images

Classifications

G—PHYSICS

G06—COMPUTING; CALCULATING; COUNTING

G06F—ELECTRIC DIGITAL DATA PROCESSING

G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements

G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer

G06F3/0487—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser

G06F3/0488—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures

G06F3/04883—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures for entering handwritten data, e.g. gestures, text

G—PHYSICS

G06—COMPUTING; CALCULATING; COUNTING

G06F—ELECTRIC DIGITAL DATA PROCESSING

G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements

G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer

G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements

G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer

G06F3/0487—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser

G06F3/0488—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures

G06F3/04886—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures by partitioning the screen or tablet into independently controllable areas, e.g. virtual keyboards, menus

Abstract

Systems and methods for gesture repetition that is easy to repeat quickly. A user is able to quickly repeat any activation of a key, including any of the alternate, overloaded functions that are associated with the key. In addition, by appropriately defining the gesture, a large number of these repetitions can be quickly generated. These repeated activations can be generated much faster than a key can be repeatedly tapped with sufficient accuracy, and far faster than an alternative function can be repeatedly generated when each repetition would require again waiting past an initial time threshold. Furthermore, the number of repetitions can be easily and precisely controlled by the user. As the desired number of repetitions is approached, the user simply begins to repeat the gesture at a slower pace until precisely the desired number of repetitions is achieved.

Description

The invention relates to a text input system and, more specifically, to a text input system where characters are entered and functions are activated by interacting with a virtual keyboard, such as a keyboard displayed on a touch screen text input system.

BACKGROUND OF THE INVENTION

The origin of the modern keyboard as the primary method for inputting text from a human to a machine dates back to early typewriters in the 19th century. As computers were developed, it was a natural evolution to adapt the typewriter keyboard to use as the primary method for inputting text. For a skilled typist, it has remained the fastest way possible to input text into the computer.

With ongoing efforts to make computers smaller and more portable, the physical keyboard has become one of the most significant limiting factors in just how small a device can become: the physical size of the human finger is not something computer designers could change. As a result, computers for certain portable applications have been designed without a physical keyboard, and use touch-screen based input methods as the primary form of human-computer interface. (This is also the case for some applications where people are physically unable to use a keyboard, such as persons with physical disabilities.)

Input methods that are designed for portable touch-screen based devices may have two primary requirements that frequently conflict with each other. The method of input should be as fast as possible and at the same time the method of input should take as little of the display screen as possible. Unfortunately, as the space taken up for input on the display screen is decreased, it can become difficult to increase speed without adversely affecting accuracy and usability.

In spite of a recent surge in the market for pen-based computing devices, many people who generate text still do so with a standard keyboard. To date, pen-based computing has not replaced conventional portable laptop computers, because text input on pen-based computers can be slow, among other reasons. Other devices sacrifice half of the area of the face of the device in order to include a “thumb” keyboard with miniature keys that are difficult to use, time-consuming, and error-prone. At the same time, such an approach means that the area relegated to a physical keyboard can no longer be used to provide a much larger display with correspondingly more efficient and enjoyable access to information. The recent success of the “iPhone” demonstrates that the “all touch-screen” device model has significant advantages in that it can offer a simpler and more intuitive user interface for many of the tasks commonly performed with handheld devices.

Analogous to one-finger typing, one type of input using a virtual keyboard is called “point and tap”. A stylus (or more and more commonly, a finger) is moved from letter to letter and a “tap” is performed by contacting the screen on the desired key to select it. This results in the need to constantly lift the stylus from the screen and set it back down again, significantly slowing input. Cursive handwriting was invented to allow a better (and faster) flow from letter to letter and reduce the number of pen (or quill) lifts. In a similar way, the current invention reduces the number of taps required when inputting using an on-screen keyboard, thus speeding text entry.

Handheld devices are designed to fit comfortably in a pocket, and are consequently limited in size. A virtual keyboard that must be displayed on the device's touch-screen is therefore restricted in size. Depending on the target language to be generated and application being used, it may be desirable to be able to generate a large variety of characters from the keyboard. There are generally three methods to achieve this: (1) Reduce the average size of each key to fit a large number of keys on a small keyboard area; (2) Add (and force the user to navigate to) other keyboard layouts from which alternate characters can be generated; or (3) Overload one or more of the keys on the primary keyboard layout so that multiple characters can be generated from a single key by activating it in different ways.

Method 1 is undesirable because the keys tend to become too small to activate easily or reliably. Method 2 is undesirable because the process of navigating between alternate keyboards is inconvenient and time-consuming. Method 3 may require that the user learn novel ways of activating keys, but once learned, this approach is fast and effective.

One very common way of over-loading a key with alternate functionality is to assign an alternative “tap-hold” behavior to a key, where the “base function” of a key is invoked by a “normal tap” of the key (where the key is contacted and the contact is terminated relatively quickly), and an alternative function is invoked by contacting the key and maintaining the contact past an initial time threshold. This approach can be extended, so that maintaining the contact past a second time threshold results in triggering a second alternative function for the key. Another method to overload a key with alternative functions is to define a set of one or more “gestures” that the user can perform following the initial contact with the key. For example, King et al (U.S. Pat. No. 5,953,541) disclose tapping a key and then moving the point of contact in a particular direction to generate one of an ambiguous set of characters assigned to the key. In another example, Buxton (U.S. Pat. No. 6,094,197) disclose tapping a key and then moving the point of contact to exit from the key in a particular direction to apply a function modifier (such as Shift or Control) to the character assigned to the key.

Another function that has been invoked based on the length of time during which contact with a key is continuously maintained is the automatic repetition of activations of the key. This is most commonly found to be useful in, for example, generating a relatively rapid sequence of a large number of activations of the backspace key to delete a relatively large piece of text. It can also be used to generate a large number of repetitions of a single character, such as simulating a horizontal line with a large number of hyphen characters. This has generally been done by providing for an initial time threshold past which contact must continue before the first repeated activation of the contacted key is generated. Additional activations are then generated at regular intervals as long as the contact with the key is continued. In some implementations, these regular intervals also become shorter over time so that additional repetitions occur more rapidly.

There are two problems with automatic key repetition based on the length of time during which contact with a key is continuously maintained. One is that these repetitions tend to be either slower than desired when a large number of repetitions is required, or conversely, faster than desired in order to accurately generate exactly the required number of repetitions (so that either too few—or worse, in the case of a repeated backspace, too many—repetitions are generated). The second problem is that this approach is not compatible with using time-based activation to generate alternate overloaded function definitions for a keyboard.

SUMMARY OF THE INVENTION

The present invention provides systems and methods for gesture repetition that is easy to repeat quickly. A user is able to quickly repeat any activation of a key, including any of the alternate, overloaded functions that are associated with the key. In addition, by appropriately defining the gesture, a large number of these repetitions can be quickly generated. These repeated activations can be generated much faster than a key can be repeatedly tapped with sufficient accuracy, and far faster than an alternative function can be repeatedly generated when each repetition would require again waiting past an initial time threshold. Furthermore, the number of repetitions can be easily and precisely controlled by the user. As the desired number of repetitions is approached, the user simply begins to repeat the gesture at a slower pace until precisely the desired number of repetitions is achieved.

BRIEF DESCRIPTION OF THE DRAWINGS

Preferred and alternative examples of the present invention are described in detail below with reference to the following drawings:

FIG. 1 shows a structural system block diagram showing the typical hardware components of a system that embodies the methods of the text editing and menu selection systems of the present invention; and

FIGS. 2 and 3 show examples of contact actions on a menu that are structured in one way that takes advantage of the method of the present invention.

DESCRIPTION OF THE INVENTION

FIG. 1 shows a simplified block diagram of the hardware components of a typical device 100 in which the present invention is implemented. The device 100 includes one or more input devices 120 that provide input to the CPU (processor) 110 notifying it of actions performed by a user, typically mediated by a hardware controller that interprets the raw signals received from the input device and communicates the information to the CPU 110 using a known communication protocol via an available data port. One example of an input device 120 is a touch-screen that provides input to the CPU 110 notifying it of contact events when the touch-screen is touched by a user. Similarly, the CPU 110 communicates with a hardware controller for a display 130 to draw on the display 130. One example of the display 130 is a touch-screen display that provides graphical and textual visual feedback to a user. Optionally, a speaker 140 is also coupled to the CPU 110 so that any appropriate auditory signals can be passed on to the user as guidance (predominantly for error signals), and a microphone 141 is also coupled to the CPU 110 so that any spoken input can be received from the user (predominantly for input action recognition systems implementing speech recognition as a method of text input by the user). The CPU 110 has access to a memory 150, which may include a combination of temporary and/or permanent storage, and both read-only and writable memory (random access memory or RAM), read-only memory (ROM), writable non-volatile memory such as FLASH memory, hard drives, floppy disks, and so forth. The memory 150 includes program memory 160 that contains all programs and software such as an operating system 161, an input action recognition system software 162, and any other application programs 163. The program memory 160 also contains at least one of a text editing system software 164 for recording and restoring the location of a text insertion position according to the method of the present invention, and a menu selection system software 165 for graphically displaying two or more choices to a user and determining a selection by a user of one of said graphically displayed choices according to the method of the present invention. The memory 150 also includes data memory 170 that includes any textual object database(s) 171 required by the input action recognition system software 162, optional storage for maintaining a record of user options and preferences 172, and any other data 173 required by any element of the device 100.

The present invention introduces a “repetition gesture” that is easy to repeat quickly. The CPU 110 allows a user to quickly repeat any activation of a key, including any of the alternate, overloaded functions that are associated with the key. In addition, by appropriately defining the gesture, a large number of these repetitions can be quickly generated. These repeated activations can be generated much faster than a key can be repeatedly tapped with sufficient accuracy, and far faster than an alternative function can be repeatedly generated when each repetition would require again waiting past an initial time threshold. Furthermore, the number of repetitions can be easily and precisely controlled by the user. As the desired number of repetitions is approached, the user simply begins to repeat the gesture at a slower pace until precisely the desired number of repetitions is achieved. In addition, the repetition gesture can be defined in such a way that it does “collide” (become easily confused with) other gestures that may be in use on the system such that it becomes a potential source of recognition errors.

For example, consider a keyboard in which the backspace key, when tapped normally, performs the expected function of deleting a single character to the left of the text cursor, and when activated with a “tap-hold” deletes the word to the left of the cursor (similar to a control-backspace on many Windows computers). By tapping the key then performing a repetition gesture before the end of the time threshold for a “tap-hold” activation, the user can precisely delete any desired number of characters to the left of the cursor. By waiting until the “tap-hold” time threshold has elapsed before performing the repetition gesture, the user can precisely delete any desired number of words to the left of the cursor.

The keyboard may be displayed on a touch-sensitive display screen (herein referred to as a touch-screen) and the user contacts the display by means of a stylus or finger. It should be understood, however, that the method can be applied to any system where the user can trace out a continuous path on a displayed keyboard, for example, a standard computer display monitor (not a touch-sensitive screen) where the point of “contact” is the location of an on-screen cursor whose location on the display is controlled by a mouse (or equivalent positional control device) and where the acts of “contacting” and “breaking contact with” the screen are indicated by closing and opening a switch. Alternatively, the beginning and end of a “contact action” is indicated by performing some other equivalent control action, such as “dwelling” near a location (maintaining the position of the mouse cursor within a defined region for a period of time longer than a selected threshold). In another embodiment, the beginning and end of a repetition gesture is defined by one or more recognizable features of a longer contact action which includes the key activation and repetition gesture. For example, in a system which includes a displayed keyboard as part of a larger display area within which the location of a mouse cursor can be controlled, a key activation is indicated by starting with the displayed cursor outside the area defined by the displayed keyboard, and moving the cursor into the keyboard area to the location of an intended key, performing a repetition gesture to effect a desired number of repetitions, then moving the cursor back outside the keyboard area. Such approaches enable the use of other positional control devices such as joysticks, eye-gaze control systems, etc., to control the movement of the displayed cursor over the keyboard, without requiring a separate control mechanism to effect the actions of “contacting” and “breaking contact with” screen.

Thus, the operation of the system will be described with reference to an aspect including a touch-screen contacted by a stylus or finger, but this should not be construed as limiting the scope of the invention, but simply as a means to provide illustrations of some of the present aspects of this method. The method simply processes a user-generated stream of location data, which begin at a known point (in the touch-screen model, the point where the stylus first contacts the screen), proceed through a sequence of data point locations which have a known relationship to a defined virtual keyboard area (in the touch-screen model, the sequence of points detected where the stylus is moved while in contact with the screen), and which end at a known point (in the touch-screen model, the point where the stylus last contacts the screen). In a touch-screen model, the keyboard and stream of data points exist in the same two-dimensional space, but the present invention is equally applicable to a system that can track movement in a three-dimensional space, wherein aspects of the repetition gesture are defined with respect to movement in the third dimension. In certain embodiments, the times when the data points are generated, or equivalently for these embodiments, the time intervals between the data points, are also processed, and can also be provided by various equivalent technologies.

In one embodiment, the repetition gesture includes moving the point of contact back and forth on the touch-screen of the device, wherein an activation of the initially contacted key is repeated for each time the movement of the point of contact changes direction. In a system where the function of a key changes in a time-based manner (as described previously for a Delete-Character/Delete-Word key), the function that is repeated is the function determined according to the time elapsed prior to initiating the repetition gesture. Thus, the repetition gesture can be performed on or near the region associated with the key being repeated.

In one embodiment, in addition to the at least one primary repetition gesture (each repetition of which results in one additional repetition of the activated key function), one or more alternate secondary repetition gestures are also defined, each repetition of which results in a pre-defined fixed number of repetitions of the key function. For example, an alternate repetition gesture includes moving the point of contact in a circular movement, where each time the point of contact moves through a complete 360-degree circle, the key function is repeated ten times. Of course, other specific gestures and other numbers of repetitions can be defined without changing the scope or spirit of the invention. The specific nature of each alternate repetition gesture, and the number of repetitions to be triggered for each performance of an alternate repetition gesture may also be implemented as options selectable by the user.

In one embodiment, the repetition gesture includes an initial component that is performed so that the gesture as a whole is distinct from other gestures that may be meaningful in the system. For example, following the initial key contact (which hereinafter will implicitly include any pause or distinguishing gesture that may be used to specify an alternate function for the contacted key), the point of contact is moved to a designated region of the screen where one or more repetition gestures are them performed. Thus, the repetition gesture can be performed at a location that is distinct from the region associated with the key being repeated. As an example, in the system disclosed by Kushler et al (U.S. Pat. No. 7,098,896), virtually any path traced on the keyboard can be interpreted as an action intended to output a word. This and other patents referenced here are incorporated by reference in their entirety.

For example, FIG. 2 shows a representative touch-screen device with a displayed virtual keyboard in region 200. FIG. 2 shows the path trace of a contact action beginning at a location 210 and ending at a location 215, wherein a user intends to enter the word “retreat” according to the system disclosed by Kushler et al (U.S. Pat. No. 7,098,896), by passing through each of the letters of the word “retreat” in sequence, terminating the path at the location of the final letter “t.” In a case where the repetition gesture is defined as simple back-and-forth movement following a key contact, the path trace shown in FIG. 2 would be ambiguous as to whether the path was intended to specify the entry of the word “retreat,” or as the activation of the “r” key four times: the initial activation followed by a repetition gesture in which the traced path changes direction three times (at locations 211, 212 and 213) to generate three additional repeated activations of the “r” key. By defining the gesture appropriately, as in the examples described below, an initial component can be included that distinguishes a key repetition gesture from any other gesture used by a system. This avoids the possible introduction of any ambiguity as to how a contact action should be interpreted.

For example, in the system disclosed by Kushler et al, FIG. 3 shows one example of an initial component of a repetition gesture that includes moving off of the keyboard (in the example of FIG. 3, to an area 301 beneath the displayed keyboard 300) following the initial key contact. Once this initial gesture component is performed, then any number of primary or secondary repetition gestures can be performed without being ambiguous with any of the other input paths that the system is designed to process. In the example of FIG. 3, the path trace shown begins at location 310 on the “r” key, moves to region 301 below the keyboard region 300, and changes direction three times at locations 311, 312 and 313 to similarly generate a total of four activations of the “r” key. In contrast to the path trace shown in FIG. 2, the path trace shown in FIG. 3 can be easily distinguished from any that would be intended to generate the output of a word according to the system disclosed by Kushler et al (U.S. Pat. No. 7,098,896). In a similar manner, an appropriate initial component can be defined for virtually any system that recognizes a set of gestures so that the key repetition gesture can be reliably recognized and distinguished from any other gesture used by the system. The initial component of the repetition gesture can also include time-based aspects, such as one or more pauses or other changes in the velocity of the movement of the point of contact.

While examples of the technology have been described herein, many changes can be made without departing from the spirit and scope of the invention. The various insights embodied in the invention enable text to be input using a touch-screen keyboard more rapidly and efficiently, without limiting the ability to specify other gesture-based methods of input, and allowing time-based key activation to be used as extensively as desired to overload a relatively small number of virtual keys with all of the functionality found to be beneficial in the system. Any type of touch-screen may be used by the system, and the input device may be a stylus, a finger, or any tool that works as an input device on the touch-sensitive screen; as described above, the method can also be used with other types of systems that include a means to control the position of a cursor within the input space. Thus the scope of the method should be determined by the appended claims and their legal equivalents, rather than by the specific aspects described above.

Claims (14)

1. A method of controlling an electronic device having one or more defined regions activated in association with a contact action to move a point of control within one or more of said regions, the method comprising:

detecting an activating contact action in one of said defined regions,

identifying a region activated by said contact action;

recognizing the subsequent movement of said point of control; and

generating one or more repetitions of the activation of said identified region, when the recognized subsequent movement comprises one or more occurrences of a defined repetition gesture.

2. The method of claim 1, further comprising:

recognizing a portion of the subsequent movement of said point of control as a defined initial component of said repetition gesture.

3. The method of claim 2, wherein said defined initial component of said repetition gesture is not associated with a repetition of the activation of said identified region.

4. The method of claim 1, further comprising:

recognizing the subsequent movement of said point of control as comprising one or more occurrences of one or more defined alternate forms of a repetition gesture; and

generating a defined number repetitions of the activation of said identified region for each detected occurrence of one of said alternate forms of said repetition gesture.

5. The method of claim 1, wherein said electronic device comprises a displayed keyboard, said displayed keyboard includes a set of one or more regions comprising keys wherein each key is associated with one or more characters or functions.

6. An electronic device having one or more defined regions activated in association with a contact action to move a point of control within one or more of said regions, the device comprising:

a means for detecting an activating contact action in one of said defined regions,

a means for identifying a region activated by said contact action;

a means for recognizing the subsequent movement of said point of control; and

a means for generating one or more repetitions of the activation of said identified region, when the recognized subsequent movement comprises one or more occurrences of a defined repetition gesture.

7. The device of claim 6, further comprising:

a means for recognizing a portion of the subsequent movement of said point of control as a defined initial component of said repetition gesture.

8. The device of claim 7, wherein said defined initial component of said repetition gesture is not associated with a repetition of the activation of said identified region.

9. The device of claim 6, further comprising:

a means for recognizing the subsequent movement of said point of control as comprising one or more occurrences of one or more defined alternate forms of a repetition gesture; and

a means for generating a defined number repetitions of the activation of said identified region for each detected occurrence of one of said alternate forms of said repetition gesture.

10. The device of claim 6, wherein said electronic device comprises a displayed keyboard, said displayed keyboard includes a set of one or more regions comprising keys wherein each key is associated with one or more characters or functions.

11. The method of claim 1, further comprising:

recognizing one or more variations of said activating contact action in one or more of said defined regions,

associating one or more characters or functions with said one or more variations of said activating contact action in one or more of said defined regions;

identifying a region activated by said contact action;

identifying which of said one or more variations of said activating contact action is performed;

identifying which of said one or more characters or functions is associated with said identified contact action variation;

recognizing the subsequent movement of said point of control; and

generating one or more repetitions of said identified one or more characters or functions, when the recognized subsequent movement comprises one or more occurrences of a defined repetition gesture.

12. The method of claim 11, further comprising:

recognizing a portion of the subsequent movement of said point of control as a defined initial component of said repetition gesture.

13. The method of claim 12, wherein said defined initial component of said repetition gesture is not associated with a repetition of the activation of said identified one or more characters or functions.

14. The method of claim 11, further comprising:

recognizing the subsequent movement of said point of control as comprising one or more occurrences of one or more defined alternate forms of a repetition gesture; and

generating a defined number repetitions of the activation of said identified one or more characters or functions for each detected occurrence of one of said alternate forms of said repetition gesture.