US20080278455A1 - User-Defined Enablement Protocol - Google Patents

User-Defined Enablement Protocol Download PDF

Info

Publication number
US20080278455A1
US20080278455A1 US12/118,047 US11804708A US2008278455A1 US 20080278455 A1 US20080278455 A1 US 20080278455A1 US 11804708 A US11804708 A US 11804708A US 2008278455 A1 US2008278455 A1 US 2008278455A1
Authority
US
United States
Prior art keywords
user
touch
enablement
enablement protocol
function
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US12/118,047
Inventor
Graham Roy Atkins
Ian Andrew Maxwell
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Zetta Research and Development LLC RPO Series
Original Assignee
RPO Pty Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Family has litigation
First worldwide family litigation filed litigation Critical https://patents.darts-ip.com/?family=39969088&utm_source=google_patent&utm_medium=platform_link&utm_campaign=public_patent_search&patent=US20080278455(A1) "Global patent litigation dataset” by Darts-ip is licensed under a Creative Commons Attribution 4.0 International License.
Priority claimed from AU2007902519A external-priority patent/AU2007902519A0/en
Application filed by RPO Pty Ltd filed Critical RPO Pty Ltd
Assigned to RPO PTY LIMITED reassignment RPO PTY LIMITED ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: ATKINS, GRAHAM ROY, MAXWELL, IAN ANDREW
Publication of US20080278455A1 publication Critical patent/US20080278455A1/en
Assigned to BRIDGE BANK, NATIONAL ASSOCIATION reassignment BRIDGE BANK, NATIONAL ASSOCIATION SECURITY AGREEMENT Assignors: RPO PTY LTD
Priority to US13/473,961 priority Critical patent/US9280279B2/en
Assigned to RPO PTY LTD reassignment RPO PTY LTD REASSIGNMENT AND RELEASE OF IP SECURITY INTEREST Assignors: BRIDGE BANK, NATIONAL ASSOCIATION
Assigned to TRINITY CAPITAL INVESTMENT LLC reassignment TRINITY CAPITAL INVESTMENT LLC ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: RPO PTY LTD
Assigned to ZETTA RESEARCH AND DEVELOPMENT LLC - RPO SERIES reassignment ZETTA RESEARCH AND DEVELOPMENT LLC - RPO SERIES ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: TRINITY CAPITAL INVESTMENT LLC
Priority to US15/056,015 priority patent/US9696847B2/en
Priority to US15/626,585 priority patent/US20170285853A1/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/03Arrangements for converting the position or the displacement of a member into a coded form
    • G06F3/041Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means
    • G06F3/0416Control or interface arrangements specially adapted for digitisers
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0487Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
    • G06F3/0488Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0487Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
    • G06F3/0488Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures
    • G06F3/04883Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures for inputting data by handwriting, e.g. gesture or text

Definitions

  • the present disclosure relates to user-defined enablement protocols for electronic devices.
  • the disclosure has been developed primarily for providing an enhanced security means for enabling or executing functions of electronic devices through an interface such as a touch-sensitive display, and will be described hereinafter with reference to this application. However it will be appreciated that the disclosure is not limited to this particular field of use.
  • Electronic devices often require some form of enablement protocol to activate them or to enable a particular function. For example to activate a mobile phone or to log on to a computer system a user may need to enter an alphanumeric code e.g. ‘123’.
  • an alphanumeric code e.g. ‘123’.
  • a known drawback here is that if an unauthorised person learns the alphanumeric code, it is simple for them to activate the mobile phone or log in to the computer system. If the device is equipped with a touch screen so that the alphanumeric keys can be shown on a display, an additional level of security is available in that the locations of the keys can be scrambled either before or during entry of the code, making it more difficult for another person to learn the code by observation. However once the code is known, it is once again simple for an unauthorised person to enter it.
  • US Patent Publication No. 2008/00823015 discloses a non-alphanumeric method for unlocking an electronic device whereby direction keys are used to traverse a path through a displayed maze, and the device is unlocked if the path matches a path preset by the user. Once again, there is no security once the unlock code (i.e. the path) is known, and the restriction to using direction keys limits the range and complexity of possible enablement protocols.
  • touch screens offer a much greater range of enablement protocols for an electronic device.
  • a user can also provide ‘gestural input’ in the form of one or more swipes or taps for example.
  • US Patent Publication No. 2007/0150842 discloses a method for unlocking an electronic device whereby a person touches a displayed image and moves it along an indicated path. Besides unlocking a device, many other functions can be enabled with gestures on a touch screen.
  • 2006/0026535 discloses various mechanisms for interacting with one or more GUI (Graphical User Interface) elements based on a user interface mode and in response to one or more detected touches.
  • the graphical elements that may be activated include a virtual scroll wheel, a virtual keyboard, a toolbar and a control panel.
  • Functions applied include rotating, translating or re-sizing a displayed image, and scrolling through a list.
  • U.S. Pat. Nos. 5,825,352 and 5,943,043 disclose various two finger gestural inputs that emulate mouse functions
  • US Patent Publication No. 2007/0177803 discloses the possibility of allowing a user to assign an arbitrary function to a predetermined gesture. However in all these cases the user is either guided by displayed material or the gesture is predetermined, so there is limited opportunity for altering the functionality provided by the device or the mechanisms by which those functions are enabled.
  • the present disclosure overcomes or ameliorates at least one of the disadvantages of the prior art, and provides a useful alternative.
  • the present disclosure provides a user interface method, comprising the steps of:
  • defining the enablement protocol for a function comprises user-engagement with an interface, which comprises rises one or more touches or near touches of a touch screen.
  • an enablement protocol by engaging other types or configurations of interfaces, for example a rectangular frame surrounding a space wherein adjacent borders of the frame include optical emitters and the opposing borders include photo-detectors, i.e. essentially a touch-sensitive screen apparatus minus the actual touch surface.
  • the enablement protocol may comprise voice recognition of pre-defined commands (audible instructions), or the tracking of eye movements, or some other type of gesture of one or more parts of the human body.
  • the user will preferably define or record that protocol in the same form.
  • the enablement protocol is a sequence of touches with a particular rhythm at the top left corner of a touch screen
  • the user will usually define or record that protocol by actually performing the sequence of touches.
  • the user could define or record the protocol by other means, such as a spoken or typed description.
  • the present disclosure provides user interface method, comprising the steps of:
  • an enablement protocol for a function of an electronic device said enablement protocol comprising at least one user-defined touch on a touch-sensitive area operatively associated with said electronic device; and ii. recording and retaining said enablement protocol of said function, such that a user enables said function by substantially reproducing said enablement protocol in the absence of spatial or temporal indication of at least a portion of said enablement protocol.
  • the enablement protocol which comprises a plurality of spatial and temporal data points (i.e. touches and/or gestures) may be selected from one or more of the following non-exhaustive list of spatial or temporal indications: a simple touch, movement of touch, direction of touch, timing or rhythm between touches, force of touch, number of sequential or simultaneous touches, position of touch or shape or change of shape of touch, etc.
  • the disclosure provides a method for user-defined enablement of a function through an interface, said method comprising the steps of:
  • the function can be enabled by one or more touches, and that the user is required to provide the enablement protocol without spatial or temporal indication of where the enablement protocol should be received to execute or enable the function.
  • the enablement protocol may comprise a plurality of user-defined touches of the interface, and at least one spatial or temporal indication of the plurality of the user-defined touches of the enablement protocol is not shown or displayed on the interface.
  • the enablement protocol may comprise a swipe of a touch screen with a 1 second dwell time at the conclusion of the swipe. The screen may only show where the swipe should be received and not indicate that the user must hold their finger down at the end of the swipe for 1 second. In this case only one of the two spatial/temporal indications is shown to the user.
  • the function may be enabled when the user input meets a predefined number of spatial or temporal indications.
  • the enablement protocol may comprise a partial swipe of a touch screen, and then a 1 second dwell time, and then a further partial swipe, i.e. swipe+dwell+swipe.
  • the function may be enabled if only two out of these three inputs are received, e.g. swipe+dwell, or swipe+swipe, or dwell+swipe.
  • the present disclosure provides a method for enablement of a function through an interface, said method comprising the steps of:
  • the user may be presented with a choice of suggested enablement protocols from which they can choose and substantially mimic, thereby recording their own specific and unique enablement protocol.
  • the user can then assign a function to their unique recorded enablement protocol and execute or enable the function when they input an enablement protocol that substantially coincides with the recorded enablement protocol, wherein the interface is without spatial or temporal indication of where at least a portion of the enablement protocol is to be received to execute or enable the function.
  • the present disclosure provides a method for userdefined enablement of a function through an interface, said method comprising the steps of:
  • the enablement protocol effectively acts as a ‘gatekeeper’ or security password control over a particular function.
  • the user records an enablement protocol, assigns a function to the enablement protocol and then when wishing to execute or enable the function is prompted to input their unique enablement protocol which, if coinciding with the recorded enablement protocol, allows the function to be executed or enabled.
  • Prior art devices typically require touches or near touches within one or more pre-existing or specified areas to enable a particular pre-defined function. For instance, a simple keypad on a touch display provides alphanumeric material at specific predefined locations, and touches or near touches at these locations must be accomplished in a specific sequence to enable the function, e.g. to activate a device containing the touch display.
  • the user-defined interface method according to the present disclosure has no such spatial or temporal indication to a user of where touches or near touches should be received to enable the function, thereby allowing a user to have complete freedom in regard to defining how the functionality of an electronic device associated with the interface is enabled by touch.
  • a simple embodiment may involve the activation of a personal electronic device such as PDA or mobile phone.
  • a security code (optionally user-defined) is required to enable the device.
  • This security code is normally a series of alphanumeric keystrokes either on a keypad separate from the display or on the display itself.
  • the present disclosure envisages no such spatial or temporal indication to a user. Rather a user can define a security code or ‘enablement protocol’ on the device or indeed on an apparatus separate from but operatively associated with the device by defining one or more touches to enable the aforementioned function.
  • One such ‘touch’ may be for instance the touch of a thumb in the upper right hand corner of the screen or other touch sensitive area, or both.
  • several touches at user-defined locations on the screen may enable the device.
  • a sequence of touches either at the user-defined locations or indeed anywhere on the screen can enable the functionality required. The timing or protocol (temporal sequence) of these touches must then be recreated to enable said function.
  • both the spatial and temporal nature of the touches are examined to ensure that they meet the user-defined enablement protocol.
  • different types of touching bodies could enable different functions by performing otherwise identical enablement protocols. For example a swipe down one side of a touch screen with a stylus enables function A, while the same swipe with a finger enables function B.
  • This embodiment expands the range of functions that can be performed with a limited range of gestures, or alternatively provides an additional security feature.
  • any ‘touch screen’ input device can have a series of functions which may only be enabled by the user who defined the aforementioned ‘enablement protocol’, since there is no spatial or temporal indication to assist another party who may wish to enable the aforementioned functions.
  • enablement protocols may be linked to the profiles of multiple users such that a single device may be enabled according to the profile of a user who defined a specific enablement protocol.
  • user A defines touches in the upper and lower left hand corners of a display to initiate a device.
  • touches are provided in those upper and lower corners the device is activated according to user A's profile.
  • User B may have an entirely separate profile and may wish the device to be set up according to their specific needs. They therefore define an enablement protocol different from that of user A; not only will other parties not be able to access the device without knowing the enablement protocol of either party, but once user B initiates their particular enablement protocol the device will be activated according to their specific profile.
  • an enablement protocol such as the locations to be touched or the particular gesture
  • other aspects such as the rhythm, pattern or sequence of the touches or their touch signature e.g. finger shape, will be different and therefore they will not be able to replicate the enablement protocol.
  • the disclosure as described is particularly useful for users with some form of physical handicap that may prevent them from using conventional touch screen devices.
  • Touches and near touches may be detected in several ways, including a touch-down, movement of touch, direction of touch, timing or rhythm (protocol) between touches i.e. temporal sequence, force of touch, number of sequential or simultaneous touches, position of touch, and shape or change of shape of touch approaching, contacting or leaving the touch sensitive area.
  • One or more of these detection mechanisms may be used to complete the enablement protocol. For instance, if the touch location is correct but the shape of the touch is incorrect the protocol will not be satisfied (i.e. there is not substantial coincidence between the received and the recorded protocol) and the associated function will not be enabled. Similarly if the timing or rhythm between touches is correct but the force applied by those touches is incorrect, once again the protocol will not be satisfied. It will be appreciated that although only selected touch technologies can directly measure the force of an applied touch, certain other touch technologies, especially those with high spatial resolution, can indirectly measure touch force from deformation of the touch object.
  • Touches or near touches may be detected by any known touch sensing technology including analogue resistive, surface capacitive, surface acoustic wave, projected capacitive, optical and bending wave technologies, bearing in mind that some touch technologies are better suited than others for distinguishing multiple simultaneous touches, should these be part of an enablement protocol.
  • Enablement protocols that include touch shape or change of shape will generally require high spatial resolution, such as may be offered by optical touch detection technologies disclosed in U.S. Pat. Nos. 5,914,709 and 7,352,940.
  • any known touch technology including those mentioned above, can be applied with the present disclosure, with enablement protocols defined within the limitations of that technology.
  • Certain touch technologies e.g.
  • resistive and bending wave require a direct touch for activation, whereas other touch technologies, such as optical touch where a sheet of light is established in front of a display, can also sense an object hovering close to the display.
  • touch sensitive device is activated by a touch or a near touch of an interface.
  • the present disclosure provides a user interface device comprising: a touch sensitive area adapted to detect touches; a display device to display one or more functions; and a storage medium including a user-defined enablement protocol for enabling functions operatively associated with said display device, said enablement protocol including one or more user-defined touches on or near said touchsensitive area, there being no spatial or temporal indication to said user of at least a portion of said enablement protocol.
  • the enablement protocol be performed on the display device, this is not absolutely necessary.
  • a user-defined interface method comprising the steps of:
  • said enablement protocol comprises one or more user-defined touches, ii. retaining said enablement protocol of said function, and iii. enabling said function by said one or more user-defined touches wherein, if a user is provided with spatial or temporal indication for a portion of said enablement protocol, said enablement protocol additionally requires detection of an additional touch parameter.
  • the additional parameter required for the enablement protocol would include one or more of touch movement, direction of touch, timing or rhythm between touches i.e. temporal sequence, force of touch, number of sequential or simultaneous touches, shape and/or change of shape of touch including in different planes e.g. vertical and horizontal, on approach or withdrawal.
  • some devices may have a spatial or temporal indication for a portion of an enablement protocol.
  • a mobile phone may have locations shown on the display that must be pressed in a certain sequence for partial enablement of the protocol.
  • One mechanism to complete the enablement protocol could be temporal i.e. the locations must be pressed in the precise timing and sequence required.
  • the size and shape of the touch may be detected to confirm enablement by an authorised user. This detection of touch size and shape may be accomplished not only in the plane of the display or touch sensitive area but also in planes above the display e.g. the shape of the touch approaching or leaving the touch sensitive area, if the touch sensor can detect near touches. This touch signature is virtually impossible to replicate by another user and therefore provides a secure individual enablement protocol. It will be clear to a person skilled in the art that there are various other options available to confirm enablement.
  • the term ‘coincide’ when used herein and when referring to a touch input being compared with a recorded or retained touch input, is meant to define that the inputs should substantially correspond.
  • the coincidence is substantial overlap in overall touch area, and in another embodiment may also include one or more other parameters, such as dwell time.
  • ‘coincide’ may refer to a predetermined number of matches between the spatial and/or temporal data points of the touch or near touch and those of the enablement protocol.
  • touch herein the term is also intended to encompass within its scope the term ‘near touch’.
  • the touch-sensitive area is larger than the area of the display device and contains the display device within its borders.
  • conventional touch-sensitive input devices include a display device that is at least as large as the touch-sensitive area.
  • the touch-sensitive area is larger than the display itself and includes the display device within its borders. This is highly unusual but provides advantages over conventional systems, for example the opportunity of producing a smaller display with consequent power reduction. Also, by including the touch-sensitive area outside the display, the display itself is not obscured while input is applied via the touch-sensitive area.
  • the touch portions/locations may be located on a first face of the touch-sensitive area, generally the display side, and/or a second opposite face of the touch-sensitive area. In one particular embodiment the touch portions/locations for enablement of the functions are arranged on the first and second opposite faces of the touch-sensitive area in such a way that they can be simultaneously operated by the thumb and fingers of a user.
  • FIGS. 1 and 2 illustrate conventional user interface devices according to the prior art
  • FIGS. 3A-3D illustrate a first embodiment of the present disclosure
  • FIGS. 4A-4C illustrate a second embodiment of the present disclosure
  • FIGS. 5A-5C illustrate a third embodiment of the present disclosure.
  • FIGS. 6 and 7 illustrate touch signature recognition according to a fourth embodiment of the present disclosure.
  • FIG. 8A-8C illustrate a fifth embodiment of the present disclosure.
  • FIGS. 9A-9C illustrate a sixth embodiment of the present disclosure.
  • FIGS. 1 and 2 provide illustrations of prior art functionality in conventional devices.
  • FIG. 1 shows a mobile phone 10 with an alphanumeric keypad 11 and a display 12 .
  • the mobile phone can only be enabled when a user-defined security code is entered via the alphanumeric keypad 11 .
  • the alphanumeric keypad may be provided on a touch-sensitive display (i.e. a touch screen) but once again the conventional device comprises a pre-defined arrangement of the enablement protocol such that if an unauthorised user were aware of the code e.g. 123, they could enable the device simply by engaging the predefined locations for numerals 1, 2 and 3.
  • FIG. 2 is an illustration of an automatic teller machine 20 with a touch sensitive display 21 .
  • This technology is similar to the mobile phone in FIG. 1 , and in this instance various functions and spatial locations for enabling those functions are clearly defined on the display.
  • the display includes indicia 22 describing the functions e.g. withdrawal, deposit and account balance, and the spatial locations for those functions are defined by the touch portions 23 . Accordingly any user can engage the functions defined on the display 21 by simply following the spatial arrangement already predefined by the device.
  • the enablement protocol of the present disclosure acts to restrict the performing of a user-instigated function. That is, a user attempts to perform (whether by touch or otherwise) a function on an electronic device, and the function is only carried out if the user performs their defined enablement protocol required for that function. Generally this enablement protocol will be performed in response to an instruction, an indication or a request for user input from the electronic device, similar to a conventional ‘password’ dialog box.
  • Each user of the device may have defined a single enablement protocol that enables all functions of the device, or they may have defined different enablement protocols for different functions.
  • a computer system may provide different levels of access to different classes of user, with one class of users having full access to a given file, a second class of users having read-only access, and other users barred or restricted access.
  • the enablement protocol is embedded in the particular touch input, such as a gesture, that instigates the function.
  • This instigating touch input may be predetermined by the electronic device or by software running on it (e.g. a ‘double tap’ gesture that opens a file), or it may be user-defined (e.g. tracing a person's initials to instigate a phone call to that person).
  • a user may attempt to open a file by double tapping on an icon, but the file will only be opened if the ‘details’ of the double tap gesture, such as the touch object size, the duration of each tap and the delay between the taps, substantially coincide with the pre-recorded gesture (user-defined protocol).
  • a user may assign one particular function to a swipe gesture when the gesture is performed with a finger, and a different function when the swipe is performed with a stylus.
  • FIGS. 3A-3D illustrate a first embodiment according to the present disclosure.
  • FIG. 3A shows a particular device 30 with a touch screen or touch-sensitive area 31 .
  • a user has defined the location of a touch or near touch at a location 32 in accordance with an enablement protocol for activation of the device.
  • any function of a device 30 may similarly include an enablement protocol comprising one or more user defined touches or near touches 32 .
  • the required touch location 32 i.e. the enablement protocol, is not shown on the device at all.
  • the dotted line 33 indicates the location for illustrative purposes only, and there is no spatial or temporal indication whatsoever on the device 30 of where or how the touches should be applied to satisfy the user-defined enablement protocol. In fact there may be no display associated with the touch-sensitive area.
  • FIG. 3B a user commences application of the enablement protocol, which requires a thumb to be placed at the location 32 .
  • FIG. 3C it can be seen that the user places their thumb 34 at the required location 32 , thereby satisfying the enablement protocol that results in activation of the device.
  • the enablement protocol comprises a single touch at the user-defined location 32 on the touch-sensitive area 31 of the particular device 30 . It will be clear, however, that the enablement protocol may comprise other forms of touch.
  • FIGS. 4A-4C displays a swipe by the user of their thumb to satisfy the enablement protocol.
  • This embodiment again comprises a device 40 with a touch-sensitive area 41 .
  • the spatial location of the enablement protocol for a certain function is shown by dotted lines 42 again for illustrative purposes only.
  • the enablement protocol requires movement of the thumb along the arc 43 .
  • the enablement protocol may also include the timing or speed of the swipe from the lower right hand corner to the upper right hand corner, the contact dwell time at the start or end of the swipe, and the shape of the thumb on the surface of the touch-sensitive area. It will be appreciated that this offers an enhanced degree of security in that even if another person knew the correct gesture, say by observation, it would be extremely difficult for them to replicate all other aspects of the enablement protocol.
  • FIG. 4B the user holds the device, swipes their thumb 44 across the area 42 in the predefined fashion required by the enablement protocol i.e. speed, timing, area covered etc such that, as shown in FIG. 4C , the device is activated.
  • FIGS. 5A-5C a multiple touch enablement protocol is shown.
  • a device 50 with a touch sensitive area 51 requires two touches at specific spatial locations to activate the device.
  • the spatial location 52 requires a touch of a left thumb 54
  • the spatial location 53 requires a swipe of a right thumb 55 along a linear path 56 .
  • the screen may be completely blank and the enablement protocol is known only to the authorised user of the device 50 . Once the touch at the location 52 and the swipe at the location 53 are accomplished in line with the protocol, the device 50 is activated.
  • the screen may display indicia suggesting where touch input is required to enable a protocol, however the indicia may be specifically configured to falsely indicate where the actual input is required to enable the function, thereby intentionally frustrating the attempts of an unauthorised person to access the electronic device.
  • FIGS. 6A-6D and 7 A- 7 D illustrate how a touch is detected and how a characteristic touch signature of an individual may be determined.
  • FIG. 6A a finger 60 is approaching the surface 61 of a touch screen.
  • FIG. 6B the finger first comes into contact with the surface and is therefore detected.
  • the area 62 of the surface covered by the finger is relatively small.
  • greater pressure is applied to the surface thereby distorting the finger such that it covers a greater area 63 of the surface, as seen in FIG. 6C .
  • 6 D the finger is withdrawn after a certain time that may itself be part of an enablement protocol.
  • this change of shape of a finger as it approaches, touches and withdraws from the surface 61 can be detected and may form part of an enablement protocol. This is particularly useful since even if the precise spatial and temporal location of an enablement protocol is known by an unauthorised user, the protocol may only be satisfied if the specific shape and change of shape of the touch matches that of the authorised user.
  • FIG. 7 gives a further illustration.
  • FIGS. 7A-7D show a finger movement or swipe on a touch-sensitive surface 70 .
  • a finger 71 approaches the surface 70
  • the finger first contacts the surface.
  • the finger then slides across the surface to the point shown in FIG. 7C and is then removed as shown in FIG. 7D .
  • the distance 72 covered by the finger may be detected and form part of an enablement protocol.
  • the change in shape of the finger as it slides may also be detected. Accordingly it can be seen that virtually any movement, shape or change of shape of a touch object can be detected and form part of an enablement protocol for enabling a particular function on a device.
  • FIGS. 8A-8C and 9 A- 9 C disclose a further embodiment of the present disclosure displaying the powerful flexibility and security available with user-defined enablement protocols.
  • FIGS. 8A-8C it can be seen that even if spatial or temporal indicators are provided on the device, an enablement protocol will still not be satisfied if it includes detection of additional touch or near touch parameters, e.g. shape or change of shape of a touch object.
  • a device 80 with a touch sensitive display 81 shows a spatial indicator 82 forming part of an enablement protocol.
  • the enablement protocol for a specific function of this device requires a single touch at the location of the indicator 82 .
  • the shape 83 of the touch object also forms part of the enablement protocol as discussed above in relation to FIGS. 6A-6D .
  • the resolution is sufficiently high to measure shape and/or change of shape on approach, at touch-down and at withdrawal of touch.
  • FIG. 8B a user holds the device 80 and commences the enablement protocol by placing the pad of their thumb 84 on the indicator 82 . This is accomplished in FIG.
  • the spatial indicator 82 gives a user some indication of the required enablement protocol
  • a touch object does not match the required shape 83
  • the device will not be enabled. For example touching the indicator 82 with a stylus or fingertip will not satisfy the enablement protocol.
  • a device 90 includes a touch sensitive display 91 showing two spatial and/or temporal indicators 92 and 93 .
  • the indicators 92 and 93 may be flashing to indicate the desired temporal sequence between touches at the indicators 92 and 93 .
  • the enablement protocol requires a simple touch at the location of the indicator 92 , but not at the location of the indicator 93 . Rather, the protocol requires a swipe along an arc 96 that begins in the lower right hand corner and terminates at the indicator 93 .
  • the device is held such that that portion of the enablement protocol requiring a left thumb touch 97 at the location of the indicator 92 is satisfied, and the right thumb 98 proceeds from the lower right hand corner up to the location of the indicator 93 within the time specified by the enablement protocol.
  • the enablement protocol enables the function of the device 90 and the device is, in this embodiment, activated.
  • an unauthorised person attempting to access the electronic device could attempt to replicate the enablement protocol of an authorised person by carefully inspecting the touch surface for residual finger marks.
  • the enablement protocol may be selected to avoid indicating, from an inspection of finger marks alone, the actual enablement protocol. For example, a circular enablement protocol would not indicate the direction which the user traced out the circle. Similarly, the direction of a swipe would be difficult to interpret from only residual finger marks, and enablement protocols with temporal features such as the rhythm of a touch sequence could not be ascertained from residual finger marks.

Abstract

A user interface method is disclosed. For a particular interface, such as a touch input device, the method involves defining an enablement protocol for a function and recording and retaining the enablement protocol of said function, such that a user enables the function by substantially reproducing the enablement protocol in the absence of spatial or temporal indication of at least a portion of the enablement protocol.

Description

    CROSS-REFERENCE TO RELATED APPLICATIONS
  • This application claims priority under the Paris Convention to Australian Provisional Patent Application No. 2007902519 filed on May 11, 2007, and Australian Provisional Patent Application No. 2007902517 filed on May 11, 2007, the contents of both of which are incorporated herein by reference.
  • FIELD OF THE DISCLOSURE
  • The present disclosure relates to user-defined enablement protocols for electronic devices. The disclosure has been developed primarily for providing an enhanced security means for enabling or executing functions of electronic devices through an interface such as a touch-sensitive display, and will be described hereinafter with reference to this application. However it will be appreciated that the disclosure is not limited to this particular field of use.
  • BACKGROUND OF THE DISCLOSURE
  • Any discussion of the prior art throughout the specification should in no way be considered as an admission that such prior art is widely known or forms part of the common general knowledge in the field.
  • Electronic devices often require some form of enablement protocol to activate them or to enable a particular function. For example to activate a mobile phone or to log on to a computer system a user may need to enter an alphanumeric code e.g. ‘123’. A known drawback here is that if an unauthorised person learns the alphanumeric code, it is simple for them to activate the mobile phone or log in to the computer system. If the device is equipped with a touch screen so that the alphanumeric keys can be shown on a display, an additional level of security is available in that the locations of the keys can be scrambled either before or during entry of the code, making it more difficult for another person to learn the code by observation. However once the code is known, it is once again simple for an unauthorised person to enter it. US Patent Publication No. 2008/00823015 discloses a non-alphanumeric method for unlocking an electronic device whereby direction keys are used to traverse a path through a displayed maze, and the device is unlocked if the path matches a path preset by the user. Once again, there is no security once the unlock code (i.e. the path) is known, and the restriction to using direction keys limits the range and complexity of possible enablement protocols.
  • Compared with a keyboard with hard keys, touch screens offer a much greater range of enablement protocols for an electronic device. Apart from the well-known ‘touch to select’ operation mode where a user selects or enables a function by touching a displayed icon, a user can also provide ‘gestural input’ in the form of one or more swipes or taps for example. US Patent Publication No. 2007/0150842 for instance discloses a method for unlocking an electronic device whereby a person touches a displayed image and moves it along an indicated path. Besides unlocking a device, many other functions can be enabled with gestures on a touch screen. US Patent Publication No. 2006/0026535 for instance discloses various mechanisms for interacting with one or more GUI (Graphical User Interface) elements based on a user interface mode and in response to one or more detected touches. The graphical elements that may be activated include a virtual scroll wheel, a virtual keyboard, a toolbar and a control panel. Functions applied include rotating, translating or re-sizing a displayed image, and scrolling through a list. U.S. Pat. Nos. 5,825,352 and 5,943,043 disclose various two finger gestural inputs that emulate mouse functions, while US Patent Publication No. 2007/0177803 discloses the possibility of allowing a user to assign an arbitrary function to a predetermined gesture. However in all these cases the user is either guided by displayed material or the gesture is predetermined, so there is limited opportunity for altering the functionality provided by the device or the mechanisms by which those functions are enabled.
  • The present disclosure overcomes or ameliorates at least one of the disadvantages of the prior art, and provides a useful alternative.
  • SUMMARY OF THE DISCLOSURE
  • In a first broad aspect, the present disclosure provides a user interface method, comprising the steps of:
  • i. defining an enablement protocol for a function; and
    ii. recording and retaining said enablement protocol of said function, such that a user enables said function by substantially reproducing said enablement protocol in the absence of spatial or temporal indication of at least a portion of said enablement protocol.
  • In one embodiment, defining the enablement protocol for a function comprises user-engagement with an interface, which comprises rises one or more touches or near touches of a touch screen. However, it will be appreciated that a user may also define an enablement protocol by engaging other types or configurations of interfaces, for example a rectangular frame surrounding a space wherein adjacent borders of the frame include optical emitters and the opposing borders include photo-detectors, i.e. essentially a touch-sensitive screen apparatus minus the actual touch surface. Alternatively, or additionally, the enablement protocol may comprise voice recognition of pre-defined commands (audible instructions), or the tracking of eye movements, or some other type of gesture of one or more parts of the human body. Whatever form a user's enablement protocol takes, the user will preferably define or record that protocol in the same form. For example if the enablement protocol is a sequence of touches with a particular rhythm at the top left corner of a touch screen, the user will usually define or record that protocol by actually performing the sequence of touches. Alternatively the user could define or record the protocol by other means, such as a spoken or typed description.
  • In a second aspect, the present disclosure provides user interface method, comprising the steps of:
  • i. defining an enablement protocol for a function of an electronic device, said enablement protocol comprising at least one user-defined touch on a touch-sensitive area operatively associated with said electronic device; and
    ii. recording and retaining said enablement protocol of said function, such that a user enables said function by substantially reproducing said enablement protocol in the absence of spatial or temporal indication of at least a portion of said enablement protocol.
  • It will be appreciated that the enablement protocol which comprises a plurality of spatial and temporal data points (i.e. touches and/or gestures) may be selected from one or more of the following non-exhaustive list of spatial or temporal indications: a simple touch, movement of touch, direction of touch, timing or rhythm between touches, force of touch, number of sequential or simultaneous touches, position of touch or shape or change of shape of touch, etc.
  • According to a third aspect the disclosure provides a method for user-defined enablement of a function through an interface, said method comprising the steps of:
  • i. recording an enablement protocol comprising at least one user-defined touch of said interface;
    ii. assigning said function to said enablement protocol; and
    iii. executing or enabling said function when one or more touches received on or near said interface substantially coincide with the recorded enablement protocol, wherein said interface is without spatial or temporal indication of where at least a portion of said enablement protocol is to be received to execute or enable said function.
  • It will be appreciated that the function can be enabled by one or more touches, and that the user is required to provide the enablement protocol without spatial or temporal indication of where the enablement protocol should be received to execute or enable the function. In one aspect, the enablement protocol may comprise a plurality of user-defined touches of the interface, and at least one spatial or temporal indication of the plurality of the user-defined touches of the enablement protocol is not shown or displayed on the interface. In such embodiments, for example, the enablement protocol may comprise a swipe of a touch screen with a 1 second dwell time at the conclusion of the swipe. The screen may only show where the swipe should be received and not indicate that the user must hold their finger down at the end of the swipe for 1 second. In this case only one of the two spatial/temporal indications is shown to the user.
  • In another somewhat related aspect, the function may be enabled when the user input meets a predefined number of spatial or temporal indications. For example, the enablement protocol may comprise a partial swipe of a touch screen, and then a 1 second dwell time, and then a further partial swipe, i.e. swipe+dwell+swipe. The function may be enabled if only two out of these three inputs are received, e.g. swipe+dwell, or swipe+swipe, or dwell+swipe.
  • According to a fourth aspect the present disclosure provides a method for enablement of a function through an interface, said method comprising the steps of:
  • i. recording an enablement protocol substantially according to a pre-defined enablement protocol, wherein said enablement protocol includes at least one user-defined touch of said interface;
    ii. assigning said function to the recorded enablement protocol; and
    iii. executing or enabling said function when one or more touches received on or near said interface substantially coincide with said recorded enablement protocol, wherein said interface is without spatial or temporal indication of where at least a portion of said enablement protocol is to be received to execute or enable said function.
  • In this aspect, the user may be presented with a choice of suggested enablement protocols from which they can choose and substantially mimic, thereby recording their own specific and unique enablement protocol. The user can then assign a function to their unique recorded enablement protocol and execute or enable the function when they input an enablement protocol that substantially coincides with the recorded enablement protocol, wherein the interface is without spatial or temporal indication of where at least a portion of the enablement protocol is to be received to execute or enable the function.
  • According to a fifth aspect the present disclosure provides a method for userdefined enablement of a function through an interface, said method comprising the steps of:
  • i. recording an enablement protocol comprising one or more user-defined touches of said interface; and
    ii. assigning said function to said enablement protocol, wherein in response to a user's request for execution or enablement of said function the user is prompted to perform said enablement protocol, wherein said function is executed or enabled when one or more touches received on or near said interface substantially coincide with the recorded enablement protocol, wherein said interface is without spatial or temporal indication of where at least a portion of said enablement protocol is to be received to execute or enable said function.
  • In this aspect the enablement protocol effectively acts as a ‘gatekeeper’ or security password control over a particular function. Essentially, the user records an enablement protocol, assigns a function to the enablement protocol and then when wishing to execute or enable the function is prompted to input their unique enablement protocol which, if coinciding with the recorded enablement protocol, allows the function to be executed or enabled.
  • Prior art devices typically require touches or near touches within one or more pre-existing or specified areas to enable a particular pre-defined function. For instance, a simple keypad on a touch display provides alphanumeric material at specific predefined locations, and touches or near touches at these locations must be accomplished in a specific sequence to enable the function, e.g. to activate a device containing the touch display. However, the user-defined interface method according to the present disclosure has no such spatial or temporal indication to a user of where touches or near touches should be received to enable the function, thereby allowing a user to have complete freedom in regard to defining how the functionality of an electronic device associated with the interface is enabled by touch.
  • To explain, a simple embodiment may involve the activation of a personal electronic device such as PDA or mobile phone. Normally a security code (optionally user-defined) is required to enable the device. This security code is normally a series of alphanumeric keystrokes either on a keypad separate from the display or on the display itself.
  • In one embodiment the present disclosure envisages no such spatial or temporal indication to a user. Rather a user can define a security code or ‘enablement protocol’ on the device or indeed on an apparatus separate from but operatively associated with the device by defining one or more touches to enable the aforementioned function. One such ‘touch’ may be for instance the touch of a thumb in the upper right hand corner of the screen or other touch sensitive area, or both. Alternatively, several touches at user-defined locations on the screen may enable the device. In another embodiment, a sequence of touches either at the user-defined locations or indeed anywhere on the screen can enable the functionality required. The timing or protocol (temporal sequence) of these touches must then be recreated to enable said function. In more complex arrangements both the spatial and temporal nature of the touches are examined to ensure that they meet the user-defined enablement protocol. In further arrangements, different types of touching bodies could enable different functions by performing otherwise identical enablement protocols. For example a swipe down one side of a touch screen with a stylus enables function A, while the same swipe with a finger enables function B. This embodiment expands the range of functions that can be performed with a limited range of gestures, or alternatively provides an additional security feature.
  • It can be seen then that any ‘touch screen’ input device can have a series of functions which may only be enabled by the user who defined the aforementioned ‘enablement protocol’, since there is no spatial or temporal indication to assist another party who may wish to enable the aforementioned functions.
  • In another embodiment, enablement protocols may be linked to the profiles of multiple users such that a single device may be enabled according to the profile of a user who defined a specific enablement protocol.
  • As an example, user A defines touches in the upper and lower left hand corners of a display to initiate a device. When touches are provided in those upper and lower corners the device is activated according to user A's profile. User B may have an entirely separate profile and may wish the device to be set up according to their specific needs. They therefore define an enablement protocol different from that of user A; not only will other parties not be able to access the device without knowing the enablement protocol of either party, but once user B initiates their particular enablement protocol the device will be activated according to their specific profile.
  • Indeed even if another person knew some aspects of an enablement protocol, such as the locations to be touched or the particular gesture, other aspects such as the rhythm, pattern or sequence of the touches or their touch signature e.g. finger shape, will be different and therefore they will not be able to replicate the enablement protocol.
  • The disclosure as described is particularly useful for users with some form of physical handicap that may prevent them from using conventional touch screen devices. By providing complete freedom to an individual user to define enablement protocols for the various functions of a device, they may define the touches in terms of spatial and temporal parameters that they can easily replicate.
  • Since there is no spatial or temporal indication of the enablement protocol to a user, even a simple single touch or near touch at a specific location may be sufficient as an enablement protocol with enhanced security. This is to be distinguished from known fingerprint readers, which require a direct touch within a specific area.
  • While enablement by a single touch or near touch is possible with the present disclosure it is generally preferred that two or more touches or near touches be used to define the enablement protocol. Touches and near touches may be detected in several ways, including a touch-down, movement of touch, direction of touch, timing or rhythm (protocol) between touches i.e. temporal sequence, force of touch, number of sequential or simultaneous touches, position of touch, and shape or change of shape of touch approaching, contacting or leaving the touch sensitive area.
  • One or more of these detection mechanisms may be used to complete the enablement protocol. For instance, if the touch location is correct but the shape of the touch is incorrect the protocol will not be satisfied (i.e. there is not substantial coincidence between the received and the recorded protocol) and the associated function will not be enabled. Similarly if the timing or rhythm between touches is correct but the force applied by those touches is incorrect, once again the protocol will not be satisfied. It will be appreciated that although only selected touch technologies can directly measure the force of an applied touch, certain other touch technologies, especially those with high spatial resolution, can indirectly measure touch force from deformation of the touch object.
  • Touches or near touches may be detected by any known touch sensing technology including analogue resistive, surface capacitive, surface acoustic wave, projected capacitive, optical and bending wave technologies, bearing in mind that some touch technologies are better suited than others for distinguishing multiple simultaneous touches, should these be part of an enablement protocol. Enablement protocols that include touch shape or change of shape will generally require high spatial resolution, such as may be offered by optical touch detection technologies disclosed in U.S. Pat. Nos. 5,914,709 and 7,352,940. However any known touch technology, including those mentioned above, can be applied with the present disclosure, with enablement protocols defined within the limitations of that technology. Certain touch technologies, e.g. resistive and bending wave, require a direct touch for activation, whereas other touch technologies, such as optical touch where a sheet of light is established in front of a display, can also sense an object hovering close to the display. The principles of the present disclosure apply whether the touch sensitive device is activated by a touch or a near touch of an interface.
  • In a further aspect the present disclosure provides a user interface device comprising: a touch sensitive area adapted to detect touches; a display device to display one or more functions; and a storage medium including a user-defined enablement protocol for enabling functions operatively associated with said display device, said enablement protocol including one or more user-defined touches on or near said touchsensitive area, there being no spatial or temporal indication to said user of at least a portion of said enablement protocol.
  • While it is preferred that the enablement protocol be performed on the display device, this is not absolutely necessary. One could envisage the enablement protocol being applied to a touch-sensitive area or ‘protocol apparatus’ separate from the display device. The user would simply follow the defined enablement protocol and the ‘protocol apparatus’ then communicates with the display device e.g. PDA or computer to enable the function held by that device. Indeed the function itself may be held by a third device, such that instigation of the enablement protocol, the display device and the various functions to be applied to the display device are all held separately.
  • In another aspect of the present disclosure there is provided a user-defined interface method, comprising the steps of:
  • i. defining for an electronic device an enablement protocol for a function of said device whereby said enablement protocol comprises one or more user-defined touches,
    ii. retaining said enablement protocol of said function, and
    iii. enabling said function by said one or more user-defined touches wherein, if a user is provided with spatial or temporal indication for a portion of said enablement protocol, said enablement protocol additionally requires detection of an additional touch parameter.
  • Preferably the additional parameter required for the enablement protocol would include one or more of touch movement, direction of touch, timing or rhythm between touches i.e. temporal sequence, force of touch, number of sequential or simultaneous touches, shape and/or change of shape of touch including in different planes e.g. vertical and horizontal, on approach or withdrawal.
  • To explain, some devices may have a spatial or temporal indication for a portion of an enablement protocol. For instance, a mobile phone may have locations shown on the display that must be pressed in a certain sequence for partial enablement of the protocol. One mechanism to complete the enablement protocol could be temporal i.e. the locations must be pressed in the precise timing and sequence required. In one alternative, if spatial and temporal data are provided e.g. by flashing of the locations, the size and shape of the touch may be detected to confirm enablement by an authorised user. This detection of touch size and shape may be accomplished not only in the plane of the display or touch sensitive area but also in planes above the display e.g. the shape of the touch approaching or leaving the touch sensitive area, if the touch sensor can detect near touches. This touch signature is virtually impossible to replicate by another user and therefore provides a secure individual enablement protocol. It will be clear to a person skilled in the art that there are various other options available to confirm enablement.
  • Accordingly it can be seen that according to several aspects of the present disclosure there is significant flexibility and security as compared with prior art systems. Unless the context clearly requires otherwise, throughout the description and the claims, the words ‘comprise’, ‘comprising’, and the like are to be construed in an inclusive sense as opposed to an exclusive or exhaustive sense; that is to say, in the sense of ‘including, but not limited to’.
  • The skilled person will appreciate that the term ‘coincide’, when used herein and when referring to a touch input being compared with a recorded or retained touch input, is meant to define that the inputs should substantially correspond. In one embodiment, as discussed above, the coincidence is substantial overlap in overall touch area, and in another embodiment may also include one or more other parameters, such as dwell time.
  • In a further embodiment ‘coincide’ may refer to a predetermined number of matches between the spatial and/or temporal data points of the touch or near touch and those of the enablement protocol.
  • It will be appreciated that when referring to ‘touch’ herein the term is also intended to encompass within its scope the term ‘near touch’.
  • In one embodiment the touch-sensitive area is larger than the area of the display device and contains the display device within its borders. Generally, conventional touch-sensitive input devices include a display device that is at least as large as the touch-sensitive area. In one embodiment the touch-sensitive area is larger than the display itself and includes the display device within its borders. This is highly unusual but provides advantages over conventional systems, for example the opportunity of producing a smaller display with consequent power reduction. Also, by including the touch-sensitive area outside the display, the display itself is not obscured while input is applied via the touch-sensitive area. The touch portions/locations may be located on a first face of the touch-sensitive area, generally the display side, and/or a second opposite face of the touch-sensitive area. In one particular embodiment the touch portions/locations for enablement of the functions are arranged on the first and second opposite faces of the touch-sensitive area in such a way that they can be simultaneously operated by the thumb and fingers of a user.
  • BRIEF DESCRIPTION OF DRAWINGS
  • Preferred embodiments of the disclosure will now be described, by way of example only, with reference to the accompanying drawings in which:
  • FIGS. 1 and 2 illustrate conventional user interface devices according to the prior art;
  • FIGS. 3A-3D illustrate a first embodiment of the present disclosure;
  • FIGS. 4A-4C illustrate a second embodiment of the present disclosure;
  • FIGS. 5A-5C illustrate a third embodiment of the present disclosure. FIGS. 6 and 7 illustrate touch signature recognition according to a fourth embodiment of the present disclosure;
  • FIG. 8A-8C illustrate a fifth embodiment of the present disclosure; and
  • FIGS. 9A-9C illustrate a sixth embodiment of the present disclosure.
  • DETAILED DESCRIPTION OF THE DISCLOSURE
  • FIGS. 1 and 2 provide illustrations of prior art functionality in conventional devices.
  • FIG. 1 shows a mobile phone 10 with an alphanumeric keypad 11 and a display 12. In many instances the mobile phone can only be enabled when a user-defined security code is entered via the alphanumeric keypad 11. Alternatively the alphanumeric keypad may be provided on a touch-sensitive display (i.e. a touch screen) but once again the conventional device comprises a pre-defined arrangement of the enablement protocol such that if an unauthorised user were aware of the code e.g. 123, they could enable the device simply by engaging the predefined locations for numerals 1, 2 and 3.
  • FIG. 2 is an illustration of an automatic teller machine 20 with a touch sensitive display 21. This technology is similar to the mobile phone in FIG. 1, and in this instance various functions and spatial locations for enabling those functions are clearly defined on the display. The display includes indicia 22 describing the functions e.g. withdrawal, deposit and account balance, and the spatial locations for those functions are defined by the touch portions 23. Accordingly any user can engage the functions defined on the display 21 by simply following the spatial arrangement already predefined by the device.
  • The present disclosure seeks to overcome some of the difficulties associated with the prior art and provides greater flexibility and security as compared the with prior art systems. In certain embodiments the enablement protocol of the present disclosure acts to restrict the performing of a user-instigated function. That is, a user attempts to perform (whether by touch or otherwise) a function on an electronic device, and the function is only carried out if the user performs their defined enablement protocol required for that function. Generally this enablement protocol will be performed in response to an instruction, an indication or a request for user input from the electronic device, similar to a conventional ‘password’ dialog box. Each user of the device may have defined a single enablement protocol that enables all functions of the device, or they may have defined different enablement protocols for different functions. It will be appreciated that the particular function can be virtually anything, such as activating a device from an ‘off’ state or a sleep mode, logging into a computer system, or accessing a file. For example a computer system may provide different levels of access to different classes of user, with one class of users having full access to a given file, a second class of users having read-only access, and other users barred or restricted access.
  • In other embodiments that are applicable to functions instigated by touch input, the enablement protocol is embedded in the particular touch input, such as a gesture, that instigates the function. This instigating touch input may be predetermined by the electronic device or by software running on it (e.g. a ‘double tap’ gesture that opens a file), or it may be user-defined (e.g. tracing a person's initials to instigate a phone call to that person). In one example a user may attempt to open a file by double tapping on an icon, but the file will only be opened if the ‘details’ of the double tap gesture, such as the touch object size, the duration of each tap and the delay between the taps, substantially coincide with the pre-recorded gesture (user-defined protocol). In another example a user may assign one particular function to a swipe gesture when the gesture is performed with a finger, and a different function when the swipe is performed with a stylus.
  • Some specific embodiments will now be described, involving certain enablement protocols that activate an electronic device. However it will be appreciated that these particular enablement protocols could be associated with other functions of the electronic device.
  • FIGS. 3A-3D illustrate a first embodiment according to the present disclosure. FIG. 3A shows a particular device 30 with a touch screen or touch-sensitive area 31. A user has defined the location of a touch or near touch at a location 32 in accordance with an enablement protocol for activation of the device. However, it will be clear from the foregoing description that any function of a device 30 may similarly include an enablement protocol comprising one or more user defined touches or near touches 32. Importantly, the required touch location 32, i.e. the enablement protocol, is not shown on the device at all. The dotted line 33 indicates the location for illustrative purposes only, and there is no spatial or temporal indication whatsoever on the device 30 of where or how the touches should be applied to satisfy the user-defined enablement protocol. In fact there may be no display associated with the touch-sensitive area.
  • This is clearly a significant advantage over the prior art since it will be virtually impossible for an unauthorised user to determine the enablement protocol without any form of spatial or temporal indication.
  • In FIG. 3B a user commences application of the enablement protocol, which requires a thumb to be placed at the location 32. In FIG. 3C it can be seen that the user places their thumb 34 at the required location 32, thereby satisfying the enablement protocol that results in activation of the device.
  • In the embodiment shown in FIGS. 3A-3D the enablement protocol comprises a single touch at the user-defined location 32 on the touch-sensitive area 31 of the particular device 30. It will be clear, however, that the enablement protocol may comprise other forms of touch.
  • For instance, the embodiment shown in FIGS. 4A-4C displays a swipe by the user of their thumb to satisfy the enablement protocol. This embodiment again comprises a device 40 with a touch-sensitive area 41. The spatial location of the enablement protocol for a certain function is shown by dotted lines 42 again for illustrative purposes only. In this instance, the enablement protocol requires movement of the thumb along the arc 43. The enablement protocol may also include the timing or speed of the swipe from the lower right hand corner to the upper right hand corner, the contact dwell time at the start or end of the swipe, and the shape of the thumb on the surface of the touch-sensitive area. It will be appreciated that this offers an enhanced degree of security in that even if another person knew the correct gesture, say by observation, it would be extremely difficult for them to replicate all other aspects of the enablement protocol.
  • As shown in FIG. 4B the user holds the device, swipes their thumb 44 across the area 42 in the predefined fashion required by the enablement protocol i.e. speed, timing, area covered etc such that, as shown in FIG. 4C, the device is activated. In FIGS. 5A-5C a multiple touch enablement protocol is shown. In this embodiment a device 50 with a touch sensitive area 51 requires two touches at specific spatial locations to activate the device. The spatial location 52 requires a touch of a left thumb 54, and the spatial location 53 requires a swipe of a right thumb 55 along a linear path 56, Once again, similar to FIGS. 3A-3D and FIGS. 4A-4C, there is no outward indication either spatially or temporally of the enablement protocol. The screen may be completely blank and the enablement protocol is known only to the authorised user of the device 50. Once the touch at the location 52 and the swipe at the location 53 are accomplished in line with the protocol, the device 50 is activated. In an alternative embodiment that provides an additional security measure, the screen may display indicia suggesting where touch input is required to enable a protocol, however the indicia may be specifically configured to falsely indicate where the actual input is required to enable the function, thereby intentionally frustrating the attempts of an unauthorised person to access the electronic device.
  • FIGS. 6A-6D and 7A-7D illustrate how a touch is detected and how a characteristic touch signature of an individual may be determined.
  • In FIG. 6A a finger 60 is approaching the surface 61 of a touch screen. In FIG. 6B the finger first comes into contact with the surface and is therefore detected. At this point it can be seen that the area 62 of the surface covered by the finger is relatively small. As the user proceeds with the touch, greater pressure is applied to the surface thereby distorting the finger such that it covers a greater area 63 of the surface, as seen in FIG. 6C. In 6D the finger is withdrawn after a certain time that may itself be part of an enablement protocol.
  • If the resolution and sensitivity of the touch detection apparatus is sufficient, this change of shape of a finger as it approaches, touches and withdraws from the surface 61 can be detected and may form part of an enablement protocol. This is particularly useful since even if the precise spatial and temporal location of an enablement protocol is known by an unauthorised user, the protocol may only be satisfied if the specific shape and change of shape of the touch matches that of the authorised user.
  • FIG. 7 gives a further illustration. FIGS. 7A-7D show a finger movement or swipe on a touch-sensitive surface 70. In FIG. 7A a finger 71 approaches the surface 70, and in FIG. 7B the finger first contacts the surface. The finger then slides across the surface to the point shown in FIG. 7C and is then removed as shown in FIG. 7D. The distance 72 covered by the finger may be detected and form part of an enablement protocol. Similarly, the change in shape of the finger as it slides may also be detected. Accordingly it can be seen that virtually any movement, shape or change of shape of a touch object can be detected and form part of an enablement protocol for enabling a particular function on a device. Unless an unauthorised user can match with sufficient precision the touch signature of the user who defined the enablement protocol, it is virtually impossible for an unauthorised user to match the enablement protocol and thereby enable the functions of the particular device. Clearly touching with an implement such as a pen will not satisfy an enablement protocol defined by a finger or thumb.
  • FIGS. 8A-8C and 9A-9C disclose a further embodiment of the present disclosure displaying the powerful flexibility and security available with user-defined enablement protocols.
  • In FIGS. 8A-8C it can be seen that even if spatial or temporal indicators are provided on the device, an enablement protocol will still not be satisfied if it includes detection of additional touch or near touch parameters, e.g. shape or change of shape of a touch object.
  • In FIG. 8A a device 80 with a touch sensitive display 81 shows a spatial indicator 82 forming part of an enablement protocol. In other words, the enablement protocol for a specific function of this device requires a single touch at the location of the indicator 82. However in this embodiment the shape 83 of the touch object, of which there is no indication given on the display, also forms part of the enablement protocol as discussed above in relation to FIGS. 6A-6D. With certain touch detection systems e.g. optical systems, the resolution is sufficiently high to measure shape and/or change of shape on approach, at touch-down and at withdrawal of touch. In FIG. 8B a user holds the device 80 and commences the enablement protocol by placing the pad of their thumb 84 on the indicator 82. This is accomplished in FIG. 8C and the function is enabled. Importantly, even though the spatial indicator 82 gives a user some indication of the required enablement protocol, if a touch object does not match the required shape 83, the device will not be enabled. For example touching the indicator 82 with a stylus or fingertip will not satisfy the enablement protocol.
  • A further embodiment can be seen in FIGS. 9A-9C. In this embodiment a device 90 includes a touch sensitive display 91 showing two spatial and/or temporal indicators 92 and 93. In this instance one or both of the indicators 92 and 93 may be flashing to indicate the desired temporal sequence between touches at the indicators 92 and 93. However, as indicated by the dotted lines 94 and 95 (illustrative only, not shown on the display) the enablement protocol requires a simple touch at the location of the indicator 92, but not at the location of the indicator 93. Rather, the protocol requires a swipe along an arc 96 that begins in the lower right hand corner and terminates at the indicator 93. Accordingly, for an authorised user to enable a specific function, the device is held such that that portion of the enablement protocol requiring a left thumb touch 97 at the location of the indicator 92 is satisfied, and the right thumb 98 proceeds from the lower right hand corner up to the location of the indicator 93 within the time specified by the enablement protocol. When both the spatial and temporal sequence as well as shape, change of shape or any other touch or near touch parameters are satisfied, the enablement protocol enables the function of the device 90 and the device is, in this embodiment, activated. There is virtually no prospect, even when the spatial and temporal indicators are provided, that an unauthorised user could satisfy all aspects of the user-defined protocol.
  • In another embodiment, an unauthorised person attempting to access the electronic device could attempt to replicate the enablement protocol of an authorised person by carefully inspecting the touch surface for residual finger marks. In cases such as these the enablement protocol may be selected to avoid indicating, from an inspection of finger marks alone, the actual enablement protocol. For example, a circular enablement protocol would not indicate the direction which the user traced out the circle. Similarly, the direction of a swipe would be difficult to interpret from only residual finger marks, and enablement protocols with temporal features such as the rhythm of a touch sequence could not be ascertained from residual finger marks.
  • It will be clear to persons skilled in the art that the proposed method and device has significant flexibility and security as compared with the prior art. In addition, since all functions of a device may be linked to user-defined enablement protocols, operation of the device is completely user-dependent. This is also particularly helpful for users of such interface devices who cannot use conventional touch techniques, e.g. handicapped persons.
  • Although the disclosure has been described with reference to specific embodiments, it will be appreciated by those skilled in the art that the disclosure may be embodied in many other forms.

Claims (15)

1. A user interface method, comprising the steps of:
i. defining an enablement protocol for a function; and
ii. recording and retaining said enablement protocol of said function, such that a user enables said id function by substantially reproducing said enablement protocol in the absence of spatial or temporal indication of at least a portion of said enablement protocol.
2. A method according to claim 1, wherein said enablement protocol comprises a single touch at a user-defined location on a touch-sensitive area of an interface.
3. A method according to claim 1, wherein said enablement protocol comprises multiple sequential or simultaneous touches at user-defined locations on a touchsensitive area of an interface.
4. A user interface method, comprising the steps of:
i. defining an enablement protocol for a function of an electronic device, said enablement protocol comprising at least one user-defined touch on a touch-sensitive area operatively associated with said electronic device; and
ii. recording and retaining said enablement protocol of said function, such that a user enables said function by substantially reproducing said enablement protocol in the absence of spatial or temporal indication of at least a portion of said enablement protocol.
5. A user interface method according to claim 1, wherein a user enables said function by substantially reproducing said enablement protocol in the absence of any spatial or temporal indication of said enablement protocol.
6. A method for user-defined enablement of a function through an interface, said method comprising the steps of:
i. recording an enablement protocol comprising at least one user-defined touch of said interface;
ii. assigning said function to said enablement protocol; and
iii. executing or enabling said function when one or more touches received on or near said interface substantially coincide with the recorded enablement protocol, wherein said interface is without spatial or temporal indication of where at least a portion of said enablement protocol is to be received to execute or enable said function.
7. A method for enablement of a function through an interface, said method comprising the steps of:
i. recording an enablement protocol substantially according to a pre-defined enablement protocol, wherein said enablement protocol includes at least one user-defined touch of said interface;
ii. assigning said function to the recorded enablement protocol; and
iii. executing or enabling said function when one or more touches received on or near said interface substantially coincide with said recorded enablement protocol, wherein said interface is without spatial or temporal indication of where at least a portion of said enablement protocol is to be received to execute or enable said function.
8. A method for user-defined enablement of a function through an interface, said method comprising the steps of:
i. recording an enablement protocol comprising one or more user-defined touches of said interface; and
ii. assigning said function to said enablement protocol, wherein in response to a user's request for execution or enablement of said function the user is prompted to perform said enablement protocol, wherein said function is executed or enabled when one or more touches received on or near said interface substantially coincide with the recorded enablement protocol, wherein said interface is without spatial or temporal indication of where at least a portion of said enablement protocol is to be received to execute or enable said function.
9. A user interface method according to claim 2, wherein said enablement protocol comprises one or more of a simple touch, movement of touch, direction of touch, timing or rhythm between touches, force of touch, number of sequential or simultaneous touches, position of touch or shape or change of shape of touch.
10. A method according to claim 6, wherein said interface is without any spatial or temporal indication of where said enablement protocol is to be received to execute or enable said function.
11. A method according to claim 2, wherein said enablement protocol comprises a plurality of user-defined touches, wherein said function is enabled when user input meets a predefined number of spatial or temporal indications of said enablement protocol.
12. A user interface device, comprising:
a touch sensitive area adapted to detect touches;
a display device to display one or more functions; and
a storage medium including a user-defined enablement protocol for enabling functions operatively associated with said display device, said enablement protocol including one or more user-defined touches on or near said touch-sensitive area, there being no spatial or temporal indication to said user of at least a portion of said enablement protocol.
13. A user-defined interface method, comprising the steps of:
i. defining for an electronic device an enablement protocol for a function of said device whereby said enablement protocol comprises one or more userdefined touches,
ii. retaining said enablement protocol of said function, and
iii. enabling said function by said one or more user-defined touches wherein, if a user is provided with spatial or temporal indication for a portion of said enablement protocol, said enablement protocol additionally requires detection of an additional touch parameter.
14. A user interface method according to claim 13, wherein said additional parameter includes one or more of touch movement, direction of touch, timing or rhythm between touches, force of touch, number of sequential or simultaneous touches, and shape or change of shape of touch.
15. A user interface method according to claim 14, wherein said additional parameter includes the change of shape of a touch in the vertical and/or horizontal planes on approach or withdrawal of said touch.
US12/118,047 2007-05-11 2008-05-09 User-Defined Enablement Protocol Abandoned US20080278455A1 (en)

Priority Applications (3)

Application Number Priority Date Filing Date Title
US13/473,961 US9280279B2 (en) 2007-05-11 2012-05-17 User-defined gesture enablement protocols for touch input devices
US15/056,015 US9696847B2 (en) 2007-05-11 2016-02-29 User-defined gesture enablement protocols for touch input device
US15/626,585 US20170285853A1 (en) 2007-05-11 2017-06-19 User-defined gesture enablement protocols for touch input device

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
AU2007/902,517 2007-05-11
AU2007/902,519 2007-05-11
AU2007902519A AU2007902519A0 (en) 2007-05-11 Gestural input
AU2007902517A AU2007902517A0 (en) 2007-05-11 Input device

Related Child Applications (1)

Application Number Title Priority Date Filing Date
US13/473,961 Continuation US9280279B2 (en) 2007-05-11 2012-05-17 User-defined gesture enablement protocols for touch input devices

Publications (1)

Publication Number Publication Date
US20080278455A1 true US20080278455A1 (en) 2008-11-13

Family

ID=39969088

Family Applications (4)

Application Number Title Priority Date Filing Date
US12/118,047 Abandoned US20080278455A1 (en) 2007-05-11 2008-05-09 User-Defined Enablement Protocol
US13/473,961 Active US9280279B2 (en) 2007-05-11 2012-05-17 User-defined gesture enablement protocols for touch input devices
US15/056,015 Active - Reinstated US9696847B2 (en) 2007-05-11 2016-02-29 User-defined gesture enablement protocols for touch input device
US15/626,585 Abandoned US20170285853A1 (en) 2007-05-11 2017-06-19 User-defined gesture enablement protocols for touch input device

Family Applications After (3)

Application Number Title Priority Date Filing Date
US13/473,961 Active US9280279B2 (en) 2007-05-11 2012-05-17 User-defined gesture enablement protocols for touch input devices
US15/056,015 Active - Reinstated US9696847B2 (en) 2007-05-11 2016-02-29 User-defined gesture enablement protocols for touch input device
US15/626,585 Abandoned US20170285853A1 (en) 2007-05-11 2017-06-19 User-defined gesture enablement protocols for touch input device

Country Status (3)

Country Link
US (4) US20080278455A1 (en)
TW (1) TW200907742A (en)
WO (1) WO2008138030A1 (en)

Cited By (122)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7593000B1 (en) * 2008-05-17 2009-09-22 David H. Chin Touch-based authentication of a mobile device through user generated pattern creation
US20090249476A1 (en) * 2008-03-26 2009-10-01 Lisa Anne Seacat Using Multi-Touch Gestures to Protect Sensitive Content Using a Completely Automated Public Turing Test to Tell Computers and Humans Apart (CAPTCHA)
US20090289916A1 (en) * 2008-05-23 2009-11-26 Hon Hai Precision Industry Co., Ltd. Electronic device and method for switching between locked state and unlocked state
US20100013780A1 (en) * 2008-07-17 2010-01-21 Sony Corporation Information processing device, information processing method, and information processing program
US20100083108A1 (en) * 2008-09-26 2010-04-01 Research In Motion Limited Touch-screen device having soft escape key
US20100162182A1 (en) * 2008-12-23 2010-06-24 Samsung Electronics Co., Ltd. Method and apparatus for unlocking electronic appliance
US20100201641A1 (en) * 2007-08-13 2010-08-12 Hideaki Tetsuhashi Contact type input device, contact type input method, and program
US20100222049A1 (en) * 2009-03-02 2010-09-02 Lg Electronics Inc. Method for releasing a locking in mobile terminal and mobile terminal using the same
US20100255885A1 (en) * 2009-04-07 2010-10-07 Samsung Electronics Co., Ltd. Input device and method for mobile terminal
US20110006996A1 (en) * 2009-07-08 2011-01-13 Smith Nathan J Private data entry
US20110019105A1 (en) * 2009-07-27 2011-01-27 Echostar Technologies L.L.C. Verification of symbols received through a touchpad of a remote control device in an electronic system to allow access to system functions
US20110025620A1 (en) * 2008-01-11 2011-02-03 Opdi Technologies A/S Touch-sensitive device
US20110032377A1 (en) * 2009-08-07 2011-02-10 Samsung Electronics Co., Ltd. Digital photographing apparatus, method of controlling the same, and recording medium storing program to execute the method
US20110041102A1 (en) * 2009-08-11 2011-02-17 Jong Hwan Kim Mobile terminal and method for controlling the same
US20110099473A1 (en) * 2009-10-23 2011-04-28 Samsung Electronics Co., Ltd. Input signal processing device for portable device and method of the same
US20110106446A1 (en) * 2008-05-26 2011-05-05 Volkswagen Ag Display method for a display system, display system and operating method for a navigation system of a vehicle
US20110161888A1 (en) * 2009-12-28 2011-06-30 Sony Corporation Operation direction determination apparatus, remote operating system, operation direction determination method and program
US20110162066A1 (en) * 2009-12-29 2011-06-30 Bizmodeline Co., Ltd. Password processing method and apparatus
US20110163967A1 (en) * 2010-01-06 2011-07-07 Imran Chaudhri Device, Method, and Graphical User Interface for Changing Pages in an Electronic Document
US20110185321A1 (en) * 2010-01-26 2011-07-28 Jay Christopher Capela Device, Method, and Graphical User Interface for Precise Positioning of Objects
US20110185300A1 (en) * 2010-01-28 2011-07-28 Microsoft Corporation Brush, carbon-copy, and fill gestures
US20110187497A1 (en) * 2008-05-17 2011-08-04 David H Chin Comparison of an applied gesture on a touch screen of a mobile device with a remotely stored security gesture
WO2011094045A2 (en) * 2010-01-28 2011-08-04 Microsoft Corporation Copy and staple gestures
US20110191719A1 (en) * 2010-02-04 2011-08-04 Microsoft Corporation Cut, Punch-Out, and Rip Gestures
US20110205176A1 (en) * 2008-11-05 2011-08-25 Takashi Okada Portable electronic device, and power saving method and power saving program for the same
US20110209103A1 (en) * 2010-02-25 2011-08-25 Hinckley Kenneth P Multi-screen hold and drag gesture
US20110265045A1 (en) * 2010-04-26 2011-10-27 Via Technologies, Inc. Electronic system and method for operating touch screen thereof
US20110282785A1 (en) * 2008-05-17 2011-11-17 Chin David H Gesture based authentication for wireless payment by a mobile electronic device
US20110283241A1 (en) * 2010-05-14 2011-11-17 Google Inc. Touch Gesture Actions From A Device's Lock Screen
US20110289423A1 (en) * 2010-05-24 2011-11-24 Samsung Electronics Co., Ltd. Method and apparatus for controlling objects of a user interface
US20120096376A1 (en) * 2010-10-14 2012-04-19 Canon Kabushiki Kaisha Display control apparatus, display control method, and storage medium
US20120133484A1 (en) * 2010-11-29 2012-05-31 Research In Motion Limited Multiple-input device lock and unlock
US20120196573A1 (en) * 2009-10-09 2012-08-02 Nec Corporation Mobile device, and security lock control method and program thereof
US8239785B2 (en) 2010-01-27 2012-08-07 Microsoft Corporation Edge gestures
US20120252410A1 (en) * 2011-03-28 2012-10-04 Htc Corporation Systems and Methods for Gesture Lock Obfuscation
US20120274583A1 (en) * 2011-02-08 2012-11-01 Ammon Haggerty Multimodal Touchscreen Interaction Apparatuses, Methods and Systems
US20130063009A1 (en) * 2011-09-13 2013-03-14 Nunoerin, Llc Interactive furniture device
US20130086673A1 (en) * 2011-09-29 2013-04-04 Cisco Technology, Inc. Techniques for securely unlocking a touch screen user device
US20130174083A1 (en) * 2011-12-28 2013-07-04 Acer Incorporated Electronic device and method of controlling the same
US8539386B2 (en) 2010-01-26 2013-09-17 Apple Inc. Device, method, and graphical user interface for selecting and moving objects
US8539384B2 (en) 2010-02-25 2013-09-17 Microsoft Corporation Multi-screen pinch and expand gestures
US20130278540A1 (en) * 2012-04-20 2013-10-24 Esat Yilmaz Inter Touch Sensor Communications
US20130298079A1 (en) * 2012-05-02 2013-11-07 Pantech Co., Ltd. Apparatus and method for unlocking an electronic device
US8587422B2 (en) 2010-03-31 2013-11-19 Tk Holdings, Inc. Occupant sensing system
US8612884B2 (en) 2010-01-26 2013-12-17 Apple Inc. Device, method, and graphical user interface for resizing objects
US20140002397A1 (en) * 2012-06-29 2014-01-02 Samsung Electronics Co., Ltd. Method and device for handling event invocation using a stylus pen
US20140006965A1 (en) * 2012-07-02 2014-01-02 Beijing Xiaomi Technology Co., Ltd. Method of entering a user interface in a device and the device thereof
US20140026105A1 (en) * 2012-07-18 2014-01-23 Research In Motion Limited Method and Apparatus Pertaining to a Gesture-Controlled Snooze Instruction
US20140028598A1 (en) * 2012-07-30 2014-01-30 Samsung Electronics Co., Ltd Apparatus and method for controlling data transmission in terminal
US20140033140A1 (en) * 2012-07-11 2014-01-30 Guang Dong Oppo Mobile Telecommunications Corp., Ltd. Quick access function setting method for a touch control device
US8707174B2 (en) 2010-02-25 2014-04-22 Microsoft Corporation Multi-screen hold and page-flip gesture
US8725230B2 (en) 2010-04-02 2014-05-13 Tk Holdings Inc. Steering wheel with hand sensors
EP2731002A1 (en) * 2012-11-09 2014-05-14 Thales Method for securing a command on a viewing device with tactile surface and related system
US20140137234A1 (en) * 2008-05-17 2014-05-15 David H. Chin Mobile device authentication through touch-based gestures
US20140145990A1 (en) * 2012-11-29 2014-05-29 Egalax_Empia Technology Inc. Method for unlocking touch screen, electronic device thereof, and recording medium thereof
US8751970B2 (en) 2010-02-25 2014-06-10 Microsoft Corporation Multi-screen synchronous slide gesture
US8766928B2 (en) 2009-09-25 2014-07-01 Apple Inc. Device, method, and graphical user interface for manipulating user interface objects
US8780069B2 (en) 2009-09-25 2014-07-15 Apple Inc. Device, method, and graphical user interface for manipulating user interface objects
US8799827B2 (en) 2010-02-19 2014-08-05 Microsoft Corporation Page manipulations using on and off-screen gestures
US8799826B2 (en) 2009-09-25 2014-08-05 Apple Inc. Device, method, and graphical user interface for moving a calendar entry in a calendar application
US8836648B2 (en) 2009-05-27 2014-09-16 Microsoft Corporation Touch pull-in gesture
US20140267064A1 (en) * 2013-03-13 2014-09-18 Htc Corporation Unlock Method and Mobile Device Using the Same
US20140298672A1 (en) * 2012-09-27 2014-10-09 Analog Devices Technology Locking and unlocking of contacless gesture-based user interface of device having contactless gesture detection system
US8863016B2 (en) 2009-09-22 2014-10-14 Apple Inc. Device, method, and graphical user interface for manipulating user interface objects
US20140347302A1 (en) * 2013-05-24 2014-11-27 Boe Technology Group Co., Ltd. Unlocking method and device for touch screen terminal and touch screen terminal
US8972879B2 (en) 2010-07-30 2015-03-03 Apple Inc. Device, method, and graphical user interface for reordering the front-to-back positions of objects
US20150066762A1 (en) * 2013-08-28 2015-03-05 Geoffrey W. Chatterton Authentication system
US9007190B2 (en) 2010-03-31 2015-04-14 Tk Holdings Inc. Steering wheel sensors
EP2835732A3 (en) * 2013-08-06 2015-04-22 Samsung Electronics Co., Ltd Electronic device provided with touch screen and operating method thereof
US9052820B2 (en) 2011-05-27 2015-06-09 Microsoft Technology Licensing, Llc Multi-application environment
US9075522B2 (en) 2010-02-25 2015-07-07 Microsoft Technology Licensing, Llc Multi-screen bookmark hold gesture
US9081494B2 (en) 2010-07-30 2015-07-14 Apple Inc. Device, method, and graphical user interface for copying formatting attributes
US20150199553A1 (en) * 2014-01-13 2015-07-16 Samsung Electronics Co., Ltd. Method for recognizing fingerprint and mobile terminal supporting the same
US9098182B2 (en) 2010-07-30 2015-08-04 Apple Inc. Device, method, and graphical user interface for copying user interface objects between content regions
US9104440B2 (en) 2011-05-27 2015-08-11 Microsoft Technology Licensing, Llc Multi-application environment
EP2919108A1 (en) * 2014-03-13 2015-09-16 LG Electronics, Inc. Mobile terminal and method of controlling the same
US9158445B2 (en) 2011-05-27 2015-10-13 Microsoft Technology Licensing, Llc Managing an immersive interface in a multi-application immersive environment
US20150317076A1 (en) * 2012-12-11 2015-11-05 University Of Washington Through Its Center For Commercialization Use of hand posture to improve text entry
WO2015149101A3 (en) * 2014-04-01 2015-11-26 Linkilike Gmbh Method for comparing user identities of databases
US9229918B2 (en) 2010-12-23 2016-01-05 Microsoft Technology Licensing, Llc Presenting an application change through a tile
EP2977884A1 (en) * 2014-07-22 2016-01-27 LG Electronics Inc. Mobile terminal and method for controlling the same
US20160026380A1 (en) * 2014-07-25 2016-01-28 Hannstar Display (Nanjing) Corporation Hand-held electronic device, computer-executed method and touch-sensing cover
US20160034172A1 (en) * 2014-07-30 2016-02-04 Wistron Corporation Touch device and control method and method for determining unlocking thereof
US9261964B2 (en) 2005-12-30 2016-02-16 Microsoft Technology Licensing, Llc Unintentional touch rejection
US20160054827A1 (en) * 2014-08-21 2016-02-25 Echostar Technologies L.L.C. Determining handedness on multi-element capacitive devices
US9274682B2 (en) 2010-02-19 2016-03-01 Microsoft Technology Licensing, Llc Off-screen gestures to create on-screen input
US20160070465A1 (en) * 2014-09-08 2016-03-10 Lenovo (Singapore) Pte, Ltd. Managing an on-screen keyboard
US9310994B2 (en) 2010-02-19 2016-04-12 Microsoft Technology Licensing, Llc Use of bezel as an input mechanism
US9354803B2 (en) 2005-12-23 2016-05-31 Apple Inc. Scrolling list with floating adjacent index symbols
US9367205B2 (en) 2010-02-19 2016-06-14 Microsoft Technolgoy Licensing, Llc Radial menus with bezel gestures
US9436374B2 (en) 2009-09-25 2016-09-06 Apple Inc. Device, method, and graphical user interface for scrolling a multi-section document
US9454304B2 (en) 2010-02-25 2016-09-27 Microsoft Technology Licensing, Llc Multi-screen dual tap gesture
US9477337B2 (en) 2014-03-14 2016-10-25 Microsoft Technology Licensing, Llc Conductive trace routing for display and bezel sensors
US9519356B2 (en) 2010-02-04 2016-12-13 Microsoft Technology Licensing, Llc Link gestures
US9582122B2 (en) 2012-11-12 2017-02-28 Microsoft Technology Licensing, Llc Touch-sensitive bezel techniques
US9658766B2 (en) 2011-05-27 2017-05-23 Microsoft Technology Licensing, Llc Edge gesture
US9696223B2 (en) 2012-09-17 2017-07-04 Tk Holdings Inc. Single layer force sensor
US9696888B2 (en) 2010-12-20 2017-07-04 Microsoft Technology Licensing, Llc Application-launching interface for multiple modes
US9727031B2 (en) 2012-04-13 2017-08-08 Tk Holdings Inc. Pressure sensor including a pressure sensitive material for use with control systems and methods of using the same
JP2017140167A (en) * 2016-02-09 2017-08-17 株式会社コナミデジタルエンタテインメント Game system, game control device, and program
US20170285921A1 (en) * 2016-03-31 2017-10-05 Brother Kogyo Kabushiki Kaisha Information processing apparatus,non-transitory computer-readable medium storing instructions therefor, and information processing method
WO2017167698A1 (en) * 2016-04-01 2017-10-05 Thales Secure method for control of a function by means of a touchpad
US9965165B2 (en) 2010-02-19 2018-05-08 Microsoft Technology Licensing, Llc Multi-finger gestures
US10156934B2 (en) * 2016-08-12 2018-12-18 Canon Kabushiki Kaisha Electronic device and control method thereof
US10254927B2 (en) 2009-09-25 2019-04-09 Apple Inc. Device, method, and graphical user interface for manipulating workspace views
US10254955B2 (en) 2011-09-10 2019-04-09 Microsoft Technology Licensing, Llc Progressively indicating new content in an application-selectable user interface
US10269156B2 (en) 2015-06-05 2019-04-23 Manufacturing Resources International, Inc. System and method for blending order confirmation over menu board background
US10313037B2 (en) 2016-05-31 2019-06-04 Manufacturing Resources International, Inc. Electronic display remote image verification system and method
US10319271B2 (en) 2016-03-22 2019-06-11 Manufacturing Resources International, Inc. Cyclic redundancy check for electronic displays
US10319408B2 (en) 2015-03-30 2019-06-11 Manufacturing Resources International, Inc. Monolithic display with separately controllable sections
US10510304B2 (en) 2016-08-10 2019-12-17 Manufacturing Resources International, Inc. Dynamic dimming LED backlight for LCD array
US10579250B2 (en) 2011-09-01 2020-03-03 Microsoft Technology Licensing, Llc Arranging tiles
US10705701B2 (en) 2009-03-16 2020-07-07 Apple Inc. Device, method, and graphical user interface for moving a current position in content at a variable scrubbing rate
US10725624B2 (en) 2015-06-05 2020-07-28 Apple Inc. Movement between multiple views
US10922736B2 (en) 2015-05-15 2021-02-16 Manufacturing Resources International, Inc. Smart electronic display for restaurants
US10969944B2 (en) 2010-12-23 2021-04-06 Microsoft Technology Licensing, Llc Application reporting in an application-selectable user interface
US10990251B1 (en) * 2019-11-08 2021-04-27 Sap Se Smart augmented reality selector
US11164508B2 (en) * 2019-05-16 2021-11-02 Asustek Computer Inc. Electronic device
US11272017B2 (en) 2011-05-27 2022-03-08 Microsoft Technology Licensing, Llc Application notifications manifest
US11294484B1 (en) * 2012-12-31 2022-04-05 Allscripts Software, Llc Method for ensuring use intentions of a touch screen device
US11301128B2 (en) * 2019-05-01 2022-04-12 Google Llc Intended input to a user interface from detected gesture positions
US11895362B2 (en) 2021-10-29 2024-02-06 Manufacturing Resources International, Inc. Proof of play for images displayed at electronic displays

Families Citing this family (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
TWI403932B (en) * 2009-09-14 2013-08-01 Acer Inc Method for operating a touch screen, method for defining a touch gesture on the touch screen, and electronic device thereof
US9213482B2 (en) * 2011-11-11 2015-12-15 Elan Microelectronics Corporation Touch control device and method
US9880653B2 (en) 2012-04-30 2018-01-30 Corning Incorporated Pressure-sensing touch system utilizing total-internal reflection
CN106873874B (en) * 2017-01-20 2021-01-08 维沃移动通信有限公司 Application program starting method and mobile terminal

Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5583946A (en) * 1993-09-30 1996-12-10 Apple Computer, Inc. Method and apparatus for recognizing gestures on a computer system
US6163616A (en) * 1995-12-29 2000-12-19 Feldman; Stephen E. System and method for verifying the identity of a person
US20050122315A1 (en) * 2001-11-16 2005-06-09 Martin Chalk Communications device and supporting network
US20050210417A1 (en) * 2004-03-23 2005-09-22 Marvit David L User definable gestures for motion controlled handheld devices
US7286063B2 (en) * 2002-11-26 2007-10-23 Asulab S.A. Method of input of a security code by means of a touch screen for access to a function, an apparatus or a given location, and device for implementing the same
US20080012835A1 (en) * 2006-07-12 2008-01-17 N-Trig Ltd. Hover and touch detection for digitizer
US20080092245A1 (en) * 2006-09-15 2008-04-17 Agent Science Technologies, Inc. Multi-touch device behaviormetric user authentication and dynamic usability system
US20080098456A1 (en) * 2006-09-15 2008-04-24 Agent Science Technologies, Inc. Continuous user identification and situation analysis with identification of anonymous users through behaviormetrics

Family Cites Families (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7170496B2 (en) * 2003-01-24 2007-01-30 Bruce Peter Middleton Zero-front-footprint compact input system
US9063647B2 (en) * 2006-05-12 2015-06-23 Microsoft Technology Licensing, Llc Multi-touch uses, gestures, and implementation
US8842074B2 (en) * 2006-09-06 2014-09-23 Apple Inc. Portable electronic device performing similar operations for different gestures
US7877707B2 (en) * 2007-01-06 2011-01-25 Apple Inc. Detecting and interpreting real-world and security gestures on touch and hover sensitive devices
US8311530B2 (en) * 2007-01-26 2012-11-13 Research In Motion Limited Touch entry of password on a mobile device
US20080204417A1 (en) * 2007-02-27 2008-08-28 Pierce Paul M Multimodal Adaptive User Interface for a Portable Electronic Device

Patent Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5583946A (en) * 1993-09-30 1996-12-10 Apple Computer, Inc. Method and apparatus for recognizing gestures on a computer system
US6163616A (en) * 1995-12-29 2000-12-19 Feldman; Stephen E. System and method for verifying the identity of a person
US20050122315A1 (en) * 2001-11-16 2005-06-09 Martin Chalk Communications device and supporting network
US7286063B2 (en) * 2002-11-26 2007-10-23 Asulab S.A. Method of input of a security code by means of a touch screen for access to a function, an apparatus or a given location, and device for implementing the same
US20050210417A1 (en) * 2004-03-23 2005-09-22 Marvit David L User definable gestures for motion controlled handheld devices
US20080012835A1 (en) * 2006-07-12 2008-01-17 N-Trig Ltd. Hover and touch detection for digitizer
US20080092245A1 (en) * 2006-09-15 2008-04-17 Agent Science Technologies, Inc. Multi-touch device behaviormetric user authentication and dynamic usability system
US20080098456A1 (en) * 2006-09-15 2008-04-24 Agent Science Technologies, Inc. Continuous user identification and situation analysis with identification of anonymous users through behaviormetrics

Cited By (212)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9354803B2 (en) 2005-12-23 2016-05-31 Apple Inc. Scrolling list with floating adjacent index symbols
US10732814B2 (en) 2005-12-23 2020-08-04 Apple Inc. Scrolling list with floating adjacent index symbols
US9594457B2 (en) 2005-12-30 2017-03-14 Microsoft Technology Licensing, Llc Unintentional touch rejection
US9946370B2 (en) 2005-12-30 2018-04-17 Microsoft Technology Licensing, Llc Unintentional touch rejection
US9952718B2 (en) 2005-12-30 2018-04-24 Microsoft Technology Licensing, Llc Unintentional touch rejection
US10019080B2 (en) 2005-12-30 2018-07-10 Microsoft Technology Licensing, Llc Unintentional touch rejection
US9261964B2 (en) 2005-12-30 2016-02-16 Microsoft Technology Licensing, Llc Unintentional touch rejection
US20100201641A1 (en) * 2007-08-13 2010-08-12 Hideaki Tetsuhashi Contact type input device, contact type input method, and program
US20110025620A1 (en) * 2008-01-11 2011-02-03 Opdi Technologies A/S Touch-sensitive device
US9740336B2 (en) 2008-01-11 2017-08-22 O-Net Wavetouch Limited Touch-sensitive device
US9342187B2 (en) 2008-01-11 2016-05-17 O-Net Wavetouch Limited Touch-sensitive device
US8214891B2 (en) * 2008-03-26 2012-07-03 International Business Machines Corporation Using multi-touch gestures to protect sensitive content using a completely automated public turing test to tell computers and humans apart (CAPTCHA)
US20090249476A1 (en) * 2008-03-26 2009-10-01 Lisa Anne Seacat Using Multi-Touch Gestures to Protect Sensitive Content Using a Completely Automated Public Turing Test to Tell Computers and Humans Apart (CAPTCHA)
US9082117B2 (en) * 2008-05-17 2015-07-14 David H. Chin Gesture based authentication for wireless payment by a mobile electronic device
US8174503B2 (en) 2008-05-17 2012-05-08 David H. Cain Touch-based authentication of a mobile device through user generated pattern creation
US20140137234A1 (en) * 2008-05-17 2014-05-15 David H. Chin Mobile device authentication through touch-based gestures
US8913028B2 (en) * 2008-05-17 2014-12-16 David H. Chin Mobile device authentication through touch-based gestures
US7593000B1 (en) * 2008-05-17 2009-09-22 David H. Chin Touch-based authentication of a mobile device through user generated pattern creation
US9024890B2 (en) 2008-05-17 2015-05-05 David H. Chin Comparison of an applied gesture on a touch screen of a mobile device with a remotely stored security gesture
US20110187497A1 (en) * 2008-05-17 2011-08-04 David H Chin Comparison of an applied gesture on a touch screen of a mobile device with a remotely stored security gesture
US20110282785A1 (en) * 2008-05-17 2011-11-17 Chin David H Gesture based authentication for wireless payment by a mobile electronic device
US20090289916A1 (en) * 2008-05-23 2009-11-26 Hon Hai Precision Industry Co., Ltd. Electronic device and method for switching between locked state and unlocked state
US20110106446A1 (en) * 2008-05-26 2011-05-05 Volkswagen Ag Display method for a display system, display system and operating method for a navigation system of a vehicle
US10330488B2 (en) * 2008-05-26 2019-06-25 Volkswagen Ag Display method for a display system, display system and operating method for a navigation system of a vehicle
US9459116B2 (en) * 2008-05-26 2016-10-04 Volkswagen Ag Display method for a display system, display system and operating method for a navigation system of a vehicle
US9411503B2 (en) * 2008-07-17 2016-08-09 Sony Corporation Information processing device, information processing method, and information processing program
US20100013780A1 (en) * 2008-07-17 2010-01-21 Sony Corporation Information processing device, information processing method, and information processing program
US20100083108A1 (en) * 2008-09-26 2010-04-01 Research In Motion Limited Touch-screen device having soft escape key
US20110205176A1 (en) * 2008-11-05 2011-08-25 Takashi Okada Portable electronic device, and power saving method and power saving program for the same
US9032337B2 (en) * 2008-12-23 2015-05-12 Samsung Electronics Co., Ltd. Method and apparatus for unlocking electronic appliance
US10175875B2 (en) 2008-12-23 2019-01-08 Samsung Electronics Co., Ltd. Method and apparatus for unlocking electronic appliance
US11137895B2 (en) 2008-12-23 2021-10-05 Samsung Electronics Co., Ltd. Method and apparatus for unlocking electronic appliance
US20100162182A1 (en) * 2008-12-23 2010-06-24 Samsung Electronics Co., Ltd. Method and apparatus for unlocking electronic appliance
KR20100098804A (en) * 2009-03-02 2010-09-10 엘지전자 주식회사 Method for releasing locking in mobile terminal and mobile terminal using the same
US20100222049A1 (en) * 2009-03-02 2010-09-02 Lg Electronics Inc. Method for releasing a locking in mobile terminal and mobile terminal using the same
US8712392B2 (en) * 2009-03-02 2014-04-29 Lg Electronics Inc. Method for releasing a locking in mobile terminal and mobile terminal using the same
KR101582687B1 (en) * 2009-03-02 2016-01-05 엘지전자 주식회사 Method for releasing locking in mobile terminal and mobile terminal using the same
US9298349B2 (en) 2009-03-02 2016-03-29 Lg Electronics Inc. Method for releasing a locking in mobile terminal and mobile terminal using the same
US10705701B2 (en) 2009-03-16 2020-07-07 Apple Inc. Device, method, and graphical user interface for moving a current position in content at a variable scrubbing rate
US11567648B2 (en) 2009-03-16 2023-01-31 Apple Inc. Device, method, and graphical user interface for moving a current position in content at a variable scrubbing rate
US11907519B2 (en) 2009-03-16 2024-02-20 Apple Inc. Device, method, and graphical user interface for moving a current position in content at a variable scrubbing rate
US20100255885A1 (en) * 2009-04-07 2010-10-07 Samsung Electronics Co., Ltd. Input device and method for mobile terminal
US8836648B2 (en) 2009-05-27 2014-09-16 Microsoft Corporation Touch pull-in gesture
US20110006996A1 (en) * 2009-07-08 2011-01-13 Smith Nathan J Private data entry
US20110019105A1 (en) * 2009-07-27 2011-01-27 Echostar Technologies L.L.C. Verification of symbols received through a touchpad of a remote control device in an electronic system to allow access to system functions
US8520113B2 (en) * 2009-08-07 2013-08-27 Samsung Electronics Co., Ltd. Digital photographing apparatus, method of controlling the same, and recording medium storing program to execute the method
US20110032377A1 (en) * 2009-08-07 2011-02-10 Samsung Electronics Co., Ltd. Digital photographing apparatus, method of controlling the same, and recording medium storing program to execute the method
US9563350B2 (en) * 2009-08-11 2017-02-07 Lg Electronics Inc. Mobile terminal and method for controlling the same
US20110041102A1 (en) * 2009-08-11 2011-02-17 Jong Hwan Kim Mobile terminal and method for controlling the same
US10282070B2 (en) 2009-09-22 2019-05-07 Apple Inc. Device, method, and graphical user interface for manipulating user interface objects
US8863016B2 (en) 2009-09-22 2014-10-14 Apple Inc. Device, method, and graphical user interface for manipulating user interface objects
US11334229B2 (en) 2009-09-22 2022-05-17 Apple Inc. Device, method, and graphical user interface for manipulating user interface objects
US10788965B2 (en) 2009-09-22 2020-09-29 Apple Inc. Device, method, and graphical user interface for manipulating user interface objects
US10564826B2 (en) 2009-09-22 2020-02-18 Apple Inc. Device, method, and graphical user interface for manipulating user interface objects
US11947782B2 (en) 2009-09-25 2024-04-02 Apple Inc. Device, method, and graphical user interface for manipulating workspace views
US8766928B2 (en) 2009-09-25 2014-07-01 Apple Inc. Device, method, and graphical user interface for manipulating user interface objects
US8780069B2 (en) 2009-09-25 2014-07-15 Apple Inc. Device, method, and graphical user interface for manipulating user interface objects
US10254927B2 (en) 2009-09-25 2019-04-09 Apple Inc. Device, method, and graphical user interface for manipulating workspace views
US9310907B2 (en) 2009-09-25 2016-04-12 Apple Inc. Device, method, and graphical user interface for manipulating user interface objects
US10928993B2 (en) 2009-09-25 2021-02-23 Apple Inc. Device, method, and graphical user interface for manipulating workspace views
US8799826B2 (en) 2009-09-25 2014-08-05 Apple Inc. Device, method, and graphical user interface for moving a calendar entry in a calendar application
US11366576B2 (en) 2009-09-25 2022-06-21 Apple Inc. Device, method, and graphical user interface for manipulating workspace views
US9436374B2 (en) 2009-09-25 2016-09-06 Apple Inc. Device, method, and graphical user interface for scrolling a multi-section document
JP2015130182A (en) * 2009-10-09 2015-07-16 レノボ・イノベーションズ・リミテッド(香港) Mobile device, security lock control method thereof, and program
EP2487622A1 (en) * 2009-10-09 2012-08-15 Nec Corporation Mobile device, and security lock control method and program thereof
US20120196573A1 (en) * 2009-10-09 2012-08-02 Nec Corporation Mobile device, and security lock control method and program thereof
EP2487622A4 (en) * 2009-10-09 2014-05-28 Nec Corp Mobile device, and security lock control method and program thereof
US20110099473A1 (en) * 2009-10-23 2011-04-28 Samsung Electronics Co., Ltd. Input signal processing device for portable device and method of the same
US20110161888A1 (en) * 2009-12-28 2011-06-30 Sony Corporation Operation direction determination apparatus, remote operating system, operation direction determination method and program
US20150365236A1 (en) * 2009-12-29 2015-12-17 Bizmodeline Co., Ltd. Password processing device
US20110162066A1 (en) * 2009-12-29 2011-06-30 Bizmodeline Co., Ltd. Password processing method and apparatus
US9830444B2 (en) * 2009-12-29 2017-11-28 Bizmodeline Co., Ltd. Password processing device
US9146669B2 (en) * 2009-12-29 2015-09-29 Bizmodeline Co., Ltd. Password processing method and apparatus
US9444626B2 (en) * 2009-12-29 2016-09-13 Bizmodeline Co., Ltd. Password processing device
US20160342785A1 (en) * 2009-12-29 2016-11-24 Bizmodeline Co., Ltd. Password processing device
US20110163967A1 (en) * 2010-01-06 2011-07-07 Imran Chaudhri Device, Method, and Graphical User Interface for Changing Pages in an Electronic Document
US8539385B2 (en) * 2010-01-26 2013-09-17 Apple Inc. Device, method, and graphical user interface for precise positioning of objects
US8539386B2 (en) 2010-01-26 2013-09-17 Apple Inc. Device, method, and graphical user interface for selecting and moving objects
US20110185321A1 (en) * 2010-01-26 2011-07-28 Jay Christopher Capela Device, Method, and Graphical User Interface for Precise Positioning of Objects
US8612884B2 (en) 2010-01-26 2013-12-17 Apple Inc. Device, method, and graphical user interface for resizing objects
US8677268B2 (en) 2010-01-26 2014-03-18 Apple Inc. Device, method, and graphical user interface for resizing objects
US8239785B2 (en) 2010-01-27 2012-08-07 Microsoft Corporation Edge gestures
WO2011094045A2 (en) * 2010-01-28 2011-08-04 Microsoft Corporation Copy and staple gestures
US8261213B2 (en) 2010-01-28 2012-09-04 Microsoft Corporation Brush, carbon-copy, and fill gestures
US9857970B2 (en) 2010-01-28 2018-01-02 Microsoft Technology Licensing, Llc Copy and staple gestures
US9411504B2 (en) 2010-01-28 2016-08-09 Microsoft Technology Licensing, Llc Copy and staple gestures
US9411498B2 (en) 2010-01-28 2016-08-09 Microsoft Technology Licensing, Llc Brush, carbon-copy, and fill gestures
WO2011094045A3 (en) * 2010-01-28 2011-10-20 Microsoft Corporation Copy and staple gestures
US10282086B2 (en) 2010-01-28 2019-05-07 Microsoft Technology Licensing, Llc Brush, carbon-copy, and fill gestures
US20110185300A1 (en) * 2010-01-28 2011-07-28 Microsoft Corporation Brush, carbon-copy, and fill gestures
US20110191719A1 (en) * 2010-02-04 2011-08-04 Microsoft Corporation Cut, Punch-Out, and Rip Gestures
US9519356B2 (en) 2010-02-04 2016-12-13 Microsoft Technology Licensing, Llc Link gestures
US8799827B2 (en) 2010-02-19 2014-08-05 Microsoft Corporation Page manipulations using on and off-screen gestures
US9310994B2 (en) 2010-02-19 2016-04-12 Microsoft Technology Licensing, Llc Use of bezel as an input mechanism
US10268367B2 (en) 2010-02-19 2019-04-23 Microsoft Technology Licensing, Llc Radial menus with bezel gestures
US9274682B2 (en) 2010-02-19 2016-03-01 Microsoft Technology Licensing, Llc Off-screen gestures to create on-screen input
US9367205B2 (en) 2010-02-19 2016-06-14 Microsoft Technolgoy Licensing, Llc Radial menus with bezel gestures
US9965165B2 (en) 2010-02-19 2018-05-08 Microsoft Technology Licensing, Llc Multi-finger gestures
US9454304B2 (en) 2010-02-25 2016-09-27 Microsoft Technology Licensing, Llc Multi-screen dual tap gesture
US20110209103A1 (en) * 2010-02-25 2011-08-25 Hinckley Kenneth P Multi-screen hold and drag gesture
US8539384B2 (en) 2010-02-25 2013-09-17 Microsoft Corporation Multi-screen pinch and expand gestures
US11055050B2 (en) 2010-02-25 2021-07-06 Microsoft Technology Licensing, Llc Multi-device pairing and combined display
US8751970B2 (en) 2010-02-25 2014-06-10 Microsoft Corporation Multi-screen synchronous slide gesture
US9075522B2 (en) 2010-02-25 2015-07-07 Microsoft Technology Licensing, Llc Multi-screen bookmark hold gesture
US8707174B2 (en) 2010-02-25 2014-04-22 Microsoft Corporation Multi-screen hold and page-flip gesture
US8473870B2 (en) 2010-02-25 2013-06-25 Microsoft Corporation Multi-screen hold and drag gesture
US8587422B2 (en) 2010-03-31 2013-11-19 Tk Holdings, Inc. Occupant sensing system
US9007190B2 (en) 2010-03-31 2015-04-14 Tk Holdings Inc. Steering wheel sensors
US8725230B2 (en) 2010-04-02 2014-05-13 Tk Holdings Inc. Steering wheel with hand sensors
US9134897B2 (en) * 2010-04-26 2015-09-15 Via Technologies, Inc. Electronic system and method for operating touch screen thereof
US20110265045A1 (en) * 2010-04-26 2011-10-27 Via Technologies, Inc. Electronic system and method for operating touch screen thereof
US8136053B1 (en) * 2010-05-14 2012-03-13 Google Inc. Direct, gesture-based actions from device's lock screen
US20110283241A1 (en) * 2010-05-14 2011-11-17 Google Inc. Touch Gesture Actions From A Device's Lock Screen
US20110289423A1 (en) * 2010-05-24 2011-11-24 Samsung Electronics Co., Ltd. Method and apparatus for controlling objects of a user interface
EP2390812B1 (en) * 2010-05-24 2016-07-20 Samsung Electronics Co., Ltd. Method and apparatus for controlling objects of a user interface
US9098182B2 (en) 2010-07-30 2015-08-04 Apple Inc. Device, method, and graphical user interface for copying user interface objects between content regions
US8972879B2 (en) 2010-07-30 2015-03-03 Apple Inc. Device, method, and graphical user interface for reordering the front-to-back positions of objects
US9626098B2 (en) 2010-07-30 2017-04-18 Apple Inc. Device, method, and graphical user interface for copying formatting attributes
US9081494B2 (en) 2010-07-30 2015-07-14 Apple Inc. Device, method, and graphical user interface for copying formatting attributes
US20120096376A1 (en) * 2010-10-14 2012-04-19 Canon Kabushiki Kaisha Display control apparatus, display control method, and storage medium
US8782544B2 (en) * 2010-10-14 2014-07-15 Canon Kabushiki Kaisha Display control apparatus, display control method, and storage medium
US20120133484A1 (en) * 2010-11-29 2012-05-31 Research In Motion Limited Multiple-input device lock and unlock
US9696888B2 (en) 2010-12-20 2017-07-04 Microsoft Technology Licensing, Llc Application-launching interface for multiple modes
US10969944B2 (en) 2010-12-23 2021-04-06 Microsoft Technology Licensing, Llc Application reporting in an application-selectable user interface
US11126333B2 (en) 2010-12-23 2021-09-21 Microsoft Technology Licensing, Llc Application reporting in an application-selectable user interface
US9229918B2 (en) 2010-12-23 2016-01-05 Microsoft Technology Licensing, Llc Presenting an application change through a tile
US20120274583A1 (en) * 2011-02-08 2012-11-01 Ammon Haggerty Multimodal Touchscreen Interaction Apparatuses, Methods and Systems
CN103534674A (en) * 2011-02-08 2014-01-22 海沃氏公司 Multimodal touchscreen interaction apparatuses, methods and systems
US20120252410A1 (en) * 2011-03-28 2012-10-04 Htc Corporation Systems and Methods for Gesture Lock Obfuscation
US9514297B2 (en) * 2011-03-28 2016-12-06 Htc Corporation Systems and methods for gesture lock obfuscation
US9535597B2 (en) 2011-05-27 2017-01-03 Microsoft Technology Licensing, Llc Managing an immersive interface in a multi-application immersive environment
US9052820B2 (en) 2011-05-27 2015-06-09 Microsoft Technology Licensing, Llc Multi-application environment
US9158445B2 (en) 2011-05-27 2015-10-13 Microsoft Technology Licensing, Llc Managing an immersive interface in a multi-application immersive environment
US9658766B2 (en) 2011-05-27 2017-05-23 Microsoft Technology Licensing, Llc Edge gesture
US9104440B2 (en) 2011-05-27 2015-08-11 Microsoft Technology Licensing, Llc Multi-application environment
US11698721B2 (en) 2011-05-27 2023-07-11 Microsoft Technology Licensing, Llc Managing an immersive interface in a multi-application immersive environment
US11272017B2 (en) 2011-05-27 2022-03-08 Microsoft Technology Licensing, Llc Application notifications manifest
US10303325B2 (en) 2011-05-27 2019-05-28 Microsoft Technology Licensing, Llc Multi-application environment
US9104307B2 (en) 2011-05-27 2015-08-11 Microsoft Technology Licensing, Llc Multi-application environment
US10579250B2 (en) 2011-09-01 2020-03-03 Microsoft Technology Licensing, Llc Arranging tiles
US10254955B2 (en) 2011-09-10 2019-04-09 Microsoft Technology Licensing, Llc Progressively indicating new content in an application-selectable user interface
US20130063009A1 (en) * 2011-09-13 2013-03-14 Nunoerin, Llc Interactive furniture device
US20130086673A1 (en) * 2011-09-29 2013-04-04 Cisco Technology, Inc. Techniques for securely unlocking a touch screen user device
US20130174083A1 (en) * 2011-12-28 2013-07-04 Acer Incorporated Electronic device and method of controlling the same
US9727031B2 (en) 2012-04-13 2017-08-08 Tk Holdings Inc. Pressure sensor including a pressure sensitive material for use with control systems and methods of using the same
US20130278540A1 (en) * 2012-04-20 2013-10-24 Esat Yilmaz Inter Touch Sensor Communications
US20130298079A1 (en) * 2012-05-02 2013-11-07 Pantech Co., Ltd. Apparatus and method for unlocking an electronic device
US11204653B2 (en) * 2012-06-29 2021-12-21 Samsung Electronics Co., Ltd Method and device for handling event invocation using a stylus pen
US10831289B2 (en) * 2012-06-29 2020-11-10 Samsung Electronics Co., Ltd Method and device for handling event invocation using a stylus pen
US9857889B2 (en) * 2012-06-29 2018-01-02 Samsung Electronic Co., Ltd Method and device for handling event invocation using a stylus pen
US20140002397A1 (en) * 2012-06-29 2014-01-02 Samsung Electronics Co., Ltd. Method and device for handling event invocation using a stylus pen
US20140006965A1 (en) * 2012-07-02 2014-01-02 Beijing Xiaomi Technology Co., Ltd. Method of entering a user interface in a device and the device thereof
US9823834B2 (en) * 2012-07-11 2017-11-21 Guang Dong Oppo Mobile Telecommunications., Ltd. Quick access gesture setting and accessing method for a touch control device
US20140033140A1 (en) * 2012-07-11 2014-01-30 Guang Dong Oppo Mobile Telecommunications Corp., Ltd. Quick access function setting method for a touch control device
US20140026105A1 (en) * 2012-07-18 2014-01-23 Research In Motion Limited Method and Apparatus Pertaining to a Gesture-Controlled Snooze Instruction
US20140028598A1 (en) * 2012-07-30 2014-01-30 Samsung Electronics Co., Ltd Apparatus and method for controlling data transmission in terminal
US9696223B2 (en) 2012-09-17 2017-07-04 Tk Holdings Inc. Single layer force sensor
US20140298672A1 (en) * 2012-09-27 2014-10-09 Analog Devices Technology Locking and unlocking of contacless gesture-based user interface of device having contactless gesture detection system
FR2998071A1 (en) * 2012-11-09 2014-05-16 Thales Sa METHOD FOR SECURING A CONTROL ON A TOUCH-SURFACE VISUALIZATION DEVICE AND SYSTEM THEREOF
EP2731002A1 (en) * 2012-11-09 2014-05-14 Thales Method for securing a command on a viewing device with tactile surface and related system
US9582122B2 (en) 2012-11-12 2017-02-28 Microsoft Technology Licensing, Llc Touch-sensitive bezel techniques
US10656750B2 (en) 2012-11-12 2020-05-19 Microsoft Technology Licensing, Llc Touch-sensitive bezel techniques
US9389762B2 (en) * 2012-11-29 2016-07-12 Egalax_Empia Technology Inc. Method for unlocking touch screen, electronic device thereof, and recording medium thereof
US20140145990A1 (en) * 2012-11-29 2014-05-29 Egalax_Empia Technology Inc. Method for unlocking touch screen, electronic device thereof, and recording medium thereof
US20150317076A1 (en) * 2012-12-11 2015-11-05 University Of Washington Through Its Center For Commercialization Use of hand posture to improve text entry
US10540083B2 (en) * 2012-12-11 2020-01-21 University Of Washington Use of hand posture to improve text entry
US11294484B1 (en) * 2012-12-31 2022-04-05 Allscripts Software, Llc Method for ensuring use intentions of a touch screen device
US9158399B2 (en) * 2013-03-13 2015-10-13 Htc Corporation Unlock method and mobile device using the same
DE102013211011B4 (en) * 2013-03-13 2017-05-24 Htc Corporation Unlock method and mobile device using same
US20140267064A1 (en) * 2013-03-13 2014-09-18 Htc Corporation Unlock Method and Mobile Device Using the Same
US20140347302A1 (en) * 2013-05-24 2014-11-27 Boe Technology Group Co., Ltd. Unlocking method and device for touch screen terminal and touch screen terminal
US9400595B2 (en) * 2013-05-24 2016-07-26 Boe Technology Group Co., Ltd. Unlocking method and device for touch screen terminal and touch screen terminal
EP2835732A3 (en) * 2013-08-06 2015-04-22 Samsung Electronics Co., Ltd Electronic device provided with touch screen and operating method thereof
US10776479B2 (en) * 2013-08-28 2020-09-15 Paypal, Inc. Authentication system
US20150066762A1 (en) * 2013-08-28 2015-03-05 Geoffrey W. Chatterton Authentication system
US20150199553A1 (en) * 2014-01-13 2015-07-16 Samsung Electronics Co., Ltd. Method for recognizing fingerprint and mobile terminal supporting the same
US10101843B2 (en) * 2014-01-13 2018-10-16 Samsung Electronics Co., Ltd. Method for recognizing fingerprint and mobile terminal supporting the same
US9521241B2 (en) 2014-03-13 2016-12-13 Lg Electronics Inc. Mobile terminal and method of controlling the same
EP2919108A1 (en) * 2014-03-13 2015-09-16 LG Electronics, Inc. Mobile terminal and method of controlling the same
CN104915133A (en) * 2014-03-13 2015-09-16 Lg电子株式会社 Mobile terminal and method of controlling the same
US9477337B2 (en) 2014-03-14 2016-10-25 Microsoft Technology Licensing, Llc Conductive trace routing for display and bezel sensors
US9946383B2 (en) 2014-03-14 2018-04-17 Microsoft Technology Licensing, Llc Conductive trace routing for display and bezel sensors
WO2015149101A3 (en) * 2014-04-01 2015-11-26 Linkilike Gmbh Method for comparing user identities of databases
US9645668B2 (en) 2014-07-22 2017-05-09 Lg Electronics Inc. Mobile terminal and method for controlling the same
EP2977884A1 (en) * 2014-07-22 2016-01-27 LG Electronics Inc. Mobile terminal and method for controlling the same
US20160026380A1 (en) * 2014-07-25 2016-01-28 Hannstar Display (Nanjing) Corporation Hand-held electronic device, computer-executed method and touch-sensing cover
US20160034172A1 (en) * 2014-07-30 2016-02-04 Wistron Corporation Touch device and control method and method for determining unlocking thereof
US9727233B2 (en) * 2014-07-30 2017-08-08 Wistron Corporation Touch device and control method and method for determining unlocking thereof
US10678381B2 (en) * 2014-08-21 2020-06-09 DISH Technologies L.L.C. Determining handedness on multi-element capacitive devices
US20160054827A1 (en) * 2014-08-21 2016-02-25 Echostar Technologies L.L.C. Determining handedness on multi-element capacitive devices
US10048862B2 (en) * 2014-09-08 2018-08-14 Lenovo (Singapore) Pte. Ltd. Managing an on-screen keyboard
US20160070465A1 (en) * 2014-09-08 2016-03-10 Lenovo (Singapore) Pte, Ltd. Managing an on-screen keyboard
US10319408B2 (en) 2015-03-30 2019-06-11 Manufacturing Resources International, Inc. Monolithic display with separately controllable sections
US10922736B2 (en) 2015-05-15 2021-02-16 Manufacturing Resources International, Inc. Smart electronic display for restaurants
US10725624B2 (en) 2015-06-05 2020-07-28 Apple Inc. Movement between multiple views
US10467610B2 (en) 2015-06-05 2019-11-05 Manufacturing Resources International, Inc. System and method for a redundant multi-panel electronic display
US10269156B2 (en) 2015-06-05 2019-04-23 Manufacturing Resources International, Inc. System and method for blending order confirmation over menu board background
JP2017140167A (en) * 2016-02-09 2017-08-17 株式会社コナミデジタルエンタテインメント Game system, game control device, and program
US10319271B2 (en) 2016-03-22 2019-06-11 Manufacturing Resources International, Inc. Cyclic redundancy check for electronic displays
US10705697B2 (en) * 2016-03-31 2020-07-07 Brother Kogyo Kabushiki Kaisha Information processing apparatus configured to edit images, non-transitory computer-readable medium storing instructions therefor, and information processing method for editing images
US20170285921A1 (en) * 2016-03-31 2017-10-05 Brother Kogyo Kabushiki Kaisha Information processing apparatus,non-transitory computer-readable medium storing instructions therefor, and information processing method
FR3049733A1 (en) * 2016-04-01 2017-10-06 Thales Sa METHOD FOR SECURELY CONTROLLING A FUNCTION USING A TOUCH SLAB
US10649649B2 (en) 2016-04-01 2020-05-12 Thales Secure method for control of a function by means of a touchpad
WO2017167698A1 (en) * 2016-04-01 2017-10-05 Thales Secure method for control of a function by means of a touchpad
US10313037B2 (en) 2016-05-31 2019-06-04 Manufacturing Resources International, Inc. Electronic display remote image verification system and method
US10756836B2 (en) 2016-05-31 2020-08-25 Manufacturing Resources International, Inc. Electronic display remote image verification system and method
US10510304B2 (en) 2016-08-10 2019-12-17 Manufacturing Resources International, Inc. Dynamic dimming LED backlight for LCD array
US10156934B2 (en) * 2016-08-12 2018-12-18 Canon Kabushiki Kaisha Electronic device and control method thereof
US11301128B2 (en) * 2019-05-01 2022-04-12 Google Llc Intended input to a user interface from detected gesture positions
US11164508B2 (en) * 2019-05-16 2021-11-02 Asustek Computer Inc. Electronic device
US10990251B1 (en) * 2019-11-08 2021-04-27 Sap Se Smart augmented reality selector
US11895362B2 (en) 2021-10-29 2024-02-06 Manufacturing Resources International, Inc. Proof of play for images displayed at electronic displays

Also Published As

Publication number Publication date
TW200907742A (en) 2009-02-16
WO2008138030A1 (en) 2008-11-20
US20170285853A1 (en) 2017-10-05
US9696847B2 (en) 2017-07-04
US20160253032A1 (en) 2016-09-01
US20120233690A1 (en) 2012-09-13
US9280279B2 (en) 2016-03-08

Similar Documents

Publication Publication Date Title
US9696847B2 (en) User-defined gesture enablement protocols for touch input device
US10228833B2 (en) Input device user interface enhancements
JP3589381B2 (en) Virtual pointing device generation method, apparatus, and computer system
EP3066605B1 (en) Directional touch unlocking for electronic devices
JP4384734B2 (en) Virtual pointing device generation instruction method, computer system, and apparatus
JP3504462B2 (en) Pointing device generation instruction method and computer system
CN103765368B (en) Mobile terminal for performing screen unlock based on motion and method thereof
EP2506178B1 (en) User interface, touch-controlled device and method for authenticating a user of a touch-controlled device
KR20120003441A (en) Bimodal touch sensitive digital notebook
JPH1063425A (en) Method for instracting generation of virtual pointing device, and computer system
JPH1063422A (en) Method for generating at least two virtual pointing device, and computer system
CN107407998B (en) Method for interacting with an electronic and/or computer device implementing a capacitive control surface and a peripheral surface, interface and device implementing the method
KR101349526B1 (en) Automatic teller machine and display method
KR20120081422A (en) Terminal having touch screen and method for inputting letter according to touch event thereof
US10237736B2 (en) Unlocking of a mobile device by a code received via a stencil on a touchscreen
KR102647593B1 (en) Method of unlocking device by performing fingerprint verification and moving gesture and the device
CN109634417A (en) A kind of processing method and electronic equipment
KR101706909B1 (en) Finger Input Devices
KR20010102638A (en) Keyboard system having easy key operation function and method thereof
TW202009687A (en) Method for unlocking a display of a multi-display device
WO2015016894A1 (en) Gesture detection

Legal Events

Date Code Title Description
AS Assignment

Owner name: RPO PTY LIMITED, AUSTRALIA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:ATKINS, GRAHAM ROY;MAXWELL, IAN ANDREW;REEL/FRAME:021279/0291

Effective date: 20080606

AS Assignment

Owner name: BRIDGE BANK, NATIONAL ASSOCIATION, CALIFORNIA

Free format text: SECURITY AGREEMENT;ASSIGNOR:RPO PTY LTD;REEL/FRAME:024838/0948

Effective date: 20100813

STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION

AS Assignment

Owner name: RPO PTY LTD, CALIFORNIA

Free format text: REASSIGNMENT AND RELEASE OF IP SECURITY INTEREST;ASSIGNOR:BRIDGE BANK, NATIONAL ASSOCIATION;REEL/FRAME:028737/0963

Effective date: 20120802

AS Assignment

Owner name: TRINITY CAPITAL INVESTMENT LLC, ARIZONA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:RPO PTY LTD;REEL/FRAME:029770/0739

Effective date: 20120628

Owner name: ZETTA RESEARCH AND DEVELOPMENT LLC - RPO SERIES, D

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:TRINITY CAPITAL INVESTMENT LLC;REEL/FRAME:029770/0778

Effective date: 20120629