US20080088602A1 - Multi-functional hand-held device - Google Patents

Multi-functional hand-held device Download PDF

Info

Publication number
US20080088602A1
US20080088602A1 US11/966,948 US96694807A US2008088602A1 US 20080088602 A1 US20080088602 A1 US 20080088602A1 US 96694807 A US96694807 A US 96694807A US 2008088602 A1 US2008088602 A1 US 2008088602A1
Authority
US
United States
Prior art keywords
hand
electronic device
display
held electronic
touch
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US11/966,948
Inventor
Steven Hotelling
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Apple Inc
Original Assignee
Apple Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Family has litigation
First worldwide family litigation filed litigation Critical https://patents.darts-ip.com/?family=36936660&utm_source=google_patent&utm_medium=platform_link&utm_campaign=public_patent_search&patent=US20080088602(A1) "Global patent litigation dataset” by Darts-ip is licensed under a Creative Commons Attribution 4.0 International License.
Priority claimed from US11/367,749 external-priority patent/US11275405B2/en
Application filed by Apple Inc filed Critical Apple Inc
Priority to US11/966,948 priority Critical patent/US20080088602A1/en
Assigned to APPLE COMPUTER, INC. reassignment APPLE COMPUTER, INC. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: HOTELLING, STEVE P.
Assigned to APPLE INC. reassignment APPLE INC. CHANGE OF NAME (SEE DOCUMENT FOR DETAILS). Assignors: APPLE COMPUTER, INC.
Publication of US20080088602A1 publication Critical patent/US20080088602A1/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F1/00Details not covered by groups G06F3/00 - G06F13/00 and G06F21/00
    • G06F1/16Constructional details or arrangements
    • G06F1/1613Constructional details or arrangements for portable computers
    • G06F1/1626Constructional details or arrangements for portable computers with a single-body enclosure integrating a flat display, e.g. Personal Digital Assistants [PDAs]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F1/00Details not covered by groups G06F3/00 - G06F13/00 and G06F21/00
    • G06F1/16Constructional details or arrangements
    • G06F1/1613Constructional details or arrangements for portable computers
    • G06F1/1633Constructional details or arrangements of portable computers not specific to the type of enclosures covered by groups G06F1/1615 - G06F1/1626
    • G06F1/1637Details related to the display arrangement, including those related to the mounting of the display in the housing
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F1/00Details not covered by groups G06F3/00 - G06F13/00 and G06F21/00
    • G06F1/16Constructional details or arrangements
    • G06F1/1613Constructional details or arrangements for portable computers
    • G06F1/1633Constructional details or arrangements of portable computers not specific to the type of enclosures covered by groups G06F1/1615 - G06F1/1626
    • G06F1/1656Details related to functional adaptations of the enclosure, e.g. to provide protection against EMI, shock, water, or to host detachable peripherals like a mouse or removable expansions units like PCMCIA cards, or to provide access to internal components for maintenance or to removable storage supports like CDs or DVDs, or to mechanically mount accessories
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F1/00Details not covered by groups G06F3/00 - G06F13/00 and G06F21/00
    • G06F1/16Constructional details or arrangements
    • G06F1/1613Constructional details or arrangements for portable computers
    • G06F1/1633Constructional details or arrangements of portable computers not specific to the type of enclosures covered by groups G06F1/1615 - G06F1/1626
    • G06F1/1684Constructional details or arrangements related to integrated I/O peripherals not covered by groups G06F1/1635 - G06F1/1675
    • G06F1/169Constructional details or arrangements related to integrated I/O peripherals not covered by groups G06F1/1635 - G06F1/1675 the I/O peripheral being an integrated pointing device, e.g. trackball in the palm rest area, mini-joystick integrated between keyboard keys, touch pads or touch stripes
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/016Input arrangements with force or tactile feedback as computer generated output to the user
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/02Input arrangements using manually operated switches, e.g. using keyboards or dials
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/03Arrangements for converting the position or the displacement of a member into a coded form
    • G06F3/041Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means
    • G06F3/0412Digitisers structurally integrated in a display
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/03Arrangements for converting the position or the displacement of a member into a coded form
    • G06F3/041Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means
    • G06F3/0414Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means using force sensing means to determine a position
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/03Arrangements for converting the position or the displacement of a member into a coded form
    • G06F3/041Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means
    • G06F3/0416Control or interface arrangements specially adapted for digitisers
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/03Arrangements for converting the position or the displacement of a member into a coded form
    • G06F3/041Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means
    • G06F3/044Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means by capacitive means
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/03Arrangements for converting the position or the displacement of a member into a coded form
    • G06F3/041Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means
    • G06F3/044Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means by capacitive means
    • G06F3/0443Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means by capacitive means using a single layer of sensing electrodes
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/03Arrangements for converting the position or the displacement of a member into a coded form
    • G06F3/041Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means
    • G06F3/044Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means by capacitive means
    • G06F3/0445Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means by capacitive means using two or more layers of sensing electrodes, e.g. using two layers of electrodes separated by a dielectric layer
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/03Arrangements for converting the position or the displacement of a member into a coded form
    • G06F3/041Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means
    • G06F3/044Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means by capacitive means
    • G06F3/0446Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means by capacitive means using a grid-like structure of electrodes in at least two directions, e.g. using row and column electrodes
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0481Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance
    • G06F3/0482Interaction with lists of selectable items, e.g. menus
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0484Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range
    • G06F3/0485Scrolling or panning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0487Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
    • G06F3/0488Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures
    • G06F3/04883Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures for inputting data by handwriting, e.g. gesture or text
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0487Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
    • G06F3/0488Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures
    • G06F3/04886Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures by partitioning the display area of the touch-screen or the surface of the digitising tablet into independently controllable areas, e.g. virtual keyboards or menus
    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09GARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
    • G09G3/00Control arrangements or circuits, of interest only in connection with visual indicators other than cathode-ray tubes
    • G09G3/20Control arrangements or circuits, of interest only in connection with visual indicators other than cathode-ray tubes for presentation of an assembly of a number of characters, e.g. a page, by composing the assembly by combination of individual elements arranged in a matrix no fixed position being assigned to or needed to be assigned to the individual characters or partial characters
    • G09G3/34Control arrangements or circuits, of interest only in connection with visual indicators other than cathode-ray tubes for presentation of an assembly of a number of characters, e.g. a page, by composing the assembly by combination of individual elements arranged in a matrix no fixed position being assigned to or needed to be assigned to the individual characters or partial characters by control of light from an independent source
    • G09G3/36Control arrangements or circuits, of interest only in connection with visual indicators other than cathode-ray tubes for presentation of an assembly of a number of characters, e.g. a page, by composing the assembly by combination of individual elements arranged in a matrix no fixed position being assigned to or needed to be assigned to the individual characters or partial characters by control of light from an independent source using liquid crystals
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2203/00Indexing scheme relating to G06F3/00 - G06F3/048
    • G06F2203/041Indexing scheme relating to G06F3/041 - G06F3/045
    • G06F2203/04105Pressure sensors for measuring the pressure or force exerted on the touch surface without providing the touch position
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2203/00Indexing scheme relating to G06F3/00 - G06F3/048
    • G06F2203/041Indexing scheme relating to G06F3/041 - G06F3/045
    • G06F2203/04106Multi-sensing digitiser, i.e. digitiser using at least two different sensing technologies simultaneously or alternatively, e.g. for detecting pen and finger, for saving power or for improving position detection
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2203/00Indexing scheme relating to G06F3/00 - G06F3/048
    • G06F2203/048Indexing scheme relating to G06F3/048
    • G06F2203/04808Several contacts: gestures triggering a specific function, e.g. scrolling, zooming, right-click, when the user establishes several contacts with the surface simultaneously; e.g. using several fingers or a combination of fingers and pen
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/03Arrangements for converting the position or the displacement of a member into a coded form
    • G06F3/041Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means
    • G06F3/0414Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means using force sensing means to determine a position
    • G06F3/04144Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means using force sensing means to determine a position using an array of force sensing means

Definitions

  • the user interface typically includes an output device in the form of a display, such as a Liquid Crystal Display (LCD), and one or more input devices, which can be mechanically actuated (e.g., switches, buttons, keys, dials, joysticks, joy pads) or electrically activated (e.g., touch pads or touch screens).
  • the display is typically configured to present visual information such as text and graphics, and the input devices are typically configured perform operations such as issuing commands, making selections or moving a cursor or selector in the electronic device.
  • the user interface is positioned on the front face (or front surface) of the hand-held device for easy viewing of the display and easy manipulation of the input devices.
  • FIGS. 1A-1F are diagrams of various hand-held electronic devices including for example a telephone 10 A ( FIG. 1A ), a PDA 10 B ( FIG. 1B ), a media player 10 C ( FIG. 1C ), a remote control 10 D ( FIG. 1D ), a camera 10 E ( FIG. 1E ), and a GPS module 10 F ( FIG. 1F ).
  • a display 12 which is secured inside the housing of the device 10 and which can be seen through an opening in the housing, is typically positioned in a first region of the electronic device 10 .
  • Each of these devices also include one or more input devices 14 , which are typically positioned in a second region of the electronic device 10 next to the display 12 .
  • the telephone 10 A typically includes a display 12 such as a character or graphical display, and input devices 14 such as a number pad and in some cases a navigation pad.
  • the PDA 10 B typically includes a display 12 such as a graphical display, and input devices 14 such as a stylus based resistive touch screen and buttons.
  • the media player 10 C typically includes a display 12 such as a character or graphic display and input devices 14 such as buttons or wheels.
  • the iPod® media player manufactured by Apple Computer, Inc. of Cupertino, Calif. is one example of a media player that includes both a display and input devices disposed next to the display.
  • the remote control 10 D typically includes an input device 14 such as a keypad and may or may not have a character display 12 .
  • the camera 10 E typically includes a display 12 such as a graphic display and input devices 14 such as buttons.
  • the GPS module 10 F typically includes a display 12 such as graphic display and input devices 14 such as buttons, and in some cases a joy pad.
  • the telephone 10 A has been combined with the PDA 10 B.
  • One problem that has been encountered is in the way inputs are made into the device.
  • Each of these devices has a particular set of input mechanisms for providing inputs into the device.
  • Some of these input mechanisms are generic to all the devices (e.g., power button) while others are not.
  • the ones that are not generic are typically dedicated to a particular functionality of the device.
  • PDAs typically include four dedicated buttons while cell phones typically include a numeric keypad and at least two dedicated buttons.
  • a multi-functional hand-held device capable of configuring user inputs based on how the device is to be used.
  • the multi-functional hand-held device has at most only a few physical buttons, keys, or switches so that its display size can be substantially increased.
  • additional surface area becomes available for a larger display.
  • this strategy would allow a substantially full screen display.
  • a full screen display is a display that consumes, or at least dominates, a surface (e.g., front surface) of an electronic device.
  • FIGS. 2-28 Various embodiments of a multi-functional hand-held device are discussed below with reference to FIGS. 2-28 . However, those skilled in the art will appreciate that the detailed description given herein with respect to these figures is exemplary and not exhaustive and that many variations on these embodiments are possible.
  • FIGS. 1A-1F are diagrams of various electronic devices.
  • FIG. 2 is a simplified diagram of a multi-functional hand-held device.
  • FIG. 3 is a perspective view of a substantially full screen hand-held device with a limited number of limited buttons.
  • FIG. 4 is a front view of a hand-held device with at least one button.
  • FIG. 5 is a diagram of a GUI separated into a standard region and a control region.
  • FIG. 6 is a diagram of a GUI separated into a standard region and a control region.
  • FIG. 7 is a diagram of a GUI separated into a standard region and a control region.
  • FIG. 8 is a diagram of a GUI separated into a standard region and a control region.
  • FIG. 9 illustrates an exemplary a GUI for a PDA.
  • FIG. 10 illustrates an exemplary GUI for a cell phone.
  • FIG. 11 illustrates an exemplary GUI for a media player.
  • FIG. 12 illustrates an exemplary GUI for a video player.
  • FIG. 13 illustrates an exemplary GUI for a game player.
  • FIG. 14 illustrates an exemplary GUI for a camera.
  • FIG. 15 illustrates an exemplary GUI for a GPS.
  • FIG. 16 illustrates an exemplary GUI for a remote control.
  • FIG. 17 illustrates an exemplary GUI for a hand top.
  • FIG. 18 illustrates an exemplary GUI for a main menu of a multi-functional hand held device.
  • FIG. 19 is a side elevation view, in cross section, of a hand-held device incorporating a force sensitive display.
  • FIG. 20 illustrates an input device that combines touch sensing and force sensing devices to provide x, y and z components when touched.
  • FIG. 21 is a side elevation view of an I/O device that combines a display with touch screen and a force sensing mechanism.
  • FIG. 22 is a side elevation view of an input device.
  • FIG. 23 is a side view, in cross section, of a hand-held device that incorporates a squeeze feature.
  • FIG. 24 is a side view, in cross section, of a hand-held electronic device.
  • FIG. 25 is a block diagram of a touch sensing method.
  • FIG. 26 is a block diagram of touch sensing method.
  • FIGS. 27 A-E are tables representing one example of a touch vocabulary associated with a music player.
  • FIG. 28 is a block diagram of an exemplary multi-functional hand-held device.
  • multi-functional is used to define a device that has the capabilities of two or more traditional devices in a single device.
  • the multi-functional device may, for example, include two or more of the following device functionalities: PDA, cell phone, music player, video player, game player, digital camera, handtop, Internet terminal, GPS or remote control.
  • combining devices may result in redundant hardware components, which allows components to be used for multiple different, device functionalities.
  • certain hardware components are distinct to each device and therefore additional space and connectivity must be made available.
  • each device functionality typically has its own programming or application software and, therefore, the multifunction device must be designed with enough memory to accommodate all the various software components.
  • a personal digital assistant is a mobile hand-held device that provides computing and information storage and retrieval capabilities for personal and/or business use.
  • PDAs are severally capable of tracking names, addresses, phone numbers and appointments. They are also often capable of taking notes, performing calculations, paging, data messaging, and electronic mail. PDAs may also include functionality for playing simple games, music, and other media files. Examples of PDAs include the Palm Pilot and Blackberry.
  • PDAs typically include a display and various input devices.
  • the input devices may include a stylus and touch screen that work in combination with a handwriting recognition program, keypads, mini-keyboards, navigation pads, and/or soft or fixed function buttons.
  • Cell phones are mobile telephones that allow a user to connect to other telephones using a cellular network.
  • Cell phones typically include a transceiver for transmitting and receiving telephone calls, controls such as a navigation pad for traversing through a display, a keypad for making numeric entries (and in some cases alphabetic entries), and soft or fixed function buttons. For example, in many cell phones one fixed function button is used for starting a call and another fixed function button is used for ending a call.
  • Music players come in a variety of forms. Music players are generally configured to store, process and output music. Music players can be based on the MP3 or MC format, which is a compression system for music. Music Players typically include a microprocessor, memory, display, audio jack, data port and playback controls. The playback controls typically include features such as menu, play/pause, next, previous, volume up, and volume down. Video players are similar to music players in most respects. In some cases, they may include a data storage device for receiving a removable storage medium such as a DVD. The iPod® media player manufactured by Apple Computer, Inc. of Cupertino, Calif. is one example of a media player.
  • Handtops are general purpose computers similar to laptops, but in a smaller form factor. Handtops typically include a display and a full keyboard.
  • FIG. 2 is a simplified diagram of a multi-functional hand-held device 100 .
  • the multi-functional hand-held device 100 integrates at least two devices 102 into a single device.
  • Each device 102 includes both hardware and software components 104 and 106 , which are integrated into multi-functional hand-held device 100 .
  • the multi-functional hand-held device 100 is not limited to only two devices, and may in fact integrate any number of devices.
  • Multi-functional device 100 also includes switch 110 , which that allows multi-functional device 100 to be switched from one device operating mode to another device operating mode.
  • switch 110 may allow a user to cycle through cell phone, media player, and PDA modes of operation.
  • the multi-functional device 100 operates as the selected device.
  • the programming related to the selected device is activated for use by the multi-functional hand-held device.
  • the programming may include reconfiguring the UI based on the selected device so that the inputs made by the user correlate to the device in use.
  • the functions of any physical buttons, switches or dials as well as soft buttons, switches or dials can be reconfigured to correspond to the selected device.
  • multi-functional hand-held device 100 need not completely independent. In many cases, it will be desirable to allow the multiple functionalities to interact with each other. For Example, a user may look up a telephone number of a contact in the PDA and pass this number to the phone to be dialed.
  • the form factor of a hand-held device is generally a device that can easily be held in one hand
  • a typical hand-held device includes a small display in an upper portion of the front surface of the device and input controls in a lower portion of the front surface of the device.
  • the device may additionally include controls and ports on the top, bottom, and side surfaces.
  • Prior art hand-held devices have typically had small displays which causes some dissatisfaction for users of these devices. It is generally preferred to have larger displays so that more information can be displayed or the information being displayed can be more easily viewable (e.g., larger text). In the case of video players and game players in particular, larger displays are much preferred over smaller ones.
  • a hand-held electronic device may be directed at one-handed operation or two-handed operation.
  • one-handed operation a single hand is used to both support the device as well as to perform operations with the user interface during use.
  • Cellular phones and media players are examples of hand-held devices are generally intended to can be operated solely with one hand.
  • a user may grasp the phone in one hand between the fingers and the palm and use the thumb to make entries using keys, buttons or a # joy pad.
  • one hand In two-handed operation, one hand is used to support the device while the other hand performs operations with a user interface during use or, alternatively, both hands support the device as well as perform operations during use.
  • PDA's and game players are examples of hand-held device that are typically operated with two hands.
  • the user may grasp the device with one hand and make entries using the other hand, for example, using a stylus.
  • the user In the case of a game player, the user typically grasps the device in both hands and make entries using either or both hands while holding the device.
  • Hand-held devices may have a variety different footprints or sizes.
  • the footprint is typically associated with how the device is going to be used. Devices such as PDAs are typically used with both hands and thus they tend to be larger. Alternatively, cell phone handsets are typically used with only one hand, and thus they tend to be smaller.
  • PDAs are typically used with both hands and thus they tend to be larger.
  • cell phone handsets are typically used with only one hand, and thus they tend to be smaller.
  • a media player is one example of a hand-held device that is sized for placement into a pocket of the user.
  • the user does not have to directly carry the device and therefore the device can be taken almost anywhere the user travels (e.g., the user is not limited by carrying a large, bulky and often heavy device, as in a laptop or notebook computer).
  • hand-held devices of the type disclosed herein have dimensions of about 5 in ⁇ 3 in ⁇ 1 in and perhaps about 4.1 in ⁇ 2.4 in ⁇ 0.75 in.
  • the footprint of a multifunctional hand-held device is substantially fixed by the intended primary use of the device, it becomes important for designers to determine an appropriate layout for the UI. For example, some devices may work better with a limited number of buttons and a large display while others may work better with a large number of buttons or a complete keyboard and a small display.
  • a preferred multi-functional hand-held device may be configured with a full screen display or a near full screen display.
  • a full screen display consumes substantially the entire front surface of the device.
  • the display may extend edge to edge or may fit within a small bezel of the housing at the edge of the device.
  • the full screen display may consumes 90% or more of a front surface of a housing for a hand-held electronic device.
  • the full screen display may have a variety of different configurations depending on the overall footprint of the device. If the device is wide, the full screen display may have a traditional aspect ratio of about 4:3. If the device is elongated, the full screen display may have an aspect ratio that is more panoramic such as 16:9.
  • the multi-functional hand-held device is preferably configured with a limited number of physical buttons. Because a limited number of physical buttons are provided, the hand-held device preferably uses a touch screen as the primary input device.
  • Touch screens are transparent touch sensing devices that are positioned over displays. They typically work in conjunction with a GUI presented on the display. For example, the GUI may present an on-screen button, and the touch screen may detect when a user presses the on-screen button (e.g., places their finger or stylus over the on-screen button). Touch screens and GUI's are described in greater detail below.
  • the hand-held device may be constructed with only cross-functional physical buttons, i.e., there are no buttons dedicated to individual devices. These type of buttons may include power buttons and hold switches. In another embodiment, the hand-held device may not include any physical buttons at all. In some embodiments, the physical buttons are limited to only the sides and back surface of the hand-held device. In other embodiments, the physical buttons of the hand-held device are limited to the upper and lower portion of the sides so that there are no buttons in the regions of the sides where a user would physically support the device (i.e., holding region). In still other embodiments, the physical buttons may be located on the front surface, but only in the bezel region surrounding the display. In some embodiments, the buttons may be located on only the top and bottom surfaces of the device.
  • FIG. 3 is a perspective view of a substantially full screen multi-functional hand-held device 120 with a limited number of buttons.
  • the front surface is used entirely for the display 122 .
  • the sides 126 are used for grasping the device 120 it may be preferred to leave the sides free from buttons to prevent accidental actions in the event a user inadvertently presses a button while supporting the device.
  • the top surface 128 and bottom surface 130 would not typically be used to hold the device, these surfaces are not ideal locations for buttons that are often actuated because it would be awkward to reach these buttons when operating the device with one hand.
  • the top surface 128 may be reserved for buttons that have limited action and generic functions that are cross-functional, for example, power and hold switches.
  • the top and bottom surfaces 128 and 130 are also well suited for placement of I/O and communication ports.
  • the top surface 128 may, for example, include a headset/microphone jack and an antenna, and the bottom surface 130 may include power and data ports.
  • buttons in the upper or lower regions of the side surfaces 126 may be placed out of the way of the grasping hand. This may be particularly well suited for elongated devices that are larger than the width of the grasping hand.
  • the hand-held device 120 includes a button 140 in the upper region on the side surface 126 of the hand-held device 120 . Because the button 140 is in the upper region, it tends to be out of the way of the grasping hand and therefore accidental activation is substantially eliminated.
  • the upper button may be configured to switch the functionality of the multi-functional device i.e., button 140 may be switch 110 of FIG. 2 . For example, by pressing the button 140 , a new device functionality is activated, and the current device functionality is deactivated.
  • the term button is used, it should be appreciated that the button 140 may correspond to a dial, wheel, switch and/or the like.
  • buttons it would be preferred, although not required, to limit the number of physical buttons to eight or fewer, and perhaps five or fewer.
  • the multi-functional hand-held device is preferably adaptable, i.e., the device is capable of changing its UI based on how the device is to be used. For example, if a cell phone functionality of the multi-functional device is to be used, the UI changes to accommodate the cell phone. Alternatively, if the PDA aspect of the multi-functional device is to be used, the UI changes to accommodate the PDA, etc. In essence, the multi-functional device is capable of reconfiguring its user interface based on the state or mode of the device.
  • Adaptability may be accomplished by reassigning the functions of the limited number of physical buttons for each device functionality. For example, a button may perform one function when a first device is activated and another when another device is activated. While this may work, it suffers from physical limitations (i.e., the number of buttons) and can be confusing to the user (who must keep track of different button meanings).
  • Alternatively adaptability may be accomplished by virtually incorporating the physical inputs for each functionality into the GUI in conjunction with a touch screen.
  • This allows the GUI to adapt to whatever device is selected, and the touch screen can receive inputs corresponding to the GUI.
  • the UI for the hand-held device adapts such that the multi-function device, in effect, becomes a particular device.
  • the GUI presents a set of virtual or soft controls that look similar to the physical controls typically used on a cell phone, such as a keypad, function buttons and possibly a navigation pad, for example.
  • the GUI 150 may be separated into a standard region 152 and a control region 154 .
  • the standard region 152 represents what would normally be displayed on the display 122 when using a particular device. That is, the standard GUI screens associated with the selected device are displayed in the standard region.
  • a main menu window with a set of icons
  • calendar address book or date book
  • the control region 154 virtually represents the physical controls that would normally be physically placed on a particular device. That is, the virtual controls that mimic the physical controls are displayed in the control region 154 .
  • the control region 154 may include virtual representations of a hand writing recognition area, a navigation pad and the standard function buttons.
  • the standard and control regions 152 and 154 can be positioned at any position on the display 122 (top, bottom, sides, center, etc.). For example, as shown in FIG. 5 , they may be positioned vertically relative to one another (one on top of the other) or as shown in FIG. 6 , they may be positioned horizontally relative to one another (side by side). These configurations can be used in either portrait or landscape modes.
  • the standard region 152 may be placed on one side and the control region may be placed on the opposite side.
  • Landscape orientation may, for example, facilitate one handed operation. In some cases, the side on which the controls are displayed is based on the handedness of the user.
  • the controls may be place on the right side for right-handed users, and the controls may be placed on the left side for left-handed users.
  • the controls may be placed on both sides as shown in FIG. 7 .
  • This arrangement is particularly well suited for game playing.
  • the amount of area dedicated to each portion may be widely varied.
  • the screen may be divided equally and in other cases one or the other portion constitutes a greater amount of the display.
  • the standard region 154 is maximized to increase the normal viewing area of the display.
  • the hand-held device loads the software for the selected functionality and configures the GUI 150 including the standard region 152 and the control region 154 .
  • the controls in the control region 154 can therefore be used to control whatever is being shown in the standard region 152 .
  • the control region 154 may even change in accordance with the needs of each displayed window for the particular device.
  • virtual controls 160 may be overlaid on top of the standard region 152 so that the standard region 152 can fully utilize the entire display 122 .
  • the virtual controls 160 may appear and disappear as needed.
  • the user may touch the screen and this may drive the device to display the controls over a portion of the display including whatever is already displayed. Examples of virtual controls that operate in this manner can be found in U.S. patent application Ser. No. 11/038,590, titled “Mode-Based Graphical User Interfaces for Touch Sensitive Input Devices,” filed on Jan. 18, 2005.
  • FIGS. 9-17 illustrate various examples of GUIs for different states or modes of the multi-functional device.
  • FIG. 9 is a diagram of a GUI 170 that is used in a PDA mode. As shown, the GUI is divided into a standard region 152 and a control region 154 . Located inside the control region 154 are a virtual handwriting pad 172 , four virtual buttons 174 and a virtual navigation pad 176 .
  • FIG. 10 is a diagram of a GUI 180 that is used in a cell phone mode. As shown, the GUI 180 is divided into a standard region 152 and a control region 154 . Located inside the control region 154 are a virtual keypad 182 , a virtual navigation pad 184 and two virtual buttons 186 .
  • FIG. 11 is a diagram of a GUI 190 that is used in a music player mode. As shown, the GUI 190 is divided into a standard region 152 and a control region 154 . Located inside the control region 154 are a virtual scroll wheel 192 and five virtual buttons 194 . Additional details on a virtual scroll wheel are provided in U.S. patent application Ser. No. 11/038,590, titled “Mode-Based Graphical User Interfaces for Touch Sensitive Input Devices,” filed on Jan. 18, 2005.
  • FIG. 12 is a diagram of a GUI 200 that is used in a video player mode. As shown, the GUI 200 is divided into a standard region 152 and a control region 154 . Located inside the control region 154 are a plurality of virtual buttons 202 . Alternatively, the controls may appears and disappears as needed since the video player is primarily used in conjunction with a full screen viewing mode.
  • FIG. 13 is a diagram of a GUI 210 that is used in a game player mode. As shown, the GUI 210 is divided into a standard region 152 and two control regions 154 A and 154 B on the sides of the standard region 152 .
  • the left side control region 154 A includes a navigation or directional pad 212
  • the right side control region includes four virtual buttons 214 (or vice versa depending on the users particular needs, left or right handed).
  • FIG. 14 is a diagram of a GUI 220 that is used in a camera mode. As shown, the GUI 220 is divided into a standard region 152 and a control region 154 .
  • the standard region 152 may represent the view finder.
  • Located inside the control region 154 are various buttons 222 including for example picture click, zoom, flash, etc.
  • a navigation pad 224 may also be included so that the pictures can be scrolled through or for menu navigation.
  • FIG. 15 is a diagram of a GUI 230 that is used in a GPS receiver mode. As shown, the GUI 230 is divided into a standard region 152 and a control region 154 . Located inside the control region 154 are various buttons 222 including for example zoom, pan, etc. A navigation pad 224 may also be included.
  • FIG. 16 is a diagram of a GUI 240 that is used in a hand top mode. As shown, the GUI 240 is divided into a standard region 152 and a control region 154 . Located inside the control region 154 is a virtual keyboard 242 .
  • FIG. 17 is a diagram of a GUI 250 that is used in a remote control mode. As shown, the GUI 250 is divided into a standard region 152 and a control region 154 . Located inside the control region 154 are various keys and buttons 252 associated with controlling a remote device such as a TV, DVD player, A/V amplifier, VHS, CD player, etc.
  • a remote device such as a TV, DVD player, A/V amplifier, VHS, CD player, etc.
  • the selection can come in a variety of forms. For example, the selection may be made via a main menu that includes soft buttons or icons that, when selected, activate the device functionality associated with the soft button. During activation, the GUI for that particular device is brought into view on the display (see FIGS. 9-17 ) and the software associated with the device is installed, loaded or activated. From that point on, the multi-functional device operates like the selected device.
  • FIG. 18 illustrated an exemplary main menu GUI 260 of a multi-functional device.
  • the GUI 260 includes icons/buttons 262 for launching each of the various device functionalities.
  • the main menu page 260 includes a PDA button 262 A, a cell phone button 262 B, a music player button 262 C, a game player button 262 D, a video player button 262 E, a GPS button 262 F, a remote control button 262 G, a camera button 262 H and a handtop button 262 I.
  • the various buttons 262 are virtual buttons.
  • the main page for the selected functionally e.g., as shown in FIGS. 9-17
  • the user simply selects a soft home button 264 located in the GUI of each device to return to the main menu page 260 , and thereafter selects the desired functionality in the main menu page 260 .
  • the selection of alternative functionalities may also be accomplished by flipping (or scrolling) through the various GUIs until the desired GUI is found.
  • the different GUIs may be incrementally brought into view page after page (or frame after frame) when a next (flip) command signal is generated (e.g., slide show effect).
  • the transition between pages may be widely varied. The transition may be from side to side, top to bottom or center to center. The transition may also include fading in and out, popping in and out, or enlarging and reducing.
  • the command signal may be generated by a physical or virtual button or wheel. Using a button, each press may cause a new page to be displayed. Using a wheel, a predetermined amount of rotation may cause a new page to be displayed.
  • the command signal may also be generated in a variety of other ways.
  • the command signal may also be generated by gestures initiated on the touch screen. For example, sliding a finger (or stylus) across the display may cause a new page to be displayed. If slid to the right, the next page may be displayed. If slid to the left, the previous page may be displayed.
  • the command signal may also be generated by 3D device gestures created when the entire hand-held device is moved spatially. By way of example, shaking the device may cause a new page to be displayed.
  • the command signal may also be generated by forces that are applied to the device. By way of example, squeezing the device may cause a new page to be displayed.
  • the command signal may also be generated by sensing the orientation of the device either relative to the ground, as sensed by accelerometers, or relative to a compass direction indicated by an internal compass. For example, if the device is at 0 degrees, a first page is displayed, at 90 degrees a second page is displayed, at 180 degrees a third page is displayed and at 270 degrees a fourth page is displayed.
  • the command signal may also be generated by monitoring a user's voice (i.e., voice recognition). If the user calls out “PHONE,” the page associated with the phone is displayed, if the user calls out “PDA,” the page associated with the PDA is displayed.
  • voice recognition i.e., voice recognition
  • the command signal may also be generated by monitoring incoming signals from other systems (whether transmitted wirelessly or via a cable). For example, if a call is received, the device may automatically configure the system as a phone. Alternatively, it may only present a control panel for taking or passing on the call.
  • the device may be configured to keep the various modes separate. That is, the device does not merge the functionality together (integrated layers and GUIs), but instead keeps them distinct from one another. In some cases, by keeping different functionalities distinct, user confusion may be reduced.
  • the user may be able to activate two or more device functionalities simultaneously.
  • the software for the multiple functionalities is activated simultaneously and the display operates in a split screen mode where the screen is parsed into different sections, each section including a particular device GUI. Generally this would require the GUI for each functionality to fit on the screen.
  • the multi-function mode may be selected in a variety of ways. In one implementation, when the user simultaneously touches two or more device icons, the device activates the multiple device functionalities and brings the appropriate GUIs into view on the screen.
  • the GUI control panels for each device functionality may be configurable by the user.
  • the user may design or customize his own UI for each device, and assign functions to the various elements of the UI.
  • the functions may include initiating commands, selecting an item, opening a file or document, launching a program, executing instructions, viewing a menu on the display screen, etc.
  • the parameters configurable by the user may include selecting the number and type of GUI elements (buttons) as well as the location of the GUI elements on the page.
  • the system may include a design palette that allows a user to review and/or customize the UI layout, i.e., the user may quickly and conveniently review preconfigured or default layout and make changes thereto. Once changed, the modified layout will be automatically saved and thereby employed to handle future events.
  • buttons there are a number of problems with current input devices for hand-held computers. For example, there is no realistic way to fit all the dedicated physical buttons that are needed for each device because more and more buttons must be fit into a small space. A related problem arises in that as more physical buttons are incorporated into the device, the buttons must get closer together thereby challenging the dexterity of the user. Furthermore, because physical buttons stick out they are often accidentally activated during normal use or while the device is transported e.g., in a user's pocket. Finally, large numbers of small physical buttons can be confusing to the user as well as aesthetically unpleasing.
  • the basic idea is to reduce the number of physical and mechanical input mechanisms (and therefore the amount of dedicated space needed on the surface of the device) and/or eliminate the physical and mechanical input mechanisms altogether.
  • the display of the electronic device can be maximized, or alternatively the size of the electronic device can be minimized.
  • the hand-held device can be configured so that it appears to only have a display and no dedicated physical input devices.
  • the alternative input means may also be hidden from view such that they are not visible to the user.
  • the hand-held electronic device may appear to only have a display with no visible buttons, switches, etc.
  • Such a device is more aesthetically pleasing (e.g., may comprise smooth surfaces with no breaks gaps or lines), and, in many cases, can be made smaller without sacrificing screen size and input functionality.
  • a touch screen in conjunction with GUI may be configured as the primary input arrangement for a hand-held device.
  • a touch screen is basically a transparent input panel positioned in front of the display.
  • a touch screen generates input signals when an object such as a finger or stylus touches or is moved across the surface of the touch screen.
  • touch screens allow a user to make selections and initiate movements in a GUI by simply touching the display screen via a finger. For example, a user may make a selection by pointing directly to a graphical object displayed on the display screen.
  • the touch screen recognizes the touch and position of the touch on the display and a controller of the hand-held electronic device interprets the touch and thereafter performs an action based on the touch event.
  • touch screen technologies including resistive, capacitive, infrared and surface acoustic wave.
  • a preferred touch screen for a multi-functional hand-held computer is a multipoint capacitive touch screen.
  • Such a touch screen comprises several independent and spatially distinct sensing points, nodes, or regions that are positioned throughout the touch screen.
  • the sensing points are dispersed about the touch screen with each sensing point representing a different position on the surface of the touch screen.
  • the sensing points may be positioned in a grid or a pixel array where each sensing point is capable of generating a signal.
  • a signal is produced each time an object is positioned over a sensing point. When an object is placed over multiple sensing points or when the object is moved between or over multiple sensing point, multiple signals can be generated.
  • the sensing points generally map the touch screen plane into a coordinate system such as a Cartesian coordinate system or polar coordinate system.
  • a coordinate system such as a Cartesian coordinate system or polar coordinate system.
  • One example of such a touch screen is disclosed in U.S. patent application Ser. No. 10/840,862, titled “Multipoint Touch Screen” filed on May 6, 2004.
  • a hand-held electronic device may also incorporate one or more touch sensitive surfaces of the device housing itself that provide either a larger surface for tracking touch inputs or smaller dedicated areas such as touch buttons for performing dedicated functions.
  • Such surfaces may be located on any surface of the housing, any side of the housing, any portion of any side of the housing or at dedicated locations on the surface of the housing.
  • the touch regions may be located on the sides or back surface of the housing, and may even be located at the bezel located at the front surface of the housing. In all of these cases, a large portion of the front surface of the housing is saved for the display so that the viewing area of the hand-held electronic device can be maximized.
  • the touch sensitive surfaces of the housing may take the form of one or more touch panels that are positioned within the housing.
  • the touch sensitive surface may be alternatively or additionally be provided directly by the housing. That is, the touch sensing components may be integrated or incorporated into or disposed underneath the housing such that the housing itself is the touch sensing device (rather than using a separate touch panel).
  • a touch sensitive housing recognizes the touch and position of a touch on the surface and a controller of the hand-held electronic device interprets the touch and thereafter performs an action based on the touch event.
  • Touch surfaces are constructed in basically the same manner as a touch screen, except the surfaces need not be substantially transparent.
  • the touch sensitive housing may generally correspond to the touch sensitive housing described in greater detail U.S. patent application Ser. No. 11/115,539, titled “Hand-Held Electronic Device with Multiple Touch Sensing Devices,” filed Apr. 26, 2005.
  • a hand-held multi-functional electronic device may also include a display actuator, which is an input device that mechanically uses the display of the hand-held device to provide inputs into the device (rather than electrically as with a touch screen).
  • the display actuator may be used separately or in combination with the touch screen.
  • the display actuator may include a movable display that causes one or more input signals to be generated when moved. The input signals then can be used to initiate commands, make selections, or control motion in a display.
  • the movable display may be configured to translate, slide, pivot, and/or rotate relative to the frame.
  • the display is typically movable relative to a frame or housing that movably supports the display in its various positions. In some cases, the display is movably coupled to the frame, and in other cases the frame movably restrains a floating display.
  • the input signals may be generated by movement indicator(s) that monitor the movements of the display and produces signals indicative of such movements.
  • the detection mechanism may, for example, one or more switches, sensors, encoders, and/or the like. Any suitable mechanical, electrical and/or optical switch, sensor or encoder may be used. For example, tact switches, force sensitive resistors, pressure sensors, proximity sensors, infrared sensors, mechanical or optical encoders and/or the like may be used.
  • the movement indicators may be placed underneath the display or at the sides of the display as appropriate. Alternatively or additionally, these movement indicators may be attached to the display or some component of the display.
  • the multi-functional hand-held electronic device described above may further include force or pressure sensing devices such as a force sensing display or housing.
  • a force sensitive display causes one or more input signals to be generated when pressure is exerted on the display screen of the device.
  • the input signals can be used to initiate commands, make selections, or control motion in a display.
  • Such displays generally provide a slight amount of flex (unnoticeable to the user) so, that any forces exerted thereon can be measured by a force detection arrangement generally provided underneath the display.
  • the force detection arrangement monitors the forces exerted on the display and produces signals indicative thereof.
  • the force detection arrangement may include one or more force sensors such as force sensitive resistors, force sensitive capacitors, load cells, pressure plates, piezoelectric transducers, strain gauges, etc.
  • the force sensors may be attached to the back of the display or to a structural platform located within the housing of the device. When a force is applied to the display, it is transmitted through the display to the force sensor located underneath the display.
  • FIG. 19 is a side elevation view, in cross section, of a hand-held device 300 incorporating a force sensitive display 302 .
  • the force sensitive display 302 includes a display 308 and one or more force sensors 310 disposed underneath the display 308 (between the display and a structural platform 306 ).
  • the force sensitive display 302 includes a plurality of sensors 310 that are laid out in an array.
  • the sensors 310 may be positioned side by side in rows and columns.
  • the force sensors 310 measure the amount of force being applied to the display and when a desired force threshold is reached a control signal is generated.
  • an elastomer 312 is placed between the display and the structural platform to help transmit the force being exerted on the surface of the display to the force sensors disposed below the display.
  • Force sensing may be provided in conjunction with a touch screen to differentiate between light and hard touches.
  • the determination of whether a touch is a light touch or a hard touch may be made by monitoring the force with the force sensors and comparing the force to a predetermined threshold. When the force threshold is not exceeded, the touch is considered a light touch. When the force threshold is exceeded, the touch is considered a hard touch.
  • Each type of touch may be used to control different aspects of the device.
  • Light touches may be associated with passive events such as navigation (e.g., cursor control scrolling, panning, zoom, rotation, etc.) and hard touches may be associated with active events such as selections or commands (e.g., button click).
  • FIG. 20 illustrates an input device 320 that combines touch sensing and force sensing devices to provide x, y and z components when touched.
  • the touch sensing device provides position sensing in the x and y directions
  • the force sensing device provides force sensing in the z direction. These devices cooperate to output x, y location and z pressure information whenever there is a touch on the touch surface.
  • FIG. 21 is a side elevation view of an I/O device 330 that combines a display 332 with touch screen 334 and a force sensing mechanism 336 .
  • the touch screen 334 provides high resolution touch locations, and the force sensing mechanism 336 provides a measure of where the force is coming from as well the total force.
  • Touch screen 334 is disposed over the display 332 , and the force sensing mechanism 336 is disposed below the display 332 although other arrangements are possible.
  • Force sensing mechanism 336 may also be widely varied.
  • the force sensing mechanism 336 is based on capacitance, and more particularly, self capacitance.
  • the illustrated force sensing mechanism 336 is formed from various layers including an electrode layer 338 , an elastomer layer 340 and a conductive layer 342 .
  • Electrode layer 338 includes a plurality of spatially separated electrodes 339 that are positioned across the bottom of the display 332 . Electrodes 339 are typically positioned in an array of rows and columns although often configuring are possible. Any number of electrodes may be used.
  • Elastomer layer 340 includes one or more elastic members 341 positioned between the electrode layer 338 and the conductive layer 342 .
  • Elastic members 341 allow the display 332 to move inwardly with a limited amount of displacement.
  • elastic members 441 are silicone patches with a thickness of about 0.2 mm.
  • Conductive layer 342 typically takes the form of a grounded metal plate 343 .
  • a capacitive circuit is formed between each of the electrodes 339 and the grounded metal plate 343 .
  • the force being exerted thereon causes the display 332 to displace inwardly against the elastic members 341 compressing the elastic members. This causes a change in the capacitance between the electrodes 339 and the metal plate 343 .
  • This change in capacitance is sensed by a control circuit operatively coupled to each of the electrodes 339 . Capacitance sensing circuits are disclosed in the various incorporated references.
  • FIG. 22 is a side elevation view of an input device 350 that may be positioned over a display.
  • the input device 350 combines touch sensing and force sensing into a single device. In this embodiment, both the touch sensing and force sensing is provided by mutual capacitance.
  • the input device 350 is formed from various layers including a top drive layer 352 , a middle sense layer 354 , and a bottom drive layer 356 .
  • the middle sense layer 354 is positioned on an elastomer layer 358 disposed between the middle sense layer 354 and the bottom drive layer 356 .
  • the top and bottom drive layers 353 and 356 include a plurality of spatially separated lines in rows and the middle sense layer 354 includes a plurality of spatially separated lines in columns.
  • the top and middle layers 352 and 354 therefore form a grid
  • the bottom and middle layers 356 and 354 form a grid.
  • the lines on the top layer 352 are scanned, and thereafter the lines on the bottom layer 356 are scanned (or vice versa).
  • the mutual capacitance measured between the top drive layer 352 and the middle sense layer 354 provide the x and y location of the touch.
  • the mutual capacitance measured between the bottom drive layer 356 and the middle sense layer 354 provide the amount of force of the touch. This particular arrangement provides a full image of force superimposed on a full image of touch.
  • the input device including the touch layers and the force layers may be operated similarly to the methods described in U.S.
  • the hand-held device may also include a force sensitive housing.
  • the force sensitive housing provides inputs when forces are applied to the housing of the hand-held device.
  • a force sensitive housing is similar to a force sensitive screen in that.
  • the housing provides a slight amount of flex (possibly unnoticeable to the user) so that any forces exerted thereon can be distributed to a force detection arrangement located within the housing.
  • the force detection arrangement monitors the forces on the housing and produces signals indicative thereof.
  • the force detection mechanism may include one or more force sensors disposed with in the housing such as force sensitive resistors, force sensitive capacitors, load cells, pressure plates, piezoelectric transducers, strain gauges and/or the like. When a force is applied to the housing (squeezing or pushing on the housing), it is transmitted through the housing to the force sensor located within the housing.
  • the force sensitive portions of the housing may be located on any surface of the housing, any side of the housing, any portion of any side of the housing or at dedicated locations on the surface of the housing,
  • the sides of the housing are ideal places for implementing a squeeze feature. This is because the users fingers are typically positioned on one side of the device and thumb on the other and therefore the hand may easily squeeze the sides via a pinching action. Because it is so convenient to activate the squeeze feature, special care must be taken when designing the squeeze feature so that it will not be accidentally activate during normal use. Thus the device needs to be able to differentiate between light and hard squeezes. If the squeeze feature is implemented using force sensitive resistors (FSRs) which exhibit, a decrease in resistance with an increase in force applied to the active surface a comparator circuit can be used to output a signal to indicate activation when a preset force threshold is reached.
  • FSRs force sensitive resistors
  • FIG. 23 is a side view, in cross section, of a hand-held device 370 that incorporates a squeeze feature.
  • the device 370 includes a housing 372 and a support platform 374 inside the housing 372 . Between the support platform 374 and the inner surface of the housing 372 are a pair of force sensors 376 .
  • the housing 372 flexes inwardly under the pressure. This causes the force sensors 376 to be sandwiched between the housing 372 and the support platform 374 .
  • the force sensors 376 measure the amount of force being exerted and when a desired force threshold is reached, the force sensors 376 generate a control signal. For example, as a result of being sandwiched, a force resistive sensor may exhibit a reduced resistance and when a desired threshold is reached, a control signal is generated.
  • the force sensitive housing may be is provided in conjunction with a touch sensitive housing as discussed above.
  • the hand-held electronic device may also include a motion actuated input device.
  • the motion actuated input device provides inputs when the hand-held device is in motion or is placed in a certain orientation.
  • a motion actuated input device typically includes a motion sensor, such as an accelerometer, that monitors the motion of the device along the x, y, and/or z axis and produces signals indicative thereof.
  • the motion sensor may, for example, include an accelerometer.
  • the motion sensor could be an orientation sensor, such as an electronic compass, that allows the device to determine its orientation in a generally horizontal plane.
  • the motion sensors may be attached to the housing or to some other structural component located within the housing of the device. When motion is applied to the device (gesturing, shaking, hand waving, etc), it is transmitted through the housing to the motion sensor.
  • the intended motion information typically must be separated from the other motion information to produce an accurate command signal.
  • large scale movements such as shaking the device will produce primarily low frequency information.
  • small scale movements such as vibrations, primarily produce high frequency information.
  • the high frequency information can be filtered out thereby leaving only low frequency information indicative of the large scale movements (e.g., shaking).
  • the filtered information can then be converted into a control signal.
  • FIG. 24 is a side view, in cross section, of a hand-held electronic device 380 .
  • the hand-held device 380 includes an accelerometer 382 that is attached to a housing 384 of the hand-held device 380 .
  • the accelerometer 382 recognizes the motion and a controller of the hand-held electronic device 380 interprets the motion and thereafter performs an action based on the motion event.
  • the hand-held device may include some number of surface mounted actuators.
  • these actuators are generic to each of the integrated devices. That is, their meaning is the same regardless of what device functionality is activated. It is also preferred that the surface mounted actuators be placed on surfaces other than the front surface, which houses the viewing region of the display, although this is not required.
  • the hold switch may be configured to activate and deactivate the primary input means, e.g., the touch screen. This permits a user to prevent unwanted entries, for example, when the device is stored inside a user's pocket.
  • the hold switch may be placed on the top surface out of the way of the grasping hand, but in a position for easy access (as opposed to the bottom surface). The hold switch can not only deactivate the touch screen but also mechanical actuators and other input and other input devices.
  • a power switch When the power switch is turned on, the device is powered up and ready to go. When the power switch is turned off, the device is shut down.
  • the power switch may be placed on the top surface out of the way of the grasping hand, but in a position for easy access (as opposed to the bottom surface).
  • the navigation pad is typically included with many hand-held devices.
  • the functionality of the navigation pad may be changed according to the current operating mode of the device.
  • the directional keys may be assigned, play/pause, next, previous, and volume up and down.
  • Other assignable buttons may also be included on the device.
  • Still another useful mechanical actuator is a switching actuator.
  • the switching actuator may be configured to change the functionality of the device, i.e., by activating the switching actuator the functionality or state of the device switches from one mode to another.
  • the switching actuator may be widely varied.
  • the switching actuator may be a dial or wheel.
  • the device By incrementally rotating the wheel, the device is incrementally switched from one device to the other (generally in some predetermined order). A full rotation of each device generally cycles through the entire group of integrated devices.
  • the wheel or dial may for example operate like a scroll wheel.
  • the switching wheel may be placed in the upper region of the sides of the device. By placing the wheel here, a users thumb may be used to easily rotate the wheel. For example, the users thumb may be extended from the grasping action so that the wheel can be rotated.
  • the switching actuator may be a button.
  • the switching button By repetitively pressing on the button, the device is switched from one device to another (generally in some predetermined order).
  • the switching button may be placed in the upper region of the sides of the device. By placing the button here, a users thumb or index finger may be used to easily press the button.
  • the hand-held device may also include any commercially available touch pad.
  • touch pads may be found in U.S. patent application Ser. No. 10/188,182, titled “Touch Pad for Handheld Device,” filed on Jul. 1, 2002, U.S. patent application Ser. No. 10/722,948, titled “Touch Pad for Handheld Device,” filed on Nov. 25, 2003, and U.S. patent application Ser. No. 10/643,256, titled “Movable Touch Pad with Added Functionality,” filed on Aug. 18, 2003.
  • the hand-held device may include a scroll wheel.
  • Scroll wheels can be used in each functionality to scroll through a window.
  • the hand-held device may also include a microphone that picks-up audio sounds.
  • the microphone may be used in conjunction with a cell phone to transmit sounds, such as the user's voice.
  • the microphone may also be used to record sounds or enter voice commands into the hand-held device.
  • voice recognition software the hand-held device may be able to recognize voice commands and generated control signals associated therewith.
  • the microphone may be placed in the bottom surface of the hand-held device or possible in the front lower bezel. This particular configuration is well suited for picking up a user's voice during a phone call.
  • a hand-held electronic device may also include an image sensor and lens related components so that the hand-held device can operate like a camera.
  • the image sensor may, for example, include a charge coupled device (CCD) camera.
  • CCD charge coupled device
  • a hand-held electronic device may be designed to recognize touch gestures applied to a touch screen and/or touch sensitive surface of the housing and thereby control aspects of the hand-held electronic device.
  • Gestures are a stylized interaction with an input device that is mapped to one or more specific computing operations.
  • the gestures may be made through various hand and finger motions.
  • Gestures generally comprise a contact chord e.g., one or more fingers, and a motion associated with the chord.
  • the gestures may be made with a stylus.
  • the input device i.e., touch screen and/or touch sensitive surface
  • receive the gestures receive the gestures and a controller of the hand-held electronic device executes instructions to carry out operations associated with the gestures.
  • the hand-held electronic device may include a touch gesture operational program, which may be part of the operating system or a separate application.
  • the gesture operation program generally includes a set of instructions that recognizes the occurrence of gestures and informs one or more software agents of the gestures and/or what action(s) to take in response to the gestures.
  • gestures that can be used are disclosed in greater detail in U.S. patent application Ser. No. 10/903,964, titled “Gestures for Touch Sensitive Input Devices,” filed on Jul. 30, 2004, and U.S. patent application Ser. No. 11/038,590, titled “Mode-Based Graphical User Interfaces for Touch Sensitive Input Devices,” filed on Jan. 18, 2005.
  • the hand-held electronic device may be designed to recognize 3-D spatial gestures measured by an accelerometer and to control aspects of the hand-held electronic device based on the 3-D spatial gestures.
  • Spatial gestures are stylized motions of the device itself that are mapped to one or more specific computing operations.
  • the 3-D spatial gestures may be made through various hand and arm motions, such as for example shaking, waving and the like.
  • the accelerometer measures the motion related to the 3-D spatial gestures and a controller recognizes the motion as a 3-D spatial gesture and thereafter executes instructions to carry out operations associated with the 3-D spatial-gestures.
  • the hand-held electronic device may include a 3-D spatial gesture operational program, which may be part of the operating system or a separate application.
  • the gesture operation program generally includes a set of instructions that recognizes the occurrence of gestures and informs one or more software agents of the gestures and/or what action(s) to take in response to the gestures.
  • One exemplary 3-D gesture is shaking the device. Shaking can cause the hand-held device causes the device to shift between modes or states. Alternatively, shaking the hand-held device can cause a selection event to occur. For example, in the case of a media player, shaking the device may cause the device to randomly select a picture or song in group of pictures or songs. Alternatively, shaking the device may cause the device to select the next picture or song in a sequence of songs.
  • Other gestures may include translating or rotating the device. Translating the hand-held device (while it is face up) from side to side may be used to initiate panning or scrolling in the device, or moving the device up and down (while it is face up) may be used to initiate zooming. Rotating the device may be used to cause the device to change modes or states.
  • the orientation of the device may correspond to a particular mode. For example, a first mode may be associated with 0 degrees, a second mode may be associated with 90 degrees, a third mode may be associated with 180 degrees and a fourth mode may be associated with 270 degrees. In all these cases, the device can be configured to keep the displayed portion upright as the device is turned. That is, it maintains an upright image no matter what orientation the device is in.
  • the 3-D spatial gestures may even be based on more complex motions such as sign language, writing motions, etc.
  • the hand-held device may be configured to receive simultaneous inputs from different inputs devices, and perform actions based on the multiple simultaneous inputs.
  • the inputs that may be combined to produce new commands may be selected from voice, 2-D touch gestures, 3-D spatial gestures, actuators, etc. For example, this feature may be helpful when making calls via voice selection.
  • a caller may verbalize “TOM,” which causes the device to dial “TOM” phone number. If “TOM” has multiple phone numbers, a user may combine the verbal command “TOM” with a 3-D spatial gesture such as shaking to select Tom's second phone number.
  • force sensing in conjunction with touch sensing facilitates two distinct types of interactions, light touches and hard touches.
  • Light touches may be used to perform passive actions such as navigating through content and content manipulation generally without causing a major event to occur. Examples of passive events include moving a cursor, scrolling, panning, etc.
  • Hard touch interactions may be used to select on screen buttons or initiate commands (e.g., causes a significant change to occur).
  • FIG. 25 is a diagram of a touch method 400 for implementing this technique.
  • the method 400 begins at block 402 where one or more touches are detected.
  • the touches include not only x any y components but also z components.
  • the x and y components may be supplied by a touch sensing device such as touch screen, touch pad, or touch housing.
  • the z component may be provided by force sensors or display actuators located behind the touch surface of the touch sensing device.
  • the method proceeds to block 404 where a determination is made as to whether the touch is a light or hard touch. The determination is generally based on the force or pressure of the touch (z component). For example, if the force of the touch is smaller than a predetermined threshold then the touch is considered a light touch and if the force of the touch is larger than the predetermined threshold then the touch is considered a hard touch. If it is determined that the touch is a light touch, the method proceeds to block 406 where a passive action associated with the touch is initiated. If it is determined that the touch is hard touch, an active action associated with the touch is performed (block 408 ).
  • the touch method may additionally include a block where the one or more touches are classified as a primary touch or a secondary touch.
  • Primary touches are touches that are intended to cause an action while secondary touches are touches that are not intended to cause an action.
  • Gestures are examples of primary touches while a thumb positioned over the touch area to hold the device is an example of a secondary touch.
  • the z pressure exerted on a touch sensing device can be combined with the x and y locations of the touch to form a new touch vocabulary.
  • touch vocabularies have only included x and y locations, not z pressure.
  • a proposed touch vocabulary includes variety of variables including the UI mode, the force of the touch (e.g., light or hard), the number of fingers used, whether or not there is any movement during the touch, the duration of the touch, and the touch location, all or some of which can be combined to form a variety of behaviors and user feedback.
  • the UI mode is generally related to the mode or state of the device.
  • Each device includes a variety of states and each state may require a different UI mode.
  • a media player (a mode) may, for example, include a set of hierarchical layers (states) with each layer requiring a different UI.
  • the force of the touch may, for example, be described as light or hard.
  • a light touch may occur when a user lightly touches the surface of the touch surface, i.e., the finger hovers on top of the surface and is primarily moved in the x and y directions.
  • a hard touch may occur when a user presses on the touch surface with a certain amount of force, i.e., the finger is primarily moved in the z direction against the touch surface.
  • Motion during the touch is used to describe whether the finger has remained stationary during a touch event or has substantially moved in the X-Y plane (e.g., translation, rotation, etc.).
  • the motion may be described as none at all or a swipe or twist in some particular direction.
  • the swipe may be up, down, right, left, or some combination thereof, and the twist may be clockwise or counterclockwise.
  • Duration is defined as the amount of time the finger stays at any one point. Duration may be variable or it may include states such as short and long.
  • the touch location may be a random point or a specific location such as an onscreen button.
  • FIG. 26 is an additional touch method 500 implementing this technique.
  • the method begins at block 502 when one or more touches are detected. Thereafter, in block 504 , the UI mode is determined. In block 506 , a determination is made as to whether the touches are light touches or hard touches. Alternatively, blocks 502 and 504 could be reversed, effectively resulting in an instance of the touch method for each mode. In block 508 , the number of distinct touches (e.g., fingers) is determined. In block 510 , a determination is made as to whether the touches are stationary or in motion. In block 512 , the duration of the touches is determined. In block 514 , the locations of the touches are determined.
  • the method proceeds to block 516 where an action is performed based on the UI mode, the pressure of the touch, the number of touches, whether or not the touch is moving, the duration of the touch, and the touch location.
  • the actions may be passive or active depending on the values of each characteristic.
  • FIGS. 27 A-E One example of a touch vocabulary associated with a music player is shown in FIGS. 27 A-E.
  • the principle output of a hand-held electronic device is typically a display.
  • the display provides visual information in the form of text, characters or graphics.
  • the display is usually a flat panel device although other types of displays may be used.
  • the display may be a liquid crystal display (LCD) such as a character LCD that is capable of presenting text and symbols or a graphical LCD that is capable of presenting images, video, and graphical user interfaces (GUI).
  • LCD liquid crystal display
  • GUI graphical user interfaces
  • the display may correspond to a display based on organic light emitting diodes (OLED), or a display that is based on electronic inks.
  • OLED organic light emitting diodes
  • the display may be configured to substantially fill the front surface of the housing.
  • the display may extend from one edge of the housing to the opposite edge of the housing, the housing may include a small bezel that surrounds the edges of the display. In either case, the display makes up a substantial portion of the front surface of the hand-held electronic device, thereby eliminating any space for buttons or switches associated with a conventional user interface.
  • the display may also act like an input device.
  • a touch screen may be positioned over the display, and/or sensors may be disposed underneath the display to sense when the device is pressed or otherwise moved.
  • the small form factor of hand-held devices requires central placement of the input interfaces to permit operation while being carried around and used by the hand.
  • the display region provides a central location, which can be accessed by both the left and right hands.
  • the display region may be formed by more than one display.
  • the display region may be formed by a pair of displays that are side by side or one on top of the other.
  • a first display may be used to present the standard screen and a second display may be used to present the control screen as described above with reference to FIGS. 5-8 .
  • a first display may be a conventional display while the second display may be a display actuator.
  • a first display may comprise of a first type and a second display may be of a second type.
  • the first display may be an LCD while the second display may be a display based on electronic inks.
  • the decision to use different types may be based on the fact that one of the displays may be dedicated to a standard viewing area while another may be dedicated to a control area. Because these two areas require different resolutions, a higher resolution screen may be used in the standard area while a lower resolution screen may be used in the control area. Alternatively or additionally, one of the displays may be selected to reduce battery consumption especially in the control area where lower resolution is acceptable. In some cases, although two different displays are used, the images displayed thereon may be combined to form a single unified image.
  • a hand-held device may also include a speaker.
  • Speakers are components that accept electronic signals representing audio information from an amplifier and converts them into sound waves.
  • the speakers may be used to listen to music in conjunction with a music player functionality or to listen to an incoming call in conjunction with a cell phone functionality.
  • the speaker may be placed on the top surface or possible on the front top bezel of the hand-held device. This arrangement works particularly well when the device is used as a cell phone.
  • a hand-held device may also include one or more indicators that provide user feedback or indicate events associated with the device.
  • the events may relate to signals, conditions or status of the device.
  • the indicators provide status of battery life or alert a user when there is an incoming call.
  • the indicators which include light sources such as light emitting diodes (LED), are typically illuminated when an event occurs, and not illuminated when the event is stopped.
  • the indicator may turn on and off (blink) or cycle with increasing or decreasing intensity, and in some cases may even change colors in order to provide more detailed information about the event that is being monitored.
  • the indicators may be conventional indicators that typically include a small clear plastic insert, which is located in front of the LED, and which is inserted within an opening in the housing thus causing it to exist at the surface of the housing.
  • the LED itself may also be placed in the opening in the housing rather than using an insert.
  • the indicator can be configured not to break the surface of the housing.
  • the light source is disposed entirely inside the housing, and is configured to illuminate a portion of the housing thereby causing the housing to change its appearance, i.e., change its color.
  • the hand-held device may include speakers or buzzers to give audio feedback to the user. These may work similarly to the indicators described above, or they may be used to enhance the feel of actuating a GUI element such as a soft button or scroll wheel.
  • the speaker may be configured to output a “clicking” noise when a user presses on a virtual button, or rotates a virtual scroll wheel. This particular feature enhances the user experience and makes the virtual UI feel more like a physical UI.
  • the hand-held device may also include a haptics mechanism.
  • Haptics is the science of applying tactile sensation and control to interaction with computer applications. Haptics essentially allows a user to feel information, i.e., signals are sent to the hand.
  • the haptics mechanisms may be widely varied. They may include motors, vibrators, electromagnets, etc., all of which are capable of providing force feedback in the form of vibration or shaking.
  • the haptics mechanisms may work similarly to the indicators described above (alert), or they may be used to enhance the feel of actuating a GUI element such as a soft button or scroll wheel.
  • the haptics mechanism may be configured to vibrate when a user presses on a virtual button, or rotates a virtual scroll wheel.
  • Haptics may also be used simultaneously with onscreen actions. For example, during movies or game playing, the haptics mechanism can simulate the action being displayed. For example, the haptics mechanism may provide force feedback in the form of vibration when a car explodes during a movie or game.
  • the haptics mechanism may be located in the region of the display and further underneath the display so as to provide force feedback directly underneath the user action.
  • multiple haptics mechanisms may be used regionally across the display to further enhance the feel. It is generally believed that the closer the vibration is to the user action, the greater the haptics effect.
  • the haptics mechanisms are spaced out in an array underneath the display. That is, they are spatially separated and placed at different locations. By way of example, they may be positioned in a 2 ⁇ 2, 2 ⁇ 4, 4 ⁇ 4, 4 ⁇ 8, 8 ⁇ 8 array and so on underneath the display.
  • Audio and/or tactile feedback may be used to alert a user that a user input has been made.
  • the haptics may provide force feedback in the form of vibration and the speaker may provide audio feedback in the form of a click.
  • the tactile audio feedback can be used in conjunction with an input event including touch events, motion events, squeeze events.
  • the feedback may provide information so that the user knows that they actually implemented an input (simulates the audio and tactile feel of a button or switch).
  • the feedback is tied to the level of force being applied to the force sensing devices. For example, when a certain force threshold is reached, the audio feedback device may create a “click” on the press and a “clock” on the release.
  • the force threshold used may be similar to the threshold used to determine whether a touch is a light or hard touch.
  • the “click” and “clock” may be used to simulate a button click when a hard touch is made.
  • the hand-held device may also include one or more connectors for receiving and transmitting data to and from the device.
  • the device may include one or more audio jacks, video jacks, data ports, docking ports, etc.
  • the hand-held device may also include one or more connectors for receiving and transmitting power to and from the hand-held device.
  • the hand-held device may include a headphone/microphone jack and a data port.
  • the jack is capable of receiving a speaker and/or microphone plug so that audio may be input to and output from the device.
  • the data port is capable of receiving a data plug/cable assembly configured for transmitting and receiving data to and from a host device, such as a general purpose computer (e.g., desktop computer, portable computer).
  • the data port may be used to upload or download data to and from the hand-held device.
  • Such data may include songs and play lists, audio books, e-books, photos, address books, documents, appointments, etc. into the hand-held device.
  • the data port may be a PS/2 port, serial port, parallel port, network interface port, USB port, Firewire port etc.
  • the hand-held device may also include a power port that receives a power plug/cable assembly configured for delivering powering to the hand-held device.
  • the data port may serve as both a data and power port by employing either standard or proprietary connector
  • the device To send and receive data wirelessly, the device generally requires a transmitter, a receiver (or a transceiver) and some sort of antenna.
  • the wireless communication link may correspond to Bluetooth, WiFi (802.11), IR (infrared), etc.
  • the antenna may be fully contained within the device or they may extend outside the device.
  • the antenna may take a variety of forms depending on the frequency to be used, etc. For example be a rugged rubber duck that consists of a coiled up element encased in rubber. Alternatively, the antenna may be printed on a circuit board within the device.
  • the hand-held device may also include a radio transceiver for communications via a cellular network or a GPS receiver.
  • a hand-held electronic device may be configured to actively look for signals in the surrounding environment, and change its mode based on the signal. That is, the device tries to match the mode with the signal. If the device receives a phone signal over the cellular network, the device may turn into a phone, i.e., the phone mode is activated or brought forward relative to the other modes. If a device receives an email, the device may turn into an email terminal. As another example, when a user walks into a home theater room, the device may sense signals from the media control unit and turn itself into a remote control including functionality to control the various devices of the home theater (TV, amp, DVD, lighting).
  • the device may sense signals, which are being broadcast in physical stores, and turn itself into a device that is well suited for that store. For example, in a bank, the device may change into a calculator or bring a money program into view, or in a grocery store, the device may turn into a money payment device or bring a grocery list into view.
  • the hand held device may additionally include one or more of the following hardware components: a controller (e.g., microprocessor, DSP, A/D, D/A, converters, codes), memory (e.g., RAM, ROM, solid state (flash), hard disk (micro-drive)), storage (SD card slots, mini-DVD), battery (e.g., lithium ion), etc.
  • a controller e.g., microprocessor, DSP, A/D, D/A, converters, codes
  • memory e.g., RAM, ROM, solid state (flash), hard disk (micro-drive)
  • storage SD card slots, mini-DVD
  • battery e.g., lithium ion
  • FIG. 28 is a block diagram of an exemplary hand-held device 600 .
  • the hand-held device 600 typically includes a controller 602 (e.g., CPU) configured to execute instructions and to carry out operations associated with the hand-held device. For example, using instructions retrieved for example from memory, the controller 602 may control the reception and manipulation of input and output data between components of the hand-held device 600 .
  • the controller 602 can be implemented on a single chip, multiple chips or multiple electrical components.
  • various architectures can be used for the controller 602 , including dedicated or embedded processor, single purpose processor, controller, ASIC, etc.
  • the controller may include microprocessors, DSP, A/D converters, D/A converters, compression, decompression, etc.
  • the controller 602 together with an operating system operates to execute computer code and produce and use data.
  • the operating system may correspond to well known operating systems such as OS/2, DOS, Unix, Linux, and Palm OS, or alternatively to special purpose operating system, such as those used for limited purpose appliance-type devices.
  • the operating system, other computer code and data may reside within a memory block 604 that is operatively coupled to the controller 602 .
  • Memory block 604 generally provides a place to store computer code and data that are used by the hand-held device.
  • the memory block 604 may include read-only memory (ROM), random-access memory (RAM), hard disk drive (e.g., a micro drive), flash memory, etc.
  • the hand-held device may include a removable storage device such as an optical disc player that receives and plays DVDs, or card slots for receiving mediums such as memory cards (or memory sticks). Because the form factor of the hand-held device is small, the optical drive may only be configured for mini DVDs.
  • a removable storage device such as an optical disc player that receives and plays DVDs, or card slots for receiving mediums such as memory cards (or memory sticks). Because the form factor of the hand-held device is small, the optical drive may only be configured for mini DVDs.
  • the hand-held device 600 also includes various input devices 606 that are operatively coupled to the controller 602 .
  • the input devices 606 are configured to transfer data from the outside world into the hand-held device 600 .
  • the input devices 606 may correspond to both data entry mechanisms and data capture mechanisms.
  • the input devices 606 may include touch sensing devices 608 such as touch screens, touch pads and touch sensing surfaces, mechanical actuators 610 such as button or wheels or hold switches ( 611 ), motion sensing devices 612 such as accelerometers, force sensing devices 614 such as force sensitive displays and housings, image sensors 616 , and microphones 618 .
  • the input devices 606 may also include a clickable display actuator 619 .
  • the hand-held device 600 also includes various output devices 620 that are operatively coupled to the controller 602 .
  • the output devices 620 are configured to transfer data from the hand-held device 600 to the outside world.
  • the output devices 620 may include a display 622 such as an LCD, speakers or jacks 624 , audio/tactile feedback devices 626 , light indicators 628 , and the like
  • the hand-held device 600 also includes various communication devices 630 that are operatively coupled to the controller 602 .
  • the communication devices 630 may, for example, include both wired and wireless connectivity selected from I/O ports 632 such as IR, USB, or Firewire ports, GPS receiver 634 , and a radio receiver 636 .
  • the hand-held device 600 also includes a battery 650 and possibly a charging system 652 .
  • the battery may be charged through a transformer and power cord or through a host device or through a docking station. In the cases of the docking station, the charging may be transmitted through electrical ports or possibly through an inductance charging means that does not require a physical electrical connection to be made.
  • the various aspects, features, embodiments or implementations of the invention described above can be used alone or in various combinations.
  • the methods of this invention can be implemented by software, hardware or a combination of hardware and software.
  • the invention can also be embodied as computer readable code on a computer readable medium.
  • the computer readable medium is any data storage device that can store data which can thereafter be read by a computer system, including both transfer and non-transfer devices as defined above. Examples of the computer readable medium include read-only memory, random access memory, CD-ROMs, flash memory cards, DVDs, magnetic tape, optical data storage devices, and carrier waves.
  • the computer readable medium can also be distributed over network-coupled computer systems so that the computer readable code is stored and executed in a distributed fashion.

Abstract

Disclosed herein is a multi-functional hand-held device capable of configuring user inputs based on how the device is to be used. Preferably, the multi-functional hand-held device has at most only a few physical buttons, keys, or switches so that its display size can be substantially increased. The multi-functional hand-held device also incorporates a variety of input mechanisms, including touch sensitive screens, touch sensitive housings, display actuators, audio input, etc. The device also incorporates a user-configurable GUI for each of the multiple functions of the devices.

Description

    CROSS-REFERENCE TO RELATED APPLICATIONS
  • This application is a continuation of, and claims the benefit of, U.S. patent application Ser. No. 11/367,749, filed Mar. 3, 2006 and published as U.S. Pub. No. 2006/0197753, which in turn is related to and claims the benefit of under 35 USC 119(e) to: (1) U.S. Provisional Patent Application No. 60/658,777, titled “Multi-Functional Hand-Held Device,” filed Mar. 4, 2005 and (2) U.S. Provisional Patent Application No. 60/663,345, titled “Multi-Functional Hand-Held Device,” filed Mar. 16, 2005, each of which is hereby incorporated by reference.
  • This application is related to the following applications, which are all herein incorporated by reference: (1) U.S. patent application Ser. No. 10/188,182, titled “Touch Pad for Handheld Device,” filed on Jul. 1, 2002; (2) U.S. patent application Ser. No. 10/722,948, titled “Touch Pad for Handheld Device,” filed on Nov. 25, 2003; (3) U.S. patent application Ser. No. 10/643,256, titled “Movable Touch Pad with Added Functionality,” filed on Aug. 18, 2003; (4) U.S. patent application Ser. No. 10/654,108, titled “Ambidextrous Mouse,” filed on Sep. 2, 2003; (5) U.S. patent application Ser. No. 10/840,862, titled “Multipoint Touch Screen,” filed on May 6, 2004; (6) U.S. patent application Ser. No. 10/903,964, titled “Gestures for Touch Sensitive Input Devices,” filed on Jul. 30, 2004; (7) U.S. patent application Ser. No. 11/038,590, titled “Mode-Based Graphical User Interfaces for Touch Sensitive Input Devices,” filed on Jan. 18, 2005; and (8) U.S. patent application Ser. No. 11/057,050, titled “Display Actuator,” filed on Feb. 11, 2005, (9) U.S. patent application Ser. No. 11/115,539, titled “Hand-Held Electronic Device with Multiple Touch Sensing Devices,” filed Apr. 26, 2005.
  • BACKGROUND
  • There exist today many types of hand-held electronic devices, each of which utilizes some sort of user interface. The user interface typically includes an output device in the form of a display, such as a Liquid Crystal Display (LCD), and one or more input devices, which can be mechanically actuated (e.g., switches, buttons, keys, dials, joysticks, joy pads) or electrically activated (e.g., touch pads or touch screens). The display is typically configured to present visual information such as text and graphics, and the input devices are typically configured perform operations such as issuing commands, making selections or moving a cursor or selector in the electronic device. Each of these well known devices has considerations such as size and shape limitations, costs, functionality, complexity, etc. that must be taken into account when designing the hand-held electronic device. In most cases, the user interface is positioned on the front face (or front surface) of the hand-held device for easy viewing of the display and easy manipulation of the input devices.
  • FIGS. 1A-1F are diagrams of various hand-held electronic devices including for example a telephone 10A (FIG. 1A), a PDA 10B (FIG. 1B), a media player 10C (FIG. 1C), a remote control 10D (FIG. 1D), a camera 10E (FIG. 1E), and a GPS module 10F (FIG. 1F). In each of these devices 10, a display 12, which is secured inside the housing of the device 10 and which can be seen through an opening in the housing, is typically positioned in a first region of the electronic device 10. Each of these devices also include one or more input devices 14, which are typically positioned in a second region of the electronic device 10 next to the display 12.
  • To elaborate, the telephone 10A typically includes a display 12 such as a character or graphical display, and input devices 14 such as a number pad and in some cases a navigation pad. The PDA 10B typically includes a display 12 such as a graphical display, and input devices 14 such as a stylus based resistive touch screen and buttons. The media player 10C typically includes a display 12 such as a character or graphic display and input devices 14 such as buttons or wheels. The iPod® media player manufactured by Apple Computer, Inc. of Cupertino, Calif. is one example of a media player that includes both a display and input devices disposed next to the display. The remote control 10D typically includes an input device 14 such as a keypad and may or may not have a character display 12. The camera 10E typically includes a display 12 such as a graphic display and input devices 14 such as buttons. The GPS module 10F typically includes a display 12 such as graphic display and input devices 14 such as buttons, and in some cases a joy pad.
  • Recently, traditionally separate hand-held electronic devices have begun to be combined in limited ways. For example, the telephone 10A has been combined with the PDA 10B. One problem that has been encountered is in the way inputs are made into the device. Each of these devices has a particular set of input mechanisms for providing inputs into the device. Some of these input mechanisms are generic to all the devices (e.g., power button) while others are not. The ones that are not generic are typically dedicated to a particular functionality of the device. By way of example, PDAs typically include four dedicated buttons while cell phones typically include a numeric keypad and at least two dedicated buttons.
  • Thus it is a challenge to design a merged device with limited input devices without adversely affecting the dedicated inputs for each device. As will be appreciated, it is preferable, not to overload the hand-held devices with a large number of input mechanisms as this tends to confuse the user and take up valuable spaces i.e., “real estate.” In the case of hand-held devices, space is at a premium because of their small size. At some point there is not enough space on the device to house all the necessary buttons and switches, etc. This is especially true when considering that all these devices need a display that typically takes up a large amount of space on its own. To increase the number of input devices beyond some level, designers would have to decrease the size of the display. However, this will often leave a negative impression on the user because the user typically desires the largest display possible. Alternatively, to accommodate more input devices designers may opt to increase the size of the device. This, too, will often leave a negative impression on a user because it would make one-handed operations difficult, and at some point, the size of the device becomes so large that it is no longer considered a hand-held device.
  • Therefore what is needed in the art is an improved user interface that works for multi-functional hand-held devices.
  • SUMMARY
  • Disclosed herein is a multi-functional hand-held device capable of configuring user inputs based on how the device is to be used. Preferable, the multi-functional hand-held device has at most only a few physical buttons, keys, or switches so that its display size can be substantially increased. In other words, by eliminating physical buttons, keys, or switches from a front surface of an electronic device, additional surface area becomes available for a larger display. Ultimately this strategy would allow a substantially full screen display. As used herein, a full screen display is a display that consumes, or at least dominates, a surface (e.g., front surface) of an electronic device.
  • Various embodiments of a multi-functional hand-held device are discussed below with reference to FIGS. 2-28. However, those skilled in the art will appreciate that the detailed description given herein with respect to these figures is exemplary and not exhaustive and that many variations on these embodiments are possible.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • The invention may best be understood by reference to the following description taken in conjunction with the accompanying drawings in which:
  • FIGS. 1A-1F are diagrams of various electronic devices.
  • FIG. 2 is a simplified diagram of a multi-functional hand-held device.
  • FIG. 3 is a perspective view of a substantially full screen hand-held device with a limited number of limited buttons.
  • FIG. 4 is a front view of a hand-held device with at least one button.
  • FIG. 5 is a diagram of a GUI separated into a standard region and a control region.
  • FIG. 6 is a diagram of a GUI separated into a standard region and a control region.
  • FIG. 7 is a diagram of a GUI separated into a standard region and a control region.
  • FIG. 8 is a diagram of a GUI separated into a standard region and a control region.
  • FIG. 9 illustrates an exemplary a GUI for a PDA.
  • FIG. 10 illustrates an exemplary GUI for a cell phone.
  • FIG. 11 illustrates an exemplary GUI for a media player.
  • FIG. 12 illustrates an exemplary GUI for a video player.
  • FIG. 13 illustrates an exemplary GUI for a game player.
  • FIG. 14 illustrates an exemplary GUI for a camera.
  • FIG. 15 illustrates an exemplary GUI for a GPS.
  • FIG. 16 illustrates an exemplary GUI for a remote control.
  • FIG. 17 illustrates an exemplary GUI for a hand top.
  • FIG. 18 illustrates an exemplary GUI for a main menu of a multi-functional hand held device.
  • FIG. 19 is a side elevation view, in cross section, of a hand-held device incorporating a force sensitive display.
  • FIG. 20 illustrates an input device that combines touch sensing and force sensing devices to provide x, y and z components when touched.
  • FIG. 21 is a side elevation view of an I/O device that combines a display with touch screen and a force sensing mechanism.
  • FIG. 22 is a side elevation view of an input device.
  • FIG. 23 is a side view, in cross section, of a hand-held device that incorporates a squeeze feature.
  • FIG. 24 is a side view, in cross section, of a hand-held electronic device.
  • FIG. 25 is a block diagram of a touch sensing method.
  • FIG. 26 is a block diagram of touch sensing method.
  • FIGS. 27A-E are tables representing one example of a touch vocabulary associated with a music player.
  • FIG. 28 is a block diagram of an exemplary multi-functional hand-held device.
  • DETAILED DESCRIPTION
  • I. Multi-Functionality
  • Electronic device manufacturers have discovered the advantages of combining separate hand-held electronic devices to form multi-function devices. By having a single multi-function device, a user is not burdened with carrying, purchasing, and maintaining multiple devices. Further, the user is not limited in the operations that can be performed, i.e., the user can perform different operations with a single device that would have otherwise required the use of a different devices.
  • As used herein, the term “multi-functional” is used to define a device that has the capabilities of two or more traditional devices in a single device. The multi-functional device may, for example, include two or more of the following device functionalities: PDA, cell phone, music player, video player, game player, digital camera, handtop, Internet terminal, GPS or remote control. For each new device functionality that is added to a single device, the complexity and size of the device tends to increase. Therefore, with hand-held devices, there is typically a trade-off between keeping the footprint small and complexity low while still maximizing the functionality of the device.
  • In some cases, combining devices may result in redundant hardware components, which allows components to be used for multiple different, device functionalities. In other cases, certain hardware components are distinct to each device and therefore additional space and connectivity must be made available. Furthermore, each device functionality typically has its own programming or application software and, therefore, the multifunction device must be designed with enough memory to accommodate all the various software components.
  • A personal digital assistant (PDA) is a mobile hand-held device that provides computing and information storage and retrieval capabilities for personal and/or business use. PDAs are severally capable of tracking names, addresses, phone numbers and appointments. They are also often capable of taking notes, performing calculations, paging, data messaging, and electronic mail. PDAs may also include functionality for playing simple games, music, and other media files. Examples of PDAs include the Palm Pilot and Blackberry.
  • Like most hand-held devices, PDAs typically include a display and various input devices. The input devices may include a stylus and touch screen that work in combination with a handwriting recognition program, keypads, mini-keyboards, navigation pads, and/or soft or fixed function buttons.
  • Cell phones are mobile telephones that allow a user to connect to other telephones using a cellular network. Cell phones typically include a transceiver for transmitting and receiving telephone calls, controls such as a navigation pad for traversing through a display, a keypad for making numeric entries (and in some cases alphabetic entries), and soft or fixed function buttons. For example, in many cell phones one fixed function button is used for starting a call and another fixed function button is used for ending a call.
  • Media players come in a variety of forms. Music players are generally configured to store, process and output music. Music players can be based on the MP3 or MC format, which is a compression system for music. Music Players typically include a microprocessor, memory, display, audio jack, data port and playback controls. The playback controls typically include features such as menu, play/pause, next, previous, volume up, and volume down. Video players are similar to music players in most respects. In some cases, they may include a data storage device for receiving a removable storage medium such as a DVD. The iPod® media player manufactured by Apple Computer, Inc. of Cupertino, Calif. is one example of a media player.
  • Handtops are general purpose computers similar to laptops, but in a smaller form factor. Handtops typically include a display and a full keyboard.
  • FIG. 2 is a simplified diagram of a multi-functional hand-held device 100. The multi-functional hand-held device 100 integrates at least two devices 102 into a single device. Each device 102 includes both hardware and software components 104 and 106, which are integrated into multi-functional hand-held device 100. It should be pointed out that the multi-functional hand-held device 100 is not limited to only two devices, and may in fact integrate any number of devices.
  • Multi-functional device 100 also includes switch 110, which that allows multi-functional device 100 to be switched from one device operating mode to another device operating mode. For example, switch 110 may allow a user to cycle through cell phone, media player, and PDA modes of operation. Once a particular operating mode is selected, the multi-functional device 100 operates as the selected device. For example, the programming related to the selected device is activated for use by the multi-functional hand-held device. The programming may include reconfiguring the UI based on the selected device so that the inputs made by the user correlate to the device in use. For example, the functions of any physical buttons, switches or dials as well as soft buttons, switches or dials can be reconfigured to correspond to the selected device.
  • However, the operating modes of multi-functional hand-held device 100 need not completely independent. In many cases, it will be desirable to allow the multiple functionalities to interact with each other. For Example, a user may look up a telephone number of a contact in the PDA and pass this number to the phone to be dialed.
  • II. Form Factor
  • The form factor of a hand-held device is generally a device that can easily be held in one hand, A typical hand-held device includes a small display in an upper portion of the front surface of the device and input controls in a lower portion of the front surface of the device. The device may additionally include controls and ports on the top, bottom, and side surfaces. Prior art hand-held devices have typically had small displays which causes some dissatisfaction for users of these devices. It is generally preferred to have larger displays so that more information can be displayed or the information being displayed can be more easily viewable (e.g., larger text). In the case of video players and game players in particular, larger displays are much preferred over smaller ones.
  • However, the use of large displays has been limited in the past because the required input controls often take up substantial portions of the available space. Furthermore, as device functionality converges, the number of controls on the device typically increases as each device includes its own dedicated controls. Therefore devices must be made larger or the displays must be smaller to accommodate all the new controls. Both of these outcomes are unsatisfactory. Larger devices are bulky and hard to use, and smaller screens are difficult to utilize for both relaying information and reading information from the screen.
  • A. One-Handed Vs. Two-Handed Operation
  • A hand-held electronic device may be directed at one-handed operation or two-handed operation. In one-handed operation, a single hand is used to both support the device as well as to perform operations with the user interface during use. Cellular phones and media players are examples of hand-held devices are generally intended to can be operated solely with one hand. In the case of a cell phone, for example, a user may grasp the phone in one hand between the fingers and the palm and use the thumb to make entries using keys, buttons or a # joy pad.
  • In two-handed operation, one hand is used to support the device while the other hand performs operations with a user interface during use or, alternatively, both hands support the device as well as perform operations during use. PDA's and game players are examples of hand-held device that are typically operated with two hands. In the case of a PDA, for example, the user may grasp the device with one hand and make entries using the other hand, for example, using a stylus. In the case of a game player, the user typically grasps the device in both hands and make entries using either or both hands while holding the device.
  • B. Footprint/Size
  • Hand-held devices may have a variety different footprints or sizes. The footprint is typically associated with how the device is going to be used. Devices such as PDAs are typically used with both hands and thus they tend to be larger. Alternatively, cell phone handsets are typically used with only one hand, and thus they tend to be smaller. When integrating multiple devices, it is important task to determine the appropriate footprint of the device. For example, some believe that PDAs are too large for phone use while cell phones are too small for PDA use. Designers generally must consider the primary use of the device and gear the footprint towards that use. Although there are different footprints, there are typically minimum and maximum footprints. If the footprint is too large or too small, the device may be difficult to use.
  • Because of their size, the smaller devices are typically placed in a pocket while the larger devices are not. A media player is one example of a hand-held device that is sized for placement into a pocket of the user. By being pocket-sized, the user does not have to directly carry the device and therefore the device can be taken almost anywhere the user travels (e.g., the user is not limited by carrying a large, bulky and often heavy device, as in a laptop or notebook computer).
  • Generally it is preferred, although not necessary, that hand-held devices of the type disclosed herein have dimensions of about 5 in ×3 in ×1 in and perhaps about 4.1 in ×2.4 in ×0.75 in.
  • C. Full Screen Display
  • Because the footprint of a multifunctional hand-held device is substantially fixed by the intended primary use of the device, it becomes important for designers to determine an appropriate layout for the UI. For example, some devices may work better with a limited number of buttons and a large display while others may work better with a large number of buttons or a complete keyboard and a small display.
  • A preferred multi-functional hand-held device may be configured with a full screen display or a near full screen display. A full screen display consumes substantially the entire front surface of the device. The display may extend edge to edge or may fit within a small bezel of the housing at the edge of the device. The full screen display may consumes 90% or more of a front surface of a housing for a hand-held electronic device.
  • The full screen display may have a variety of different configurations depending on the overall footprint of the device. If the device is wide, the full screen display may have a traditional aspect ratio of about 4:3. If the device is elongated, the full screen display may have an aspect ratio that is more panoramic such as 16:9.
  • D. Limited Number of Mechanical Actuators
  • To accommodate a full screen display, the multi-functional hand-held device is preferably configured with a limited number of physical buttons. Because a limited number of physical buttons are provided, the hand-held device preferably uses a touch screen as the primary input device. Touch screens are transparent touch sensing devices that are positioned over displays. They typically work in conjunction with a GUI presented on the display. For example, the GUI may present an on-screen button, and the touch screen may detect when a user presses the on-screen button (e.g., places their finger or stylus over the on-screen button). Touch screens and GUI's are described in greater detail below.
  • The hand-held device may be constructed with only cross-functional physical buttons, i.e., there are no buttons dedicated to individual devices. These type of buttons may include power buttons and hold switches. In another embodiment, the hand-held device may not include any physical buttons at all. In some embodiments, the physical buttons are limited to only the sides and back surface of the hand-held device. In other embodiments, the physical buttons of the hand-held device are limited to the upper and lower portion of the sides so that there are no buttons in the regions of the sides where a user would physically support the device (i.e., holding region). In still other embodiments, the physical buttons may be located on the front surface, but only in the bezel region surrounding the display. In some embodiments, the buttons may be located on only the top and bottom surfaces of the device.
  • FIG. 3 is a perspective view of a substantially full screen multi-functional hand-held device 120 with a limited number of buttons. There are no physical buttons on the front and side surfaces 124 and 126. The front surface is used entirely for the display 122. Further, because the sides 126 are used for grasping the device 120 it may be preferred to leave the sides free from buttons to prevent accidental actions in the event a user inadvertently presses a button while supporting the device. Although the top surface 128 and bottom surface 130 would not typically be used to hold the device, these surfaces are not ideal locations for buttons that are often actuated because it would be awkward to reach these buttons when operating the device with one hand.
  • The top surface 128 may be reserved for buttons that have limited action and generic functions that are cross-functional, for example, power and hold switches. The top and bottom surfaces 128 and 130 are also well suited for placement of I/O and communication ports. The top surface 128 may, for example, include a headset/microphone jack and an antenna, and the bottom surface 130 may include power and data ports.
  • In some cases, it may be desirable to place buttons in the upper or lower regions of the side surfaces 126 out of the way of the grasping hand. This may be particularly well suited for elongated devices that are larger than the width of the grasping hand. As shown in FIG. 4, the hand-held device 120 includes a button 140 in the upper region on the side surface 126 of the hand-held device 120. Because the button 140 is in the upper region, it tends to be out of the way of the grasping hand and therefore accidental activation is substantially eliminated. The upper button may be configured to switch the functionality of the multi-functional device i.e., button 140 may be switch 110 of FIG. 2. For example, by pressing the button 140, a new device functionality is activated, and the current device functionality is deactivated. Although the term button is used, it should be appreciated that the button 140 may correspond to a dial, wheel, switch and/or the like.
  • Generally it would be preferred, although not required, to limit the number of physical buttons to eight or fewer, and perhaps five or fewer.
  • III. Adaptability
  • To limit the number of physical controls on the device (thereby maximizing the display area), the multi-functional hand-held device is preferably adaptable, i.e., the device is capable of changing its UI based on how the device is to be used. For example, if a cell phone functionality of the multi-functional device is to be used, the UI changes to accommodate the cell phone. Alternatively, if the PDA aspect of the multi-functional device is to be used, the UI changes to accommodate the PDA, etc. In essence, the multi-functional device is capable of reconfiguring its user interface based on the state or mode of the device.
  • Adaptability may be accomplished by reassigning the functions of the limited number of physical buttons for each device functionality. For example, a button may perform one function when a first device is activated and another when another device is activated. While this may work, it suffers from physical limitations (i.e., the number of buttons) and can be confusing to the user (who must keep track of different button meanings).
  • Alternatively adaptability may be accomplished by virtually incorporating the physical inputs for each functionality into the GUI in conjunction with a touch screen. This allows the GUI to adapt to whatever device is selected, and the touch screen can receive inputs corresponding to the GUI. With a GUI for each functionality, the UI for the hand-held device adapts such that the multi-function device, in effect, becomes a particular device. For example, if the cell phone functionality is selected, the GUI presents a set of virtual or soft controls that look similar to the physical controls typically used on a cell phone, such as a keypad, function buttons and possibly a navigation pad, for example.
  • The entire display may be used to present this information, or only a portion of the display may be used for the GUI controls. In the later case, referring to FIGS. 5-8, the GUI 150 may be separated into a standard region 152 and a control region 154. The standard region 152 represents what would normally be displayed on the display 122 when using a particular device. That is, the standard GUI screens associated with the selected device are displayed in the standard region. For example, in the case of the PDA, a main menu (window with a set of icons), calendar, address book or date book may be displayed in the standard region 152. The control region 154, on the other hand, virtually represents the physical controls that would normally be physically placed on a particular device. That is, the virtual controls that mimic the physical controls are displayed in the control region 154. For example, in the case of the PDA, the control region 154 may include virtual representations of a hand writing recognition area, a navigation pad and the standard function buttons.
  • The standard and control regions 152 and 154 can be positioned at any position on the display 122 (top, bottom, sides, center, etc.). For example, as shown in FIG. 5, they may be positioned vertically relative to one another (one on top of the other) or as shown in FIG. 6, they may be positioned horizontally relative to one another (side by side). These configurations can be used in either portrait or landscape modes. By way of example, in cases where the device is operated in landscape mode, the standard region 152 may be placed on one side and the control region may be placed on the opposite side. Landscape orientation may, for example, facilitate one handed operation. In some cases, the side on which the controls are displayed is based on the handedness of the user. For example, the controls may be place on the right side for right-handed users, and the controls may be placed on the left side for left-handed users. Alternatively, the controls may be placed on both sides as shown in FIG. 7. This arrangement is particularly well suited for game playing. Furthermore, the amount of area dedicated to each portion may be widely varied. For example, the screen may be divided equally and in other cases one or the other portion constitutes a greater amount of the display. In some cases, the standard region 154 is maximized to increase the normal viewing area of the display.
  • When a particular functionality is selected for use, the hand-held device loads the software for the selected functionality and configures the GUI 150 including the standard region 152 and the control region 154. The controls in the control region 154 can therefore be used to control whatever is being shown in the standard region 152. In some cases, the control region 154 may even change in accordance with the needs of each displayed window for the particular device.
  • Alternatively, as shown in FIG. 8, virtual controls 160 may be overlaid on top of the standard region 152 so that the standard region 152 can fully utilize the entire display 122. In fact, the virtual controls 160 may appear and disappear as needed. For example, the user may touch the screen and this may drive the device to display the controls over a portion of the display including whatever is already displayed. Examples of virtual controls that operate in this manner can be found in U.S. patent application Ser. No. 11/038,590, titled “Mode-Based Graphical User Interfaces for Touch Sensitive Input Devices,” filed on Jan. 18, 2005.
  • A. GUI Based on Functionality
  • FIGS. 9-17 illustrate various examples of GUIs for different states or modes of the multi-functional device.
  • FIG. 9 is a diagram of a GUI 170 that is used in a PDA mode. As shown, the GUI is divided into a standard region 152 and a control region 154. Located inside the control region 154 are a virtual handwriting pad 172, four virtual buttons 174 and a virtual navigation pad 176.
  • FIG. 10 is a diagram of a GUI 180 that is used in a cell phone mode. As shown, the GUI 180 is divided into a standard region 152 and a control region 154. Located inside the control region 154 are a virtual keypad 182, a virtual navigation pad 184 and two virtual buttons 186.
  • FIG. 11 is a diagram of a GUI 190 that is used in a music player mode. As shown, the GUI 190 is divided into a standard region 152 and a control region 154. Located inside the control region 154 are a virtual scroll wheel 192 and five virtual buttons 194. Additional details on a virtual scroll wheel are provided in U.S. patent application Ser. No. 11/038,590, titled “Mode-Based Graphical User Interfaces for Touch Sensitive Input Devices,” filed on Jan. 18, 2005.
  • FIG. 12 is a diagram of a GUI 200 that is used in a video player mode. As shown, the GUI 200 is divided into a standard region 152 and a control region 154. Located inside the control region 154 are a plurality of virtual buttons 202. Alternatively, the controls may appears and disappears as needed since the video player is primarily used in conjunction with a full screen viewing mode.
  • FIG. 13 is a diagram of a GUI 210 that is used in a game player mode. As shown, the GUI 210 is divided into a standard region 152 and two control regions 154A and 154B on the sides of the standard region 152. The left side control region 154A includes a navigation or directional pad 212, and the right side control region includes four virtual buttons 214 (or vice versa depending on the users particular needs, left or right handed).
  • FIG. 14 is a diagram of a GUI 220 that is used in a camera mode. As shown, the GUI 220 is divided into a standard region 152 and a control region 154. The standard region 152 may represent the view finder. Located inside the control region 154 are various buttons 222 including for example picture click, zoom, flash, etc. A navigation pad 224 may also be included so that the pictures can be scrolled through or for menu navigation.
  • FIG. 15 is a diagram of a GUI 230 that is used in a GPS receiver mode. As shown, the GUI 230 is divided into a standard region 152 and a control region 154. Located inside the control region 154 are various buttons 222 including for example zoom, pan, etc. A navigation pad 224 may also be included.
  • FIG. 16 is a diagram of a GUI 240 that is used in a hand top mode. As shown, the GUI 240 is divided into a standard region 152 and a control region 154. Located inside the control region 154 is a virtual keyboard 242.
  • FIG. 17 is a diagram of a GUI 250 that is used in a remote control mode. As shown, the GUI 250 is divided into a standard region 152 and a control region 154. Located inside the control region 154 are various keys and buttons 252 associated with controlling a remote device such as a TV, DVD player, A/V amplifier, VHS, CD player, etc.
  • B. Switching Between Devices (GUI)
  • Before a particular device functionality can be used, it typically must be selected for use. The selection can come in a variety of forms. For example, the selection may be made via a main menu that includes soft buttons or icons that, when selected, activate the device functionality associated with the soft button. During activation, the GUI for that particular device is brought into view on the display (see FIGS. 9-17) and the software associated with the device is installed, loaded or activated. From that point on, the multi-functional device operates like the selected device.
  • FIG. 18 illustrated an exemplary main menu GUI 260 of a multi-functional device. As shown, the GUI 260 includes icons/buttons 262 for launching each of the various device functionalities. In this particular example, the main menu page 260 includes a PDA button 262A, a cell phone button 262B, a music player button 262C, a game player button 262D, a video player button 262E, a GPS button 262F, a remote control button 262G, a camera button 262H and a handtop button 262I. The various buttons 262 are virtual buttons. When a button is pressed, the main page for the selected functionally (e.g., as shown in FIGS. 9-17) is brought into view on the display. To select another device, the user simply selects a soft home button 264 located in the GUI of each device to return to the main menu page 260, and thereafter selects the desired functionality in the main menu page 260.
  • The selection of alternative functionalities may also be accomplished by flipping (or scrolling) through the various GUIs until the desired GUI is found. For example, the different GUIs may be incrementally brought into view page after page (or frame after frame) when a next (flip) command signal is generated (e.g., slide show effect). The transition between pages may be widely varied. The transition may be from side to side, top to bottom or center to center. The transition may also include fading in and out, popping in and out, or enlarging and reducing. The command signal may be generated by a physical or virtual button or wheel. Using a button, each press may cause a new page to be displayed. Using a wheel, a predetermined amount of rotation may cause a new page to be displayed.
  • The command signal may also be generated in a variety of other ways. For example, the command signal may also be generated by gestures initiated on the touch screen. For example, sliding a finger (or stylus) across the display may cause a new page to be displayed. If slid to the right, the next page may be displayed. If slid to the left, the previous page may be displayed. The command signal may also be generated by 3D device gestures created when the entire hand-held device is moved spatially. By way of example, shaking the device may cause a new page to be displayed.
  • The command signal may also be generated by forces that are applied to the device. By way of example, squeezing the device may cause a new page to be displayed. The command signal may also be generated by sensing the orientation of the device either relative to the ground, as sensed by accelerometers, or relative to a compass direction indicated by an internal compass. For example, if the device is at 0 degrees, a first page is displayed, at 90 degrees a second page is displayed, at 180 degrees a third page is displayed and at 270 degrees a fourth page is displayed.
  • The command signal may also be generated by monitoring a user's voice (i.e., voice recognition). If the user calls out “PHONE,” the page associated with the phone is displayed, if the user calls out “PDA,” the page associated with the PDA is displayed.
  • The command signal may also be generated by monitoring incoming signals from other systems (whether transmitted wirelessly or via a cable). For example, if a call is received, the device may automatically configure the system as a phone. Alternatively, it may only present a control panel for taking or passing on the call.
  • As an alternative to integrating functionalities, the device may be configured to keep the various modes separate. That is, the device does not merge the functionality together (integrated layers and GUIs), but instead keeps them distinct from one another. In some cases, by keeping different functionalities distinct, user confusion may be reduced.
  • C. Operating at Least Two Functionalities Simultaneously
  • Preferably, the user may be able to activate two or more device functionalities simultaneously. In such a case, the software for the multiple functionalities is activated simultaneously and the display operates in a split screen mode where the screen is parsed into different sections, each section including a particular device GUI. Generally this would require the GUI for each functionality to fit on the screen. The multi-function mode may be selected in a variety of ways. In one implementation, when the user simultaneously touches two or more device icons, the device activates the multiple device functionalities and brings the appropriate GUIs into view on the screen.
  • D. Configurable GUI (User Preferences)
  • The GUI control panels for each device functionality may be configurable by the user. For example, the user may design or customize his own UI for each device, and assign functions to the various elements of the UI. The functions may include initiating commands, selecting an item, opening a file or document, launching a program, executing instructions, viewing a menu on the display screen, etc. The parameters configurable by the user may include selecting the number and type of GUI elements (buttons) as well as the location of the GUI elements on the page. In some cases, the system may include a design palette that allows a user to review and/or customize the UI layout, i.e., the user may quickly and conveniently review preconfigured or default layout and make changes thereto. Once changed, the modified layout will be automatically saved and thereby employed to handle future events.
  • IV. Input Devices
  • There are a number of problems with current input devices for hand-held computers. For example, there is no realistic way to fit all the dedicated physical buttons that are needed for each device because more and more buttons must be fit into a small space. A related problem arises in that as more physical buttons are incorporated into the device, the buttons must get closer together thereby challenging the dexterity of the user. Furthermore, because physical buttons stick out they are often accidentally activated during normal use or while the device is transported e.g., in a user's pocket. Finally, large numbers of small physical buttons can be confusing to the user as well as aesthetically unpleasing.
  • To overcome these and other problems a variety of alternative inputs arrangements are proposed (in addition to or in place of physical buttons, switches, etc.). The basic idea is to reduce the number of physical and mechanical input mechanisms (and therefore the amount of dedicated space needed on the surface of the device) and/or eliminate the physical and mechanical input mechanisms altogether. By reducing or eliminating the physical input devices, the display of the electronic device can be maximized, or alternatively the size of the electronic device can be minimized. Furthermore, such a device is more aesthetically pleasing. In some cases, the hand-held device can be configured so that it appears to only have a display and no dedicated physical input devices.
  • The alternative input means may also be hidden from view such that they are not visible to the user. As a result, the hand-held electronic device may appear to only have a display with no visible buttons, switches, etc. Such a device is more aesthetically pleasing (e.g., may comprise smooth surfaces with no breaks gaps or lines), and, in many cases, can be made smaller without sacrificing screen size and input functionality.
  • A. Touch Screen
  • One particularly suitable input arrangement is a touch screen. A touch screen in conjunction with GUI may be configured as the primary input arrangement for a hand-held device. A touch screen is basically a transparent input panel positioned in front of the display. A touch screen generates input signals when an object such as a finger or stylus touches or is moved across the surface of the touch screen. In most cases, touch screens allow a user to make selections and initiate movements in a GUI by simply touching the display screen via a finger. For example, a user may make a selection by pointing directly to a graphical object displayed on the display screen. Corresponding to an on-screen button for performing specific actions in the hand-held electronic device. In general, the touch screen recognizes the touch and position of the touch on the display and a controller of the hand-held electronic device interprets the touch and thereafter performs an action based on the touch event. There are several types of touch screen technologies including resistive, capacitive, infrared and surface acoustic wave.
  • A preferred touch screen for a multi-functional hand-held computer is a multipoint capacitive touch screen. Such a touch screen comprises several independent and spatially distinct sensing points, nodes, or regions that are positioned throughout the touch screen. The sensing points are dispersed about the touch screen with each sensing point representing a different position on the surface of the touch screen. The sensing points may be positioned in a grid or a pixel array where each sensing point is capable of generating a signal. A signal is produced each time an object is positioned over a sensing point. When an object is placed over multiple sensing points or when the object is moved between or over multiple sensing point, multiple signals can be generated. The sensing points generally map the touch screen plane into a coordinate system such as a Cartesian coordinate system or polar coordinate system. One example of such a touch screen is disclosed in U.S. patent application Ser. No. 10/840,862, titled “Multipoint Touch Screen” filed on May 6, 2004.
  • B. Touch Sensitive Housing
  • A hand-held electronic device may also incorporate one or more touch sensitive surfaces of the device housing itself that provide either a larger surface for tracking touch inputs or smaller dedicated areas such as touch buttons for performing dedicated functions. Such surfaces may be located on any surface of the housing, any side of the housing, any portion of any side of the housing or at dedicated locations on the surface of the housing. For example, the touch regions may be located on the sides or back surface of the housing, and may even be located at the bezel located at the front surface of the housing. In all of these cases, a large portion of the front surface of the housing is saved for the display so that the viewing area of the hand-held electronic device can be maximized. The touch sensitive surfaces of the housing may take the form of one or more touch panels that are positioned within the housing. The touch sensitive surface may be alternatively or additionally be provided directly by the housing. That is, the touch sensing components may be integrated or incorporated into or disposed underneath the housing such that the housing itself is the touch sensing device (rather than using a separate touch panel). Similarly to a touch screen, a touch sensitive housing recognizes the touch and position of a touch on the surface and a controller of the hand-held electronic device interprets the touch and thereafter performs an action based on the touch event. Touch surfaces are constructed in basically the same manner as a touch screen, except the surfaces need not be substantially transparent.
  • By way of example, the touch sensitive housing may generally correspond to the touch sensitive housing described in greater detail U.S. patent application Ser. No. 11/115,539, titled “Hand-Held Electronic Device with Multiple Touch Sensing Devices,” filed Apr. 26, 2005.
  • C. Display Actuator
  • A hand-held multi-functional electronic device may also include a display actuator, which is an input device that mechanically uses the display of the hand-held device to provide inputs into the device (rather than electrically as with a touch screen). The display actuator may be used separately or in combination with the touch screen. The display actuator may include a movable display that causes one or more input signals to be generated when moved. The input signals then can be used to initiate commands, make selections, or control motion in a display.
  • The movable display may be configured to translate, slide, pivot, and/or rotate relative to the frame. The display is typically movable relative to a frame or housing that movably supports the display in its various positions. In some cases, the display is movably coupled to the frame, and in other cases the frame movably restrains a floating display.
  • The input signals may be generated by movement indicator(s) that monitor the movements of the display and produces signals indicative of such movements. The detection mechanism may, for example, one or more switches, sensors, encoders, and/or the like. Any suitable mechanical, electrical and/or optical switch, sensor or encoder may be used. For example, tact switches, force sensitive resistors, pressure sensors, proximity sensors, infrared sensors, mechanical or optical encoders and/or the like may be used. The movement indicators may be placed underneath the display or at the sides of the display as appropriate. Alternatively or additionally, these movement indicators may be attached to the display or some component of the display.
  • An exemplary display actuator is disclosed in U.S. patent application Ser. No. 11/057,050, titled “Display Actuator,” filed on Feb. 11, 2005.
  • D. Pressure or Force Sensing Devices
  • The multi-functional hand-held electronic device described above may further include force or pressure sensing devices such as a force sensing display or housing.
  • 1. Force Sensitive Display
  • A force sensitive display causes one or more input signals to be generated when pressure is exerted on the display screen of the device. The input signals can be used to initiate commands, make selections, or control motion in a display. Such displays generally provide a slight amount of flex (unnoticeable to the user) so, that any forces exerted thereon can be measured by a force detection arrangement generally provided underneath the display. The force detection arrangement monitors the forces exerted on the display and produces signals indicative thereof. The force detection arrangement may include one or more force sensors such as force sensitive resistors, force sensitive capacitors, load cells, pressure plates, piezoelectric transducers, strain gauges, etc. The force sensors may be attached to the back of the display or to a structural platform located within the housing of the device. When a force is applied to the display, it is transmitted through the display to the force sensor located underneath the display.
  • FIG. 19 is a side elevation view, in cross section, of a hand-held device 300 incorporating a force sensitive display 302. The force sensitive display 302 includes a display 308 and one or more force sensors 310 disposed underneath the display 308 (between the display and a structural platform 306). In most cases, the force sensitive display 302 includes a plurality of sensors 310 that are laid out in an array. For example, the sensors 310 may be positioned side by side in rows and columns. The force sensors 310 measure the amount of force being applied to the display and when a desired force threshold is reached a control signal is generated. In some cases, an elastomer 312 is placed between the display and the structural platform to help transmit the force being exerted on the surface of the display to the force sensors disposed below the display.
  • Force sensing may be provided in conjunction with a touch screen to differentiate between light and hard touches. The determination of whether a touch is a light touch or a hard touch may be made by monitoring the force with the force sensors and comparing the force to a predetermined threshold. When the force threshold is not exceeded, the touch is considered a light touch. When the force threshold is exceeded, the touch is considered a hard touch. Each type of touch may be used to control different aspects of the device. Light touches may be associated with passive events such as navigation (e.g., cursor control scrolling, panning, zoom, rotation, etc.) and hard touches may be associated with active events such as selections or commands (e.g., button click).
  • FIG. 20 illustrates an input device 320 that combines touch sensing and force sensing devices to provide x, y and z components when touched. The touch sensing device provides position sensing in the x and y directions, and the force sensing device provides force sensing in the z direction. These devices cooperate to output x, y location and z pressure information whenever there is a touch on the touch surface.
  • FIG. 21 is a side elevation view of an I/O device 330 that combines a display 332 with touch screen 334 and a force sensing mechanism 336. The touch screen 334 provides high resolution touch locations, and the force sensing mechanism 336 provides a measure of where the force is coming from as well the total force. Touch screen 334 is disposed over the display 332, and the force sensing mechanism 336 is disposed below the display 332 although other arrangements are possible.
  • Force sensing mechanism 336 may also be widely varied. In the illustrated embodiment, the force sensing mechanism 336 is based on capacitance, and more particularly, self capacitance. The illustrated force sensing mechanism 336 is formed from various layers including an electrode layer 338, an elastomer layer 340 and a conductive layer 342.
  • Electrode layer 338 includes a plurality of spatially separated electrodes 339 that are positioned across the bottom of the display 332. Electrodes 339 are typically positioned in an array of rows and columns although often configuring are possible. Any number of electrodes may be used.
  • Elastomer layer 340 includes one or more elastic members 341 positioned between the electrode layer 338 and the conductive layer 342. Elastic members 341 allow the display 332 to move inwardly with a limited amount of displacement. In one implementation, elastic members 441 are silicone patches with a thickness of about 0.2 mm.
  • Conductive layer 342 typically takes the form of a grounded metal plate 343. A capacitive circuit is formed between each of the electrodes 339 and the grounded metal plate 343. When a user pushes down on the display 332, the force being exerted thereon causes the display 332 to displace inwardly against the elastic members 341 compressing the elastic members. This causes a change in the capacitance between the electrodes 339 and the metal plate 343. This change in capacitance is sensed by a control circuit operatively coupled to each of the electrodes 339. Capacitance sensing circuits are disclosed in the various incorporated references.
  • FIG. 22 is a side elevation view of an input device 350 that may be positioned over a display. The input device 350 combines touch sensing and force sensing into a single device. In this embodiment, both the touch sensing and force sensing is provided by mutual capacitance. As shown, the input device 350 is formed from various layers including a top drive layer 352, a middle sense layer 354, and a bottom drive layer 356. Furthermore, the middle sense layer 354 is positioned on an elastomer layer 358 disposed between the middle sense layer 354 and the bottom drive layer 356. The top and bottom drive layers 353 and 356 include a plurality of spatially separated lines in rows and the middle sense layer 354 includes a plurality of spatially separated lines in columns. The top and middle layers 352 and 354 therefore form a grid, and the bottom and middle layers 356 and 354 form a grid.
  • During operation, the lines on the top layer 352 are scanned, and thereafter the lines on the bottom layer 356 are scanned (or vice versa). When there is a touch, the mutual capacitance measured between the top drive layer 352 and the middle sense layer 354 provide the x and y location of the touch. In addition, the mutual capacitance measured between the bottom drive layer 356 and the middle sense layer 354 provide the amount of force of the touch. This particular arrangement provides a full image of force superimposed on a full image of touch. The input device including the touch layers and the force layers may be operated similarly to the methods described in U.S. patent application Ser. No. 10/840,862, titled “Multipoint Touch Screen,” filed on May 6, 2004.
  • 2. Force Sensitive Housing
  • The hand-held device may also include a force sensitive housing. The force sensitive housing provides inputs when forces are applied to the housing of the hand-held device. A force sensitive housing is similar to a force sensitive screen in that. The housing provides a slight amount of flex (possibly unnoticeable to the user) so that any forces exerted thereon can be distributed to a force detection arrangement located within the housing. The force detection arrangement monitors the forces on the housing and produces signals indicative thereof. As with the force sensitive display discussed above, the force detection mechanism may include one or more force sensors disposed with in the housing such as force sensitive resistors, force sensitive capacitors, load cells, pressure plates, piezoelectric transducers, strain gauges and/or the like. When a force is applied to the housing (squeezing or pushing on the housing), it is transmitted through the housing to the force sensor located within the housing.
  • The force sensitive portions of the housing may be located on any surface of the housing, any side of the housing, any portion of any side of the housing or at dedicated locations on the surface of the housing, The sides of the housing are ideal places for implementing a squeeze feature. This is because the users fingers are typically positioned on one side of the device and thumb on the other and therefore the hand may easily squeeze the sides via a pinching action. Because it is so convenient to activate the squeeze feature, special care must be taken when designing the squeeze feature so that it will not be accidentally activate during normal use. Thus the device needs to be able to differentiate between light and hard squeezes. If the squeeze feature is implemented using force sensitive resistors (FSRs) which exhibit, a decrease in resistance with an increase in force applied to the active surface a comparator circuit can be used to output a signal to indicate activation when a preset force threshold is reached.
  • FIG. 23 is a side view, in cross section, of a hand-held device 370 that incorporates a squeeze feature. As shown, the device 370 includes a housing 372 and a support platform 374 inside the housing 372. Between the support platform 374 and the inner surface of the housing 372 are a pair of force sensors 376. When a force is applied to the housing 372 as for example by the pinching nature of the hand, the housing 372 flexes inwardly under the pressure. This causes the force sensors 376 to be sandwiched between the housing 372 and the support platform 374. The force sensors 376 measure the amount of force being exerted and when a desired force threshold is reached, the force sensors 376 generate a control signal. For example, as a result of being sandwiched, a force resistive sensor may exhibit a reduced resistance and when a desired threshold is reached, a control signal is generated.
  • The force sensitive housing may be is provided in conjunction with a touch sensitive housing as discussed above.
  • E. Motion Actuated Input Device
  • The hand-held electronic device may also include a motion actuated input device. The motion actuated input device provides inputs when the hand-held device is in motion or is placed in a certain orientation. A motion actuated input device typically includes a motion sensor, such as an accelerometer, that monitors the motion of the device along the x, y, and/or z axis and produces signals indicative thereof. The motion sensor may, for example, include an accelerometer. Alternatively, the motion sensor could be an orientation sensor, such as an electronic compass, that allows the device to determine its orientation in a generally horizontal plane. The motion sensors may be attached to the housing or to some other structural component located within the housing of the device. When motion is applied to the device (gesturing, shaking, hand waving, etc), it is transmitted through the housing to the motion sensor.
  • Because motion sensors typically measure all motion, not just intended motion, the intended motion information typically must be separated from the other motion information to produce an accurate command signal. For example, large scale movements such as shaking the device will produce primarily low frequency information. Conversely, small scale movements, such as vibrations, primarily produce high frequency information. The high frequency information can be filtered out thereby leaving only low frequency information indicative of the large scale movements (e.g., shaking). The filtered information can then be converted into a control signal.
  • FIG. 24 is a side view, in cross section, of a hand-held electronic device 380. The hand-held device 380 includes an accelerometer 382 that is attached to a housing 384 of the hand-held device 380. When the device 380 is moved about by the user, the accelerometer 382 recognizes the motion and a controller of the hand-held electronic device 380 interprets the motion and thereafter performs an action based on the motion event.
  • F. Mechanical Actuators
  • While one would like to eliminate all surface mounted actuators such as buttons and wheels, it is sometimes impractical. Therefore the hand-held device may include some number of surface mounted actuators. Preferably, these actuators are generic to each of the integrated devices. That is, their meaning is the same regardless of what device functionality is activated. It is also preferred that the surface mounted actuators be placed on surfaces other than the front surface, which houses the viewing region of the display, although this is not required.
  • One particularly useful mechanical actuator is a hold switch. The hold switch may be configured to activate and deactivate the primary input means, e.g., the touch screen. This permits a user to prevent unwanted entries, for example, when the device is stored inside a user's pocket. In one implementation, the hold switch may be placed on the top surface out of the way of the grasping hand, but in a position for easy access (as opposed to the bottom surface). The hold switch can not only deactivate the touch screen but also mechanical actuators and other input and other input devices.
  • Another particularly useful mechanical actuator is a power switch. When the power switch is turned on, the device is powered up and ready to go. When the power switch is turned off, the device is shut down. In one implementation, the power switch may be placed on the top surface out of the way of the grasping hand, but in a position for easy access (as opposed to the bottom surface).
  • Another useful mechanical actuator is a navigation pad. The navigation pad is typically included with many hand-held devices. The functionality of the navigation pad may be changed according to the current operating mode of the device. In the case of a music player, for example, the directional keys may be assigned, play/pause, next, previous, and volume up and down. Other assignable buttons may also be included on the device.
  • Still another useful mechanical actuator is a switching actuator. The switching actuator may be configured to change the functionality of the device, i.e., by activating the switching actuator the functionality or state of the device switches from one mode to another. The switching actuator may be widely varied.
  • For example, the switching actuator may be a dial or wheel. By incrementally rotating the wheel, the device is incrementally switched from one device to the other (generally in some predetermined order). A full rotation of each device generally cycles through the entire group of integrated devices. The wheel or dial may for example operate like a scroll wheel. Although the placement may be widely varied, the switching wheel may be placed in the upper region of the sides of the device. By placing the wheel here, a users thumb may be used to easily rotate the wheel. For example, the users thumb may be extended from the grasping action so that the wheel can be rotated.
  • Alternatively, the switching actuator may be a button. By repetitively pressing on the button, the device is switched from one device to another (generally in some predetermined order). Although the placement may be widely varied, the switching button may be placed in the upper region of the sides of the device. By placing the button here, a users thumb or index finger may be used to easily press the button.
  • The hand-held device may also include any commercially available touch pad. Several examples of touch pads may be found in U.S. patent application Ser. No. 10/188,182, titled “Touch Pad for Handheld Device,” filed on Jul. 1, 2002, U.S. patent application Ser. No. 10/722,948, titled “Touch Pad for Handheld Device,” filed on Nov. 25, 2003, and U.S. patent application Ser. No. 10/643,256, titled “Movable Touch Pad with Added Functionality,” filed on Aug. 18, 2003.
  • In another embodiment, the hand-held device may include a scroll wheel. Scroll wheels can be used in each functionality to scroll through a window.
  • G. Microphone
  • The hand-held device may also include a microphone that picks-up audio sounds. The microphone may be used in conjunction with a cell phone to transmit sounds, such as the user's voice. The microphone may also be used to record sounds or enter voice commands into the hand-held device. For example, using voice recognition software, the hand-held device may be able to recognize voice commands and generated control signals associated therewith. The microphone may be placed in the bottom surface of the hand-held device or possible in the front lower bezel. This particular configuration is well suited for picking up a user's voice during a phone call.
  • H. Image Sensor
  • A hand-held electronic device may also include an image sensor and lens related components so that the hand-held device can operate like a camera. The image sensor may, for example, include a charge coupled device (CCD) camera.
  • I. Input Device Functionality
  • 1. Touch Gestures
  • A hand-held electronic device may be designed to recognize touch gestures applied to a touch screen and/or touch sensitive surface of the housing and thereby control aspects of the hand-held electronic device. Gestures are a stylized interaction with an input device that is mapped to one or more specific computing operations. The gestures may be made through various hand and finger motions. Gestures generally comprise a contact chord e.g., one or more fingers, and a motion associated with the chord. Alternatively or additionally, the gestures may be made with a stylus. In all of these cases, the input device i.e., touch screen and/or touch sensitive surface) receive the gestures and a controller of the hand-held electronic device executes instructions to carry out operations associated with the gestures. The hand-held electronic device may include a touch gesture operational program, which may be part of the operating system or a separate application. The gesture operation program generally includes a set of instructions that recognizes the occurrence of gestures and informs one or more software agents of the gestures and/or what action(s) to take in response to the gestures. For example, gestures that can be used are disclosed in greater detail in U.S. patent application Ser. No. 10/903,964, titled “Gestures for Touch Sensitive Input Devices,” filed on Jul. 30, 2004, and U.S. patent application Ser. No. 11/038,590, titled “Mode-Based Graphical User Interfaces for Touch Sensitive Input Devices,” filed on Jan. 18, 2005.
  • 2. 3-D Spatial Gestures
  • In accordance with one embodiment, the hand-held electronic device may be designed to recognize 3-D spatial gestures measured by an accelerometer and to control aspects of the hand-held electronic device based on the 3-D spatial gestures. Spatial gestures are stylized motions of the device itself that are mapped to one or more specific computing operations. The 3-D spatial gestures may be made through various hand and arm motions, such as for example shaking, waving and the like. The accelerometer measures the motion related to the 3-D spatial gestures and a controller recognizes the motion as a 3-D spatial gesture and thereafter executes instructions to carry out operations associated with the 3-D spatial-gestures. The hand-held electronic device may include a 3-D spatial gesture operational program, which may be part of the operating system or a separate application. The gesture operation program generally includes a set of instructions that recognizes the occurrence of gestures and informs one or more software agents of the gestures and/or what action(s) to take in response to the gestures.
  • One exemplary 3-D gesture is shaking the device. Shaking can cause the hand-held device causes the device to shift between modes or states. Alternatively, shaking the hand-held device can cause a selection event to occur. For example, in the case of a media player, shaking the device may cause the device to randomly select a picture or song in group of pictures or songs. Alternatively, shaking the device may cause the device to select the next picture or song in a sequence of songs.
  • Other gestures may include translating or rotating the device. Translating the hand-held device (while it is face up) from side to side may be used to initiate panning or scrolling in the device, or moving the device up and down (while it is face up) may be used to initiate zooming. Rotating the device may be used to cause the device to change modes or states. In some cases, for example, the orientation of the device may correspond to a particular mode. For example, a first mode may be associated with 0 degrees, a second mode may be associated with 90 degrees, a third mode may be associated with 180 degrees and a fourth mode may be associated with 270 degrees. In all these cases, the device can be configured to keep the displayed portion upright as the device is turned. That is, it maintains an upright image no matter what orientation the device is in.
  • The 3-D spatial gestures may even be based on more complex motions such as sign language, writing motions, etc.
  • 3. Perform Action Based on Multiple Inputs
  • Because the device may have multiple input modes, the hand-held device may be configured to receive simultaneous inputs from different inputs devices, and perform actions based on the multiple simultaneous inputs. The inputs that may be combined to produce new commands may be selected from voice, 2-D touch gestures, 3-D spatial gestures, actuators, etc. For example, this feature may be helpful when making calls via voice selection. A caller may verbalize “TOM,” which causes the device to dial “TOM” phone number. If “TOM” has multiple phone numbers, a user may combine the verbal command “TOM” with a 3-D spatial gesture such as shaking to select Tom's second phone number. Various other possibilities will be appreciated by those skilled in the art.
  • 4. Differentiating Between Light and Hard Touches
  • As noted above, force sensing in conjunction with touch sensing facilitates two distinct types of interactions, light touches and hard touches. Light touches may be used to perform passive actions such as navigating through content and content manipulation generally without causing a major event to occur. Examples of passive events include moving a cursor, scrolling, panning, etc. Hard touch interactions may be used to select on screen buttons or initiate commands (e.g., causes a significant change to occur).
  • FIG. 25 is a diagram of a touch method 400 for implementing this technique. The method 400 begins at block 402 where one or more touches are detected. The touches include not only x any y components but also z components. The x and y components may be supplied by a touch sensing device such as touch screen, touch pad, or touch housing. The z component may be provided by force sensors or display actuators located behind the touch surface of the touch sensing device.
  • Following block 402, the method proceeds to block 404 where a determination is made as to whether the touch is a light or hard touch. The determination is generally based on the force or pressure of the touch (z component). For example, if the force of the touch is smaller than a predetermined threshold then the touch is considered a light touch and if the force of the touch is larger than the predetermined threshold then the touch is considered a hard touch. If it is determined that the touch is a light touch, the method proceeds to block 406 where a passive action associated with the touch is initiated. If it is determined that the touch is hard touch, an active action associated with the touch is performed (block 408).
  • The touch method may additionally include a block where the one or more touches are classified as a primary touch or a secondary touch. Primary touches are touches that are intended to cause an action while secondary touches are touches that are not intended to cause an action. Gestures are examples of primary touches while a thumb positioned over the touch area to hold the device is an example of a secondary touch. Once the touches are classified as primary or secondary, the secondary touches are filtered out, and the determination of whether a touch is a light or hard touch is made with the primary touches.
  • 5. Example of a New Touch Vocabulary
  • The z pressure exerted on a touch sensing device can be combined with the x and y locations of the touch to form a new touch vocabulary. As should be appreciated, up to this point touch vocabularies have only included x and y locations, not z pressure. A proposed touch vocabulary includes variety of variables including the UI mode, the force of the touch (e.g., light or hard), the number of fingers used, whether or not there is any movement during the touch, the duration of the touch, and the touch location, all or some of which can be combined to form a variety of behaviors and user feedback.
  • The UI mode is generally related to the mode or state of the device. Each device includes a variety of states and each state may require a different UI mode. A media player (a mode) may, for example, include a set of hierarchical layers (states) with each layer requiring a different UI.
  • As noted above, the force of the touch may, for example, be described as light or hard. A light touch may occur when a user lightly touches the surface of the touch surface, i.e., the finger hovers on top of the surface and is primarily moved in the x and y directions. A hard touch may occur when a user presses on the touch surface with a certain amount of force, i.e., the finger is primarily moved in the z direction against the touch surface.
  • Motion during the touch is used to describe whether the finger has remained stationary during a touch event or has substantially moved in the X-Y plane (e.g., translation, rotation, etc.). The motion may be described as none at all or a swipe or twist in some particular direction. By way of example, the swipe may be up, down, right, left, or some combination thereof, and the twist may be clockwise or counterclockwise.
  • Duration is defined as the amount of time the finger stays at any one point. Duration may be variable or it may include states such as short and long. The touch location may be a random point or a specific location such as an onscreen button.
  • FIG. 26 is an additional touch method 500 implementing this technique. The method begins at block 502 when one or more touches are detected. Thereafter, in block 504, the UI mode is determined. In block 506, a determination is made as to whether the touches are light touches or hard touches. Alternatively, blocks 502 and 504 could be reversed, effectively resulting in an instance of the touch method for each mode. In block 508, the number of distinct touches (e.g., fingers) is determined. In block 510, a determination is made as to whether the touches are stationary or in motion. In block 512, the duration of the touches is determined. In block 514, the locations of the touches are determined. Following blocks 502-514, the method proceeds to block 516 where an action is performed based on the UI mode, the pressure of the touch, the number of touches, whether or not the touch is moving, the duration of the touch, and the touch location. The actions may be passive or active depending on the values of each characteristic.
  • One example of a touch vocabulary associated with a music player is shown in FIGS. 27A-E.
  • V. Output Devices
  • A. Display
  • The principle output of a hand-held electronic device is typically a display. The display provides visual information in the form of text, characters or graphics. The display is usually a flat panel device although other types of displays may be used. The display may be a liquid crystal display (LCD) such as a character LCD that is capable of presenting text and symbols or a graphical LCD that is capable of presenting images, video, and graphical user interfaces (GUI). Alternatively, the display may correspond to a display based on organic light emitting diodes (OLED), or a display that is based on electronic inks.
  • Preferably, the display may be configured to substantially fill the front surface of the housing. The display may extend from one edge of the housing to the opposite edge of the housing, the housing may include a small bezel that surrounds the edges of the display. In either case, the display makes up a substantial portion of the front surface of the hand-held electronic device, thereby eliminating any space for buttons or switches associated with a conventional user interface.
  • As mentioned above, besides outputting visual information, the display may also act like an input device. For example, a touch screen may be positioned over the display, and/or sensors may be disposed underneath the display to sense when the device is pressed or otherwise moved. In most cases, the small form factor of hand-held devices requires central placement of the input interfaces to permit operation while being carried around and used by the hand. The display region provides a central location, which can be accessed by both the left and right hands.
  • The display region may be formed by more than one display. For example, the display region may be formed by a pair of displays that are side by side or one on top of the other. A first display may be used to present the standard screen and a second display may be used to present the control screen as described above with reference to FIGS. 5-8. Furthermore, a first display may be a conventional display while the second display may be a display actuator. Moreover, a first display may comprise of a first type and a second display may be of a second type. For example, the first display may be an LCD while the second display may be a display based on electronic inks.
  • The decision to use different types may be based on the fact that one of the displays may be dedicated to a standard viewing area while another may be dedicated to a control area. Because these two areas require different resolutions, a higher resolution screen may be used in the standard area while a lower resolution screen may be used in the control area. Alternatively or additionally, one of the displays may be selected to reduce battery consumption especially in the control area where lower resolution is acceptable. In some cases, although two different displays are used, the images displayed thereon may be combined to form a single unified image.
  • B. Speaker
  • A hand-held device may also include a speaker. Speakers are components that accept electronic signals representing audio information from an amplifier and converts them into sound waves. The speakers may be used to listen to music in conjunction with a music player functionality or to listen to an incoming call in conjunction with a cell phone functionality. The speaker may be placed on the top surface or possible on the front top bezel of the hand-held device. This arrangement works particularly well when the device is used as a cell phone.
  • C. Indicator (LED)
  • A hand-held device may also include one or more indicators that provide user feedback or indicate events associated with the device. The events may relate to signals, conditions or status of the device. For example, the indicators provide status of battery life or alert a user when there is an incoming call. The indicators, which include light sources such as light emitting diodes (LED), are typically illuminated when an event occurs, and not illuminated when the event is stopped. Furthermore, the indicator may turn on and off (blink) or cycle with increasing or decreasing intensity, and in some cases may even change colors in order to provide more detailed information about the event that is being monitored.
  • The indicators may be conventional indicators that typically include a small clear plastic insert, which is located in front of the LED, and which is inserted within an opening in the housing thus causing it to exist at the surface of the housing. The LED itself may also be placed in the opening in the housing rather than using an insert. Alternatively, the indicator can be configured not to break the surface of the housing. In this configuration, the light source is disposed entirely inside the housing, and is configured to illuminate a portion of the housing thereby causing the housing to change its appearance, i.e., change its color.
  • D. Audio/Tactile Feedback Devices
  • The hand-held device may include speakers or buzzers to give audio feedback to the user. These may work similarly to the indicators described above, or they may be used to enhance the feel of actuating a GUI element such as a soft button or scroll wheel. For example, the speaker may be configured to output a “clicking” noise when a user presses on a virtual button, or rotates a virtual scroll wheel. This particular feature enhances the user experience and makes the virtual UI feel more like a physical UI.
  • The hand-held device may also include a haptics mechanism. Haptics is the science of applying tactile sensation and control to interaction with computer applications. Haptics essentially allows a user to feel information, i.e., signals are sent to the hand. The haptics mechanisms may be widely varied. They may include motors, vibrators, electromagnets, etc., all of which are capable of providing force feedback in the form of vibration or shaking. The haptics mechanisms may work similarly to the indicators described above (alert), or they may be used to enhance the feel of actuating a GUI element such as a soft button or scroll wheel. For example, the haptics mechanism may be configured to vibrate when a user presses on a virtual button, or rotates a virtual scroll wheel. This particular feature enhances the users experience and makes the virtual UI feel more like a physical UI. Haptics may also be used simultaneously with onscreen actions. For example, during movies or game playing, the haptics mechanism can simulate the action being displayed. For example, the haptics mechanism may provide force feedback in the form of vibration when a car explodes during a movie or game.
  • In cases where haptics is used to enhance the feel of actuating a GUI element, such as a soft button or scroll wheel, the haptics mechanism may be located in the region of the display and further underneath the display so as to provide force feedback directly underneath the user action. In fact, multiple haptics mechanisms may be used regionally across the display to further enhance the feel. It is generally believed that the closer the vibration is to the user action, the greater the haptics effect. In one implementation, the haptics mechanisms are spaced out in an array underneath the display. That is, they are spatially separated and placed at different locations. By way of example, they may be positioned in a 2×2, 2×4, 4×4, 4×8, 8×8 array and so on underneath the display.
  • Audio and/or tactile feedback may be used to alert a user that a user input has been made. For example, in response to touching a virtual button on the GUI, the haptics may provide force feedback in the form of vibration and the speaker may provide audio feedback in the form of a click. The tactile audio feedback can be used in conjunction with an input event including touch events, motion events, squeeze events. The feedback may provide information so that the user knows that they actually implemented an input (simulates the audio and tactile feel of a button or switch). In one implementation, the feedback is tied to the level of force being applied to the force sensing devices. For example, when a certain force threshold is reached, the audio feedback device may create a “click” on the press and a “clock” on the release. The force threshold used may be similar to the threshold used to determine whether a touch is a light or hard touch. The “click” and “clock” may be used to simulate a button click when a hard touch is made.
  • VI. Communication Devices
  • A. Wired
  • The hand-held device may also include one or more connectors for receiving and transmitting data to and from the device. By way of example, the device may include one or more audio jacks, video jacks, data ports, docking ports, etc. The hand-held device may also include one or more connectors for receiving and transmitting power to and from the hand-held device.
  • The hand-held device may include a headphone/microphone jack and a data port. The jack is capable of receiving a speaker and/or microphone plug so that audio may be input to and output from the device. The data port is capable of receiving a data plug/cable assembly configured for transmitting and receiving data to and from a host device, such as a general purpose computer (e.g., desktop computer, portable computer). For example, the data port may be used to upload or download data to and from the hand-held device. Such data may include songs and play lists, audio books, e-books, photos, address books, documents, appointments, etc. into the hand-held device. The data port may be a PS/2 port, serial port, parallel port, network interface port, USB port, Firewire port etc. The hand-held device may also include a power port that receives a power plug/cable assembly configured for delivering powering to the hand-held device. In some cases, the data port may serve as both a data and power port by employing either standard or proprietary connectors.
  • B. Wireless
  • To send and receive data wirelessly, the device generally requires a transmitter, a receiver (or a transceiver) and some sort of antenna. The wireless communication link may correspond to Bluetooth, WiFi (802.11), IR (infrared), etc. The antenna may be fully contained within the device or they may extend outside the device. The antenna may take a variety of forms depending on the frequency to be used, etc. For example be a rugged rubber duck that consists of a coiled up element encased in rubber. Alternatively, the antenna may be printed on a circuit board within the device.
  • The hand-held device may also include a radio transceiver for communications via a cellular network or a GPS receiver.
  • C. Change UI Based on Received Communication Signals
  • A hand-held electronic device may be configured to actively look for signals in the surrounding environment, and change its mode based on the signal. That is, the device tries to match the mode with the signal. If the device receives a phone signal over the cellular network, the device may turn into a phone, i.e., the phone mode is activated or brought forward relative to the other modes. If a device receives an email, the device may turn into an email terminal. As another example, when a user walks into a home theater room, the device may sense signals from the media control unit and turn itself into a remote control including functionality to control the various devices of the home theater (TV, amp, DVD, lighting). In other cases, the device may sense signals, which are being broadcast in physical stores, and turn itself into a device that is well suited for that store. For example, in a bank, the device may change into a calculator or bring a money program into view, or in a grocery store, the device may turn into a money payment device or bring a grocery list into view.
  • VII. Other Components of Hand-Held Device
  • The hand held device may additionally include one or more of the following hardware components: a controller (e.g., microprocessor, DSP, A/D, D/A, converters, codes), memory (e.g., RAM, ROM, solid state (flash), hard disk (micro-drive)), storage (SD card slots, mini-DVD), battery (e.g., lithium ion), etc.
  • VIII. Overall Block Diagram
  • FIG. 28 is a block diagram of an exemplary hand-held device 600. The hand-held device 600 typically includes a controller 602 (e.g., CPU) configured to execute instructions and to carry out operations associated with the hand-held device. For example, using instructions retrieved for example from memory, the controller 602 may control the reception and manipulation of input and output data between components of the hand-held device 600. The controller 602 can be implemented on a single chip, multiple chips or multiple electrical components. For example, various architectures can be used for the controller 602, including dedicated or embedded processor, single purpose processor, controller, ASIC, etc. By way of example, the controller may include microprocessors, DSP, A/D converters, D/A converters, compression, decompression, etc.
  • In most cases, the controller 602 together with an operating system operates to execute computer code and produce and use data. The operating system may correspond to well known operating systems such as OS/2, DOS, Unix, Linux, and Palm OS, or alternatively to special purpose operating system, such as those used for limited purpose appliance-type devices. The operating system, other computer code and data may reside within a memory block 604 that is operatively coupled to the controller 602. Memory block 604 generally provides a place to store computer code and data that are used by the hand-held device. By way of example, the memory block 604 may include read-only memory (ROM), random-access memory (RAM), hard disk drive (e.g., a micro drive), flash memory, etc. In conjunction with the memory block 604, the hand-held device may include a removable storage device such as an optical disc player that receives and plays DVDs, or card slots for receiving mediums such as memory cards (or memory sticks). Because the form factor of the hand-held device is small, the optical drive may only be configured for mini DVDs.
  • The hand-held device 600 also includes various input devices 606 that are operatively coupled to the controller 602. The input devices 606 are configured to transfer data from the outside world into the hand-held device 600. As shown, the input devices 606 may correspond to both data entry mechanisms and data capture mechanisms. In particular, the input devices 606 may include touch sensing devices 608 such as touch screens, touch pads and touch sensing surfaces, mechanical actuators 610 such as button or wheels or hold switches (611), motion sensing devices 612 such as accelerometers, force sensing devices 614 such as force sensitive displays and housings, image sensors 616, and microphones 618. The input devices 606 may also include a clickable display actuator 619.
  • The hand-held device 600 also includes various output devices 620 that are operatively coupled to the controller 602. The output devices 620 are configured to transfer data from the hand-held device 600 to the outside world. The output devices 620 may include a display 622 such as an LCD, speakers or jacks 624, audio/tactile feedback devices 626, light indicators 628, and the like
  • The hand-held device 600 also includes various communication devices 630 that are operatively coupled to the controller 602. The communication devices 630 may, for example, include both wired and wireless connectivity selected from I/O ports 632 such as IR, USB, or Firewire ports, GPS receiver 634, and a radio receiver 636.
  • The hand-held device 600 also includes a battery 650 and possibly a charging system 652. The battery may be charged through a transformer and power cord or through a host device or through a docking station. In the cases of the docking station, the charging may be transmitted through electrical ports or possibly through an inductance charging means that does not require a physical electrical connection to be made.
  • The various aspects, features, embodiments or implementations of the invention described above can be used alone or in various combinations. The methods of this invention can be implemented by software, hardware or a combination of hardware and software. The invention can also be embodied as computer readable code on a computer readable medium. The computer readable medium is any data storage device that can store data which can thereafter be read by a computer system, including both transfer and non-transfer devices as defined above. Examples of the computer readable medium include read-only memory, random access memory, CD-ROMs, flash memory cards, DVDs, magnetic tape, optical data storage devices, and carrier waves. The computer readable medium can also be distributed over network-coupled computer systems so that the computer readable code is stored and executed in a distributed fashion.
  • While this invention has been described in terms of several preferred embodiments, there are alterations, permutations, and equivalents, which fall within the scope of this invention. It should also be noted that there are many alternative ways of implementing the methods and apparatuses of the present invention. It is therefore intended that the following appended claims be interpreted as including all such alterations, permutations, and equivalents as fall within the true spirit and scope of the present invention.

Claims (20)

1. A hand-held multi-functional electronic device, comprising:
a combined touch screen and user interface display; and
a processing unit operatively connected to said touch screen and display, said processing unit capable of identifying and tracking a plurality of concurrent moving touch inputs from a user via said touch screen and discriminating a user requested action from the plurality of touch inputs;
wherein the touch screen is adapted for recognition of a plurality of concurrent touchdown locations anywhere on the surface.
2. A hand-held electronic device as recited in claim 1, wherein said hand-held electronic device includes two or more of the following functionality modes: PDA, mobile phone, music player, camera, video player, game player, handtop, Internet terminal, GPS receiver, and remote control.
3. A hand-held electronic device as recited in claim 2, wherein said hand-held electronic device is capable of reconfiguring or adapting the user interface based on the functionality mode of said hand-held electronic device.
4. A hand-held electronic device as recited in claim 1, wherein said display is a substantially full screen display.
5. A hand-held electronic device as recited in claim 1, wherein said multi-functional device is adapted for activation of two or more device functionality modes simultaneously and said display operates in a split screen mode when two or functionality modes are activated.
6. A hand-held electronic device as recited in claim 1, wherein said touch screen serves as a primary input mechanism for interacting with said hand-held electronic device.
7. A hand-held electronic device as recited in claim 1, further comprising one or more further input mechanisms including at least one cross-functional physical button located on a top or side of said device.
8. A hand-held electronic device as recited in claim 1, wherein said multipoint touch screen is a multipoint capacitive touch screen.
9. A hand-held electronic device as recited in claim 1, wherein said hand-held electronic device is operable to recognize touch gestures applied to said touch screen and wherein the touch gestures are used to control one or more of translation, scaling or rotation of an image displayed on said hand-held electronic device.
10. A hand-held electronic device as recited in claim 1, wherein said hand-held electronic device is operable to receive simultaneous inputs from different input mechanisms.
11. A hand-held electronic device as recited in claim 10, wherein a signal from at least one of the input mechanisms has a different meaning or output based on the functionality mode of said hand-held electronic device.
12. A hand-held electronic device as recited in claim 10, wherein said user interface display comprises a standard region and a control region, the standard region being used to display data, and the control region including one or more virtual controls for user interaction.
13. A hand-held electronic device as recited in claim 2, wherein each functionality mode presents a mode specific GUI control panel and at least one of the mode specific GUI control panels is user configurable.
14. A hand-held electronic device as recited in claim 1, wherein said touch screen comprises a force sensitivity arrangement, said force sensitive sensitivity arrangement producing one or more input signals when force is exerted thereon.
15. A hand-held electronic device as recited in claim 14, wherein said force sensitivity arrangement senses a force indication, and wherein said hand-held electronic device distinguishes the force indication into at least a first touch type and a second touch type.
16. A hand-held electronic device as recited in claim 1, further comprising a motion sensor whereby the device is able to detect its orientation in a generally horizontal plane.
17. A hand-held electronic device as recited in claim 1, wherein said hand-held electronic device provides audio or tactile feedback to a user based on user inputs made with respect to said hand-held electronic device.
18. A hand-held electronic device as recited in claim 1, wherein said hand-held electronic device is configurable to actively look for signals in a surrounding environment, and change user interface or functionality mode based on the signals.
19. A hand-held electronic device as recited in claim 1, wherein the display further comprises a display actuator and the device is configured to generate a signal when the display is moved relative to the housing.
20. A hand-held electronic device, comprising:
a display;
a multipoint touch screen; and
a processing unit operatively connected to said multipoint touch screen, said multipoint touch screen and processing unit adapted to translate, rotate and/or scale images on the display based on a plurality of concurrent touch input gestures applied by a user to the multipoint touch screen.
US11/966,948 2005-03-04 2007-12-28 Multi-functional hand-held device Abandoned US20080088602A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US11/966,948 US20080088602A1 (en) 2005-03-04 2007-12-28 Multi-functional hand-held device

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
US65877705P 2005-03-04 2005-03-04
US66334505P 2005-03-16 2005-03-16
US11/367,749 US11275405B2 (en) 2005-03-04 2006-03-03 Multi-functional hand-held device
US11/966,948 US20080088602A1 (en) 2005-03-04 2007-12-28 Multi-functional hand-held device

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
US11/367,749 Continuation US11275405B2 (en) 2002-07-01 2006-03-03 Multi-functional hand-held device

Publications (1)

Publication Number Publication Date
US20080088602A1 true US20080088602A1 (en) 2008-04-17

Family

ID=36936660

Family Applications (4)

Application Number Title Priority Date Filing Date
US11/966,948 Abandoned US20080088602A1 (en) 2005-03-04 2007-12-28 Multi-functional hand-held device
US14/595,032 Abandoned US20150153895A1 (en) 2005-03-04 2015-01-12 Multi-functional hand-held device
US14/940,010 Abandoned US20160070399A1 (en) 2005-03-04 2015-11-12 Multi-functional hand-held device
US15/933,196 Abandoned US20180217709A1 (en) 2005-03-04 2018-03-22 Multi-functional hand-held device

Family Applications After (3)

Application Number Title Priority Date Filing Date
US14/595,032 Abandoned US20150153895A1 (en) 2005-03-04 2015-01-12 Multi-functional hand-held device
US14/940,010 Abandoned US20160070399A1 (en) 2005-03-04 2015-11-12 Multi-functional hand-held device
US15/933,196 Abandoned US20180217709A1 (en) 2005-03-04 2018-03-22 Multi-functional hand-held device

Country Status (10)

Country Link
US (4) US20080088602A1 (en)
EP (2) EP4177708A1 (en)
JP (8) JP5550211B2 (en)
KR (15) KR101984833B1 (en)
CN (2) CN103336562A (en)
AU (4) AU2006218381B8 (en)
CA (1) CA2600326A1 (en)
DE (2) DE202006020369U1 (en)
HK (7) HK1122461A2 (en)
WO (1) WO2006094308A2 (en)

Cited By (148)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20030095155A1 (en) * 2001-11-16 2003-05-22 Johnson Michael J. Method and apparatus for displaying images on a display
US20050179650A1 (en) * 2004-02-13 2005-08-18 Ludwig Lester F. Extended parameter-set mouse-based user interface device offering offset, warping, and mixed-reference features
US20060238517A1 (en) * 2005-03-04 2006-10-26 Apple Computer, Inc. Electronic Device Having Display and Surrounding Touch Sensitive Bezel for User Interface and Control
US20060263068A1 (en) * 2005-05-19 2006-11-23 Sony Corporation Reproducing apparatus, program, and reproduction control method
US20070010293A1 (en) * 2005-07-08 2007-01-11 Pchome Online Inc. Phone connected to a personal computer
US20070126731A1 (en) * 2005-12-06 2007-06-07 Sabram Stephen R Graphical architecture for handheld measurement system
US20070156679A1 (en) * 2005-12-20 2007-07-05 Kretz Martin H Electronic equipment with shuffle operation
US20070213090A1 (en) * 2006-03-07 2007-09-13 Sony Ericsson Mobile Communications Ab Programmable keypad
US20080030456A1 (en) * 2006-07-19 2008-02-07 Sony Ericsson Mobile Communications Ab Apparatus and Methods for Providing Motion Responsive Output Modifications in an Electronic Device
US20080165160A1 (en) * 2007-01-07 2008-07-10 Kenneth Kocienda Portable Multifunction Device, Method, and Graphical User Interface for Interpreting a Finger Gesture on a Touch Screen Display
US20080168396A1 (en) * 2007-01-07 2008-07-10 Michael Matas Portable Multifunction Device, Method, and Graphical User Interface for Providing Maps and Directions
US20080294984A1 (en) * 2007-05-25 2008-11-27 Immersion Corporation Customizing Haptic Effects On An End User Device
US20090111517A1 (en) * 2007-10-31 2009-04-30 Mitac International Corp. Handheld device capable of independently actuating camera and method thereof
US20090135164A1 (en) * 2007-11-26 2009-05-28 Ki Uk Kyung Pointing apparatus capable of providing haptic feedback, and haptic interaction system and method using the same
US20090143109A1 (en) * 2007-12-03 2009-06-04 Semiconductor Energy Laboratory Co., Ltd. Mobile phone
US20090150659A1 (en) * 2007-12-06 2009-06-11 Fuccello James R Device and Method for Adjusting a Converged Device
US20090174680A1 (en) * 2008-01-06 2009-07-09 Freddy Allen Anzures Portable Multifunction Device, Method, and Graphical User Interface for Viewing and Managing Electronic Calendars
US20090262074A1 (en) * 2007-01-05 2009-10-22 Invensense Inc. Controlling and accessing content using motion processing on mobile devices
US20090289911A1 (en) * 2008-05-20 2009-11-26 Canon Kabushiki Kaisha Information processing apparatus and control method thereof
US20090300498A1 (en) * 2008-05-29 2009-12-03 Telcordia Technologies, Inc. Method and System for Generating and Presenting Mobile Content Summarization
US20090300530A1 (en) * 2008-05-29 2009-12-03 Telcordia Technologies, Inc. Method and system for multi-touch-based browsing of media summarizations on a handheld device
US20090313660A1 (en) * 2008-06-16 2009-12-17 Imu Solutions, Inc. Home entertainment system and operating method thereof
US20100037180A1 (en) * 2008-08-08 2010-02-11 At&T Intellectual Property I, L.P. Touch Screen Remote Control with Dynamic Keypad Layouts
US20100039244A1 (en) * 2008-08-18 2010-02-18 Michael Dean Tschirhart Haptic effect control system
US20100060588A1 (en) * 2008-09-09 2010-03-11 Microsoft Corporation Temporally separate touch input
US20100060586A1 (en) * 2008-09-05 2010-03-11 Pisula Charles J Portable touch screen device, method, and graphical user interface for providing workout support
US20100088630A1 (en) * 2008-10-07 2010-04-08 Microsoft Corporation Content aware adaptive display
US20100097347A1 (en) * 2008-10-21 2010-04-22 Tzu-Chih Lin Pressure Detection Module, and Touch Panel with Pressure Detection Module
WO2010045498A1 (en) * 2008-10-15 2010-04-22 Invensense Inc. Mobile devices with motion gesture recognition
US20100129782A1 (en) * 2008-11-25 2010-05-27 Sony Corporation And Sony Electronics Inc. Electronic book with enhanced features
US20100127995A1 (en) * 2008-11-26 2010-05-27 Panasonic Corporation System and method for differentiating between intended and unintended user input on a touchpad
US20100177034A1 (en) * 2009-01-12 2010-07-15 Samsung Electronics Co., Ltd. Portable storage device having user interface and method of controlling the user interface
US20100188349A1 (en) * 2007-09-14 2010-07-29 Yannick Molard Control panels for onboard instruments
US20100195452A1 (en) * 2005-07-06 2010-08-05 Sony Corporation Contents data reproduction apparatus and contents data reproduction method
US20100214216A1 (en) * 2007-01-05 2010-08-26 Invensense, Inc. Motion sensing and processing on mobile devices
US20100229121A1 (en) * 2009-03-09 2010-09-09 Telcordia Technologies, Inc. System and method for capturing, aggregating and presenting attention hotspots in shared media
US7796872B2 (en) 2007-01-05 2010-09-14 Invensense, Inc. Method and apparatus for producing a sharp image from a handheld device containing a gyroscope
US20100238109A1 (en) * 2007-09-18 2010-09-23 Thomson Licensing User interface for set top box
WO2010111668A1 (en) * 2009-03-26 2010-09-30 Cypress Semiconductor Multi-functional capacitance sensing circuit with a current conveyor
US20100268940A1 (en) * 2009-04-15 2010-10-21 Wyse Technology Inc. Method and apparatus for portability of a remote session
US20100268813A1 (en) * 2009-04-15 2010-10-21 Wyse Technology Inc. System and method for handling remote drawing commands
US20100269046A1 (en) * 2009-04-15 2010-10-21 Wyse Technology Inc. Sever-side computing from a remote client device
US20100269039A1 (en) * 2009-04-15 2010-10-21 Wyse Technology Inc. Custom pointer features for touch-screen on remote client devices
US20100268762A1 (en) * 2009-04-15 2010-10-21 Wyse Technology Inc. System and method for scrolling a remote application
US20100277337A1 (en) * 2009-05-01 2010-11-04 Apple Inc. Directional touch remote
US20100289760A1 (en) * 2007-09-14 2010-11-18 Kyocera Corporation Electronic apparatus
US20100312368A1 (en) * 2009-06-05 2010-12-09 Anthony Rodriguez Aural Audio Player
US20100309149A1 (en) * 2009-06-07 2010-12-09 Chris Blumenberg Device, Method, and Graphical User Interface for Providing Maps, Directions, and Location-Based Information
US20100328224A1 (en) * 2009-06-25 2010-12-30 Apple Inc. Playback control using a touch interface
US7870496B1 (en) 2009-01-29 2011-01-11 Jahanzeb Ahmed Sherwani System using touchscreen user interface of a mobile device to remotely control a host computer
US20110018695A1 (en) * 2009-07-24 2011-01-27 Research In Motion Limited Method and apparatus for a touch-sensitive display
US20110074827A1 (en) * 2009-09-25 2011-03-31 Research In Motion Limited Electronic device including touch-sensitive input device and method of controlling same
US7934423B2 (en) 2007-12-10 2011-05-03 Invensense, Inc. Vertically integrated 3-axis MEMS angular accelerometer with integrated electronics
US20110105186A1 (en) * 2009-10-29 2011-05-05 Research In Motion Limited Systems and methods for providing direct and indirect navigation modes for touchscreen devices
US20110140867A1 (en) * 2008-08-14 2011-06-16 Fm Marketing Gmbh Remote control and method for the remote control of multimedia appliances
US20110148788A1 (en) * 2009-12-17 2011-06-23 Shenzhen Futaihong Precision Industry Co., Ltd. Touch screen device with coordinate correction module
US20110157231A1 (en) * 2009-12-30 2011-06-30 Cywee Group Limited Electronic control apparatus and method for responsively controlling media content displayed on portable electronic device
US20110167058A1 (en) * 2010-01-06 2011-07-07 Van Os Marcel Device, Method, and Graphical User Interface for Mapping Directions Between Search Results
US20110163874A1 (en) * 2010-01-06 2011-07-07 Van Os Marcel Device, Method, and Graphical User Interface for Tracking Movement on a Map
US20110191692A1 (en) * 2010-02-03 2011-08-04 Oto Technologies, Llc System and method for e-book contextual communication
US20110201379A1 (en) * 2010-02-17 2011-08-18 Harris Corporation Communication device with a multi-functional control
US8020441B2 (en) 2008-02-05 2011-09-20 Invensense, Inc. Dual mode sensing for vibratory gyroscope
US20110231474A1 (en) * 2010-03-22 2011-09-22 Howard Locker Audio Book and e-Book Synchronization
US8040321B2 (en) 2006-07-10 2011-10-18 Cypress Semiconductor Corporation Touch-sensor with shared capacitive sensors
US8047075B2 (en) 2007-06-21 2011-11-01 Invensense, Inc. Vertically integrated 3-axis MEMS accelerometer with electronics
US20110267026A1 (en) * 2010-04-30 2011-11-03 Lenovo (Singapore) Pte, Ltd. Method and Apparatus for Modifying a Transition to an Altered Power State of an Electronic Device Based on Accelerometer Output
US8058937B2 (en) 2007-01-30 2011-11-15 Cypress Semiconductor Corporation Setting a discharge rate and a charge rate of a relaxation oscillator circuit
US8059015B2 (en) 2006-05-25 2011-11-15 Cypress Semiconductor Corporation Capacitance sensing matrix for keyboard architecture
US20110281533A1 (en) * 2010-05-11 2011-11-17 Harris Corporation Electronic device with a multi-functional control
US20110279382A1 (en) * 2010-05-14 2011-11-17 Research In Motion Limited Electronic device including tactile touch-sensitive display and method of controlling same
US20110285636A1 (en) * 2010-05-20 2011-11-24 Howard John W Touch screen with virtual joystick and methods for use therewith
US20110291958A1 (en) * 2010-05-28 2011-12-01 Fih (Hong Kong) Limited Touch-type transparent keyboard
US20120062473A1 (en) * 2010-09-15 2012-03-15 Verizon Patent And Licensing Inc. Media experience for touch screen devices
US8141424B2 (en) 2008-09-12 2012-03-27 Invensense, Inc. Low inertia frame for detecting coriolis acceleration
US8144125B2 (en) 2006-03-30 2012-03-27 Cypress Semiconductor Corporation Apparatus and method for reducing average scan rate to detect a conductive object on a sensing device
US20120079265A1 (en) * 2009-06-16 2012-03-29 Bran Ferren Multi-mode handheld wireless device
US20120102439A1 (en) * 2010-10-22 2012-04-26 April Slayden Mitchell System and method of modifying the display content based on sensor input
US20120139953A1 (en) * 2006-08-02 2012-06-07 Research In Motion Limited System and method for adjusting presentation of text and images on an electronic device according to an orientation of the device
US8250921B2 (en) 2007-07-06 2012-08-28 Invensense, Inc. Integrated motion processing unit (MPU) with MEMS inertial sensing and embedded digital electronics
US8258986B2 (en) 2007-07-03 2012-09-04 Cypress Semiconductor Corporation Capacitive-matrix keyboard with multiple touch detection
US20120233226A1 (en) * 2011-03-10 2012-09-13 Chi Mei Communication Systems, Inc. Electronic device and file management method
US20120299839A1 (en) * 2011-05-27 2012-11-29 Honeywell International Inc. Aircraft user interfaces with multi-mode haptics
CN103034394A (en) * 2011-10-06 2013-04-10 东芝三星存储技术韩国株式会社 Multimedia device having detachable controller
US20130097032A1 (en) * 2007-11-30 2013-04-18 Blaze Mobile, Inc. Utilizing shopping lists for nfc transactions
US20130099903A1 (en) * 2011-10-24 2013-04-25 Hon Hai Precision Industry Co., Ltd. Remote controller
US20130135205A1 (en) * 2010-08-19 2013-05-30 Beijing Lenovo Software Ltd. Display Method And Terminal Device
US20130141854A1 (en) * 2008-04-01 2013-06-06 Yves Behar Portable computer with multiple display configurations
US8487639B1 (en) 2008-11-21 2013-07-16 Cypress Semiconductor Corporation Receive demodulator for capacitive sensing
US8508039B1 (en) 2008-05-08 2013-08-13 Invensense, Inc. Wafer scale chip scale packaging of vertically integrated MEMS sensors with electronics
US20130246565A1 (en) * 2011-09-19 2013-09-19 Qualcomn Incorporated Sending human input device commands over internet protocol
JP2014016989A (en) * 2012-07-05 2014-01-30 Samsung Electronics Co Ltd Input sensing method, and electronic apparatus for processing said method
US8654524B2 (en) 2009-08-17 2014-02-18 Apple Inc. Housing as an I/O device
US20140160893A1 (en) * 2012-12-10 2014-06-12 Apple Inc. Mobile device with ultrasound ranging
US8754873B2 (en) * 2009-12-14 2014-06-17 Synaptics Incorporated System and method for measuring individual force in multi-object sensing
US8762890B2 (en) 2010-07-27 2014-06-24 Telcordia Technologies, Inc. System and method for interactive projection and playback of relevant media segments onto the facets of three-dimensional shapes
US8775966B2 (en) 2011-06-29 2014-07-08 Motorola Mobility Llc Electronic device and method with dual mode rear TouchPad
US20140232675A1 (en) * 2011-09-22 2014-08-21 Toyota Jidosha Kabushiki Kaisha Input apparatus and input apparatus processing method
US8854802B2 (en) 2010-10-22 2014-10-07 Hewlett-Packard Development Company, L.P. Display with rotatable display screen
US8897897B2 (en) 2011-05-25 2014-11-25 Remote Technologies, Inc. Companion control interface for smart devices
US8952832B2 (en) 2008-01-18 2015-02-10 Invensense, Inc. Interfacing application programs and motion sensors of a device
US8976124B1 (en) 2007-05-07 2015-03-10 Cypress Semiconductor Corporation Reducing sleep current in a capacitance sensing system
USRE45559E1 (en) 1997-10-28 2015-06-09 Apple Inc. Portable computers
US9052791B2 (en) 2011-12-16 2015-06-09 Panasonic Intellectual Property Corporation Of America Touch panel and electronic device
US9110499B2 (en) * 2006-08-02 2015-08-18 Blackberry Limited System and method for adjusting presentation of moving images on an electronic device according to an orientation of the device
US9164581B2 (en) 2010-10-22 2015-10-20 Hewlett-Packard Development Company, L.P. Augmented reality display system and method of display
US20160011699A1 (en) * 2008-08-15 2016-01-14 Intel Corporation Techniques For Automatically Distinguishing Between Users Of A Handheld Device
US9268441B2 (en) 2011-04-05 2016-02-23 Parade Technologies, Ltd. Active integrator for a capacitive sense array
US9298293B2 (en) 2012-04-19 2016-03-29 Panasonic Intellectual Property Corporation Of America Portable electronic apparatus
WO2015184116A3 (en) * 2014-05-28 2016-06-02 Sioux Automation Center Inc Electronic remote control mixing system, associated devices and methods
US9372615B2 (en) 2010-12-28 2016-06-21 Samsung Electronics Co., Ltd. Method and apparatus for touch input for electronic device
US9417754B2 (en) * 2011-08-05 2016-08-16 P4tents1, LLC User interface system, method, and computer program product
US20160266653A1 (en) * 2013-04-15 2016-09-15 Zte Corporation Gesture control method, apparatus and system
US9448712B2 (en) 2007-01-07 2016-09-20 Apple Inc. Application programming interfaces for scrolling operations
US9495070B2 (en) 2008-04-01 2016-11-15 Litl Llc Method and apparatus for managing digital media content
US9517411B2 (en) 2010-04-28 2016-12-13 Kabushiki Kaisha Square Enix Transparent user interface game control processing method, apparatus, and medium
US9537935B2 (en) 2013-09-30 2017-01-03 Eric Trent Dryden Consumer PC in the cloud
US9553953B2 (en) 2009-04-15 2017-01-24 Dell Products L.P. Method and apparatus for extending capabilities of a virtualization domain to support features available in a normal desktop application
US9578113B2 (en) 2009-04-15 2017-02-21 Wyse Technology L.L.C. Method and apparatus for transferring remote session data
US20170115867A1 (en) * 2015-10-27 2017-04-27 Yahoo! Inc. Method and system for interacting with a touch screen
US20170150201A1 (en) * 2011-06-20 2017-05-25 Enseo, Inc. Set Top/Back Box, System and Method for Providing a Remote Control Device
US9667434B2 (en) 2015-04-03 2017-05-30 Sonix Technology Co., Ltd. Information transmission system and method
US9785258B2 (en) 2003-09-02 2017-10-10 Apple Inc. Ambidextrous mouse
US9817486B2 (en) 2013-07-11 2017-11-14 Samsung Electronics Co., Ltd. User terminal device for displaying contents and methods thereof
EP3258679A1 (en) * 2009-01-23 2017-12-20 Samsung Electronics Co., Ltd. Mobile terminal having dual touch screen and method of controlling content therein
US9880715B2 (en) 2008-04-01 2018-01-30 Litl Llc System and method for streamlining user interaction with electronic content
WO2018148503A1 (en) * 2017-02-09 2018-08-16 Nextinput, Inc. Integrated digital force sensors and related methods of manufacture
US10148998B2 (en) * 2011-06-20 2018-12-04 Enseo, Inc. Set-top box with enhanced functionality and system and method for use of same
US10149005B2 (en) * 2011-06-20 2018-12-04 Enseo, Inc. Set-top box with enhanced content and system and method for use of same
US10168900B2 (en) * 2010-08-20 2019-01-01 Sony Corporation Information processing apparatus, program, and operation control me
US10178208B2 (en) 2012-01-07 2019-01-08 Samsung Electronics Co., Ltd. Method and apparatus for providing event of portable device having flexible display unit
US10242397B2 (en) * 2009-07-29 2019-03-26 Paypal, Inc. No authentication payment and seamless authentication
US10372212B2 (en) 2015-05-29 2019-08-06 Google Llc Techniques for simulated physical interaction between users via their mobile computing devices
US10379806B2 (en) 2016-11-04 2019-08-13 International Business Machines Corporation Dynamic selection for touch sensor
US20190335224A1 (en) * 2011-06-20 2019-10-31 Enseo, Inc. Television and system and method for providing a remote control device
US10466119B2 (en) 2015-06-10 2019-11-05 Nextinput, Inc. Ruggedized wafer level MEMS force sensor with a tolerance trench
US10686930B2 (en) 2007-06-22 2020-06-16 Apple Inc. Touch screen device, method, and graphical user interface for providing maps, directions, and location based information
US10776763B2 (en) 2017-09-12 2020-09-15 Paypal, Inc. System and method for performing gesture-based unified multi-marketplace transactions
US10962427B2 (en) 2019-01-10 2021-03-30 Nextinput, Inc. Slotted MEMS force sensor
US11063464B2 (en) 2016-07-07 2021-07-13 Samsung Electronics Co., Ltd. Apparatus and method for altering wireless charging mode
US11221263B2 (en) 2017-07-19 2022-01-11 Nextinput, Inc. Microelectromechanical force sensor having a strain transfer layer arranged on the sensor die
US11243126B2 (en) 2017-07-27 2022-02-08 Nextinput, Inc. Wafer bonded piezoresistive and piezoelectric force sensor and related methods of manufacture
US11243125B2 (en) 2017-02-09 2022-02-08 Nextinput, Inc. Integrated piezoresistive and piezoelectric fusion force sensor
US11275405B2 (en) 2005-03-04 2022-03-15 Apple Inc. Multi-functional hand-held device
US11385108B2 (en) 2017-11-02 2022-07-12 Nextinput, Inc. Sealed force sensor with etch stop layer
US11423686B2 (en) 2017-07-25 2022-08-23 Qorvo Us, Inc. Integrated fingerprint and force sensor
US11481050B1 (en) * 2021-07-30 2022-10-25 Primax Electronics Ltd. Function parameter switching memory circuit for peripheral device and operating method thereof
US11579028B2 (en) 2017-10-17 2023-02-14 Nextinput, Inc. Temperature coefficient of offset compensation for force sensor and strain gauge
US11874185B2 (en) 2017-11-16 2024-01-16 Nextinput, Inc. Force attenuator for force sensor

Families Citing this family (294)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7840912B2 (en) * 2006-01-30 2010-11-23 Apple Inc. Multi-touch gesture dictionary
US8018440B2 (en) 2005-12-30 2011-09-13 Microsoft Corporation Unintentional touch rejection
US8564543B2 (en) 2006-09-11 2013-10-22 Apple Inc. Media player with imaged based browsing
US8736557B2 (en) 2006-09-11 2014-05-27 Apple Inc. Electronic device with image based browsers
KR101144423B1 (en) * 2006-11-16 2012-05-10 엘지전자 주식회사 Mobile phone and display method of the same
KR101533465B1 (en) 2006-12-27 2015-07-02 임머숀 코퍼레이션 Virtual detents through vibrotactile feedback
KR100831721B1 (en) 2006-12-29 2008-05-22 엘지전자 주식회사 Apparatus and method for displaying of mobile terminal
US7802201B2 (en) 2006-12-29 2010-09-21 Research In Motion Limited System and method for panning and zooming an image on a display of a handheld electronic device
US9311528B2 (en) 2007-01-03 2016-04-12 Apple Inc. Gesture learning
US7924271B2 (en) * 2007-01-05 2011-04-12 Apple Inc. Detecting gestures on multi-event sensitive devices
US20080168478A1 (en) 2007-01-07 2008-07-10 Andrew Platzer Application Programming Interfaces for Scrolling
US20080168402A1 (en) 2007-01-07 2008-07-10 Christopher Blumenberg Application Programming Interfaces for Gesture Operations
US9001047B2 (en) 2007-01-07 2015-04-07 Apple Inc. Modal change based on orientation of a portable multifunction device
US20080207254A1 (en) * 2007-02-27 2008-08-28 Pierce Paul M Multimodal Adaptive User Interface for a Portable Electronic Device
US20080204463A1 (en) * 2007-02-27 2008-08-28 Adam Cybart Adaptable User Interface and Mechanism for a Title Portable Electronic Device
US20080249682A1 (en) * 2007-04-06 2008-10-09 Visteon Global Technologies, Inc. Touch control bezel for display devices
TWI351638B (en) * 2007-04-27 2011-11-01 Htc Corp Touch-based tab navigation method and related devi
US8902152B2 (en) 2007-04-30 2014-12-02 Motorola Mobility Llc Dual sided electrophoretic display
US20080300016A1 (en) * 2007-05-31 2008-12-04 Vocollect, Inc. Aural feedback apparatus for user controls
US9052817B2 (en) * 2007-06-13 2015-06-09 Apple Inc. Mode sensitive processing of touch data
US9122092B2 (en) 2007-06-22 2015-09-01 Google Technology Holdings LLC Colored morphing apparatus for an electronic device
US8432365B2 (en) 2007-08-30 2013-04-30 Lg Electronics Inc. Apparatus and method for providing feedback for three-dimensional touchscreen
US8219936B2 (en) * 2007-08-30 2012-07-10 Lg Electronics Inc. User interface for a mobile device using a user's gesture in the proximity of an electronic device
US8116831B2 (en) 2007-11-29 2012-02-14 Motorola Mobility, Inc. Hand-held communication device with auxiliary input apparatus, and method
TWI362673B (en) 2007-12-31 2012-04-21 Htc Corp Touch sensor keypad with tactile feedback mechanisms
US8413075B2 (en) 2008-01-04 2013-04-02 Apple Inc. Gesture movies
US8059232B2 (en) 2008-02-08 2011-11-15 Motorola Mobility, Inc. Electronic device and LC shutter for polarization-sensitive switching between transparent and diffusive states
US8717305B2 (en) * 2008-03-04 2014-05-06 Apple Inc. Touch event model for web pages
US8416196B2 (en) 2008-03-04 2013-04-09 Apple Inc. Touch event model programming interface
US8174502B2 (en) * 2008-03-04 2012-05-08 Apple Inc. Touch event processing for web pages
US8645827B2 (en) 2008-03-04 2014-02-04 Apple Inc. Touch event model
US20090225026A1 (en) * 2008-03-06 2009-09-10 Yaron Sheba Electronic device for selecting an application based on sensed orientation and methods for use therewith
KR100952699B1 (en) * 2008-03-10 2010-04-13 한국표준과학연구원 Full-browsing display method in touchscreen apparatus using tactile sensors
KR20100000514A (en) * 2008-06-25 2010-01-06 엘지전자 주식회사 Image display device with touch screen and method of controlling the same
KR101493089B1 (en) * 2008-07-10 2015-02-12 주식회사 케이티 Method For Providing User Interface In Touch Input Recognizing Apparatus And Touch Input Recognizing Device Performing The Same
JP5690726B2 (en) 2008-07-15 2015-03-25 イマージョン コーポレーションImmersion Corporation System and method for haptic messaging based on physical laws
KR101467881B1 (en) 2008-08-18 2014-12-02 엘지전자 주식회사 Controlling a Mobile Terminal with at least two display area
US20100064261A1 (en) * 2008-09-09 2010-03-11 Microsoft Corporation Portable electronic device with relative gesture recognition mode
KR101522883B1 (en) * 2008-10-07 2015-05-26 엘지전자 주식회사 Mobile terminal and mobile terminal control method
JP5228796B2 (en) * 2008-10-28 2013-07-03 ソニー株式会社 Portable information terminal, hold state switching method, and hold state switching program
JP4853507B2 (en) 2008-10-30 2012-01-11 ソニー株式会社 Information processing apparatus, information processing method, and program
US8856690B2 (en) 2008-10-31 2014-10-07 Sprint Communications Company L.P. Associating gestures on a touch screen with characters
US20100110020A1 (en) * 2008-10-31 2010-05-06 Sprint Communications Company L.P. Virtual press number pad
KR101530546B1 (en) * 2008-11-28 2015-06-23 삼성전자주식회사 Input Device For Portable Device And Method thereof
US20100153313A1 (en) * 2008-12-15 2010-06-17 Symbol Technologies, Inc. Interface adaptation system
GB2466077A (en) * 2008-12-15 2010-06-16 Symbian Software Ltd Emulator for multiple computing device inputs
US20100166404A1 (en) * 2008-12-31 2010-07-01 Lombardi Michael J Device and Method Using a Touch-Detecting Surface
JP4880003B2 (en) * 2009-01-20 2012-02-22 シャープ株式会社 Information processing apparatus, control method, and program
KR101538705B1 (en) * 2009-01-29 2015-07-22 삼성전자주식회사 System and method for controlling function of portable terminal
TWI378382B (en) 2009-02-13 2012-12-01 Htc Corp Method, apparatus and computer program product for preventing on-screen buttons from being mistakenly touched
US9311112B2 (en) 2009-03-16 2016-04-12 Apple Inc. Event recognition
US8566045B2 (en) 2009-03-16 2013-10-22 Apple Inc. Event recognition
US8566044B2 (en) 2009-03-16 2013-10-22 Apple Inc. Event recognition
US8285499B2 (en) 2009-03-16 2012-10-09 Apple Inc. Event recognition
US9684521B2 (en) 2010-01-26 2017-06-20 Apple Inc. Systems having discrete and continuous gesture recognizers
JP5620134B2 (en) * 2009-03-30 2014-11-05 アバイア インク. A system and method for managing trust relationships in a communication session using a graphical display.
US20150208012A1 (en) * 2009-05-15 2015-07-23 Tara Chand Singhal Multi-purpose handheld electronic command and control device with ergonomic features
US8836648B2 (en) 2009-05-27 2014-09-16 Microsoft Corporation Touch pull-in gesture
CN102460510B (en) * 2009-05-27 2016-04-20 奥布隆工业有限公司 For the space multi-mode opertaing device used together with spatial operation system
JP2011039765A (en) * 2009-08-11 2011-02-24 Nec Corp Function starting apparatus and function starting method
KR20110031797A (en) * 2009-09-21 2011-03-29 삼성전자주식회사 Input device for portable device and method including the same
US8780055B2 (en) 2009-10-02 2014-07-15 Blackberry Limited Low power wakeup detection circuit and a portable electronic device having a low power wakeup detection circuit
KR101069496B1 (en) * 2009-10-09 2011-09-30 엘지이노텍 주식회사 E-book
US20110087963A1 (en) * 2009-10-09 2011-04-14 At&T Mobility Ii Llc User Interface Control with Edge Finger and Motion Sensing
US8432367B2 (en) 2009-11-19 2013-04-30 Google Inc. Translating user interaction with a touch screen into input commands
CN103903418A (en) * 2009-12-14 2014-07-02 义隆电子股份有限公司 Multifunctional touchpad remote controller and control method thereof
KR20110069476A (en) * 2009-12-17 2011-06-23 주식회사 아이리버 Hand hrld electronic device to reflecting grip of user and method thereof
US9489073B2 (en) 2009-12-21 2016-11-08 Promethean Limited Multi-point contacts with pressure data on an interactive surface
US8736561B2 (en) 2010-01-06 2014-05-27 Apple Inc. Device, method, and graphical user interface with content display modes and display rotation heuristics
JP2011145829A (en) * 2010-01-13 2011-07-28 Buffalo Inc Operation input device
US8334840B2 (en) 2010-01-19 2012-12-18 Visteon Global Technologies, Inc. System and method of screen manipulation using haptic enable controller
US8261213B2 (en) 2010-01-28 2012-09-04 Microsoft Corporation Brush, carbon-copy, and fill gestures
US9411504B2 (en) 2010-01-28 2016-08-09 Microsoft Technology Licensing, Llc Copy and staple gestures
US9519356B2 (en) 2010-02-04 2016-12-13 Microsoft Technology Licensing, Llc Link gestures
US8799827B2 (en) 2010-02-19 2014-08-05 Microsoft Corporation Page manipulations using on and off-screen gestures
US9367205B2 (en) 2010-02-19 2016-06-14 Microsoft Technolgoy Licensing, Llc Radial menus with bezel gestures
US9310994B2 (en) 2010-02-19 2016-04-12 Microsoft Technology Licensing, Llc Use of bezel as an input mechanism
US9274682B2 (en) * 2010-02-19 2016-03-01 Microsoft Technology Licensing, Llc Off-screen gestures to create on-screen input
US9965165B2 (en) * 2010-02-19 2018-05-08 Microsoft Technology Licensing, Llc Multi-finger gestures
JP5403271B2 (en) * 2010-02-24 2014-01-29 日本電気株式会社 Portable terminal device, calculation method, and calculation program
US9075522B2 (en) 2010-02-25 2015-07-07 Microsoft Technology Licensing, Llc Multi-screen bookmark hold gesture
US8707174B2 (en) 2010-02-25 2014-04-22 Microsoft Corporation Multi-screen hold and page-flip gesture
US9454304B2 (en) 2010-02-25 2016-09-27 Microsoft Technology Licensing, Llc Multi-screen dual tap gesture
US8751970B2 (en) 2010-02-25 2014-06-10 Microsoft Corporation Multi-screen synchronous slide gesture
US8473870B2 (en) 2010-02-25 2013-06-25 Microsoft Corporation Multi-screen hold and drag gesture
US8539384B2 (en) 2010-02-25 2013-09-17 Microsoft Corporation Multi-screen pinch and expand gestures
JP2011198009A (en) * 2010-03-19 2011-10-06 Sony Corp Electro-optical device with input function
US8781420B2 (en) 2010-04-13 2014-07-15 Apple Inc. Adjustable wireless circuitry with antenna-based proximity detector
US20130050129A1 (en) * 2010-05-04 2013-02-28 Nokia Corporation Responding to touch inputs
WO2011139194A1 (en) * 2010-05-07 2011-11-10 Maquet Critical Care Ab User interface for breathing apparatus
US20110296333A1 (en) * 2010-05-25 2011-12-01 Bateman Steven S User interaction gestures with virtual keyboard
US8552999B2 (en) 2010-06-14 2013-10-08 Apple Inc. Control selection approximation
EP2612212A1 (en) 2010-08-30 2013-07-10 Telefonaktiebolaget L M Ericsson (publ) Methods of launching applications responsive to device orientation and related electronic devices
TW201214237A (en) * 2010-09-16 2012-04-01 Asustek Comp Inc Touch display device and control method thereof
DE112011103173T5 (en) 2010-09-24 2013-08-14 Qnx Software Systems Limited Transitional view on a portable electronic device
DE112011101209T5 (en) 2010-09-24 2013-01-17 Qnx Software Systems Ltd. Alert Display on a portable electronic device
ES2900188T3 (en) 2010-09-24 2022-03-16 Huawei Tech Co Ltd Portable electronic device and method for controlling the same
DE102010042527A1 (en) * 2010-10-15 2012-04-19 Siemens Aktiengesellschaft Device with a display and control unit
US20120092279A1 (en) * 2010-10-18 2012-04-19 Qualcomm Mems Technologies, Inc. Touch sensor with force-actuated switched capacitor
KR102188757B1 (en) * 2010-11-18 2020-12-08 구글 엘엘씨 Surfacing off-screen visible objects
US8982045B2 (en) * 2010-12-17 2015-03-17 Microsoft Corporation Using movement of a computing device to enhance interpretation of input events produced when interacting with the computing device
US20120159395A1 (en) 2010-12-20 2012-06-21 Microsoft Corporation Application-launching interface for multiple modes
CN102566806A (en) * 2010-12-20 2012-07-11 联胜(中国)科技有限公司 Touch panel and touch display panel with same
JP2012133453A (en) * 2010-12-20 2012-07-12 Sony Corp Information processing device, information processing method and program
US8612874B2 (en) 2010-12-23 2013-12-17 Microsoft Corporation Presenting an application change through a tile
US8689123B2 (en) 2010-12-23 2014-04-01 Microsoft Corporation Application reporting in an application-selectable user interface
US10042546B2 (en) 2011-01-07 2018-08-07 Qualcomm Incorporated Systems and methods to present multiple frames on a touch screen
JP2012168932A (en) * 2011-02-10 2012-09-06 Sony Computer Entertainment Inc Input device, information processing device and input value acquisition method
US8970230B2 (en) 2011-02-28 2015-03-03 Cypress Semiconductor Corporation Capacitive sensing button on chip
EP2508969A1 (en) * 2011-04-05 2012-10-10 Research In Motion Limited Electronic device and method of controlling same
US8872773B2 (en) 2011-04-05 2014-10-28 Blackberry Limited Electronic device and method of controlling same
US9298363B2 (en) 2011-04-11 2016-03-29 Apple Inc. Region activation for touch sensitive surface
US9658766B2 (en) 2011-05-27 2017-05-23 Microsoft Technology Licensing, Llc Edge gesture
US9104440B2 (en) 2011-05-27 2015-08-11 Microsoft Technology Licensing, Llc Multi-application environment
US9104307B2 (en) 2011-05-27 2015-08-11 Microsoft Technology Licensing, Llc Multi-application environment
US9158445B2 (en) 2011-05-27 2015-10-13 Microsoft Technology Licensing, Llc Managing an immersive interface in a multi-application immersive environment
US8893033B2 (en) 2011-05-27 2014-11-18 Microsoft Corporation Application notifications
JP2011227917A (en) * 2011-06-24 2011-11-10 Kyocera Corp Portable terminal
DE102011052636A1 (en) * 2011-08-12 2013-02-14 Medion Ag Input device for the remote control of electronic devices
US20130057587A1 (en) 2011-09-01 2013-03-07 Microsoft Corporation Arranging tiles
US9541966B2 (en) * 2011-09-07 2017-01-10 Htc Corporation Systems and methods for utilizing acceleration event signatures
US9146670B2 (en) 2011-09-10 2015-09-29 Microsoft Technology Licensing, Llc Progressively indicating new content in an application-selectable user interface
CN102999198B (en) * 2011-09-16 2016-03-30 宸鸿科技(厦门)有限公司 Touch panel edge holds detection method and the device of touch
WO2013040824A1 (en) * 2011-09-23 2013-03-28 Xu Minglu Tablet pc applying virtual frames
US8836654B2 (en) * 2011-10-04 2014-09-16 Qualcomm Incorporated Application window position and size control in (multi-fold) multi-display devices
DE202011110369U1 (en) * 2011-12-01 2013-09-26 Jürgen Habenstein digital camera
DE102011056940A1 (en) 2011-12-22 2013-06-27 Bauhaus Universität Weimar A method of operating a multi-touch display and device having a multi-touch display
CN103257807A (en) * 2012-02-17 2013-08-21 林卓毅 State switching method of mobile communication device and portable electronic device
US20130219340A1 (en) * 2012-02-21 2013-08-22 Sap Ag Navigation on a Portable Electronic Device
CN103324329B (en) * 2012-03-23 2016-07-06 联想(北京)有限公司 A kind of method of toch control and device
US9024910B2 (en) 2012-04-23 2015-05-05 Qualcomm Mems Technologies, Inc. Touchscreen with bridged force-sensitive resistors
DE102012008681A1 (en) * 2012-04-28 2013-10-31 Audi Ag Multifunction control device, in particular for a motor vehicle
DE112013002288T5 (en) 2012-05-03 2015-04-16 Apple Inc. Moment compensated bending beam sensor for load measurement on a bending beam supported platform
KR101670570B1 (en) 2012-05-09 2016-10-28 애플 인크. Device, method, and graphical user interface for selecting user interface objects
EP2847662B1 (en) 2012-05-09 2020-02-19 Apple Inc. Device, method, and graphical user interface for providing feedback for changing activation states of a user interface object
WO2013169842A2 (en) 2012-05-09 2013-11-14 Yknots Industries Llc Device, method, and graphical user interface for selecting object within a group of objects
WO2013169843A1 (en) 2012-05-09 2013-11-14 Yknots Industries Llc Device, method, and graphical user interface for manipulating framed graphical objects
WO2013169851A2 (en) 2012-05-09 2013-11-14 Yknots Industries Llc Device, method, and graphical user interface for facilitating user interaction with controls in a user interface
WO2013169865A2 (en) 2012-05-09 2013-11-14 Yknots Industries Llc Device, method, and graphical user interface for moving a user interface object based on an intensity of a press input
WO2013169870A1 (en) 2012-05-09 2013-11-14 Yknots Industries Llc Device, method, and graphical user interface for transitioning between display states in response to gesture
CN104487929B (en) 2012-05-09 2018-08-17 苹果公司 For contacting the equipment for carrying out display additional information, method and graphic user interface in response to user
WO2013169849A2 (en) 2012-05-09 2013-11-14 Industries Llc Yknots Device, method, and graphical user interface for displaying user interface objects corresponding to an application
JP2015519656A (en) 2012-05-09 2015-07-09 アップル インコーポレイテッド Device, method and graphical user interface for moving and dropping user interface objects
US9459781B2 (en) 2012-05-09 2016-10-04 Apple Inc. Context-specific user interfaces for displaying animated sequences
WO2013169875A2 (en) 2012-05-09 2013-11-14 Yknots Industries Llc Device, method, and graphical user interface for displaying content associated with a corresponding affordance
WO2013169845A1 (en) 2012-05-09 2013-11-14 Yknots Industries Llc Device, method, and graphical user interface for scrolling nested regions
EP3185116B1 (en) 2012-05-09 2019-09-11 Apple Inc. Device, method and graphical user interface for providing tactile feedback for operations performed in a user interface
EP2852882B1 (en) 2012-05-21 2021-08-18 Samsung Electronics Co., Ltd. Method and apparatus of controlling user interface using touch screen
EP2859427B1 (en) * 2012-06-12 2019-08-07 BlackBerry Limited Electronic device and method of control of displays
CN103593078B (en) * 2012-08-14 2017-02-15 宸鸿科技(厦门)有限公司 Touch panel
US9116546B2 (en) 2012-08-29 2015-08-25 Immersion Corporation System for haptically representing sensor input
CN102902467B (en) * 2012-09-13 2015-03-18 广东欧珀移动通信有限公司 Text cursor positioning method of terminal equipment and terminal equipment
US10671165B2 (en) * 2012-09-25 2020-06-02 Nokia Technologies Oy Method and display device with tactile feedback
US20140085217A1 (en) * 2012-09-27 2014-03-27 Htc Corporation Electronic apparatus and control method thereof
JP2014081732A (en) * 2012-10-15 2014-05-08 Ntt Docomo Inc Portable electronic device and display method
US9582122B2 (en) 2012-11-12 2017-02-28 Microsoft Technology Licensing, Llc Touch-sensitive bezel techniques
EP2924958B1 (en) * 2012-11-20 2017-12-27 NEC Corporation Portable electronic device, method for controlling same, and program
WO2014097930A1 (en) * 2012-12-17 2014-06-26 三菱電機株式会社 Video processing device and video processing system
WO2014098946A1 (en) 2012-12-17 2014-06-26 Changello Enterprise Llc Force detection in touch devices using piezoelectric sensors
US9645678B2 (en) * 2012-12-18 2017-05-09 Seiko Epson Corporation Display device, and method of controlling display device
JP2014126949A (en) * 2012-12-25 2014-07-07 Kyocera Corp Portable terminal equipment, screen control method and program
CN105264479B (en) 2012-12-29 2018-12-25 苹果公司 Equipment, method and graphic user interface for navigating to user interface hierarchical structure
WO2014105279A1 (en) 2012-12-29 2014-07-03 Yknots Industries Llc Device, method, and graphical user interface for switching between user interfaces
EP2939095B1 (en) 2012-12-29 2018-10-03 Apple Inc. Device, method, and graphical user interface for moving a cursor according to a change in an appearance of a control icon with simulated three-dimensional characteristics
EP3564806B1 (en) * 2012-12-29 2024-02-21 Apple Inc. Device, method and graphical user interface for determining whether to scroll or select contents
KR101958517B1 (en) 2012-12-29 2019-03-14 애플 인크. Device, method, and graphical user interface for transitioning between touch input to display output relationships
EP2912542B1 (en) 2012-12-29 2022-07-13 Apple Inc. Device and method for forgoing generation of tactile output for a multi-contact gesture
KR20140138224A (en) 2013-01-06 2014-12-03 인텔 코오퍼레이션 A method, apparatus, and system for distributed pre-processing of touch data and display region control
KR102004884B1 (en) * 2013-01-07 2019-07-29 삼성전자주식회사 Method and apparatus for controlling animated image in an electronic device
CN104035648A (en) * 2013-03-04 2014-09-10 联想(北京)有限公司 State switchover method and electronic terminal
WO2014149023A1 (en) 2013-03-15 2014-09-25 Rinand Solutions Llc Force sensing of inputs through strain analysis
CN103197887B (en) * 2013-03-23 2016-06-08 苏州佳世达电通有限公司 Portable electric device
JP2014222388A (en) * 2013-05-13 2014-11-27 信越ポリマー株式会社 Light-emitting capacitance type input device
US10073587B2 (en) 2013-05-29 2018-09-11 Beijing Lenovo Software Ltd. Information prompt method and electronic apparatus
CN104216640B (en) * 2013-05-29 2019-12-24 联想(北京)有限公司 Physical key position prompting method and electronic equipment
WO2014192125A1 (en) * 2013-05-30 2014-12-04 株式会社 東芝 Electronic device and processing method
US9733716B2 (en) 2013-06-09 2017-08-15 Apple Inc. Proxy gesture recognizer
US9671889B1 (en) 2013-07-25 2017-06-06 Apple Inc. Input member with capacitive sensor
KR20150019165A (en) * 2013-08-12 2015-02-25 엘지전자 주식회사 Mobile terminal and method for controlling the same
US11068128B2 (en) 2013-09-03 2021-07-20 Apple Inc. User interface object manipulations in a user interface
KR20180128091A (en) 2013-09-03 2018-11-30 애플 인크. User interface for manipulating user interface objects with magnetic properties
WO2015057936A1 (en) * 2013-10-18 2015-04-23 Citrix Systems, Inc. Providing enhanced application interoperability
US10120478B2 (en) 2013-10-28 2018-11-06 Apple Inc. Piezo based force sensing
US20150153854A1 (en) * 2013-12-03 2015-06-04 Lenovo (Singapore) Pte. Ltd. Extension of wearable information handling device user interface
KR102131828B1 (en) * 2013-12-03 2020-07-09 엘지전자 주식회사 Terminal and method for controlling the same
AU2015100011B4 (en) 2014-01-13 2015-07-16 Apple Inc. Temperature compensating transparent force sensor
CN104796599B (en) * 2014-01-17 2019-01-01 宏达国际电子股份有限公司 Electronic device and its control method with a switch button
EP3100144A4 (en) * 2014-01-31 2017-08-23 Hewlett-Packard Development Company, L.P. Touch sensor
WO2015123322A1 (en) 2014-02-12 2015-08-20 Apple Inc. Force determination employing sheet sensor and capacitive array
US9398456B2 (en) 2014-03-07 2016-07-19 Apple Inc. Electronic device with accessory-based transmit power control
US9477337B2 (en) 2014-03-14 2016-10-25 Microsoft Technology Licensing, Llc Conductive trace routing for display and bezel sensors
US9798399B2 (en) * 2014-06-02 2017-10-24 Synaptics Incorporated Side sensing for electronic devices
JP6050282B2 (en) * 2014-06-09 2016-12-21 富士フイルム株式会社 Electronics
US9444425B2 (en) 2014-06-20 2016-09-13 Apple Inc. Electronic device with adjustable wireless circuitry
AU2015279545B2 (en) 2014-06-27 2018-02-22 Apple Inc. Manipulation of calendar application in device with touch screen
US10135905B2 (en) 2014-07-21 2018-11-20 Apple Inc. Remote user interface
WO2016027779A1 (en) * 2014-08-22 2016-02-25 シャープ株式会社 Touch panel device
WO2016036541A2 (en) 2014-09-02 2016-03-10 Apple Inc. Phone user interface
KR102255143B1 (en) 2014-09-02 2021-05-25 삼성전자주식회사 Potable terminal device comprisings bended display and method for controlling thereof
US10073590B2 (en) 2014-09-02 2018-09-11 Apple Inc. Reduced size user interface
WO2016036414A1 (en) 2014-09-02 2016-03-10 Apple Inc. Button functionality
WO2016036509A1 (en) 2014-09-02 2016-03-10 Apple Inc. Electronic mail user interface
DE202014011147U1 (en) * 2014-10-16 2018-02-16 "Durable" Hunke & Jochheim Gmbh & Co. Kommanditgesellschaft lighting system
US9652057B2 (en) * 2014-12-31 2017-05-16 Synaptics Incorporated Top mount clickpad module for bi-level basin
KR20160099399A (en) * 2015-02-12 2016-08-22 엘지전자 주식회사 Watch type terminal
US10108269B2 (en) 2015-03-06 2018-10-23 Align Technology, Inc. Intraoral scanner with touch sensitive input
US10006937B2 (en) 2015-03-06 2018-06-26 Apple Inc. Capacitive sensors for electronic devices and methods of forming the same
US9990107B2 (en) 2015-03-08 2018-06-05 Apple Inc. Devices, methods, and graphical user interfaces for displaying and using menus
US10095396B2 (en) 2015-03-08 2018-10-09 Apple Inc. Devices, methods, and graphical user interfaces for interacting with a control object while dragging another object
US9632664B2 (en) 2015-03-08 2017-04-25 Apple Inc. Devices, methods, and graphical user interfaces for manipulating user interface objects with visual and/or haptic feedback
US9645732B2 (en) 2015-03-08 2017-05-09 Apple Inc. Devices, methods, and graphical user interfaces for displaying and using menus
US10048757B2 (en) 2015-03-08 2018-08-14 Apple Inc. Devices and methods for controlling media presentation
US9785305B2 (en) 2015-03-19 2017-10-10 Apple Inc. Touch input cursor manipulation
US9639184B2 (en) 2015-03-19 2017-05-02 Apple Inc. Touch input cursor manipulation
US20170045981A1 (en) 2015-08-10 2017-02-16 Apple Inc. Devices and Methods for Processing Touch Inputs Based on Their Intensities
TWI575414B (en) * 2015-04-01 2017-03-21 晨星半導體股份有限公司 Mode distinguishing method, touch-point locating method and touch control circuit
US10067653B2 (en) 2015-04-01 2018-09-04 Apple Inc. Devices and methods for processing touch inputs based on their intensities
CN104935725B (en) * 2015-04-23 2016-07-27 努比亚技术有限公司 Mobile terminal and utilize the method that virtual frame region realizes function point analysis
US10200598B2 (en) 2015-06-07 2019-02-05 Apple Inc. Devices and methods for capturing and interacting with enhanced digital images
US9830048B2 (en) 2015-06-07 2017-11-28 Apple Inc. Devices and methods for processing touch inputs with instructions in a web page
US9860451B2 (en) 2015-06-07 2018-01-02 Apple Inc. Devices and methods for capturing and interacting with enhanced digital images
US10346030B2 (en) 2015-06-07 2019-07-09 Apple Inc. Devices and methods for navigating between user interfaces
US9674426B2 (en) 2015-06-07 2017-06-06 Apple Inc. Devices and methods for capturing and interacting with enhanced digital images
US9891811B2 (en) 2015-06-07 2018-02-13 Apple Inc. Devices and methods for navigating between user interfaces
CN104866207A (en) * 2015-06-09 2015-08-26 努比亚技术有限公司 Method and device for acquiring auxiliary information of application program
CN106662971A (en) * 2015-07-03 2017-05-10 华为技术有限公司 Operation method applied to terminal device, and terminal device
US9612170B2 (en) 2015-07-21 2017-04-04 Apple Inc. Transparent strain sensors in an electronic device
US10055048B2 (en) 2015-07-31 2018-08-21 Apple Inc. Noise adaptive force touch
US10416800B2 (en) 2015-08-10 2019-09-17 Apple Inc. Devices, methods, and graphical user interfaces for adjusting user interface objects
US10235035B2 (en) 2015-08-10 2019-03-19 Apple Inc. Devices, methods, and graphical user interfaces for content navigation and manipulation
US9880735B2 (en) 2015-08-10 2018-01-30 Apple Inc. Devices, methods, and graphical user interfaces for manipulating user interface objects with visual and/or haptic feedback
US10248308B2 (en) 2015-08-10 2019-04-02 Apple Inc. Devices, methods, and graphical user interfaces for manipulating user interfaces with physical gestures
US10753874B2 (en) 2015-09-07 2020-08-25 Mobileodt Ltd. Handheld mobile device for detection of biological evidence
US9874965B2 (en) 2015-09-11 2018-01-23 Apple Inc. Transparent strain sensors in an electronic device
JP2017058719A (en) * 2015-09-14 2017-03-23 株式会社東海理化電機製作所 Operation device
US9886118B2 (en) 2015-09-30 2018-02-06 Apple Inc. Transparent force sensitive structures in an electronic device
CN106610777A (en) * 2015-10-23 2017-05-03 小米科技有限责任公司 Application starting method and device and mobile terminal
CN105320408B (en) * 2015-11-27 2018-09-21 福建捷联电子有限公司 Adjustable part touch module and its touch-control control method
CN105573545A (en) * 2015-11-27 2016-05-11 努比亚技术有限公司 Gesture correction method, apparatus and gesture input processing method
CN106814903B (en) * 2015-12-01 2020-01-14 小米科技有限责任公司 Touch control method and device of function key
CN105653027B (en) * 2015-12-24 2019-08-02 小米科技有限责任公司 Page zoom-in and zoom-out method and device
CN105511683B (en) * 2015-12-31 2019-03-12 厦门天马微电子有限公司 A kind of touch control display apparatus
US11335302B2 (en) 2016-01-15 2022-05-17 Google Llc Adaptable user interface with dual screen device
CN106997264B (en) * 2016-01-26 2020-11-24 百度在线网络技术(北京)有限公司 Method and device for changing picture typesetting
KR102456154B1 (en) 2016-01-29 2022-10-19 삼성디스플레이 주식회사 Sensor, touch sensor and display device
CN107145253B (en) * 2016-02-19 2020-06-09 苹果公司 Force sensing architecture
NZ720939A (en) * 2016-03-02 2017-12-22 Resene Paints Ltd Touch sensitive control system for non-electronic display substrate surfaces
US10006820B2 (en) 2016-03-08 2018-06-26 Apple Inc. Magnetic interference avoidance in resistive sensors
CN105808145A (en) * 2016-03-28 2016-07-27 努比亚技术有限公司 Method and terminal for achieving image processing
US10209830B2 (en) 2016-03-31 2019-02-19 Apple Inc. Electronic device having direction-dependent strain elements
KR101886209B1 (en) * 2016-04-19 2018-08-08 (주)휴맥스 Apparatus and method of providing media service
KR101853961B1 (en) 2016-04-19 2018-06-20 (주)휴맥스 Apparatus and method of providing media service
CN109416582B (en) * 2016-05-27 2022-05-27 西北大学 Tactile touch screen and operation method thereof
US10423228B2 (en) 2016-05-27 2019-09-24 Northwestern University Haptic touch screen and method of operating the same
RU170022U1 (en) * 2016-06-10 2017-04-12 Вячеслав Александрович Матвеев Triple Side Sensor
DK201770423A1 (en) 2016-06-11 2018-01-15 Apple Inc Activity and workout updates
US11216119B2 (en) 2016-06-12 2022-01-04 Apple Inc. Displaying a predetermined view of an application
KR102521192B1 (en) * 2016-06-28 2023-04-13 삼성전자주식회사 Electronic apparatus and operating method thereof
KR102544780B1 (en) 2016-07-04 2023-06-19 삼성전자주식회사 Method for controlling user interface according to handwriting input and electronic device for the same
KR102499364B1 (en) 2016-07-08 2023-02-14 삼성디스플레이 주식회사 Touch sensor
KR102552283B1 (en) 2016-07-15 2023-07-10 삼성디스플레이 주식회사 Pressure sensor and display device including the same
KR102552294B1 (en) 2016-07-15 2023-07-10 삼성디스플레이 주식회사 Pressure sensor and display device including the same
US9870033B1 (en) * 2016-08-30 2018-01-16 Apple Inc. Sensor assemblies for electronic devices
US10133418B2 (en) 2016-09-07 2018-11-20 Apple Inc. Force sensing in an electronic device using a single layer of strain-sensitive structures
DE102016010920A1 (en) 2016-09-08 2018-03-08 Fm Marketing Gmbh Smart touch
KR102610415B1 (en) 2016-11-30 2023-12-08 삼성디스플레이 주식회사 Touch sensor, display device including the same and method for driving touch sensor
KR20180090589A (en) * 2017-02-03 2018-08-13 엘지전자 주식회사 Mobile terminal and method for controlling of the same
US10444091B2 (en) 2017-04-11 2019-10-15 Apple Inc. Row column architecture for strain sensing
DK179412B1 (en) 2017-05-12 2018-06-06 Apple Inc Context-Specific User Interfaces
EP3593235A1 (en) * 2017-05-16 2020-01-15 Apple Inc. Devices, methods, and graphical user interfaces for touch input processing
DK179979B1 (en) 2017-05-16 2019-11-27 Apple Inc. Devices, methods, and graphical user interfaces for touch input processing
TWI612456B (en) 2017-06-13 2018-01-21 友達光電股份有限公司 Sensor element and electronic device having the same
US10309846B2 (en) 2017-07-24 2019-06-04 Apple Inc. Magnetic field cancellation for strain sensors
US11209927B2 (en) * 2017-12-11 2021-12-28 Google Llc Apparatus for sensing user input
CN108600419A (en) * 2018-03-09 2018-09-28 广东欧珀移动通信有限公司 Electronic device and its manufacturing method
WO2020018592A1 (en) 2018-07-17 2020-01-23 Methodical Mind, Llc. Graphical user interface system
US10866683B2 (en) 2018-08-27 2020-12-15 Apple Inc. Force or touch sensing on a mobile device using capacitive or pressure sensing
US10782818B2 (en) 2018-08-29 2020-09-22 Apple Inc. Load cell array for detection of force input to an electronic device enclosure
US11435830B2 (en) 2018-09-11 2022-09-06 Apple Inc. Content-based tactile outputs
CN109865283A (en) * 2019-03-05 2019-06-11 网易(杭州)网络有限公司 Virtual role face method of adjustment, device, electronic equipment and medium in game
CN109847347B (en) * 2019-03-13 2022-11-04 网易(杭州)网络有限公司 Method, device, medium and electronic equipment for controlling virtual operation in game
US11216149B2 (en) * 2019-03-15 2022-01-04 Samsung Electronics Co., Ltd. 360° video viewer control using smart device
CN109976637A (en) * 2019-03-27 2019-07-05 网易(杭州)网络有限公司 Dialog box method of adjustment, dialog box adjustment device, electronic equipment and storage medium
CN111845580A (en) * 2019-04-29 2020-10-30 北京新能源汽车股份有限公司 Vehicle-mounted gesture recognition device, vehicle-mounted central control system and vehicle
US20200375681A1 (en) * 2019-06-03 2020-12-03 Bausch & Lomb Incorporated Detachable Finger-Activated Surgical Control Device
CN113407053B (en) * 2019-10-15 2022-04-19 荣耀终端有限公司 Touch screen, electronic equipment and display control method
US11921998B2 (en) 2020-05-11 2024-03-05 Apple Inc. Editing features of an avatar
US11853478B2 (en) 2020-07-23 2023-12-26 Control Bionics Limited Multi-modal switching controller for communication and control
US11927989B2 (en) 2020-08-11 2024-03-12 Google Llc Computing device with rotatable keyboard
WO2022049362A1 (en) * 2020-09-02 2022-03-10 Peratech Holdco Ltd Sensing device
USD970953S1 (en) 2021-05-03 2022-11-29 Conair Llc Toaster
USD988771S1 (en) 2021-05-03 2023-06-13 Conair Llc Toaster
US11714536B2 (en) 2021-05-21 2023-08-01 Apple Inc. Avatar sticker editor user interfaces
US11893212B2 (en) 2021-06-06 2024-02-06 Apple Inc. User interfaces for managing application widgets

Citations (98)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US3798370A (en) * 1972-04-17 1974-03-19 Elographics Inc Electrographic sensor for determining planar coordinates
US4246452A (en) * 1979-01-05 1981-01-20 Mattel, Inc. Switch apparatus
US4566001A (en) * 1983-02-08 1986-01-21 Northern Telecom Limited Touch strip input for display terminal
US4719524A (en) * 1984-10-08 1988-01-12 Sony Corporation Signal reproduction apparatus including touched state pattern recognition speed control
US4733222A (en) * 1983-12-27 1988-03-22 Integrated Touch Arrays, Inc. Capacitance-variation-sensitive touch sensing array system
US4734685A (en) * 1983-07-28 1988-03-29 Canon Kabushiki Kaisha Position control apparatus
US4806846A (en) * 1987-07-06 1989-02-21 Kerber George L High accuracy direct reading capacitance-to-voltage converter
US4891508A (en) * 1988-06-30 1990-01-02 Hewlett-Packard Company Precision infrared position detector apparatus for touch screen system
US4896370A (en) * 1987-03-27 1990-01-23 Teletec Corporation Control system for microprocessor and software enhanced communications equipment
US4898555A (en) * 1989-03-23 1990-02-06 Bell & Howell Publication Systems Company Display screen bezel and assembly method
US5003519A (en) * 1988-05-26 1991-03-26 Eta S.A. Fabriques D'ebauches Alarm arrangement for a timepiece
US5178477A (en) * 1991-06-06 1993-01-12 Gambaro Thomas L Ergonomic keyboard input device
US5179648A (en) * 1986-03-24 1993-01-12 Hauck Lane T Computer auxiliary viewing system
US5186646A (en) * 1992-01-16 1993-02-16 Pederson William A Connector device for computers
US5189403A (en) * 1989-09-26 1993-02-23 Home Row, Inc. Integrated keyboard and pointing device system with automatic mode change
US5194862A (en) * 1990-06-29 1993-03-16 U.S. Philips Corporation Touch sensor array systems and display systems incorporating such
US5281966A (en) * 1992-01-31 1994-01-25 Walsh A Peter Method of encoding alphabetic characters for a chord keyboard
US5379057A (en) * 1988-11-14 1995-01-03 Microslate, Inc. Portable computer with touch screen and computer system employing same
US5398310A (en) * 1992-04-13 1995-03-14 Apple Computer, Incorporated Pointing gesture based computer note pad paging and scrolling interface
US5483261A (en) * 1992-02-14 1996-01-09 Itu Research, Inc. Graphical input controller and method with rear screen image detection
US5488204A (en) * 1992-06-08 1996-01-30 Synaptics, Incorporated Paintbrush stylus for capacitive touch sensor pad
US5495566A (en) * 1994-11-22 1996-02-27 Microsoft Corporation Scrolling contents of a window
US5495077A (en) * 1992-06-08 1996-02-27 Synaptics, Inc. Object position and proximity detector
US5565658A (en) * 1992-07-13 1996-10-15 Cirque Corporation Capacitance-based proximity with interference rejection apparatus and methods
US5592566A (en) * 1992-05-27 1997-01-07 Apple Computer, Incorporated Method and apparatus for computerized recognition
US5594471A (en) * 1992-01-09 1997-01-14 Casco Development, Inc. Industrial touchscreen workstation with programmable interface and method
US5594810A (en) * 1993-09-30 1997-01-14 Apple Computer, Inc. Method and apparatus for recognizing gestures on a computer system
US5596694A (en) * 1992-05-27 1997-01-21 Apple Computer, Inc. Method and apparatus for indicating a change in status of an object and its disposition using animation
US5596347A (en) * 1994-01-27 1997-01-21 Microsoft Corporation System and method for computer cursor control
US5598527A (en) * 1992-11-12 1997-01-28 Sextant Avionique Compact and ergonomic communications terminal equipped with proximity detection surfaces
US5611040A (en) * 1995-04-05 1997-03-11 Microsoft Corporation Method and system for activating double click applications with a single click
US5611060A (en) * 1995-02-22 1997-03-11 Microsoft Corporation Auto-scrolling during a drag and drop operation
US5612719A (en) * 1992-12-03 1997-03-18 Apple Computer, Inc. Gesture sensitive buttons for graphical user interfaces
US5710844A (en) * 1992-05-27 1998-01-20 Apple Computer Method for searching and displaying results in a pen-based computer system
US5856822A (en) * 1995-10-27 1999-01-05 02 Micro, Inc. Touch-pad digital computer pointing-device
US5931906A (en) * 1997-04-18 1999-08-03 Creative Communications Group System for creating a multimedia presentation by integrating local program materials with remotely accessible program materials
US6020881A (en) * 1993-05-24 2000-02-01 Sun Microsystems Graphical user interface with method and apparatus for interfacing to remote devices
US6031518A (en) * 1997-05-30 2000-02-29 Microsoft Corporation Ergonomic input device
US6031524A (en) * 1995-06-07 2000-02-29 Intermec Ip Corp. Hand-held portable data terminal having removably interchangeable, washable, user-replaceable components with liquid-impervious seal
US6181322B1 (en) * 1997-11-07 2001-01-30 Netscape Communications Corp. Pointing device having selection buttons operable from movement of a palm portion of a person's hands
US6179496B1 (en) * 1999-12-28 2001-01-30 Shin Jiuh Corp. Computer keyboard with turnable knob
US6188393B1 (en) * 1998-10-05 2001-02-13 Sysgration Ltd. Scroll bar input device for mouse
US6188389B1 (en) * 1998-12-10 2001-02-13 Chic Technology Corp. Third axis input device for a computer mouse
US6188391B1 (en) * 1998-07-09 2001-02-13 Synaptics, Inc. Two-layer capacitive touchpad and method of making same
US6191774B1 (en) * 1995-11-17 2001-02-20 Immersion Corporation Mouse interface for providing force feedback
US6337919B1 (en) * 1999-04-28 2002-01-08 Intel Corporation Fingerprint detecting mouse
US6340800B1 (en) * 2000-05-27 2002-01-22 International Business Machines Corporation Multiplexing control device and method for electronic systems
US20020008691A1 (en) * 1998-01-16 2002-01-24 Mitsuru Hanajima Information processing apparatus and display control method of the same information processing apparatus
US20020015024A1 (en) * 1998-01-26 2002-02-07 University Of Delaware Method and apparatus for integrating manual input
US6347290B1 (en) * 1998-06-24 2002-02-12 Compaq Information Technologies Group, L.P. Apparatus and method for detecting and executing positional and gesture commands corresponding to movement of handheld computing device
US6504530B1 (en) * 1999-09-07 2003-01-07 Elo Touchsystems, Inc. Touch confirming touchscreen utilizing plural touch sensors
US6505088B1 (en) * 2000-03-07 2003-01-07 Mental Models, Inc. Electronic controller
US20030006974A1 (en) * 2001-07-03 2003-01-09 James Clough Methods and systems for increasing the input efficiency of personal digital assistants and other handheld stylus-engagable computing devices
US20030011574A1 (en) * 2001-03-31 2003-01-16 Goodman Joshua T. Out-of-vocabulary word determination and user interface for text input via reduced keypad keys
US6515669B1 (en) * 1998-10-23 2003-02-04 Olympus Optical Co., Ltd. Operation input device applied to three-dimensional input device
US6513717B2 (en) * 2000-12-07 2003-02-04 Digimarc Corporation Integrated cursor control and scanner device
US20030025735A1 (en) * 2001-07-31 2003-02-06 Eastman Kodak Company User interface including portable display for use with multiple electronic devices
US6525749B1 (en) * 1993-12-30 2003-02-25 Xerox Corporation Apparatus and method for supporting the implicit structure of freeform lists, outlines, text, tables and diagrams in a gesture-based input system and editing system
US20040003947A1 (en) * 2002-07-02 2004-01-08 Kesselman Joshua D. Handheld electronic scale with touch screen
US6677934B1 (en) * 1999-07-30 2004-01-13 L-3 Communications Infrared touch panel with improved sunlight rejection
US6677927B1 (en) * 1999-08-23 2004-01-13 Microsoft Corporation X-Y navigation input device
US6677932B1 (en) * 2001-01-28 2004-01-13 Finger Works, Inc. System and method for recognizing touch typing under limited tactile feedback conditions
US20040012572A1 (en) * 2002-03-16 2004-01-22 Anthony Sowden Display and touch screen method and apparatus
US6686904B1 (en) * 2001-03-30 2004-02-03 Microsoft Corporation Wheel reporting method for a personal computer keyboard interface
US20040021643A1 (en) * 2002-08-02 2004-02-05 Takeshi Hoshino Display unit with touch panel and information processing method
US6690387B2 (en) * 2001-12-28 2004-02-10 Koninklijke Philips Electronics N.V. Touch-screen image scrolling system and method
US6842672B1 (en) * 2002-02-28 2005-01-11 Garmin International, Inc. Cockpit instrument panel systems and methods with redundant flight data display
US6844872B1 (en) * 2000-01-12 2005-01-18 Apple Computer, Inc. Computer mouse having side areas to maintain a depressed button position
US6844871B1 (en) * 1999-11-05 2005-01-18 Microsoft Corporation Method and apparatus for computer input using six degrees of freedom
US20050012723A1 (en) * 2003-07-14 2005-01-20 Move Mobile Systems, Inc. System and method for a portable multimedia client
US6848014B2 (en) * 2001-01-23 2005-01-25 Symbol Technologies, Inc. Adapter unit for a personal digital assistant having automatically configurable application buttons
US6847352B2 (en) * 2001-09-06 2005-01-25 Juan J Lantigua Stress relieving mouse
US6853850B2 (en) * 2000-12-04 2005-02-08 Mobigence, Inc. Automatic speaker volume and microphone gain control in a portable handheld radiotelephone with proximity sensors
US20050030278A1 (en) * 2003-08-08 2005-02-10 Liang Fu Adjustable pointing and control device with automatic handedness switch
US6856259B1 (en) * 2004-02-06 2005-02-15 Elo Touchsystems, Inc. Touch sensor system to detect multiple touch events
US20050035955A1 (en) * 2002-06-06 2005-02-17 Carter Dale J. Method of determining orientation and manner of holding a mobile telephone
US20050043060A1 (en) * 2000-04-04 2005-02-24 Wireless Agents, Llc Method and apparatus for scheduling presentation of digital content on a personal communication device
US6985801B1 (en) * 2002-02-28 2006-01-10 Garmin International, Inc. Cockpit instrument panel systems and methods with redundant flight data display
US20060010400A1 (en) * 2004-06-28 2006-01-12 Microsoft Corporation Recognizing gestures and using gestures for interacting with software applications
US6992659B2 (en) * 2001-05-22 2006-01-31 Palmone, Inc. High transparency integrated enclosure touch screen assembly for a portable hand held device
US20060026535A1 (en) * 2004-07-30 2006-02-02 Apple Computer Inc. Mode-based graphical user interfaces for touch sensitive input devices
US20060022955A1 (en) * 2004-07-30 2006-02-02 Apple Computer, Inc. Visual expander
US20060022956A1 (en) * 2003-09-02 2006-02-02 Apple Computer, Inc. Touch-sensitive electronic apparatus for media applications, and methods therefor
US6995744B1 (en) * 2000-09-28 2006-02-07 Immersion Corporation Device and assembly for providing linear tactile sensations
US20060033724A1 (en) * 2004-07-30 2006-02-16 Apple Computer, Inc. Virtual input device placement on a touch screen user interface
US20060032680A1 (en) * 2004-08-16 2006-02-16 Fingerworks, Inc. Method of increasing the spatial resolution of touch sensitive devices
US20060073272A1 (en) * 2004-09-28 2006-04-06 Clearpath Technologies Corp. Control panel and method of manufacturing
US20060181518A1 (en) * 2005-02-14 2006-08-17 Chia Shen Spatial multiplexing to mediate direct-touch input on large displays
US7164412B2 (en) * 2001-03-30 2007-01-16 Ado Max Technology Co., Ltd. Key free mouse
US7168047B1 (en) * 2002-05-28 2007-01-23 Apple Computer, Inc. Mouse having a button-less panning and scrolling switch
US7170488B2 (en) * 2000-12-22 2007-01-30 Logitech Europe S.A. Pointing device with solid-state roller
US7170496B2 (en) * 2003-01-24 2007-01-30 Bruce Peter Middleton Zero-front-footprint compact input system
US20080012838A1 (en) * 2006-07-13 2008-01-17 N-Trig Ltd. User specific recognition of intended user interaction with a digitizer
US20080012835A1 (en) * 2006-07-12 2008-01-17 N-Trig Ltd. Hover and touch detection for digitizer
US7495659B2 (en) * 2003-11-25 2009-02-24 Apple Inc. Touch pad for handheld device
US7653607B2 (en) * 2000-03-24 2010-01-26 Symyx Solutions, Inc. Remote execution of materials library designs
US7652589B2 (en) * 2005-02-04 2010-01-26 Hewlett-Packard Development Company, L.P. Selecting a function of components based on orientation
US20110012835A1 (en) * 2003-09-02 2011-01-20 Steve Hotelling Ambidextrous mouse

Family Cites Families (90)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US3859005A (en) 1973-08-13 1975-01-07 Albert L Huebner Erosion reduction in wet turbines
US5027690A (en) * 1986-02-12 1991-07-02 Nippon Gakki Seizo Kabushiki Kaisha Electronic musical keyboard instruments with variable touch sensitivity
JPS63257824A (en) 1987-04-15 1988-10-25 Fujitsu Ltd Touch panel
US4933660A (en) * 1989-10-27 1990-06-12 Elographics, Inc. Touch sensor with touch pressure capability
KR940001227A (en) * 1992-06-15 1994-01-11 에프. 제이. 스미트 Touch screen devices
JPH06131437A (en) * 1992-10-20 1994-05-13 Hitachi Ltd Method for instructing operation in composite form
JP3300819B2 (en) * 1993-01-11 2002-07-08 株式会社日立製作所 Automatic teller machine
BE1007462A3 (en) 1993-08-26 1995-07-04 Philips Electronics Nv Data processing device with touch sensor and power.
JPH07182101A (en) 1993-10-26 1995-07-21 Itu Res Inc Apparatus and method for input of graphic, operating method of graphic object and supply method of graphic input signal
US6211860B1 (en) * 1994-07-06 2001-04-03 Hewlett-Packard Company Pressure sensitive electronic device
JPH08166866A (en) * 1994-10-14 1996-06-25 Hitachi Ltd Editing support system equipped with interactive interface
US5825352A (en) * 1996-01-04 1998-10-20 Logitech, Inc. Multiple fingers contact sensing method for emulating mouse buttons and mouse operations on a touch sensor pad
JP3028047U (en) * 1996-02-16 1996-08-30 株式会社エフ・エー・システム ダイワ Touch coordinate detection device for touch panel
JPH09244810A (en) * 1996-03-07 1997-09-19 Toshiba Corp Touch panel controller and process monitor and control equipment
US5705005A (en) 1996-05-01 1998-01-06 Davidson Textron Inc. Fabrication of multiple color, gauge, and texture interior auto components
JPH10149258A (en) * 1996-11-15 1998-06-02 Omron Corp Input operation device
US6222528B1 (en) * 1997-03-07 2001-04-24 Cirque Corporation Method and apparatus for data input
JPH11119911A (en) * 1997-10-15 1999-04-30 Fujitsu Ten Ltd Switching device
JPH11215217A (en) 1998-01-23 1999-08-06 Toshiba Corp Portable information terminal equipment
US6421048B1 (en) * 1998-07-17 2002-07-16 Sensable Technologies, Inc. Systems and methods for interacting with virtual objects in a haptic virtual reality environment
JP2000293289A (en) * 1999-04-09 2000-10-20 Hitachi Ltd Portable terminal device
US6792165B1 (en) 1999-05-25 2004-09-14 Silverbrook Research Pty Ltd Sensing device
JP2001086233A (en) * 1999-07-13 2001-03-30 Denso Corp Portable set with lighting function
US20020173721A1 (en) * 1999-08-20 2002-11-21 Novasonics, Inc. User interface for handheld imaging devices
US6492979B1 (en) * 1999-09-07 2002-12-10 Elo Touchsystems, Inc. Dual sensor touchscreen utilizing projective-capacitive and force touch sensors
KR100367586B1 (en) * 1999-10-08 2003-01-10 엘지전자 주식회사 Flat type monitor
KR20010047975A (en) * 1999-11-24 2001-06-15 유경운 A System and Terminal of Multimedia with Multi-Used
US6498601B1 (en) * 1999-11-29 2002-12-24 Xerox Corporation Method and apparatus for selecting input modes on a palmtop computer
JP2001275095A (en) 2000-03-27 2001-10-05 Sony Corp Communication method, communication service device, communication terminal device, and communication system
JP4039001B2 (en) * 2000-05-30 2008-01-30 ソニー株式会社 Electronics
JP2001345913A (en) * 2000-06-05 2001-12-14 Hitachi Kokusai Electric Inc Telephone set
JP2001356878A (en) 2000-06-14 2001-12-26 Hitachi Ltd Icon control method
JP4803868B2 (en) * 2000-08-10 2011-10-26 キヤノン株式会社 Information processing apparatus and function list display method
KR20020016080A (en) * 2000-08-24 2002-03-04 정장호 folder typed phone having touch pad
EP2133777B1 (en) * 2000-10-27 2011-10-12 Tyco Electronics Corporation Dual sensor touchscreen utilizing projective-capacitive and force touch sensors
JP2002140166A (en) * 2000-11-02 2002-05-17 Kddi Corp Device for giving feel to finger touching touch panel display
JP2002185630A (en) * 2000-12-12 2002-06-28 Aiwa Co Ltd Communication terminal equipment
JP2002215304A (en) * 2001-01-17 2002-08-02 Sharp Corp Touch panel computer
JP2002229719A (en) * 2001-02-05 2002-08-16 Matsushita Electric Ind Co Ltd Touch panel type cellphone and operating method for changing its function
US7183948B2 (en) 2001-04-13 2007-02-27 3M Innovative Properties Company Tangential force control in a touch location device
JP2002342034A (en) * 2001-05-16 2002-11-29 Fujikura Ltd Touch panel
US6498590B1 (en) * 2001-05-24 2002-12-24 Mitsubishi Electric Research Laboratories, Inc. Multi-user touch surface
CN1241109C (en) * 2001-07-17 2006-02-08 仁宝电脑工业股份有限公司 Touch display able to control amplificatino rabio by pressure
JP4587616B2 (en) * 2001-08-21 2010-11-24 シャープ株式会社 Mobile device
JP2003122506A (en) * 2001-10-10 2003-04-25 Canon Inc Coordinate input and operational method directing device
US6995752B2 (en) * 2001-11-08 2006-02-07 Koninklijke Philips Electronics N.V. Multi-point touch pad
JP2003157144A (en) * 2001-11-20 2003-05-30 Sony Corp Character input device, character input method, character input program storage medium and character input program
US7009599B2 (en) * 2001-11-20 2006-03-07 Nokia Corporation Form factor for portable device
JP2005301322A (en) * 2002-02-07 2005-10-27 Kathenas Inc Input device, cellular phone, and portable information device
JP2003241872A (en) * 2002-02-20 2003-08-29 Ricoh Co Ltd Drawing processing method, program thereby, and storage medium storing its program
JP3847641B2 (en) * 2002-02-28 2006-11-22 株式会社ソニー・コンピュータエンタテインメント Information processing apparatus, information processing program, computer-readable recording medium storing information processing program, and information processing method
JP2003275484A (en) * 2002-03-25 2003-09-30 Brother Ind Ltd Screen setting device for sewing machine
US20050229103A1 (en) * 2002-03-25 2005-10-13 King David M Gui and support hardware for maintaining long-term personal access to the world
JP2003323259A (en) 2002-05-02 2003-11-14 Nec Corp Information processing apparatus
US7746325B2 (en) * 2002-05-06 2010-06-29 3M Innovative Properties Company Method for improving positioned accuracy for a determined touch input
JP3852368B2 (en) 2002-05-16 2006-11-29 ソニー株式会社 Input method and data processing apparatus
WO2003098421A1 (en) * 2002-05-16 2003-11-27 Sony Corporation Inputting method and inputting apparatus
JP2003344086A (en) * 2002-05-28 2003-12-03 Pioneer Electronic Corp Touch panel device and display input device for car
US7542052B2 (en) * 2002-05-31 2009-06-02 Hewlett-Packard Development Company, L.P. System and method of switching viewing orientations of a display
JP2004012919A (en) * 2002-06-07 2004-01-15 Clarion Co Ltd Information display device
JP2004070920A (en) * 2002-06-11 2004-03-04 Sony Computer Entertainment Inc Information processing program, computer readable recording medium recording information processing program, information processing method and information processor
JP2004021933A (en) * 2002-06-20 2004-01-22 Casio Comput Co Ltd Input device and input method
JP4207496B2 (en) * 2002-08-22 2009-01-14 ソニー株式会社 Portable information processing apparatus, information processing method, and program
US20040046739A1 (en) * 2002-09-11 2004-03-11 Palm, Inc. Pliable device navigation method and apparatus
JP2004110388A (en) * 2002-09-18 2004-04-08 Sharp Corp Device having touch panel
JP2004177993A (en) * 2002-11-22 2004-06-24 Panasonic Mobile Communications Co Ltd Mobile terminal with pressure sensor, and program executable by mobile terminal with pressure sensor
JP4022468B2 (en) * 2002-12-26 2007-12-19 株式会社ワコム Input device and input system
JP2004226715A (en) * 2003-01-23 2004-08-12 Nec Corp Automatic top and bottom display control system, method and program
JP2004302734A (en) * 2003-03-31 2004-10-28 Mitsubishi Electric Corp Information terminal and program for making computer execute its operation changeover
JP4341298B2 (en) * 2003-05-23 2009-10-07 ソニー株式会社 Information processing apparatus and method, and program
JP4161814B2 (en) * 2003-06-16 2008-10-08 ソニー株式会社 Input method and input device
US20040263484A1 (en) * 2003-06-25 2004-12-30 Tapio Mantysalo Multifunctional UI input device for moblie terminals
JP2005056232A (en) * 2003-08-06 2005-03-03 Matsushita Electric Ind Co Ltd Method for displaying operation manual of electronic apparatus
US20050048955A1 (en) * 2003-09-03 2005-03-03 Steffen Ring Method and apparatus for initiating a call from a communication device
US7411575B2 (en) * 2003-09-16 2008-08-12 Smart Technologies Ulc Gesture recognition method and touch system incorporating the same
US20050183035A1 (en) * 2003-11-20 2005-08-18 Ringel Meredith J. Conflict resolution for graphic multi-user interface
JP2005346244A (en) * 2004-06-01 2005-12-15 Nec Corp Information display unit and operation method therefor
US8232969B2 (en) * 2004-10-08 2012-07-31 Immersion Corporation Haptic feedback for button and scrolling action simulation in touch input devices
US7683889B2 (en) * 2004-12-21 2010-03-23 Microsoft Corporation Pressure based selection
JP4622590B2 (en) 2005-03-08 2011-02-02 ソニー株式会社 Input device, touch panel and electronic device
US8924892B2 (en) * 2008-08-22 2014-12-30 Fuji Xerox Co., Ltd. Multiple selection on devices with many gestures
JP5635274B2 (en) * 2010-01-27 2014-12-03 京セラ株式会社 Tactile sensation presentation apparatus and tactile sensation presentation method
US20110209098A1 (en) * 2010-02-19 2011-08-25 Hinckley Kenneth P On and Off-Screen Gesture Combinations
EP3316956B1 (en) 2015-06-30 2020-04-29 Unity Health Toronto Cardiac injury device
US10188182B2 (en) 2016-05-05 2019-01-29 M&Y Trading Corp. Clasp for beaded jewelry
US10643256B2 (en) 2016-09-16 2020-05-05 International Business Machines Corporation Configuring a self-driving vehicle for charitable donations pickup and delivery
WO2018175806A1 (en) 2017-03-24 2018-09-27 Intel IP Corporation Techniques to enable physical downlink control channel communications
JP6521127B2 (en) 2017-04-21 2019-05-29 株式会社タンガロイ Coated cutting tool
TWI640378B (en) 2017-11-15 2018-11-11 鴻安國際興業有限公司 Inverted tooth rotary knife
CN110932673A (en) 2018-09-19 2020-03-27 恩智浦美国有限公司 Chopper-stabilized amplifier containing shunt notch filter

Patent Citations (103)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US3798370A (en) * 1972-04-17 1974-03-19 Elographics Inc Electrographic sensor for determining planar coordinates
US4246452A (en) * 1979-01-05 1981-01-20 Mattel, Inc. Switch apparatus
US4566001A (en) * 1983-02-08 1986-01-21 Northern Telecom Limited Touch strip input for display terminal
US4734685A (en) * 1983-07-28 1988-03-29 Canon Kabushiki Kaisha Position control apparatus
US4733222A (en) * 1983-12-27 1988-03-22 Integrated Touch Arrays, Inc. Capacitance-variation-sensitive touch sensing array system
US4719524A (en) * 1984-10-08 1988-01-12 Sony Corporation Signal reproduction apparatus including touched state pattern recognition speed control
US5179648A (en) * 1986-03-24 1993-01-12 Hauck Lane T Computer auxiliary viewing system
US4896370A (en) * 1987-03-27 1990-01-23 Teletec Corporation Control system for microprocessor and software enhanced communications equipment
US4806846A (en) * 1987-07-06 1989-02-21 Kerber George L High accuracy direct reading capacitance-to-voltage converter
US5003519A (en) * 1988-05-26 1991-03-26 Eta S.A. Fabriques D'ebauches Alarm arrangement for a timepiece
US4891508A (en) * 1988-06-30 1990-01-02 Hewlett-Packard Company Precision infrared position detector apparatus for touch screen system
US5379057A (en) * 1988-11-14 1995-01-03 Microslate, Inc. Portable computer with touch screen and computer system employing same
US4898555A (en) * 1989-03-23 1990-02-06 Bell & Howell Publication Systems Company Display screen bezel and assembly method
US5189403A (en) * 1989-09-26 1993-02-23 Home Row, Inc. Integrated keyboard and pointing device system with automatic mode change
US5194862A (en) * 1990-06-29 1993-03-16 U.S. Philips Corporation Touch sensor array systems and display systems incorporating such
US5178477A (en) * 1991-06-06 1993-01-12 Gambaro Thomas L Ergonomic keyboard input device
US5594471A (en) * 1992-01-09 1997-01-14 Casco Development, Inc. Industrial touchscreen workstation with programmable interface and method
US5186646A (en) * 1992-01-16 1993-02-16 Pederson William A Connector device for computers
US5281966A (en) * 1992-01-31 1994-01-25 Walsh A Peter Method of encoding alphabetic characters for a chord keyboard
US5483261A (en) * 1992-02-14 1996-01-09 Itu Research, Inc. Graphical input controller and method with rear screen image detection
US5398310A (en) * 1992-04-13 1995-03-14 Apple Computer, Incorporated Pointing gesture based computer note pad paging and scrolling interface
US5592566A (en) * 1992-05-27 1997-01-07 Apple Computer, Incorporated Method and apparatus for computerized recognition
US5710844A (en) * 1992-05-27 1998-01-20 Apple Computer Method for searching and displaying results in a pen-based computer system
US5596694A (en) * 1992-05-27 1997-01-21 Apple Computer, Inc. Method and apparatus for indicating a change in status of an object and its disposition using animation
US5495077A (en) * 1992-06-08 1996-02-27 Synaptics, Inc. Object position and proximity detector
US5488204A (en) * 1992-06-08 1996-01-30 Synaptics, Incorporated Paintbrush stylus for capacitive touch sensor pad
US5565658A (en) * 1992-07-13 1996-10-15 Cirque Corporation Capacitance-based proximity with interference rejection apparatus and methods
US5598527A (en) * 1992-11-12 1997-01-28 Sextant Avionique Compact and ergonomic communications terminal equipped with proximity detection surfaces
US5612719A (en) * 1992-12-03 1997-03-18 Apple Computer, Inc. Gesture sensitive buttons for graphical user interfaces
US6020881A (en) * 1993-05-24 2000-02-01 Sun Microsystems Graphical user interface with method and apparatus for interfacing to remote devices
US5594810A (en) * 1993-09-30 1997-01-14 Apple Computer, Inc. Method and apparatus for recognizing gestures on a computer system
US6525749B1 (en) * 1993-12-30 2003-02-25 Xerox Corporation Apparatus and method for supporting the implicit structure of freeform lists, outlines, text, tables and diagrams in a gesture-based input system and editing system
US5596347A (en) * 1994-01-27 1997-01-21 Microsoft Corporation System and method for computer cursor control
US5598183A (en) * 1994-01-27 1997-01-28 Microsoft Corporation System and method for computer cursor control
US5495566A (en) * 1994-11-22 1996-02-27 Microsoft Corporation Scrolling contents of a window
US5726687A (en) * 1995-02-22 1998-03-10 Microsoft Corporation Auto-scrolling with mouse speed computation during dragging
US5611060A (en) * 1995-02-22 1997-03-11 Microsoft Corporation Auto-scrolling during a drag and drop operation
US5611040A (en) * 1995-04-05 1997-03-11 Microsoft Corporation Method and system for activating double click applications with a single click
US6031524A (en) * 1995-06-07 2000-02-29 Intermec Ip Corp. Hand-held portable data terminal having removably interchangeable, washable, user-replaceable components with liquid-impervious seal
US5856822A (en) * 1995-10-27 1999-01-05 02 Micro, Inc. Touch-pad digital computer pointing-device
US6191774B1 (en) * 1995-11-17 2001-02-20 Immersion Corporation Mouse interface for providing force feedback
US5931906A (en) * 1997-04-18 1999-08-03 Creative Communications Group System for creating a multimedia presentation by integrating local program materials with remotely accessible program materials
US6031518A (en) * 1997-05-30 2000-02-29 Microsoft Corporation Ergonomic input device
US6181322B1 (en) * 1997-11-07 2001-01-30 Netscape Communications Corp. Pointing device having selection buttons operable from movement of a palm portion of a person's hands
US20020008691A1 (en) * 1998-01-16 2002-01-24 Mitsuru Hanajima Information processing apparatus and display control method of the same information processing apparatus
US20020015024A1 (en) * 1998-01-26 2002-02-07 University Of Delaware Method and apparatus for integrating manual input
US6347290B1 (en) * 1998-06-24 2002-02-12 Compaq Information Technologies Group, L.P. Apparatus and method for detecting and executing positional and gesture commands corresponding to movement of handheld computing device
US6188391B1 (en) * 1998-07-09 2001-02-13 Synaptics, Inc. Two-layer capacitive touchpad and method of making same
US6188393B1 (en) * 1998-10-05 2001-02-13 Sysgration Ltd. Scroll bar input device for mouse
US6515669B1 (en) * 1998-10-23 2003-02-04 Olympus Optical Co., Ltd. Operation input device applied to three-dimensional input device
US6188389B1 (en) * 1998-12-10 2001-02-13 Chic Technology Corp. Third axis input device for a computer mouse
US6337919B1 (en) * 1999-04-28 2002-01-08 Intel Corporation Fingerprint detecting mouse
US6677934B1 (en) * 1999-07-30 2004-01-13 L-3 Communications Infrared touch panel with improved sunlight rejection
US6677927B1 (en) * 1999-08-23 2004-01-13 Microsoft Corporation X-Y navigation input device
US6504530B1 (en) * 1999-09-07 2003-01-07 Elo Touchsystems, Inc. Touch confirming touchscreen utilizing plural touch sensors
US6844871B1 (en) * 1999-11-05 2005-01-18 Microsoft Corporation Method and apparatus for computer input using six degrees of freedom
US6179496B1 (en) * 1999-12-28 2001-01-30 Shin Jiuh Corp. Computer keyboard with turnable knob
US6844872B1 (en) * 2000-01-12 2005-01-18 Apple Computer, Inc. Computer mouse having side areas to maintain a depressed button position
US6505088B1 (en) * 2000-03-07 2003-01-07 Mental Models, Inc. Electronic controller
US7653607B2 (en) * 2000-03-24 2010-01-26 Symyx Solutions, Inc. Remote execution of materials library designs
US20050043060A1 (en) * 2000-04-04 2005-02-24 Wireless Agents, Llc Method and apparatus for scheduling presentation of digital content on a personal communication device
US6340800B1 (en) * 2000-05-27 2002-01-22 International Business Machines Corporation Multiplexing control device and method for electronic systems
US6995744B1 (en) * 2000-09-28 2006-02-07 Immersion Corporation Device and assembly for providing linear tactile sensations
US6853850B2 (en) * 2000-12-04 2005-02-08 Mobigence, Inc. Automatic speaker volume and microphone gain control in a portable handheld radiotelephone with proximity sensors
US6513717B2 (en) * 2000-12-07 2003-02-04 Digimarc Corporation Integrated cursor control and scanner device
US7170488B2 (en) * 2000-12-22 2007-01-30 Logitech Europe S.A. Pointing device with solid-state roller
US6848014B2 (en) * 2001-01-23 2005-01-25 Symbol Technologies, Inc. Adapter unit for a personal digital assistant having automatically configurable application buttons
US6677932B1 (en) * 2001-01-28 2004-01-13 Finger Works, Inc. System and method for recognizing touch typing under limited tactile feedback conditions
US7164412B2 (en) * 2001-03-30 2007-01-16 Ado Max Technology Co., Ltd. Key free mouse
US6686904B1 (en) * 2001-03-30 2004-02-03 Microsoft Corporation Wheel reporting method for a personal computer keyboard interface
US20030011574A1 (en) * 2001-03-31 2003-01-16 Goodman Joshua T. Out-of-vocabulary word determination and user interface for text input via reduced keypad keys
US6992659B2 (en) * 2001-05-22 2006-01-31 Palmone, Inc. High transparency integrated enclosure touch screen assembly for a portable hand held device
US20030006974A1 (en) * 2001-07-03 2003-01-09 James Clough Methods and systems for increasing the input efficiency of personal digital assistants and other handheld stylus-engagable computing devices
US20030025735A1 (en) * 2001-07-31 2003-02-06 Eastman Kodak Company User interface including portable display for use with multiple electronic devices
US6847352B2 (en) * 2001-09-06 2005-01-25 Juan J Lantigua Stress relieving mouse
US6690387B2 (en) * 2001-12-28 2004-02-10 Koninklijke Philips Electronics N.V. Touch-screen image scrolling system and method
US7184064B2 (en) * 2001-12-28 2007-02-27 Koninklijke Philips Electronics N.V. Touch-screen image scrolling system and method
US6985801B1 (en) * 2002-02-28 2006-01-10 Garmin International, Inc. Cockpit instrument panel systems and methods with redundant flight data display
US6842672B1 (en) * 2002-02-28 2005-01-11 Garmin International, Inc. Cockpit instrument panel systems and methods with redundant flight data display
US20040012572A1 (en) * 2002-03-16 2004-01-22 Anthony Sowden Display and touch screen method and apparatus
US7168047B1 (en) * 2002-05-28 2007-01-23 Apple Computer, Inc. Mouse having a button-less panning and scrolling switch
US20050035955A1 (en) * 2002-06-06 2005-02-17 Carter Dale J. Method of determining orientation and manner of holding a mobile telephone
US20040003947A1 (en) * 2002-07-02 2004-01-08 Kesselman Joshua D. Handheld electronic scale with touch screen
US20040021643A1 (en) * 2002-08-02 2004-02-05 Takeshi Hoshino Display unit with touch panel and information processing method
US7170496B2 (en) * 2003-01-24 2007-01-30 Bruce Peter Middleton Zero-front-footprint compact input system
US20050012723A1 (en) * 2003-07-14 2005-01-20 Move Mobile Systems, Inc. System and method for a portable multimedia client
US20050030278A1 (en) * 2003-08-08 2005-02-10 Liang Fu Adjustable pointing and control device with automatic handedness switch
US20110012835A1 (en) * 2003-09-02 2011-01-20 Steve Hotelling Ambidextrous mouse
US20060022956A1 (en) * 2003-09-02 2006-02-02 Apple Computer, Inc. Touch-sensitive electronic apparatus for media applications, and methods therefor
US7495659B2 (en) * 2003-11-25 2009-02-24 Apple Inc. Touch pad for handheld device
US6856259B1 (en) * 2004-02-06 2005-02-15 Elo Touchsystems, Inc. Touch sensor system to detect multiple touch events
US20060010400A1 (en) * 2004-06-28 2006-01-12 Microsoft Corporation Recognizing gestures and using gestures for interacting with software applications
US20060026535A1 (en) * 2004-07-30 2006-02-02 Apple Computer Inc. Mode-based graphical user interfaces for touch sensitive input devices
US20060033724A1 (en) * 2004-07-30 2006-02-16 Apple Computer, Inc. Virtual input device placement on a touch screen user interface
US20060022955A1 (en) * 2004-07-30 2006-02-02 Apple Computer, Inc. Visual expander
US20060026536A1 (en) * 2004-07-30 2006-02-02 Apple Computer, Inc. Gestures for touch sensitive input devices
US20060026521A1 (en) * 2004-07-30 2006-02-02 Apple Computer, Inc. Gestures for touch sensitive input devices
US20060032680A1 (en) * 2004-08-16 2006-02-16 Fingerworks, Inc. Method of increasing the spatial resolution of touch sensitive devices
US20060073272A1 (en) * 2004-09-28 2006-04-06 Clearpath Technologies Corp. Control panel and method of manufacturing
US7652589B2 (en) * 2005-02-04 2010-01-26 Hewlett-Packard Development Company, L.P. Selecting a function of components based on orientation
US20060181518A1 (en) * 2005-02-14 2006-08-17 Chia Shen Spatial multiplexing to mediate direct-touch input on large displays
US20080012835A1 (en) * 2006-07-12 2008-01-17 N-Trig Ltd. Hover and touch detection for digitizer
US20080012838A1 (en) * 2006-07-13 2008-01-17 N-Trig Ltd. User specific recognition of intended user interaction with a digitizer

Cited By (378)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
USRE46548E1 (en) 1997-10-28 2017-09-12 Apple Inc. Portable computers
USRE45559E1 (en) 1997-10-28 2015-06-09 Apple Inc. Portable computers
US7714880B2 (en) * 2001-11-16 2010-05-11 Honeywell International Inc. Method and apparatus for displaying images on a display
US20030095155A1 (en) * 2001-11-16 2003-05-22 Johnson Michael J. Method and apparatus for displaying images on a display
US9983742B2 (en) 2002-07-01 2018-05-29 Apple Inc. Electronic device having display and surrounding touch sensitive bezel for user interface and control
US10474251B2 (en) 2003-09-02 2019-11-12 Apple Inc. Ambidextrous mouse
US9785258B2 (en) 2003-09-02 2017-10-10 Apple Inc. Ambidextrous mouse
US10156914B2 (en) 2003-09-02 2018-12-18 Apple Inc. Ambidextrous mouse
US11314340B2 (en) 2004-02-13 2022-04-26 Chemtron Research Llc User interface device with touch sensor
US11797107B2 (en) 2004-02-13 2023-10-24 Chemtron Research Llc Method and user interface device with touch sensor for controlling applications
US20050179650A1 (en) * 2004-02-13 2005-08-18 Ludwig Lester F. Extended parameter-set mouse-based user interface device offering offset, warping, and mixed-reference features
US20100060607A1 (en) * 2004-02-13 2010-03-11 Ludwig Lester F User interface mouse with touchpad responsive to gestures and multi-touch
US9417716B2 (en) 2004-02-13 2016-08-16 Chemtron Research Llc Mouse-based user interface device employing user-removable modules
US20050179652A1 (en) * 2004-02-13 2005-08-18 Ludwig Lester F. Mouse-based user interface device employing user-removable modules
US20110128224A1 (en) * 2004-02-13 2011-06-02 Ludwig Lester F User interface device, such as a mouse or trackball, with a high-dimension joystick providing at least three independently adjustable parameters
US8816956B2 (en) 2004-02-13 2014-08-26 Bioram Tech L.L.C. Mouse-based user interface device employing user-removable modules
US20100064238A1 (en) * 2004-02-13 2010-03-11 Lester Frank Ludwig Electronic document editing employing multiple cursors
US11809643B2 (en) 2004-02-13 2023-11-07 Chemtron Research Llc Methods and systems for controlling applications using user interface device with touch sensor
US11275405B2 (en) 2005-03-04 2022-03-15 Apple Inc. Multi-functional hand-held device
US10921941B2 (en) 2005-03-04 2021-02-16 Apple Inc. Electronic device having display and surrounding touch sensitive surfaces for user interface and control
US11360509B2 (en) 2005-03-04 2022-06-14 Apple Inc. Electronic device having display and surrounding touch sensitive surfaces for user interface and control
US9047009B2 (en) 2005-03-04 2015-06-02 Apple Inc. Electronic device having display and surrounding touch sensitive bezel for user interface and control
US7656393B2 (en) 2005-03-04 2010-02-02 Apple Inc. Electronic device having display and surrounding touch sensitive bezel for user interface and control
US20060238517A1 (en) * 2005-03-04 2006-10-26 Apple Computer, Inc. Electronic Device Having Display and Surrounding Touch Sensitive Bezel for User Interface and Control
US10386980B2 (en) 2005-03-04 2019-08-20 Apple Inc. Electronic device having display and surrounding touch sensitive surfaces for user interface and control
US20060263068A1 (en) * 2005-05-19 2006-11-23 Sony Corporation Reproducing apparatus, program, and reproduction control method
US20100195452A1 (en) * 2005-07-06 2010-08-05 Sony Corporation Contents data reproduction apparatus and contents data reproduction method
US20070010293A1 (en) * 2005-07-08 2007-01-11 Pchome Online Inc. Phone connected to a personal computer
US20070126731A1 (en) * 2005-12-06 2007-06-07 Sabram Stephen R Graphical architecture for handheld measurement system
US20090239573A1 (en) * 2005-12-20 2009-09-24 Sony Ericsson Mobile Communications Ab Electronic equipment with shuffle operation
US7882435B2 (en) * 2005-12-20 2011-02-01 Sony Ericsson Mobile Communications Ab Electronic equipment with shuffle operation
US20070156679A1 (en) * 2005-12-20 2007-07-05 Kretz Martin H Electronic equipment with shuffle operation
US8068605B2 (en) * 2006-03-07 2011-11-29 Sony Ericsson Mobile Communications Ab Programmable keypad
US20070213090A1 (en) * 2006-03-07 2007-09-13 Sony Ericsson Mobile Communications Ab Programmable keypad
US8493351B2 (en) 2006-03-30 2013-07-23 Cypress Semiconductor Corporation Apparatus and method for reducing average scan rate to detect a conductive object on a sensing device
US9152284B1 (en) 2006-03-30 2015-10-06 Cypress Semiconductor Corporation Apparatus and method for reducing average scan rate to detect a conductive object on a sensing device
US8144125B2 (en) 2006-03-30 2012-03-27 Cypress Semiconductor Corporation Apparatus and method for reducing average scan rate to detect a conductive object on a sensing device
US8482437B1 (en) 2006-05-25 2013-07-09 Cypress Semiconductor Corporation Capacitance sensing matrix for keyboard architecture
US8059015B2 (en) 2006-05-25 2011-11-15 Cypress Semiconductor Corporation Capacitance sensing matrix for keyboard architecture
US9019133B1 (en) 2006-05-25 2015-04-28 Cypress Semiconductor Corporation Low pin count solution using capacitance sensing matrix for keyboard architecture
US8040321B2 (en) 2006-07-10 2011-10-18 Cypress Semiconductor Corporation Touch-sensor with shared capacitive sensors
US20080030456A1 (en) * 2006-07-19 2008-02-07 Sony Ericsson Mobile Communications Ab Apparatus and Methods for Providing Motion Responsive Output Modifications in an Electronic Device
US9367097B2 (en) * 2006-08-02 2016-06-14 Blackberry Limited System and method for adjusting presentation of text and images on an electronic device according to an orientation of the device
US9110499B2 (en) * 2006-08-02 2015-08-18 Blackberry Limited System and method for adjusting presentation of moving images on an electronic device according to an orientation of the device
US20120139953A1 (en) * 2006-08-02 2012-06-07 Research In Motion Limited System and method for adjusting presentation of text and images on an electronic device according to an orientation of the device
US20090262074A1 (en) * 2007-01-05 2009-10-22 Invensense Inc. Controlling and accessing content using motion processing on mobile devices
US7907838B2 (en) 2007-01-05 2011-03-15 Invensense, Inc. Motion sensing and processing on mobile devices
US7796872B2 (en) 2007-01-05 2010-09-14 Invensense, Inc. Method and apparatus for producing a sharp image from a handheld device containing a gyroscope
US8462109B2 (en) 2007-01-05 2013-06-11 Invensense, Inc. Controlling and accessing content using motion processing on mobile devices
US8351773B2 (en) 2007-01-05 2013-01-08 Invensense, Inc. Motion sensing and processing on mobile devices
US9292102B2 (en) 2007-01-05 2016-03-22 Invensense, Inc. Controlling and accessing content using motion processing on mobile devices
US20100214216A1 (en) * 2007-01-05 2010-08-26 Invensense, Inc. Motion sensing and processing on mobile devices
US8607167B2 (en) 2007-01-07 2013-12-10 Apple Inc. Portable multifunction device, method, and graphical user interface for providing maps and directions
US9760272B2 (en) 2007-01-07 2017-09-12 Apple Inc. Application programming interfaces for scrolling operations
US10817162B2 (en) 2007-01-07 2020-10-27 Apple Inc. Application programming interfaces for scrolling operations
US10481785B2 (en) 2007-01-07 2019-11-19 Apple Inc. Application programming interfaces for scrolling operations
US20080165160A1 (en) * 2007-01-07 2008-07-10 Kenneth Kocienda Portable Multifunction Device, Method, and Graphical User Interface for Interpreting a Finger Gesture on a Touch Screen Display
US20080168396A1 (en) * 2007-01-07 2008-07-10 Michael Matas Portable Multifunction Device, Method, and Graphical User Interface for Providing Maps and Directions
US8519963B2 (en) 2007-01-07 2013-08-27 Apple Inc. Portable multifunction device, method, and graphical user interface for interpreting a finger gesture on a touch screen display
US9448712B2 (en) 2007-01-07 2016-09-20 Apple Inc. Application programming interfaces for scrolling operations
US8058937B2 (en) 2007-01-30 2011-11-15 Cypress Semiconductor Corporation Setting a discharge rate and a charge rate of a relaxation oscillator circuit
US10788937B2 (en) 2007-05-07 2020-09-29 Cypress Semiconductor Corporation Reducing sleep current in a capacitance sensing system
US8976124B1 (en) 2007-05-07 2015-03-10 Cypress Semiconductor Corporation Reducing sleep current in a capacitance sensing system
US20080294984A1 (en) * 2007-05-25 2008-11-27 Immersion Corporation Customizing Haptic Effects On An End User Device
US9990038B2 (en) 2007-05-25 2018-06-05 Immersion Corporation Customizing haptic effects on an end user device
US8621348B2 (en) * 2007-05-25 2013-12-31 Immersion Corporation Customizing haptic effects on an end user device
US8047075B2 (en) 2007-06-21 2011-11-01 Invensense, Inc. Vertically integrated 3-axis MEMS accelerometer with electronics
US10686930B2 (en) 2007-06-22 2020-06-16 Apple Inc. Touch screen device, method, and graphical user interface for providing maps, directions, and location based information
US11849063B2 (en) 2007-06-22 2023-12-19 Apple Inc. Touch screen device, method, and graphical user interface for providing maps, directions, and location-based information
US8258986B2 (en) 2007-07-03 2012-09-04 Cypress Semiconductor Corporation Capacitive-matrix keyboard with multiple touch detection
US10288427B2 (en) 2007-07-06 2019-05-14 Invensense, Inc. Integrated motion processing unit (MPU) with MEMS inertial sensing and embedded digital electronics
US8997564B2 (en) 2007-07-06 2015-04-07 Invensense, Inc. Integrated motion processing unit (MPU) with MEMS inertial sensing and embedded digital electronics
US8250921B2 (en) 2007-07-06 2012-08-28 Invensense, Inc. Integrated motion processing unit (MPU) with MEMS inertial sensing and embedded digital electronics
US20100188349A1 (en) * 2007-09-14 2010-07-29 Yannick Molard Control panels for onboard instruments
US8446381B2 (en) * 2007-09-14 2013-05-21 Delphi Technologies, Inc. Control panels for onboard instruments
US20100289760A1 (en) * 2007-09-14 2010-11-18 Kyocera Corporation Electronic apparatus
US20100238109A1 (en) * 2007-09-18 2010-09-23 Thomson Licensing User interface for set top box
US9167209B2 (en) * 2007-09-18 2015-10-20 Thomson Licensing Method for a touch screen that interfaces with a set top box
US20090111517A1 (en) * 2007-10-31 2009-04-30 Mitac International Corp. Handheld device capable of independently actuating camera and method thereof
US8253686B2 (en) * 2007-11-26 2012-08-28 Electronics And Telecommunications Research Institute Pointing apparatus capable of providing haptic feedback, and haptic interaction system and method using the same
US20090135164A1 (en) * 2007-11-26 2009-05-28 Ki Uk Kyung Pointing apparatus capable of providing haptic feedback, and haptic interaction system and method using the same
US9015064B2 (en) * 2007-11-30 2015-04-21 Michelle Fisher Utilizing a secure element for NFC transactions which includes response data during induction
US20130097032A1 (en) * 2007-11-30 2013-04-18 Blaze Mobile, Inc. Utilizing shopping lists for nfc transactions
US11272050B2 (en) 2007-12-03 2022-03-08 Semiconductor Energy Laboratory Co., Ltd. Mobile phone
US11843714B2 (en) 2007-12-03 2023-12-12 Semiconductor Energy Laboratory Co., Ltd. Mobile phone
US10375231B2 (en) 2007-12-03 2019-08-06 Semiconductor Energy Laboratory Co., Ltd. Mobile phone
US9883024B2 (en) 2007-12-03 2018-01-30 Semiconductor Energy Laboratory Co., Ltd. Mobile phone
US9917944B2 (en) 2007-12-03 2018-03-13 Semiconductor Energy Laboratory Co., Ltd. Mobile phone
US9118777B2 (en) 2007-12-03 2015-08-25 Semiconductor Energy Laboratory Co., Ltd. Mobile phone
US20090143109A1 (en) * 2007-12-03 2009-06-04 Semiconductor Energy Laboratory Co., Ltd. Mobile phone
US8478346B2 (en) 2007-12-03 2013-07-02 Semiconductor Energy Laboratory Co., Ltd. Mobile phone
US20090150659A1 (en) * 2007-12-06 2009-06-11 Fuccello James R Device and Method for Adjusting a Converged Device
US8122237B2 (en) * 2007-12-06 2012-02-21 Symbol Technologies, Inc. Device and method for adjusting a converged device having barcode reading and RFID functionalities
US9846175B2 (en) 2007-12-10 2017-12-19 Invensense, Inc. MEMS rotation sensor with integrated electronics
US8960002B2 (en) 2007-12-10 2015-02-24 Invensense, Inc. Vertically integrated 3-axis MEMS angular accelerometer with integrated electronics
US7934423B2 (en) 2007-12-10 2011-05-03 Invensense, Inc. Vertically integrated 3-axis MEMS angular accelerometer with integrated electronics
US9330381B2 (en) 2008-01-06 2016-05-03 Apple Inc. Portable multifunction device, method, and graphical user interface for viewing and managing electronic calendars
US8327272B2 (en) 2008-01-06 2012-12-04 Apple Inc. Portable multifunction device, method, and graphical user interface for viewing and managing electronic calendars
US10503366B2 (en) 2008-01-06 2019-12-10 Apple Inc. Portable multifunction device, method, and graphical user interface for viewing and managing electronic calendars
US11126326B2 (en) 2008-01-06 2021-09-21 Apple Inc. Portable multifunction device, method, and graphical user interface for viewing and managing electronic calendars
US10521084B2 (en) 2008-01-06 2019-12-31 Apple Inc. Portable multifunction device, method, and graphical user interface for viewing and managing electronic calendars
US9792001B2 (en) 2008-01-06 2017-10-17 Apple Inc. Portable multifunction device, method, and graphical user interface for viewing and managing electronic calendars
US20090174680A1 (en) * 2008-01-06 2009-07-09 Freddy Allen Anzures Portable Multifunction Device, Method, and Graphical User Interface for Viewing and Managing Electronic Calendars
US9342154B2 (en) 2008-01-18 2016-05-17 Invensense, Inc. Interfacing application programs and motion sensors of a device
US9811174B2 (en) 2008-01-18 2017-11-07 Invensense, Inc. Interfacing application programs and motion sensors of a device
US8952832B2 (en) 2008-01-18 2015-02-10 Invensense, Inc. Interfacing application programs and motion sensors of a device
US8020441B2 (en) 2008-02-05 2011-09-20 Invensense, Inc. Dual mode sensing for vibratory gyroscope
US11687212B2 (en) 2008-04-01 2023-06-27 Litl Llc Method and apparatus for managing digital media content
US10289154B2 (en) 2008-04-01 2019-05-14 Litl Llc Portable computer with multiple display configurations
US9563229B2 (en) * 2008-04-01 2017-02-07 Litl Llc Portable computer with multiple display configurations
US10564818B2 (en) 2008-04-01 2020-02-18 Litl Llc System and method for streamlining user interaction with electronic content
US10684743B2 (en) 2008-04-01 2020-06-16 Litl Llc Method and apparatus for managing digital media content
US10782733B2 (en) 2008-04-01 2020-09-22 Litl Llc Portable computer with multiple display configurations
US9880715B2 (en) 2008-04-01 2018-01-30 Litl Llc System and method for streamlining user interaction with electronic content
US20130141854A1 (en) * 2008-04-01 2013-06-06 Yves Behar Portable computer with multiple display configurations
US9495070B2 (en) 2008-04-01 2016-11-15 Litl Llc Method and apparatus for managing digital media content
US11604566B2 (en) 2008-04-01 2023-03-14 Litl Llc System and method for streamlining user interaction with electronic content
US11853118B2 (en) 2008-04-01 2023-12-26 Litl Llc Portable computer with multiple display configurations
US9927835B2 (en) 2008-04-01 2018-03-27 Litl Llc Portable computer with multiple display configurations
US8508039B1 (en) 2008-05-08 2013-08-13 Invensense, Inc. Wafer scale chip scale packaging of vertically integrated MEMS sensors with electronics
US20090289911A1 (en) * 2008-05-20 2009-11-26 Canon Kabushiki Kaisha Information processing apparatus and control method thereof
US8917245B2 (en) * 2008-05-20 2014-12-23 Canon Kabushiki Kaisha Information processing apparatus and control method thereof
US20090300530A1 (en) * 2008-05-29 2009-12-03 Telcordia Technologies, Inc. Method and system for multi-touch-based browsing of media summarizations on a handheld device
WO2009155092A3 (en) * 2008-05-29 2010-02-18 Telcordia Technologies, Inc. Method and system for multi-touch-based browsing of media summarizations on a handheld device
US20090300498A1 (en) * 2008-05-29 2009-12-03 Telcordia Technologies, Inc. Method and System for Generating and Presenting Mobile Content Summarization
US8171410B2 (en) 2008-05-29 2012-05-01 Telcordia Technologies, Inc. Method and system for generating and presenting mobile content summarization
WO2009155092A2 (en) * 2008-05-29 2009-12-23 Telcordia Technologies, Inc. Method and system for multi-touch-based browsing of media summarizations on a handheld device
US8584048B2 (en) 2008-05-29 2013-11-12 Telcordia Technologies, Inc. Method and system for multi-touch-based browsing of media summarizations on a handheld device
US20090313660A1 (en) * 2008-06-16 2009-12-17 Imu Solutions, Inc. Home entertainment system and operating method thereof
US20100037180A1 (en) * 2008-08-08 2010-02-11 At&T Intellectual Property I, L.P. Touch Screen Remote Control with Dynamic Keypad Layouts
US8723655B2 (en) * 2008-08-14 2014-05-13 Fm Marketing Gmbh Remote control and method for the remote control of multimedia appliances
US20110140867A1 (en) * 2008-08-14 2011-06-16 Fm Marketing Gmbh Remote control and method for the remote control of multimedia appliances
US20160011699A1 (en) * 2008-08-15 2016-01-14 Intel Corporation Techniques For Automatically Distinguishing Between Users Of A Handheld Device
US7969287B2 (en) 2008-08-18 2011-06-28 Visteon Global Technologies, Inc. Haptic effect control system
US20100039244A1 (en) * 2008-08-18 2010-02-18 Michael Dean Tschirhart Haptic effect control system
US20100060586A1 (en) * 2008-09-05 2010-03-11 Pisula Charles J Portable touch screen device, method, and graphical user interface for providing workout support
US8341557B2 (en) 2008-09-05 2012-12-25 Apple Inc. Portable touch screen device, method, and graphical user interface for providing workout support
US20100060588A1 (en) * 2008-09-09 2010-03-11 Microsoft Corporation Temporally separate touch input
US8141424B2 (en) 2008-09-12 2012-03-27 Invensense, Inc. Low inertia frame for detecting coriolis acceleration
US8539835B2 (en) 2008-09-12 2013-09-24 Invensense, Inc. Low inertia frame for detecting coriolis acceleration
US8255825B2 (en) 2008-10-07 2012-08-28 Microsoft Corporation Content aware adaptive display
US20100088630A1 (en) * 2008-10-07 2010-04-08 Microsoft Corporation Content aware adaptive display
WO2010045498A1 (en) * 2008-10-15 2010-04-22 Invensense Inc. Mobile devices with motion gesture recognition
US20100097347A1 (en) * 2008-10-21 2010-04-22 Tzu-Chih Lin Pressure Detection Module, and Touch Panel with Pressure Detection Module
US8487639B1 (en) 2008-11-21 2013-07-16 Cypress Semiconductor Corporation Receive demodulator for capacitive sensing
US8575947B1 (en) 2008-11-21 2013-11-05 Cypress Semiconductor Corporation Receive demodulator for capacitive sensing
US20100129782A1 (en) * 2008-11-25 2010-05-27 Sony Corporation And Sony Electronics Inc. Electronic book with enhanced features
US8502787B2 (en) * 2008-11-26 2013-08-06 Panasonic Corporation System and method for differentiating between intended and unintended user input on a touchpad
US20100127995A1 (en) * 2008-11-26 2010-05-27 Panasonic Corporation System and method for differentiating between intended and unintended user input on a touchpad
US20100177034A1 (en) * 2009-01-12 2010-07-15 Samsung Electronics Co., Ltd. Portable storage device having user interface and method of controlling the user interface
US11334239B2 (en) * 2009-01-23 2022-05-17 Samsung Electronics Co., Ltd. Mobile terminal having dual touch screen and method of controlling content therein
EP3258679A1 (en) * 2009-01-23 2017-12-20 Samsung Electronics Co., Ltd. Mobile terminal having dual touch screen and method of controlling content therein
US10705722B2 (en) 2009-01-23 2020-07-07 Samsung Electronics Co., Ltd. Mobile terminal having dual touch screen and method of controlling content therein
US20110093822A1 (en) * 2009-01-29 2011-04-21 Jahanzeb Ahmed Sherwani Image Navigation for Touchscreen User Interface
US7870496B1 (en) 2009-01-29 2011-01-11 Jahanzeb Ahmed Sherwani System using touchscreen user interface of a mobile device to remotely control a host computer
US8276085B2 (en) 2009-01-29 2012-09-25 Iteleport, Inc. Image navigation for touchscreen user interface
US20100229121A1 (en) * 2009-03-09 2010-09-09 Telcordia Technologies, Inc. System and method for capturing, aggregating and presenting attention hotspots in shared media
US8296675B2 (en) 2009-03-09 2012-10-23 Telcordia Technologies, Inc. System and method for capturing, aggregating and presenting attention hotspots in shared media
US9442146B2 (en) 2009-03-26 2016-09-13 Parade Technologies, Ltd. Multi-mode capacitive sensing device and method with current conveyor
WO2010111668A1 (en) * 2009-03-26 2010-09-30 Cypress Semiconductor Multi-functional capacitance sensing circuit with a current conveyor
US8866500B2 (en) * 2009-03-26 2014-10-21 Cypress Semiconductor Corporation Multi-functional capacitance sensing circuit with a current conveyor
US9185172B2 (en) 2009-04-15 2015-11-10 Wyse Technology L.L.C. System and method for rendering a remote view at a client device
US20100268941A1 (en) * 2009-04-15 2010-10-21 Wyse Technology Inc. Remote-session-to-go method and apparatus
US20100268940A1 (en) * 2009-04-15 2010-10-21 Wyse Technology Inc. Method and apparatus for portability of a remote session
US9448815B2 (en) 2009-04-15 2016-09-20 Wyse Technology L.L.C. Server-side computing from a remote client device
US20100268939A1 (en) * 2009-04-15 2010-10-21 Wyse Technology Inc. Method and apparatus for authentication of a remote session
US8863237B2 (en) 2009-04-15 2014-10-14 Wyse Technology L.L.C. Remote-session-to-go method and apparatus
US20100268813A1 (en) * 2009-04-15 2010-10-21 Wyse Technology Inc. System and method for handling remote drawing commands
US20100269046A1 (en) * 2009-04-15 2010-10-21 Wyse Technology Inc. Sever-side computing from a remote client device
US20100269048A1 (en) * 2009-04-15 2010-10-21 Wyse Technology Inc. Method and system of specifying application user interface of a remote client device
US9553953B2 (en) 2009-04-15 2017-01-24 Dell Products L.P. Method and apparatus for extending capabilities of a virtualization domain to support features available in a normal desktop application
US20100269039A1 (en) * 2009-04-15 2010-10-21 Wyse Technology Inc. Custom pointer features for touch-screen on remote client devices
US9578113B2 (en) 2009-04-15 2017-02-21 Wyse Technology L.L.C. Method and apparatus for transferring remote session data
US20100269047A1 (en) * 2009-04-15 2010-10-21 Wyse Technology Inc. System and method for rendering a composite view at a client device
US20100268828A1 (en) * 2009-04-15 2010-10-21 Wyse Technology Inc. Method and apparatus for transferring remote session data
US20100268762A1 (en) * 2009-04-15 2010-10-21 Wyse Technology Inc. System and method for scrolling a remote application
US9189124B2 (en) 2009-04-15 2015-11-17 Wyse Technology L.L.C. Custom pointer features for touch-screen on remote client devices
US20100269057A1 (en) * 2009-04-15 2010-10-21 Wyse Technology Inc. System and method for communicating events at a server to a remote device
US20100269152A1 (en) * 2009-04-15 2010-10-21 Wyse Technology Inc. Method and system for rendering composite view of an application
US9106696B2 (en) 2009-04-15 2015-08-11 Wyse Technology L.L.C. Method and apparatus for portability of a remote session
US9444894B2 (en) * 2009-04-15 2016-09-13 Wyse Technology Llc System and method for communicating events at a server to a remote device
US9413831B2 (en) 2009-04-15 2016-08-09 Wyse Technology L.L.C. Method and apparatus for authentication of a remote session
US9384526B2 (en) 2009-04-15 2016-07-05 Wyse Technology L.L.C. System and method for handling remote drawing commands
US9374426B2 (en) 2009-04-15 2016-06-21 Wyse Technology L.L.C. Remote-session-to-go method and apparatus
US10244056B2 (en) 2009-04-15 2019-03-26 Wyse Technology L.L.C. Method and apparatus for transferring remote session data
US8676926B2 (en) 2009-04-15 2014-03-18 Wyse Technology L.L.C. System and method for handling remote drawing commands
US9191449B2 (en) 2009-04-15 2015-11-17 Wyse Technology L.L.C. System and method for communicating events at a server to a remote device
US9185171B2 (en) 2009-04-15 2015-11-10 Wyse Technology L.L.C. Method and system of specifying application user interface of a remote client device
US8869239B2 (en) 2009-04-15 2014-10-21 Wyse Technology L.L.C. Method and system for rendering composite view of an application
US9191448B2 (en) 2009-04-15 2015-11-17 Wyse Technology L.L.C. System and method for rendering a composite view at a client device
AU2010241911C1 (en) * 2009-05-01 2014-05-22 Apple Inc. Directional touch remote
US11792256B2 (en) 2009-05-01 2023-10-17 Apple Inc. Directional touch remote
US20100277337A1 (en) * 2009-05-01 2010-11-04 Apple Inc. Directional touch remote
US10958707B2 (en) 2009-05-01 2021-03-23 Apple Inc. Directional touch remote
CN102460367A (en) * 2009-05-01 2012-05-16 苹果公司 Directional touch remote
US8742885B2 (en) * 2009-05-01 2014-06-03 Apple Inc. Directional touch remote
AU2010241911B2 (en) * 2009-05-01 2013-12-19 Apple Inc. Directional touch remote
US20100312368A1 (en) * 2009-06-05 2010-12-09 Anthony Rodriguez Aural Audio Player
US8464182B2 (en) 2009-06-07 2013-06-11 Apple Inc. Device, method, and graphical user interface for providing maps, directions, and location-based information
US20100309149A1 (en) * 2009-06-07 2010-12-09 Chris Blumenberg Device, Method, and Graphical User Interface for Providing Maps, Directions, and Location-Based Information
US8909915B2 (en) * 2009-06-16 2014-12-09 Intel Corporation Multi-mode handheld wireless device with shared mode to support cross-mode communications
US20120079265A1 (en) * 2009-06-16 2012-03-29 Bran Ferren Multi-mode handheld wireless device
US8407623B2 (en) 2009-06-25 2013-03-26 Apple Inc. Playback control using a touch interface
US20100328224A1 (en) * 2009-06-25 2010-12-30 Apple Inc. Playback control using a touch interface
US20110018695A1 (en) * 2009-07-24 2011-01-27 Research In Motion Limited Method and apparatus for a touch-sensitive display
US8378798B2 (en) * 2009-07-24 2013-02-19 Research In Motion Limited Method and apparatus for a touch-sensitive display
US10242397B2 (en) * 2009-07-29 2019-03-26 Paypal, Inc. No authentication payment and seamless authentication
US11182842B2 (en) 2009-07-29 2021-11-23 Paypal, Inc. No authentication payment and seamless authentication
US10248221B2 (en) 2009-08-17 2019-04-02 Apple Inc. Housing as an I/O device
US8654524B2 (en) 2009-08-17 2014-02-18 Apple Inc. Housing as an I/O device
US9600037B2 (en) 2009-08-17 2017-03-21 Apple Inc. Housing as an I/O device
US11644865B2 (en) 2009-08-17 2023-05-09 Apple Inc. Housing as an I/O device
US10739868B2 (en) 2009-08-17 2020-08-11 Apple Inc. Housing as an I/O device
US20110074827A1 (en) * 2009-09-25 2011-03-31 Research In Motion Limited Electronic device including touch-sensitive input device and method of controlling same
US20110105186A1 (en) * 2009-10-29 2011-05-05 Research In Motion Limited Systems and methods for providing direct and indirect navigation modes for touchscreen devices
US20160116986A1 (en) * 2009-10-29 2016-04-28 Blackberry Limited Systems and methods for providing direct and indirect navigation modes for touchscreen devices
US20160274710A1 (en) * 2009-12-14 2016-09-22 Synaptics Incorporated System and method for measuring individual force in multi-object sensing
US9720538B2 (en) * 2009-12-14 2017-08-01 Synaptics Incorporated System and method for measuring individual force in multi-object sensing
US8754873B2 (en) * 2009-12-14 2014-06-17 Synaptics Incorporated System and method for measuring individual force in multi-object sensing
US9377888B2 (en) * 2009-12-14 2016-06-28 Synaptics Incorporated System and method for measuring individual force in multi-object sensing
US20140240288A1 (en) * 2009-12-14 2014-08-28 Synaptics Incorporated System and method for measuring individual force in multi-object sensing
US20110148788A1 (en) * 2009-12-17 2011-06-23 Shenzhen Futaihong Precision Industry Co., Ltd. Touch screen device with coordinate correction module
US9564075B2 (en) * 2009-12-30 2017-02-07 Cyweemotion Hk Limited Electronic control apparatus and method for responsively controlling media content displayed on portable electronic device
US20110157231A1 (en) * 2009-12-30 2011-06-30 Cywee Group Limited Electronic control apparatus and method for responsively controlling media content displayed on portable electronic device
US9798395B2 (en) 2009-12-30 2017-10-24 Cm Hk Limited Electronic control apparatus and method for responsively controlling media content displayed on portable electronic device
US8862576B2 (en) 2010-01-06 2014-10-14 Apple Inc. Device, method, and graphical user interface for mapping directions between search results
US10169431B2 (en) 2010-01-06 2019-01-01 Apple Inc. Device, method, and graphical user interface for mapping directions between search results
US20110167058A1 (en) * 2010-01-06 2011-07-07 Van Os Marcel Device, Method, and Graphical User Interface for Mapping Directions Between Search Results
US8456297B2 (en) 2010-01-06 2013-06-04 Apple Inc. Device, method, and graphical user interface for tracking movement on a map
US20110163874A1 (en) * 2010-01-06 2011-07-07 Van Os Marcel Device, Method, and Graphical User Interface for Tracking Movement on a Map
US20110191692A1 (en) * 2010-02-03 2011-08-04 Oto Technologies, Llc System and method for e-book contextual communication
US8463316B2 (en) * 2010-02-17 2013-06-11 Harris Corporation Communication device with a multi-functional control
US20110201379A1 (en) * 2010-02-17 2011-08-18 Harris Corporation Communication device with a multi-functional control
US20110231474A1 (en) * 2010-03-22 2011-09-22 Howard Locker Audio Book and e-Book Synchronization
US9323756B2 (en) * 2010-03-22 2016-04-26 Lenovo (Singapore) Pte. Ltd. Audio book and e-book synchronization
US10751615B2 (en) 2010-04-28 2020-08-25 Kabushiki Kaisha Square Enix User interface processing apparatus, method of processing user interface, and non-transitory computer-readable medium embodying computer program for processing user interface having variable transparency
US9517411B2 (en) 2010-04-28 2016-12-13 Kabushiki Kaisha Square Enix Transparent user interface game control processing method, apparatus, and medium
US20110267026A1 (en) * 2010-04-30 2011-11-03 Lenovo (Singapore) Pte, Ltd. Method and Apparatus for Modifying a Transition to an Altered Power State of an Electronic Device Based on Accelerometer Output
US9122735B2 (en) * 2010-04-30 2015-09-01 Lenovo (Singapore) Pte. Ltd. Method and apparatus for modifying a transition to an altered power state of an electronic device based on accelerometer output
AU2011253295B2 (en) * 2010-05-11 2014-09-18 L3Harris Technologies, Inc. Electronic device with a multi-functional control
KR101495411B1 (en) 2010-05-11 2015-02-24 해리스 코포레이션 Electronic device with a multi-functional control
US8532585B2 (en) * 2010-05-11 2013-09-10 Harris Corporation Electronic device with rotary knob multi-functional control
US20110281533A1 (en) * 2010-05-11 2011-11-17 Harris Corporation Electronic device with a multi-functional control
US8736560B2 (en) * 2010-05-14 2014-05-27 Blackberry Limited Electronic device including tactile touch-sensitive display and method of controlling same
US20110279382A1 (en) * 2010-05-14 2011-11-17 Research In Motion Limited Electronic device including tactile touch-sensitive display and method of controlling same
US10413821B2 (en) 2010-05-20 2019-09-17 John W. Howard Touch screen with virtual joystick and methods for use therewith
US10646779B2 (en) 2010-05-20 2020-05-12 John W. Howard Touch screen with virtual joystick and methods for use therewith
US20110285636A1 (en) * 2010-05-20 2011-11-24 Howard John W Touch screen with virtual joystick and methods for use therewith
US9262073B2 (en) * 2010-05-20 2016-02-16 John W. Howard Touch screen with virtual joystick and methods for use therewith
US20110291958A1 (en) * 2010-05-28 2011-12-01 Fih (Hong Kong) Limited Touch-type transparent keyboard
US8762890B2 (en) 2010-07-27 2014-06-24 Telcordia Technologies, Inc. System and method for interactive projection and playback of relevant media segments onto the facets of three-dimensional shapes
US20130135205A1 (en) * 2010-08-19 2013-05-30 Beijing Lenovo Software Ltd. Display Method And Terminal Device
US10353580B2 (en) 2010-08-20 2019-07-16 Sony Corporation Information processing apparatus, program, and operation control method
US10649651B2 (en) 2010-08-20 2020-05-12 Sony Corporation Information processing apparatus, program, and operation control method
US10168900B2 (en) * 2010-08-20 2019-01-01 Sony Corporation Information processing apparatus, program, and operation control me
US9538229B2 (en) * 2010-09-15 2017-01-03 Verizon Patent And Licensing Inc. Media experience for touch screen devices
US20120062473A1 (en) * 2010-09-15 2012-03-15 Verizon Patent And Licensing Inc. Media experience for touch screen devices
US20120102439A1 (en) * 2010-10-22 2012-04-26 April Slayden Mitchell System and method of modifying the display content based on sensor input
US8854802B2 (en) 2010-10-22 2014-10-07 Hewlett-Packard Development Company, L.P. Display with rotatable display screen
US9164581B2 (en) 2010-10-22 2015-10-20 Hewlett-Packard Development Company, L.P. Augmented reality display system and method of display
US9372615B2 (en) 2010-12-28 2016-06-21 Samsung Electronics Co., Ltd. Method and apparatus for touch input for electronic device
US8521791B2 (en) * 2011-03-10 2013-08-27 Chi Mei Communication Systems, Inc. Electronic device and file management method
US20120233226A1 (en) * 2011-03-10 2012-09-13 Chi Mei Communication Systems, Inc. Electronic device and file management method
TWI483171B (en) * 2011-03-10 2015-05-01 Chi Mei Comm Systems Inc File management system and method of an electronic device
US9268441B2 (en) 2011-04-05 2016-02-23 Parade Technologies, Ltd. Active integrator for a capacitive sense array
US8897897B2 (en) 2011-05-25 2014-11-25 Remote Technologies, Inc. Companion control interface for smart devices
US20120299839A1 (en) * 2011-05-27 2012-11-29 Honeywell International Inc. Aircraft user interfaces with multi-mode haptics
US10180722B2 (en) * 2011-05-27 2019-01-15 Honeywell International Inc. Aircraft user interfaces with multi-mode haptics
US11503359B2 (en) 2011-06-20 2022-11-15 Enseo, Llc Set top/back box, system and method for providing a remote control device
US11765420B2 (en) 2011-06-20 2023-09-19 Enseo, Llc Television and system and method for providing a remote control device
US20190110101A1 (en) * 2011-06-20 2019-04-11 Enseo, Inc. Set-Top Box with Enhanced Content and System and Method for Use of Same
US20190110098A1 (en) * 2011-06-20 2019-04-11 Enseo, Inc. Set-Top Box with Enhanced Functionality and System and Method for Use of Same
US11039197B2 (en) 2011-06-20 2021-06-15 Enseo, Llc Set top/back box, system and method for providing a remote control device
US10798443B2 (en) * 2011-06-20 2020-10-06 Enseo, Inc. Set-top box with enhanced content and system and method for use of same
US20190335224A1 (en) * 2011-06-20 2019-10-31 Enseo, Inc. Television and system and method for providing a remote control device
US10791359B2 (en) * 2011-06-20 2020-09-29 Enseo, Inc. Set-top box with enhanced functionality and system and method for use of same
US11722724B2 (en) 2011-06-20 2023-08-08 Enseo, Llc Set top/back box, system and method for providing a remote control device
US11516530B2 (en) 2011-06-20 2022-11-29 Enseo, Llc Television and system and method for providing a remote control device
US11051065B2 (en) * 2011-06-20 2021-06-29 Enseo, Llc Television and system and method for providing a remote control device
US20170150201A1 (en) * 2011-06-20 2017-05-25 Enseo, Inc. Set Top/Back Box, System and Method for Providing a Remote Control Device
US11153638B2 (en) 2011-06-20 2021-10-19 Enseo, Llc Set-top box with enhanced content and system and method for use of same
US10187685B2 (en) * 2011-06-20 2019-01-22 Enseo, Inc. Set top/back box, system and method for providing a remote control device
US11223872B2 (en) 2011-06-20 2022-01-11 Enseo, Llc Set-top box with enhanced functionality and system and method for use of same
US10149005B2 (en) * 2011-06-20 2018-12-04 Enseo, Inc. Set-top box with enhanced content and system and method for use of same
US10148998B2 (en) * 2011-06-20 2018-12-04 Enseo, Inc. Set-top box with enhanced functionality and system and method for use of same
US10136176B2 (en) * 2011-06-20 2018-11-20 Enseo, Inc. Set top/back box, system and method for providing a remote control device
US8775966B2 (en) 2011-06-29 2014-07-08 Motorola Mobility Llc Electronic device and method with dual mode rear TouchPad
US10656754B1 (en) 2011-08-05 2020-05-19 P4tents1, LLC Devices and methods for navigating between user interfaces
US10838542B1 (en) 2011-08-05 2020-11-17 P4tents1, LLC Gesture-equipped touch screen system, method, and computer program product
US10649571B1 (en) 2011-08-05 2020-05-12 P4tents1, LLC Devices, methods, and graphical user interfaces for manipulating user interface objects with visual and/or haptic feedback
US10649581B1 (en) 2011-08-05 2020-05-12 P4tents1, LLC Devices, methods, and graphical user interfaces for manipulating user interface objects with visual and/or haptic feedback
US10649578B1 (en) 2011-08-05 2020-05-12 P4tents1, LLC Gesture-equipped touch screen system, method, and computer program product
US10649579B1 (en) 2011-08-05 2020-05-12 P4tents1, LLC Devices, methods, and graphical user interfaces for manipulating user interface objects with visual and/or haptic feedback
US10642413B1 (en) 2011-08-05 2020-05-05 P4tents1, LLC Gesture-equipped touch screen system, method, and computer program product
US10656757B1 (en) 2011-08-05 2020-05-19 P4tents1, LLC Gesture-equipped touch screen system, method, and computer program product
US10656758B1 (en) 2011-08-05 2020-05-19 P4tents1, LLC Gesture-equipped touch screen system, method, and computer program product
US10656752B1 (en) 2011-08-05 2020-05-19 P4tents1, LLC Gesture-equipped touch screen system, method, and computer program product
US9417754B2 (en) * 2011-08-05 2016-08-16 P4tents1, LLC User interface system, method, and computer program product
US10656755B1 (en) 2011-08-05 2020-05-19 P4tents1, LLC Gesture-equipped touch screen system, method, and computer program product
US10656753B1 (en) 2011-08-05 2020-05-19 P4tents1, LLC Gesture-equipped touch screen system, method, and computer program product
US10656756B1 (en) 2011-08-05 2020-05-19 P4tents1, LLC Gesture-equipped touch screen system, method, and computer program product
US10656759B1 (en) 2011-08-05 2020-05-19 P4tents1, LLC Devices, methods, and graphical user interfaces for manipulating user interface objects with visual and/or haptic feedback
US10664097B1 (en) 2011-08-05 2020-05-26 P4tents1, LLC Devices, methods, and graphical user interfaces for manipulating user interface objects with visual and/or haptic feedback
US10671213B1 (en) 2011-08-05 2020-06-02 P4tents1, LLC Devices, methods, and graphical user interfaces for manipulating user interface objects with visual and/or haptic feedback
US10671212B1 (en) 2011-08-05 2020-06-02 P4tents1, LLC Gesture-equipped touch screen system, method, and computer program product
US10606396B1 (en) 2011-08-05 2020-03-31 P4tents1, LLC Gesture-equipped touch screen methods for duration-based functions
US10649580B1 (en) 2011-08-05 2020-05-12 P4tents1, LLC Devices, methods, and graphical use interfaces for manipulating user interface objects with visual and/or haptic feedback
US11061503B1 (en) 2011-08-05 2021-07-13 P4tents1, LLC Devices, methods, and graphical user interfaces for manipulating user interface objects with visual and/or haptic feedback
US10592039B1 (en) 2011-08-05 2020-03-17 P4tents1, LLC Gesture-equipped touch screen system, method, and computer program product for displaying multiple active applications
US10725581B1 (en) 2011-08-05 2020-07-28 P4tents1, LLC Devices, methods and graphical user interfaces for manipulating user interface objects with visual and/or haptic feedback
US10996787B1 (en) 2011-08-05 2021-05-04 P4tents1, LLC Gesture-equipped touch screen system, method, and computer program product
US10551966B1 (en) 2011-08-05 2020-02-04 P4tents1, LLC Gesture-equipped touch screen system, method, and computer program product
US10534474B1 (en) 2011-08-05 2020-01-14 P4tents1, LLC Gesture-equipped touch screen system, method, and computer program product
US11740727B1 (en) 2011-08-05 2023-08-29 P4Tents1 Llc Devices, methods, and graphical user interfaces for manipulating user interface objects with visual and/or haptic feedback
US10782819B1 (en) 2011-08-05 2020-09-22 P4tents1, LLC Gesture-equipped touch screen system, method, and computer program product
US10936114B1 (en) 2011-08-05 2021-03-02 P4tents1, LLC Gesture-equipped touch screen system, method, and computer program product
US10788931B1 (en) 2011-08-05 2020-09-29 P4tents1, LLC Devices, methods, and graphical user interfaces for manipulating user interface objects with visual and/or haptic feedback
US9106651B2 (en) * 2011-09-19 2015-08-11 Qualcomm Incorporated Sending human input device commands over internet protocol
US20130246565A1 (en) * 2011-09-19 2013-09-19 Qualcomn Incorporated Sending human input device commands over internet protocol
US20140232675A1 (en) * 2011-09-22 2014-08-21 Toyota Jidosha Kabushiki Kaisha Input apparatus and input apparatus processing method
CN103034394A (en) * 2011-10-06 2013-04-10 东芝三星存储技术韩国株式会社 Multimedia device having detachable controller
US9075575B2 (en) * 2011-10-06 2015-07-07 Toshiba Samsung Storage Technology Korea Corporation Multimedia device having detachable controller functioning as remote control while detached
US20130088333A1 (en) * 2011-10-06 2013-04-11 Byung-youn Song Multimedia device having detachable controller
US20130099903A1 (en) * 2011-10-24 2013-04-25 Hon Hai Precision Industry Co., Ltd. Remote controller
US8779904B2 (en) * 2011-10-24 2014-07-15 Hon Hai Precision Industry Co., Ltd. Multimode remote controller comprising an accelerometer, a gyroscope, a capacitive pressure transducer, and a touch pad
US9182869B2 (en) 2011-12-16 2015-11-10 Panasonic Intellectual Property Corporation Of America Touch panel and electronic device
US9182870B2 (en) 2011-12-16 2015-11-10 Panasonic Intellectual Property Corporation Of America Touch panel and electronic device
US9052791B2 (en) 2011-12-16 2015-06-09 Panasonic Intellectual Property Corporation Of America Touch panel and electronic device
US11165896B2 (en) 2012-01-07 2021-11-02 Samsung Electronics Co., Ltd. Method and apparatus for providing event of portable device having flexible display unit
US10244091B2 (en) 2012-01-07 2019-03-26 Samsung Electronics Co., Ltd. Method and apparatus for providing event of portable device having flexible display unit
US10178208B2 (en) 2012-01-07 2019-01-08 Samsung Electronics Co., Ltd. Method and apparatus for providing event of portable device having flexible display unit
US9298293B2 (en) 2012-04-19 2016-03-29 Panasonic Intellectual Property Corporation Of America Portable electronic apparatus
JP2014016989A (en) * 2012-07-05 2014-01-30 Samsung Electronics Co Ltd Input sensing method, and electronic apparatus for processing said method
US11023080B2 (en) 2012-07-05 2021-06-01 Samsung Electronics Co., Ltd. Apparatus and method for detecting an input to a terminal
US10437392B2 (en) 2012-07-05 2019-10-08 Samsung Electronics Co., Ltd. Apparatus and method for detecting hard and soft touch by using acoustic sensors
US20140160893A1 (en) * 2012-12-10 2014-06-12 Apple Inc. Mobile device with ultrasound ranging
US10735106B2 (en) * 2012-12-10 2020-08-04 Apple Inc. Mobile device with ultrasound ranging
US10013067B2 (en) * 2013-04-15 2018-07-03 Zte Corporation Gesture control method, apparatus and system
US20160266653A1 (en) * 2013-04-15 2016-09-15 Zte Corporation Gesture control method, apparatus and system
US9817486B2 (en) 2013-07-11 2017-11-14 Samsung Electronics Co., Ltd. User terminal device for displaying contents and methods thereof
US9977516B2 (en) 2013-07-11 2018-05-22 Samsung Electronics Co., Ltd. User terminal device for displaying application and methods thereof
US10691313B2 (en) 2013-07-11 2020-06-23 Samsung Electronics Co., Ltd. User terminal device for displaying contents and methods thereof
US11675391B2 (en) 2013-07-11 2023-06-13 Samsung Electronics Co., Ltd. User terminal device for displaying contents and methods thereof
US9823756B2 (en) 2013-07-11 2017-11-21 Samsung Electronics Co., Ltd. User terminal device for supporting user interaction and methods thereof
US10318120B2 (en) 2013-07-11 2019-06-11 Samsung Electronics Co., Ltd. User terminal device for displaying contents and methods thereof
US11409327B2 (en) 2013-07-11 2022-08-09 Samsung Electronics Co., Ltd. User terminal device for displaying contents and methods thereof
US9537935B2 (en) 2013-09-30 2017-01-03 Eric Trent Dryden Consumer PC in the cloud
WO2015184116A3 (en) * 2014-05-28 2016-06-02 Sioux Automation Center Inc Electronic remote control mixing system, associated devices and methods
US9667434B2 (en) 2015-04-03 2017-05-30 Sonix Technology Co., Ltd. Information transmission system and method
US10901512B1 (en) 2015-05-29 2021-01-26 Google Llc Techniques for simulated physical interaction between users via their mobile computing devices
US10372212B2 (en) 2015-05-29 2019-08-06 Google Llc Techniques for simulated physical interaction between users via their mobile computing devices
US10466119B2 (en) 2015-06-10 2019-11-05 Nextinput, Inc. Ruggedized wafer level MEMS force sensor with a tolerance trench
US11182068B2 (en) * 2015-10-27 2021-11-23 Verizon Patent And Licensing Inc. Method and system for interacting with a touch screen
US20170115867A1 (en) * 2015-10-27 2017-04-27 Yahoo! Inc. Method and system for interacting with a touch screen
US11063464B2 (en) 2016-07-07 2021-07-13 Samsung Electronics Co., Ltd. Apparatus and method for altering wireless charging mode
US10620909B2 (en) 2016-11-04 2020-04-14 International Business Machines Corporation Dynamic selection for touch sensor
US10379806B2 (en) 2016-11-04 2019-08-13 International Business Machines Corporation Dynamic selection for touch sensor
WO2018148503A1 (en) * 2017-02-09 2018-08-16 Nextinput, Inc. Integrated digital force sensors and related methods of manufacture
US11946817B2 (en) 2017-02-09 2024-04-02 DecaWave, Ltd. Integrated digital force sensors and related methods of manufacture
US11604104B2 (en) 2017-02-09 2023-03-14 Qorvo Us, Inc. Integrated piezoresistive and piezoelectric fusion force sensor
US11243125B2 (en) 2017-02-09 2022-02-08 Nextinput, Inc. Integrated piezoresistive and piezoelectric fusion force sensor
US11808644B2 (en) 2017-02-09 2023-11-07 Qorvo Us, Inc. Integrated piezoresistive and piezoelectric fusion force sensor
US11255737B2 (en) 2017-02-09 2022-02-22 Nextinput, Inc. Integrated digital force sensors and related methods of manufacture
US11221263B2 (en) 2017-07-19 2022-01-11 Nextinput, Inc. Microelectromechanical force sensor having a strain transfer layer arranged on the sensor die
US11423686B2 (en) 2017-07-25 2022-08-23 Qorvo Us, Inc. Integrated fingerprint and force sensor
US11946816B2 (en) 2017-07-27 2024-04-02 Nextinput, Inc. Wafer bonded piezoresistive and piezoelectric force sensor and related methods of manufacture
US11243126B2 (en) 2017-07-27 2022-02-08 Nextinput, Inc. Wafer bonded piezoresistive and piezoelectric force sensor and related methods of manufacture
US11609131B2 (en) 2017-07-27 2023-03-21 Qorvo Us, Inc. Wafer bonded piezoresistive and piezoelectric force sensor and related methods of manufacture
US10776763B2 (en) 2017-09-12 2020-09-15 Paypal, Inc. System and method for performing gesture-based unified multi-marketplace transactions
US11875318B2 (en) 2017-09-12 2024-01-16 Paypal, Inc. Initiating mobile transactions via gestures
US11282047B2 (en) 2017-09-12 2022-03-22 Paypal, Inc. System and method for performing gesture-based unified multi-marketplace transactions
US11898918B2 (en) 2017-10-17 2024-02-13 Nextinput, Inc. Temperature coefficient of offset compensation for force sensor and strain gauge
US11579028B2 (en) 2017-10-17 2023-02-14 Nextinput, Inc. Temperature coefficient of offset compensation for force sensor and strain gauge
US11385108B2 (en) 2017-11-02 2022-07-12 Nextinput, Inc. Sealed force sensor with etch stop layer
US11874185B2 (en) 2017-11-16 2024-01-16 Nextinput, Inc. Force attenuator for force sensor
US11698310B2 (en) 2019-01-10 2023-07-11 Nextinput, Inc. Slotted MEMS force sensor
US10962427B2 (en) 2019-01-10 2021-03-30 Nextinput, Inc. Slotted MEMS force sensor
US11481050B1 (en) * 2021-07-30 2022-10-25 Primax Electronics Ltd. Function parameter switching memory circuit for peripheral device and operating method thereof

Also Published As

Publication number Publication date
KR20150011015A (en) 2015-01-29
KR101604565B1 (en) 2016-03-17
KR101492761B1 (en) 2015-02-12
KR101354316B1 (en) 2014-01-22
HK1249950A1 (en) 2018-11-16
CN103336562A (en) 2013-10-02
JP5922481B2 (en) 2016-05-24
AU2022201036B2 (en) 2024-02-22
KR20130019018A (en) 2013-02-25
JP5550211B2 (en) 2014-07-16
HK1223696A1 (en) 2017-08-04
US20150153895A1 (en) 2015-06-04
WO2006094308A2 (en) 2006-09-08
US20160070399A1 (en) 2016-03-10
CN101432677A (en) 2009-05-13
KR20220020995A (en) 2022-02-21
KR102095691B1 (en) 2020-03-31
HK1199767A1 (en) 2015-07-17
AU2007100828B4 (en) 2010-04-22
HK1257711A1 (en) 2019-10-25
HK1219318A1 (en) 2017-03-31
AU2007101054B4 (en) 2010-04-22
JP2020038706A (en) 2020-03-12
HK1122460A2 (en) 2009-05-15
JP2014042334A (en) 2014-03-06
KR101257964B1 (en) 2013-04-30
AU2007100828A5 (en) 2007-09-20
AU2022201036A1 (en) 2022-03-10
KR20160150116A (en) 2016-12-28
KR102358378B1 (en) 2022-02-08
KR20190141268A (en) 2019-12-23
AU2006218381B2 (en) 2012-01-12
KR102246065B1 (en) 2021-04-29
JP2018063720A (en) 2018-04-19
DE202006020369U1 (en) 2008-05-21
KR102058832B1 (en) 2019-12-23
KR20160104094A (en) 2016-09-02
JP2015133722A (en) 2015-07-23
JP2012142033A (en) 2012-07-26
KR101651975B1 (en) 2016-08-29
KR101984833B1 (en) 2019-06-03
KR20190061099A (en) 2019-06-04
KR102305019B1 (en) 2021-09-27
KR20210118964A (en) 2021-10-01
US20180217709A1 (en) 2018-08-02
CN101432677B (en) 2012-10-10
KR20150069036A (en) 2015-06-22
AU2006218381A1 (en) 2006-09-08
JP6444758B2 (en) 2018-12-26
KR20190109583A (en) 2019-09-25
KR20180014842A (en) 2018-02-09
EP1877889A2 (en) 2008-01-16
KR20070110114A (en) 2007-11-15
KR102402377B1 (en) 2022-05-25
HK1122461A2 (en) 2009-05-15
AU2007101054A4 (en) 2007-12-13
KR20210048583A (en) 2021-05-03
EP4177708A1 (en) 2023-05-10
KR20130108470A (en) 2013-10-02
CA2600326A1 (en) 2006-09-08
DE212006000028U1 (en) 2007-12-20
AU2006218381B8 (en) 2012-02-16
JP2008537615A (en) 2008-09-18
KR101690164B1 (en) 2016-12-27
JP2022009398A (en) 2022-01-14
KR20200035472A (en) 2020-04-03
JP2017162482A (en) 2017-09-14
WO2006094308A3 (en) 2006-12-14

Similar Documents

Publication Publication Date Title
US20220187870A1 (en) Multi-functional hand-held device
AU2022201036B2 (en) Multi-functional hand-held device
EP3301540B1 (en) Multi-functional hand-held device
AU2019261790B2 (en) Multi-functional hand-held device
AU2012202122B2 (en) Multi-functional hand-held device

Legal Events

Date Code Title Description
AS Assignment

Owner name: APPLE INC., CALIFORNIA

Free format text: CHANGE OF NAME;ASSIGNOR:APPLE COMPUTER, INC.;REEL/FRAME:020345/0285

Effective date: 20070109

Owner name: APPLE COMPUTER, INC., CALIFORNIA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:HOTELLING, STEVE P.;REEL/FRAME:020345/0258

Effective date: 20060505

STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION