WO2008132724A1 - A method and apparatus for three dimensional interaction with autosteroscopic displays - Google Patents

A method and apparatus for three dimensional interaction with autosteroscopic displays Download PDF

Info

Publication number
WO2008132724A1
WO2008132724A1 PCT/IL2008/000530 IL2008000530W WO2008132724A1 WO 2008132724 A1 WO2008132724 A1 WO 2008132724A1 IL 2008000530 W IL2008000530 W IL 2008000530W WO 2008132724 A1 WO2008132724 A1 WO 2008132724A1
Authority
WO
WIPO (PCT)
Prior art keywords
viewer
virtual
display
data
perceived
Prior art date
Application number
PCT/IL2008/000530
Other languages
French (fr)
Other versions
WO2008132724A4 (en
Inventor
Eyal Gordon
Gur Arie Bittan
Original Assignee
Mantisvision Ltd.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Mantisvision Ltd. filed Critical Mantisvision Ltd.
Publication of WO2008132724A1 publication Critical patent/WO2008132724A1/en
Publication of WO2008132724A4 publication Critical patent/WO2008132724A4/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/011Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
    • G06F3/013Eye tracking input arrangements
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/017Gesture based interaction, e.g. based on a set of recognized hand gestures
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/03Arrangements for converting the position or the displacement of a member into a coded form
    • G06F3/0304Detection arrangements using opto-electronic means
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/30Image reproducers
    • H04N13/302Image reproducers for viewing without the aid of special glasses, i.e. using autostereoscopic displays
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/30Image reproducers
    • H04N13/366Image reproducers using viewer tracking

Definitions

  • the present invention relates to an apparatus and a method for three dimensional interaction between a viewer and a virtual scene displayed on an autostereoscopic display, and more particularly but not exclusively, to the touching of and/or control of objects in the autostereoscopic scene.
  • Stereoscopic systems have gained in popularity in recent years as the capabilities of computer processing power have grown enormously along with advances in three dimensional screen technologies.
  • Autostereoscopic (AS) display technology allows the viewer to experience the sensation that 3D objects are floating in front of him without the use of any visual aids.
  • the area in front of the display is divided into viewing zones, wherein each zone is typically several centimeters wide.
  • Fig 1 shows a typical AS display along with the viewing zones, where zones numbered 1-6 show the optimal viewing distance from the screen, called the eyebox.
  • a viewer's position in front of the display is defined by any two adjacent zones, each eye falling in a separate viewing zone.
  • Each eye in each zone perceives a different image from the AS display, and the two images together give the viewer the 3D sensation, called stereo parallax.
  • Fig I 5 the viewer may experience limited motion parallax as well as stereo parallax.
  • Motion parallax gives the viewer the sensation of actually seeing a different angle of the object as he moves his position in relation to the object. That is to say, as the viewer moves laterally in front of the screen, he views different pairs of images for each pair of adjacent viewing zones. As a result, he may perceive the virtual object displayed on the AS display at various angles as he moves laterally in front of the screen.
  • Figs 2A and 2B we refer to Figs 2A and 2B. In Fig 2A, the viewer's two eyes are found in viewing zones 1 and 2 in relation to AS display 14.
  • each of the viewer's eyes will see a mixture of three different images, and the viewing experience is hampered. The same phenomenon will occur if he steps farther away from the screen, outside of the eyebox diamonds numbered 1-6.
  • the viewing zones 1-6 are therefore the optimal viewing zones from the AS display in which the viewer perceives clear stereoscopic images. Although only x-direction division of the screen resolution is illustrated for simplicity in the figures, viewing zones may be allocated according to both x and y direction. As such, content perceived by a viewer may change in accordance with both horizontal and vertical position in relation to the display.
  • the multiple viewing zones are a result of display optics, typically achieved by use of lenticular lenses or parallex barriers over the flat panel display screen.
  • the zones are necessary to allow for perception of a different image in each eye, the drawback is the significant decrease in overall screen resolution.
  • the resolution for any one view is 1/6 of the overall screen resolution.
  • One approach to compensate for the decrease in screen resolution is the use of adjustable optical filters. Through a tracking mechanism that determines the viewer's eyes position in relation to the display, the optical filters adjust the location of the viewing zones in accordance with the viewer's eyes location. Then, only the appropriate two images for the viewer's current position are displayed. In such a manner, the screen resolution may be divided into two zones and the decrease in resolution caused by multiple zones may be significantly decreased. Such a method is described in US Patent No. 6,075,557.
  • Autostereoscopic displays ultimately provide a more sensational viewing experience, as the viewer has the sensation that the objects in the displayed scene are floating in front of his eyes without the use of any visual aid or tracking device.
  • a virtual hologram experience may give the viewer the perception of 3D into the screen as well as the sensation that objects are floating in front of the screen.
  • the perceived object has a set of coordinates that are unique to the viewing zone in which the viewer's eyes are found and the viewer's position within the viewing zone.
  • known devices for interacting with displayed technology whether 2D or 3D, generally require cumbersome 'user control' tracking device(s) either held by the viewer or mounted to one or more locations of the viewer's hand.
  • Such systems may include equipment that the viewer must wear, such as gloves and helmets having tracking devices thereupon.
  • an apparatus for providing an interactive human computer interface to a viewer comprising: a) a data storage configured to store three-dimensional virtual environment representation data including at least one three-dimensional virtual object within the virtual environment; b) an autostereoscopic (AS) display configured to display simultaneous perspectives of virtual objects of the 3D virtual environment representation data to spatially separated viewing zones located in front of the AS display c) an anatomy tracking system including at least one 3D camera, the anatomy tracking system operative to determine respective real-world locations of: i) a viewer's eyes; and ii) an anatomical part on the viewer's body d) a 3D registration engine configured to generate a 3D volume comprising 3D location data of the viewer's anatomical part and 3D location data of the virtual objects perceived by the viewer in accordance with the viewer's eye location, e) an anatomical part-virtual object relation computation engine operative to determine a relation between the virtual object and the anatomical part in accordance with output of the registration
  • an apparatus for providing an interactive human computer interface to a viewer comprising: a) a data storage configured to store three-dimensional virtual environment representation data including at least one three-dimensional virtual object within the virtual environment; b) an autostereoscopic (AS) display configured to display simultaneous perspectives of virtual objects of the 3D virtual environment representation data to spatially separated viewing zones located in front of the AS display c) an anatomy tracking system including at least one 3D camera, the anatomy tracking system operative to determine respective real-world locations of: i) a viewer's eyes; and ii) an anatomical part on the viewer's body d) an anatomical part- virtual object relation computation engine operative to determine a relation between a perceived virtual object by the viewer and the anatomical part e) a rule enforcement engine operative to modify the three-dimensional environment representation data in accordance with the determined anatomical part- virtual object relation and the virtual environment data.
  • AS autostereoscopic
  • an autostereoscopic (AS) display configured to display simultaneous perspectives of virtual objects of the 3D virtual environment representation data to spatially separated viewing zones located in front of the AS display, the AS display having an acoustic lens with electrically controlled refractive index to dynamically adapt the viewing zone locations.
  • AS autostereoscopic
  • a method for providing an interactive human computer interface to a viewer comprising: a) storing three-dimensional virtual environment representation data including at least one three-dimensional virtual object within the virtual environment; b) displaying on an autostereoscopic (AS) display simultaneous perspectives of virtual objects in the 3D virtual environment representation data to spatially separated viewing zones located in front of the AS display c) tracking the anatomy of a viewer to determine respective real- world locations of: i) a viewer's eyes; and ii) an anatomical part on the viewer's body d) generating a registered 3D volume comprising 3D location data of the viewer's anatomical part and 3D location data of the virtual objects perceived by the viewer in accordance with the viewer's eye location, e) determining a relation between the virtual object and the anatomical part in accordance with the registration, f) modifying based on interactive rules the three-dimensional environment representation data in accordance with the determined anatomical part- virtual object relation and the virtual environment data.
  • AS autostereoscopic
  • a system for interactive human computer interface comprising: a self-contained autostereoscopic (AS) display configured to render 3D virtual objects into neighboring viewing zones associated with the display, an eye location tracking system, comprising at least one 3D video camera, for continuously determining: 1) a viewer perceived three dimensional space in relation to the display, and
  • AS autostereoscopic
  • an anatomy location and configuration system comprising at least one 3D video camera, for continuously determining a 3D mapping of viewer anatomy in relation to the display, and an interactive application that defines interactive rules and displayed content to the user, and an interaction processing engine configured to receive information from 1) the eye location tracking system 2) the anatomy location and configuration system, and
  • selected steps of the invention could be implemented as a plurality of software instructions being executed by a computer using any suitable operating system.
  • selected stages of the method and system of the , invention could be described as being performed by a data processor, such as a computing platform for executing a plurality of instructions.
  • FIG. 1 is a simplified illustration of a multi view autosterescopic screen along with associated viewing zones.
  • FIG. 2A is simplified illustration showing the viewing zones associated with a particular viewer location in relation to an autostereoscopic display.
  • FIG. 2B is simplified illustration showing the viewing zones associated with a different viewer location in relation to an autosterescopic display.
  • FIG. 3 A is a simplified flow chart illustrating a generalized embodiment of the invention.
  • FIG 3 B is a simplified flow chart of the interaction processing engine.
  • FIG. 4A is a simplified illustration of the viewer interacting with a virtual object in one location in relation to the AS display according to preferred embodiments of the present invention.
  • FIG. 4B is a simplified illustration of the viewer interacting with a virtual object in another location in relation to the AS display according to preferred embodiments of the present invention.
  • FIG 4C is a simplified illustration of the processing engines in preferred embodiments.
  • FIG. 5 illustrates several visual examples of interaction between a real object and a virtual 3D object.
  • FIG. 6 is a preferred embodiment showing a three dimensional autosterescopic menu driven interface.
  • FIG. 7 shows a preferred embodiment wherein viewer classification rules enable the system to identify various characteristics of the viewer and even identification and association with a stored historical profile
  • FIG. 8 shows the generalized embodiment of Fig. 3 with the addition of a personal profile engine.
  • FIG. 9 is a preferred embodiment of the present invention illustrating gesture- based interaction with 3D objects on an AS display.
  • FIG. 10 illustrates exemplary gesture based interactions.
  • FIG. 11 is a simplified illustration of an interface with the virtual scene based on a scaled model of the user seen in the virtual space.
  • FIG. 12 is a illustrates a viewer or anatomical part of a viewer and the parallel
  • FIG. 13A is a simplified illustration showing the 3D camera field of view according to preferred embodiments.
  • FIG. 13B is a simplified illustration showing the process of eye tracking and position finding.
  • FIG. 13C is a simplified illustration showing the process of 3D skeleton modeling.
  • FIG.14 is a simplified diagram illustrating a lenticular lens based autosterescopic display used in preferred embodiments of the present invention.
  • FIG 15 is a simplified illustration of the perceived 3D space in relation to the pixels on the AS display in which the viewer may observe and interact with virtual objects and scenes.
  • FIG. 16 is a simplified illustration of the interaction zone showing the overlap between the perceived space and the camera field of view for an exemplary case.
  • FIG. 17 shows an alternative embodiment in which the viewer's eyes position data obtained from the 3D camera may also be input to a screen steering device.
  • FIG. 18 shows a network of 3D virtual touch human interface systems.
  • An apparatus and a method are now disclosed for three dimensional interaction between a viewer and virtual object(s) displayed on an autostereoscopic (AS) display.
  • AS autostereoscopic
  • the viewer is able to experience the sensation of touching and/or control of objects perceived on the autostereoscopic display solely through use of an anatomical part of the viewer interfacing with the display.
  • the present embodiments provide an immersive and three dimensional real time interactive environment through the use of one or more 3D cameras.
  • a virtual object(s) is displayed on an AS display and appears to the viewer in a defined region in front of the display determined by the viewer's eyes position in relation to the display, as explained above.
  • the viewer may manipulate the object in various ways and may interact with the object in real time. These manipulations are then reflected in subsequent frames rendered and perceived by the viewer in the region for virtual objects in front of the display defined by the user's eyes location.
  • the process creates an "immersive" interactive environment between the viewer and the virtual scene perceived on. the auto stereoscopic display.
  • a viewer may be identified according to his or her stored profile.
  • the interaction process can be suited towards that particular viewer's historical preference data and other unique characteristics associated with that viewer in particular.
  • System 200 contains anatomy tracking system 18, object interaction processing engine 42, virtual environment data module 32, and scene presentation engine 44.
  • Anatomy tracking system 18 is comprised of at least one 3D video camera 22 and a 2D/3D image processing unit 26.
  • the 3D camera(s) may be situated in relation to the display in a variety of different configurations.
  • the camera(s) 22 which can preferably acquire depth information in motion, allow for the derivation of a depth map of the viewer positioned at some location typically in front of the display.
  • This depth map is typically in the form of a three dimensional point cloud.
  • the point cloud may then be input to a face and eyes finding algorithm contained in the image processing unit.
  • the viewer's eyes location determines, as described above, the mapping of the perceived coordinates ([XJ], [yj, [Zj]) of the displayed virtual objects.
  • the viewer eyes location determines the field of view or perceived space of the virtual scene seen by the viewer.
  • the perceived virtual objects may be either static or dynamic.
  • Anatomy tracking system 18 is comprised of at least one 3D camera 22 and 2D/3D image processing unit 26. It is understood that multiple 3D cameras may be present.
  • the 3D cameras together with the image processing unit 26 provide 3D coordinates ([Xi], [Yi], [Zi]) of the viewer's anatomy and/or movement of the anatomy in the camera field of view, typically the hand, arm, and face. Again, this set of data is typically a depth map in the form of a three dimensional point cloud.
  • the 3D camera preferably is a 3D video camera capable of 3D motion capture, and as such, a point cloud for each frame in a captured video sequence is preferably acquired.
  • the series of point clouds may then be input to a human finding engine and skeleton registering unit contained in image processing unit 38.
  • skeleton registering units or 3D classifiers are known in the art and are not discussed herein.
  • the anatomy tracking system provides location data of the user's eyes.
  • the object interaction processing engine 42 accepts the output from the anatomy location tracking system as well as virtual environment data module 32.
  • This virtual environment data is typically stored on a storage device and represents three dimensional representations displayed to the user according to techniques known in the art relating to AS displays.
  • the anatomical part-virtual object registration engine 34 accepts and processes information from both the anatomy tracking system 18 and virtual environment data module 32. Information concerning the eyes location from the anatomy tracking system 18 together with the virtual environment data from module 32 enables the registration engine 34 to determine the perceived virtual object 3D coordinates for the viewer's eye position ([XJ], [Vj], [ZJ]). Additionally, the perceived AS display field of view as seen by the viewer according to his eyes location is determined (see Figs 4A-4B below).
  • the information from the anatomy tracking system 18 pertaining to movement of the viewer's anatomical parts such as his or her hand, arm, and/or head, enables a 3D mapping based on the acquired point cloud coordinates ([Xj], [Yj], [Zj]). These sets of data are registered to determine a single 3D volume by the registration engine.
  • This registered data set from the registration engine 32 is then sent to the anatomical part-virtual object relation computation engine 34.
  • the relation computation engine determines if a collision occurs between the anatomical part and the virtual objects seen by the viewer. This would indicate that the viewer has "touched” the perceived virtual object(s).
  • This information from relation computation engine 36 is then sent to the rule enforcement engine 38.
  • This engine receives interactive application rules from 30 contained in the virtual environment data module 32.
  • the interactive application rules module 30 contains all rules for determining subsequent AS display frames as a function of user input.
  • the rule enforcement engine determines what the next presented scene on the AS display should be. For instance, through registration between the data sets (registration engine 34) a collision may be detected (relation computation engine 36), in which case the next frame displayed on the AS display may show the perceived virtual object, as a result of the "virtual touch", moved to a different location (rule enforcement engine 38).
  • rule enforcement engine 38 may decide that the virtual object should disappear as a result of the "virtual touch". In still further embodiments, other resultant actions may occur to the virtual object as a result of interactivity between the viewer and the virtual object.
  • the rule enforcement engine output is thus the output of the interaction processing engine 42. This output is sent to scene presentation engine 44 to render the next multiple frame content to be displayed on the AS display.
  • Data storage of the virtual environment data may be implemented in any combination of volatile (for example, RAM, registers, etc) and/or non-volatile memory (for example, flash memory, magnetic medium such as disk, etc). Data storage may be provided in a single location and/or may be distributed among multiple locations, for example, in a local and/or wide-area computer network.
  • volatile for example, RAM, registers, etc
  • non-volatile memory for example, flash memory, magnetic medium such as disk, etc.
  • Data storage may be provided in a single location and/or may be distributed among multiple locations, for example, in a local and/or wide-area computer network.
  • the viewer's hand acts as a "gravitational" source on the virtual object.
  • various gestures made by the viewer and captured by the 3D camera or 3D camera array signify various applications on the virtual object.
  • registration between the 3D data sets need not be carried out, since the relative locations between the virtual object and viewer anatomy data sets is not taken into account.
  • the object interaction processing unit based on the detected 3D movement of the viewer's anatomy, determines the effect on the currently displayed scene.
  • the interaction may combine both "touch” and “gesture” applications.
  • the interaction processing engine only contains the relation computation engine 36 and the rules enforcement engine 38.
  • the above process thus allows the viewer to touch, manipulate, and move virtual objects with his hand, watch virtual 3D objects fly into the palm of his hand, or control virtual scenes through gestures of his hand that are captured by the 3D camera.
  • the 2D/3D camera system is preferably composed of a structured light active triangulation based imaging apparatus, such as detailed in Appendix 1.
  • Such imaging devices utilize triangulation methods to determine three dimensional coordinates of imaged real world objects.
  • any type of 3D camera is appropriate, including but not limited to so-called 'time of flight' cameras, 2D cameras utilizing stereo correspondence algorithms, and triangulation based 3D cameras.
  • the present embodiments preferably utilize 3D video cameras in order to capture 3D motion.
  • the 3D imaging device(s) may be situated in relation to the display in a variety of different configurations, capture depth information in motion, and additionally capture 2D texture data of the imaged scene.
  • the 3D video camera contains both sensing and projector elements, preferably in a single housing.
  • a pattern containing distinct features is projected onto imaged elements in the camera field of view.
  • the pattern features, once reflected from the imaged elements, are captured in an image on the camera sensor.
  • the location of the features in the image is correlated, through triangulation techniques, to 3D spatial coordinates on the imaged element.
  • the totality of the 3D spatial coordinates comprises a point cloud of 3D points, which give a geometric shape of the imaged elements.
  • the point cloud may further be processed to arrive at a 3D polygonal mesh.
  • the above process is carried out per frame for every frame of the video sequence, so that 3D modeling over time of imaged objects in both a static and dynamic scene may be implemented. If several cameras are used, then for each frame, triangulation may be carried out separately for each camera and additionally through stereoscopic comparison of the two images and triangulation of common points in both images. In such a case of multiple cameras, multiple point clouds are obtained per frame. These multiple point clouds are then registered to a unified point cloud set.
  • the system may include an autostereoscopic (AS) display 14 and an array of 3D cameras 22 optionally situated on the top side of the display.
  • AS autostereoscopic
  • 3D cameras 22 optionally situated on the top side of the display.
  • Other configurations of the 3D camera(s) are understood to be possible, and the particular configuration shown is merely for illustration purposes.
  • Shaded region 52 shows the perceived field of view in which the viewer 56 sees virtual objects given the location of his eyes in front of the AS display. That is to say, the viewer only perceives virtual objects and scenes that are within the shaded triangular region 52. This region is defined by light rays from adjacent pixel columns of the AS FLAT PANEL DISPLAY display that are perceived by each eye.
  • the viewer's eye coordinates are denoted (X a ), (Y a ), (Z a ) and (X b ), (Y b ), (Z b ) for the left and right eyes respectively.
  • Fig 15 further illustrates the perceived space in which virtual objects are seen by a viewer of an AS display.
  • the viewer eyes location leads to a mapping of the virtual object 60 coordinates in 3D space ([Xj 3 ], [yj 3 ], [Zj 3 ]) seen by the viewer at this position.
  • the 3D motion capture video cameras capture data from which the 3D coordinates ([X m ] 9 [Y m ] 5 [Z m ]) of the viewer's hand and arm 62 in the camera field of view may be derived.
  • a moving skeleton model is obtained of the viewer's anatomy part.
  • a matching or collision between coordinate sets ([Xj 3 ], [yi 3 ], [Zj 3 ]) and ([ ⁇ m] 9 [Y m ] 5 [Z m ]) in a given video frame indicates that the viewer is "touching and/or manipulating" and therefore interacting with the virtual object.
  • the viewer With reference to Fig 4B, the viewer is seen moving laterally to the left in relation to the AS display. As a result, the viewer's perceived field of view of virtual scenes 64 on the AS display is shifted in relation to Fig 4A. Furthermore, the virtual object 60 now has a different mapping of coordinates ([Xj 4 ], [yj4], [Zj 4 ]) than that of Fig 4A, that correspond to the viewer's new location in relation to the display. To interact with the virtual object in 3D space, the viewer now moves his hand and arm to a different location in 3D space within the perceived field of view of virtual objects.
  • the skeleton mapping in this case is now based on a different captured point cloud series, namely ([X n ], [Yn], [Z n ]). Again, when a collision occurs between the data representing the 3D virtual object as perceived by the viewer and the data representing the human anatomy movement, some form of interaction occurs.
  • Fig 4C is a simplified illustration of the processing engines in preferred embodiments.
  • 3D cameras 22 send to the anatomy tracking system 18 depth data from which to derive viewer eyes location data.
  • the depth data is used to derive the viewer anatomy location and movement.
  • the anatomy tracking unit sends data to the object interaction processing engine 42.
  • Engine 42 additionally receives information on the interactive application from virtual environment module 32.
  • the virtual coordinate mapping for virtual object 68 is determined based on the viewer eye location data.
  • the object interaction processing engine registers the sets of virtual coordinate data and the anatomy location data.
  • the next frame in the AS video sequence is based on the interactive rules processed in the interaction processing engine.
  • the virtual environment data module 32 preferably contains the displayed content for current frames and the interactive characteristics and functionality of the displayed program. For instance, the interactive application determines the types of hand movements the viewer may make when he touches the object and whether the displayed object moves, twists, rotates, disappears, the manner in which the object performs these actions, and any other interactive functions.
  • viewer 56 may grasp virtual coke bottle 68 having initial virtual coordinate mapping ([xj], [yi], [ZJ]) at tl and rotate his hand and arm 62 to simulate a pouring action.
  • the 3D video cameras 48 capture depth information of the moving hand and arm and send this data to the anatomy tracking system 18.
  • the system 18 then computes the point clouds and subsequent skeleton model of the hand and arm for each frame. For instance, in the figure, the hand and arm have point cloud coordinates ([X p ], [Y P ], [Z p ]) at tl.
  • This engine 18 outputs to the object interaction processing engine 42 a skeleton model showing a hand in a pouring motion.
  • the interaction engine after processing this information together with the interactive application rules, tells the scene rendering unit 44 to display the next virtual coke bottle in a tilted position. That is to say, the virtual coke bottle 68 in the next frame at t2 has a new mapping of perceived coordinates ([X j ], [y j ], [z j ]) as a result of the interaction.
  • Box 86 illustrates three different 3D coordinate mappings for three frames in which the virtual coke bottle 68 is displayed. In each successive frame, the perceived coordinates of the virtual coke bottle change in 3D space as the bottle appears to slowly tilt as if it is being poured by the viewer.
  • Box 72 illustrates three different 3D coordinate mappings of the user's hand and arm for three successive time intervals, tl- t3. It is understood that certain areas of the virtual coke bottle are occluded from the viewer's view when he grasps the bottle, since in essence he is occluding his vision of the AS display at those points in space. Furthermore, the configuration and numbers of processing units in the diagram is for illustration purposes only and is in no way limiting.
  • Fig 5 illustrates several visual examples of interaction between a real object, such as the viewer's hand, and a virtual 3D object displayed on an AS screen.
  • Floating virtual beach ball 86 is displayed in the 3D space perceived by the viewer in front of the AS display 90.
  • the viewer may poke the virtual ball with his hand 94.
  • This poking action is detected by the camera(s) and processing devices of the anatomy location and configuration tracking system.
  • a 3D mapping of the viewer's hand is represented by mapping ([X j ], [Yj], [Z j ]).
  • the object interaction processing engine determines the resultant scene to be displayed based on the virtual environment data and interaction rules as a result of the poking action.
  • the virtual beach ball is displayed in a new 3D location in space in the following frame displayed to the viewer. That is to say, frame 1 has virtual object coordinate mapping ([x m ], [y m ], [z m ]) and frame 2 has mapping ([x n ], [ ⁇ n ], [z n ]), closer to the AS display. Therefore, the displayed frame 2 on the AS display is a result of interaction between the viewer's hand and the virtual beach ball. It is understood that other actions, such as swinging at the virtual ball and reactions to the above described interaction are possible.
  • the interactive application rule may dictate that the ball disappear, pop out of sight, fade out of sight, or any other desired interactive functionality as a result of the touch interaction or collision between the two sets of 3D coordinates.
  • FIG 5 Another interactive example shown in Fig 5 is virtual 3D bird 98 perching in the palm of viewer hand 102.
  • the viewer's hand is typically modeled according to a skeleton modeling algorithm utilizing the point cloud coordinates derived from information captured by the 3D camera.
  • skeleton modeling algorithms are seen for example in "Markerless Human Motion Transfer” by Cheung et al., "Image Based Markerless 3D Human Motion Capture using Multiple Cues" by Azad, et al., “Markerless Motion Capture from Single or Multi Camera Video Sequence” by Remondino et al., and others.
  • the interactive processing engine unit based on the interactive application rules and virtual environment data content, recognizes the skeleton model of an outstretched viewer's palm and sends the scene rendering unit 44 the next set of frames so that the virtual bird flies to and then rests on the viewer's hand.
  • the mapping of coordinates of the virtual bird for each frame approaches a collision with the coordinates for the skeleton model of the outstretched hand.
  • the viewer may attempt to catch a flying virtual bird 106 that appears in the virtual space that he perceives according to his position in relation to the screen.
  • Other examples in the figure show the viewer using his hands to interact with a virtual money bill 108 and a virtual leaf 110.
  • the collision of a set of 3D coordinates representing the human anatomy part with a set of 3D coordinates representing a displayed virtual object indicate touch and/or manipulation of virtual objects.
  • the interactive application rules together with the input of these two sets of coordinates determine the resultant next frame to be displayed.
  • the virtual object may be a direct input device whose interaction rules are derived from existing I/O devices such as a virtual 2D touch screen, multi-touch screen, keyboard, or other known input devices.
  • Collision detection methods for interactive environments are known in the art and thus are not discussed herein. The reader is referred to "Collision Detection In 3D Interactive Environments" by van den Bergen as an exemplary way of performing 3D collision detection.
  • Fig 6 is a preferred embodiment showing a three dimensional autosterescopic menu driven interface.
  • the floating balls or nodes 112 appear to the viewer in the 3D space perceived by him in which he may view virtual 3D objects.
  • the floating balls comprise a hierarchical menu.
  • the first stage shows 4 nodes that represent the operating system desired by the viewer.
  • a second tier menu appears to him.
  • the second tier represents the desired application software and the third tier the functionality therein.
  • the viewer chooses Windows® and Outlook® and then
  • FIG. 7 shows a preferred embodiment wherein viewer classification rules enable the system to identify various characteristics of the viewer and even identification and association with a stored historical profile.
  • a viewer imaging device 250 images the viewer and sends the image to the viewer 3D processing engine 252.
  • the 3D processing engine is controlled by the viewer classification rules module 254.
  • the viewer classification rules module 254 preferably defines the characteristics of the viewer that are to be determined based on the interactive application 256. For instance, the system may want to determine if a viewer is wearing a watch and then offer various products and services associated with various watch types. The system may want to determine the height of the person or his clothing size to present interactive content that is appropriate.
  • an on-line viewer profile is optionally obtained through an on-line viewer database 258 containing the historical profile of viewers.
  • the processing of the viewer may be iterative until an appropriate model is obtained, at which point the 3D modeling data is sent to the interaction engine 260.
  • the interaction engine receives the interactive application information 256 as well.
  • the interaction engine decides upon the 3D displayed content on the display 262 based on the viewer classification and/or identification and the interactive application.
  • buttons and other virtual objects such as seen in Fig. 7 can be adapted to display to the user a probable selection based on his historical profile. For instance, if he is known to pick Windows often, then in stage 1 he may see the 3D Windows icon enlarged for him for ease of use.
  • the 3D camera image and skeleton model enable the viewer to appear as a mirror image to himself.
  • the displayed AS content is a mirror virtual image of the viewer himself.
  • This image may utilize the viewers shape and/or motion in the form of an avatar or an actual image of the viewer.
  • Such an image can be used to display to the user how he or she would look with various hairstyles, clothing, etc.
  • this avatar can be transmitted to from one user to the next for user interaction between them.
  • Fig 8 is an illustration parallel to Fig. 3 above showing the additional personal profile engine.
  • the personal profile contains 3D camera 272, processing engine unit based on rules 274 and on-line database 276. It is understood that the 3D cameras of the anatomy tracking system may be used for the personal profile engine.
  • the personal profile engine is based on the viewer classification rules that typically call for an iterative analysis process to determine desired characteristics of the viewer in front of the screen. In order to carry out this process, the personal profile engine typically takes input from the interactive application. The application may call for the identification and presentation of content to children that is different than content to adults.
  • the interactive engine 42 receives information from the interactive application rules module as well. The interactive engine 42 then processes all four inputs to arrive at 3D content to be displayed.
  • Fig 9 is a preferred embodiment of the present invention illustrating gesture-based interaction with 3D objects on an AS display.
  • the viewer gestures typically with his arm and hand to manipulate a 3D virtual object perceived by him in the region 114 in which he may view virtual objects.
  • Various motions of the hand and arm signify interactive functions with displayed 3D virtual objects.
  • User hand and arm 116 is shown making a pulling gesture to indicate the viewer's desire to pull the virtual beach ball 118 towards him.
  • the viewer's hand and arm have real object mapping ([X k ], [Y k ], [Z k ]) in the form of a three dimensional point cloud for each frame. This point cloud is processed by image processing units as described above and the skeleton model of the moving anatomy is determined over time.
  • the virtual beach ball has virtual object mapping ([xj 5 ], [yis], [Zj 5 ]) for frame 1 at time 1.
  • the interaction engine identifies the pulling motion from the skeleton model. Moreover, the interaction engine rules may determine that the pulling motion of the hand cause the virtual object to move in the next displayed frame closer to the 3D location of the viewer's hand.
  • the precise eye location of the viewer may not be necessary for determination of the precise virtual object mapping scene by the viewer, since no actual collision algorithms are used to determine virtual touch.
  • Fig 10 shows a variety of other gesture-based interactive functions. All capital coordinate sets denoted ([X], [Y], [Z]) indicate real objects, while all lower case coordinate sets denoted ([x], [y], [z]) indicate virtual objects.
  • the viewer may point his finger at the ball, shown as action 120, push the ball, shown as action 122, or pull the ball, shown as action 124.
  • the next displayed frame is a function of the interaction application.
  • the interactive application may be a "magnetic" effect, whereby the closer one is to the virtual object the stronger and faster the action.
  • the ball may pop, disappear, be dragged to and away from the viewer, among other functions.
  • the viewer may also hold virtual objects in his hand such as virtual wrench 126 to manipulate other virtual objects in the viewer perceived space.
  • virtual wrench 126 can be twisted and turned to tighten or loosen virtual screw 128. It is understood that under certain hand positions, the viewer's hand may occlude the virtual tool from appearing to the viewer, and the use of virtual tools may need to be modified to allow viewer visibility.
  • Menu driven interface shown in Fig 6 may additionally be operated in gesture mode wherein the viewer merely gestures to choose the appropriate virtual ball.
  • gesture interactive motions 130 such as thumbs up, thumbs down, finger and hand movements are shown. Gestures may be earned out by a viewer to manipulate objects or send interactive commands without looking at his hands so long as they are within the camera(s) field of view. For instance, a tap of the viewer's left hand downwards may signify a lowering of the volume on the AS display, while upwards may cause an increase in display volume. Such an interaction is similar to touch typing on a computer keyboard, where the viewer is interacting with the computer while not moving his eyes from the screen.
  • This scaled version 134 is typically miniature in size and sometimes referred to herein as the "virtual viewer” or "ghost".
  • the 3D camera captures either the viewer 138 or an anatomical part of the viewer and, upon image processing of the captured real coordinates ([XJ, [YJ, [ZJ), a scaled or miniature and preferably semi-transparent version of the viewer and his movements is displayed in the perceived space 142.
  • the moving "ghost” has perceived virtual object coordinates ([Xi 7 ], [Vj 7 ], [zj 7 ]). Additionally, virtual object 146 with 3D perceived coordinates ([Xj 8 ], [yi 8 ], [zis]) is displayed in the viewer perceived space. Interaction may now occur between the "virtual viewer” and the virtual object.
  • the interaction engine preferably receives inputs from the anatomy tracking system and determines the interactive functionality. The interaction engine then sends commands to the scene rendering unit.
  • Such a “ghost” may be utilized in PDAs or other portable devices with a small
  • the miniature version of the viewer may be his entire body or just his hand, or even a three dimensional puppet or ghost-like image that does not resemble the viewer but moves in accordance with the viewer's body movements.
  • the miniature can be in a non proportional scale and may move faster and slower in accordance with the interactive settings. That is to say, the sensitivity to movement of the virtual miniature version of the viewer may depend on speed of the viewer's movement, similar to a computer mouse.
  • the present embodiment can also be implemented as a seamless combination of both touch interaction, as described above, and the "ghost" interaction.
  • the viewer's anatomy is within the space in which he is capable of viewing virtual objects, the viewer may virtually touch perceived virtual objects in the virtual scene.
  • a seamless transition can occur where the viewer sees the "ghost" interface at a certain distance from the display.
  • FIG 12 illustrates a viewer or anatomical part of a viewer and the parallel "virtual viewer” or “virtual anatomical part” being displayed in the viewer perceived space.
  • Viewer 150 sees a "ghost” 154 of himself in miniature form in the perceived virtual object space in front of the display 156.
  • the "ghost” appears to the viewer with a virtual “dashboard” 158 on his back that indicates to the viewer various states and modes of interaction, similar to CAPS LOCK on a standard computer keyboard.
  • Viewer hand 160 is mapped into the perceived virtual object space and may take the form of various shaped virtual hands such as 162 and 164.
  • the virtual hand may additionally manipulate scaled virtual objects such as hammer 166 and beach ball 168.
  • Fig 13A-13C illustrates the 3D camera and post processing operations as described in the above embodiments.
  • Fig 13 A illustrates the 3D camera and post processing operations as described in the above embodiments.
  • Only one 3D camera is shown in Fig 13 A with field of view 170.
  • the camera is comprised of a two dimensional camera 172 with vertically attached projector 174. Additional projectors and various configurations of projectors in relation to the AS display may be implemented as well in alternate embodiments.
  • the configuration of projectors for the current embodiment is for illustration purposes only.
  • the projections may be separated, for example, by different wavelengths, polarities, or time multiplexing.
  • the camera is typically positioned in a suitable position in front of the viewer and/or attached to the AS display 176.
  • various adjustments can be made to the camera and projector lenses, angles, and positions, aperture settings, as well as the number of cameras and/or the number of projectors to better suit the desired application.
  • the projector is placed vertically to the camera purely for ease of illustration.
  • the camera captures the viewer's depth map 1773 in the form of a 3D point cloud.
  • the depth map provides the point cloud to a face finding engine or 3D classifier 180.
  • 3D classifiers are known in the art, such as described in "Automatic Classification of Objects in 3D Laser Range Scan" by Nuchter et.al..
  • the face finding engine then is used to determine precise eyes location 182 and optionally the direction in space at which the viewer is focusing. This optional element is denoted as eye direction tracker 184. Eye direction tracking may be used to determine the position in space at which the viewer is looking at a given moment in time.
  • the eye position detection is denoted by dashed lines 178 in Fig 13 A.
  • the eyes finding engine preferably operates on a continuous basis to determine the viewer perceived 3D space as the viewer, and thus his eyes location, moves position in front of the AS display. Furthermore, the camera captures the viewer's body or parts thereof in three dimensional positional coordinates, and typically at least the viewer's hand and arm 186 and face 188.
  • Fig 13C shows a typical flow diagram of skeleton modeling based on depth capture.
  • the sampled depth data 190 in addition to optional texture data 192, is input to a viewer finding engine 194 and from there to one or more post processing units to determine his body contour from which a skeleton model is registered 196. As stated, the depth data may be combined with textural information captured simultaneously. Texture may be captured through use of an additional camera or CCD.
  • the 3D data provide dynamic video images, meaning through 3D motion capture methods the 3D positional data of the viewer's anatomy part is known at any given point in time.
  • the 3D skeleton model may be captured from various types of 3D cameras, such as stereo correspondence, time of flight or triangulation based cameras. As discussed above, the 3D skeleton model of the viewer's anatomical part is used to determine interaction with the virtual scene.
  • Fig 14 shows the basic characteristics and functionality of an exemplary automultiscopic/stereoscopic display as used in the previously described embodiments.
  • AS displays may have differing characteristics, such as screen width, number of views, and optimal viewing distance, but the display principles typically behind such displays is the same- the division of the overall resolution of the screen to a defined number of "viewing zones" through use of some type of optical filter.
  • Display 200 has lenticular lenses 204 placed over a flat panel screen 206 divided into a series of two views 210 and 214. Each pair of views on the flat panel display is covered by a separate half cylindrical shaped lenticular lens. Each view is made up of a pixel column on the fiat panel display and each lenticular lens projects both views simultaneously to a viewer.
  • the optical characteristics of the lenticular lens are responsible for the fact that at varying positions in front of such a screen, each eye 216 of a viewer sees one of the two views. By viewing two different images simultaneously, the viewer experiences the three dimensional effect called stereo parallax.
  • each zone may display to a viewer therein different content.
  • Fig 15 is a simplified illustration of the perceived 3D space in relation to the AS display in which the viewer may observe and interact with virtual objects and scenes.
  • the viewer 218 position in relation to the display 220 determines which pixel columns each eye views.
  • the display is typically divided into a repeating series of pixel columns in accordance with the number of viewing zones. In the figure, the display is divided into a series of 4 columns corresponding to 4 possible viewing zones. Under each lenticular lens a group of 4 columns exists. Example columns 222 and 224 are shown, and these pixel columns display images seen in the number 2 and 3 viewing zones (not shown).
  • columns 222 project light rays 226 and 228 into the right and left eye respectively of viewer 218 whose eyes lie in zones 2 and 3 (not shown).
  • the entire 3D space in which the viewer may perceive 3D objects from that location with respect to the display is denoted by shaded region 230.
  • Region 230 is the 3D space in front of the screen in which both eyes view light rays from pixels on the screen. It is understood that the light projections shown from the AS display are for illustration purposes.
  • Fig 16 is a simplified illustration showing the case where the camera field of view does not cover all of the viewer perceived space in front of the AS display.
  • area 232 represents what may be termed an interaction zone.
  • the interaction zone represents the overlap between the 3D camera field of view and the space within which the viewer interacts and sees virtual objects.
  • interactivity is determined by the registration between the 3D coordinates of the virtual scene (displayed object(s)) perceived by the viewer and the 3D positional data of the viewer's anatomical part. It is seen that should the viewer hand be outside the camera's field of view but still within the region in which the viewer sees virtual objects, region 234, then even if the viewer perceives the virtual touch the camera does not.
  • the interaction zone is thus the totality of 3D points in space in which virtual objects are perceived by the viewer and that is also captured by the camera. It is understood that various layouts and numbers of cameras and projectors may provide for interaction zones of varying volumes.
  • Fig 17, show an alternative embodiment in which the viewer's eyes position data obtained from the 3D camera may also be input to a screen steering device 236. In such an embodiment, the screen display's angle and location adjusts to the viewer's precise location through use of a moving optical filter that alters the optical viewing space.
  • the screen then needs only to render the two appropriate views to the viewer's location, and the decrease in screen resolution due to the partition into repeating series of pixel columns for each view is obviated.
  • Such a display device is described for instance in US Patent No. 6,075,557.
  • the exact positioning of the individual viewing zones, seen in Fig 1, can be refined and tuned to the precise viewer location rather than being in a fixed cubic area in front of the AS display.
  • a slight rotation and/or translation of the lenticular lens over the flat panel display allows for adjustment of the viewing zones.
  • the occlusion of light is determined based on the user's position so as to maximize the zone locations.
  • Various lens types are discussed in "A Survey of 3DTV Displays: Techniques and Technologies" by Benzie, P, et.al. Circuits and Systems for Video Technology, IEEE Transactions on Volume 17, Issue 11, Nov. 2007 Page(s):1647 - 1658.
  • acoustic optics may be used to control the refractive index. Such an acoustic optical covering may then be used to dynamically adjust the viewing zones.
  • Fig 18 shows a network of 3D virtual touch human interface systems as described herein. Each system is denoted as 200 and has the characteristics described in Fig 3 A above.
  • a network of such systems may allow for a multi-user interface wherein one or more users interact with each other based on content created at each viewer's individual station.
  • the scene presentation engine outputs from each system and is sent to other stations over the network.
  • each user hits the ball in his 3D volume in front of his AS display.
  • the resultant virtual environment information is then sent to the second user over the network.
  • the second user sees the virtual ball in front of his AS display as a result of interaction of the first user with the ball.
  • a multi user game is possible.
  • a unified 3D volume exists of all user that registers the individual local registered environments into a single registered environment of all users.
  • Multiuser 3D video videoconference is another implementation using such a network of systems.

Abstract

A method and apparatus for interactive human computer interface using a self-contained single housing autostereoscopic (AS) display configured to render 3D virtual objects into fixed viewing zones. The system contains an eye location tracking system for continuously determining both a viewer perceived three dimensional space in relation to the zones and a 3D mapping of the rendered virtual objects in the perceived space in accordance with a viewer eyes position. Additionally, one or more 3D cameras determines anatomy location and configuration of the viewer in real time in relation to said display. An interactive application that defines interactive rules and displayed content to the viewer is present. Furthermore, an interaction processing engine receives information from the eye location tracking system, the anatomy location and configuration system, and the interactive application to determine interaction data of the viewer anatomy with the rendered virtual objects from the AS display.

Description

A METHOD AND APPARATUS FOR THREE DIMENSIONAL INTERACTION WITH AUTOSTEROSCOPIC DISPLAYS
FIELD AND BACKGROUND OF THE INVENTION
The present invention relates to an apparatus and a method for three dimensional interaction between a viewer and a virtual scene displayed on an autostereoscopic display, and more particularly but not exclusively, to the touching of and/or control of objects in the autostereoscopic scene. Stereoscopic systems have gained in popularity in recent years as the capabilities of computer processing power have grown enormously along with advances in three dimensional screen technologies. Autostereoscopic (AS) display technology allows the viewer to experience the sensation that 3D objects are floating in front of him without the use of any visual aids. As a result of display optics, the area in front of the display is divided into viewing zones, wherein each zone is typically several centimeters wide. Fig 1 shows a typical AS display along with the viewing zones, where zones numbered 1-6 show the optimal viewing distance from the screen, called the eyebox. A viewer's position in front of the display is defined by any two adjacent zones, each eye falling in a separate viewing zone. Each eye in each zone perceives a different image from the AS display, and the two images together give the viewer the 3D sensation, called stereo parallax.
Furthermore, with a multi-viewing-zone screen, as in Fig I5 the viewer may experience limited motion parallax as well as stereo parallax. Motion parallax gives the viewer the sensation of actually seeing a different angle of the object as he moves his position in relation to the object. That is to say, as the viewer moves laterally in front of the screen, he views different pairs of images for each pair of adjacent viewing zones. As a result, he may perceive the virtual object displayed on the AS display at various angles as he moves laterally in front of the screen. To illustrate, we refer to Figs 2A and 2B. In Fig 2A, the viewer's two eyes are found in viewing zones 1 and 2 in relation to AS display 14. As a result, he perceives virtual coke bottle 16 as having a particular mapping of virtual 3D coordinates. This mapping is denoted by ([xπ], [yπ], [Zj1]) in the figure and represents a set of 3D coordinates in space that comprise the virtual coke bottle. In Fig 2B, we see the viewer has moved and his eyes now are found in viewing zones 2 and 3. As a result, he perceives the virtual coke bottle at a different angle, giving him the feeling of motion parallax. Furthermore, the set of 3D coordinates in space that comprise the perceived virtual coke bottle has changed, and is now ([Xj2], [ya], [ZQ])- SO therefore, for each pair of viewing zones in front of the AS display, a mapping exists of perceived coordinates of the displayed virtual object(s). It is understood that the motion parallax described with reference to the viewing zones is often not seamless and the motion parallax only occurs when the viewer jumps zones. That is to say, movement within the same zones may not be seen as motion parallax in less sophisticated systems.
Referring back to Fig 1 , as the viewer approaches close to the screen, his left eye may be found in diamond 345 and his right eye in diamond 234, for instance. These zones represent mixtures of three images each. In such a case, each of the viewer's eyes will see a mixture of three different images, and the viewing experience is hampered. The same phenomenon will occur if he steps farther away from the screen, outside of the eyebox diamonds numbered 1-6. The viewing zones 1-6 are therefore the optimal viewing zones from the AS display in which the viewer perceives clear stereoscopic images. Although only x-direction division of the screen resolution is illustrated for simplicity in the figures, viewing zones may be allocated according to both x and y direction. As such, content perceived by a viewer may change in accordance with both horizontal and vertical position in relation to the display.
As stated, the multiple viewing zones are a result of display optics, typically achieved by use of lenticular lenses or parallex barriers over the flat panel display screen. Although the zones are necessary to allow for perception of a different image in each eye, the drawback is the significant decrease in overall screen resolution. For instance, in Fig 1, for a 6 viewing-zone display, the resolution for any one view is 1/6 of the overall screen resolution. One approach to compensate for the decrease in screen resolution is the use of adjustable optical filters. Through a tracking mechanism that determines the viewer's eyes position in relation to the display, the optical filters adjust the location of the viewing zones in accordance with the viewer's eyes location. Then, only the appropriate two images for the viewer's current position are displayed. In such a manner, the screen resolution may be divided into two zones and the decrease in resolution caused by multiple zones may be significantly decreased. Such a method is described in US Patent No. 6,075,557.
Autostereoscopic displays ultimately provide a more sensational viewing experience, as the viewer has the sensation that the objects in the displayed scene are floating in front of his eyes without the use of any visual aid or tracking device.
Additionally, a virtual hologram experience may give the viewer the perception of 3D into the screen as well as the sensation that objects are floating in front of the screen.
When the objects of a scene are indeed floating in front of one's eyes, there is a natural tendency to want to touch and otherwise interact with the virtual object. However, several problems exist that heretofore have not been solved. Firstly, as explained above, the perceived object has a set of coordinates that are unique to the viewing zone in which the viewer's eyes are found and the viewer's position within the viewing zone. Secondly, known devices for interacting with displayed technology, whether 2D or 3D, generally require cumbersome 'user control' tracking device(s) either held by the viewer or mounted to one or more locations of the viewer's hand. Such systems may include equipment that the viewer must wear, such as gloves and helmets having tracking devices thereupon. Unfortunately, these systems do not allow for ease-of-use, quick response, and simplicity so necessary in interactive environments, such as video games. Therefore, there is an unmet need for, and it would be highly useful to have an enhanced system and method to allow for interaction in 3D environments to overcome the above drawbacks. SUMMARY OF THE INVENTION
According to one aspect of the present invention there is provided an apparatus for providing an interactive human computer interface to a viewer, the apparatus comprising: a) a data storage configured to store three-dimensional virtual environment representation data including at least one three-dimensional virtual object within the virtual environment; b) an autostereoscopic (AS) display configured to display simultaneous perspectives of virtual objects of the 3D virtual environment representation data to spatially separated viewing zones located in front of the AS display c) an anatomy tracking system including at least one 3D camera, the anatomy tracking system operative to determine respective real-world locations of: i) a viewer's eyes; and ii) an anatomical part on the viewer's body d) a 3D registration engine configured to generate a 3D volume comprising 3D location data of the viewer's anatomical part and 3D location data of the virtual objects perceived by the viewer in accordance with the viewer's eye location, e) an anatomical part-virtual object relation computation engine operative to determine a relation between the virtual object and the anatomical part in accordance with output of the registration engine: f) a rule enforcement engine operative to modify the three-dimensional environment representation data in accordance with the determined anatomical part- virtual object relation and the virtual environment data.
According to another aspect of the present invention there is an apparatus for providing an interactive human computer interface to a viewer, the apparatus comprising: a) a data storage configured to store three-dimensional virtual environment representation data including at least one three-dimensional virtual object within the virtual environment; b) an autostereoscopic (AS) display configured to display simultaneous perspectives of virtual objects of the 3D virtual environment representation data to spatially separated viewing zones located in front of the AS display c) an anatomy tracking system including at least one 3D camera, the anatomy tracking system operative to determine respective real-world locations of: i) a viewer's eyes; and ii) an anatomical part on the viewer's body d) an anatomical part- virtual object relation computation engine operative to determine a relation between a perceived virtual object by the viewer and the anatomical part e) a rule enforcement engine operative to modify the three-dimensional environment representation data in accordance with the determined anatomical part- virtual object relation and the virtual environment data.
According to another aspect of the present invention there is an autostereoscopic (AS) display configured to display simultaneous perspectives of virtual objects of the 3D virtual environment representation data to spatially separated viewing zones located in front of the AS display, the AS display having an acoustic lens with electrically controlled refractive index to dynamically adapt the viewing zone locations. According to another aspect of the present invention there is provided a method for providing an interactive human computer interface to a viewer, the method comprising: a) storing three-dimensional virtual environment representation data including at least one three-dimensional virtual object within the virtual environment; b) displaying on an autostereoscopic (AS) display simultaneous perspectives of virtual objects in the 3D virtual environment representation data to spatially separated viewing zones located in front of the AS display c) tracking the anatomy of a viewer to determine respective real- world locations of: i) a viewer's eyes; and ii) an anatomical part on the viewer's body d) generating a registered 3D volume comprising 3D location data of the viewer's anatomical part and 3D location data of the virtual objects perceived by the viewer in accordance with the viewer's eye location, e) determining a relation between the virtual object and the anatomical part in accordance with the registration, f) modifying based on interactive rules the three-dimensional environment representation data in accordance with the determined anatomical part- virtual object relation and the virtual environment data. .
According to another aspect of the present invention there is provided a system for interactive human computer interface, the system comprising: a self-contained autostereoscopic (AS) display configured to render 3D virtual objects into neighboring viewing zones associated with the display, an eye location tracking system, comprising at least one 3D video camera, for continuously determining: 1) a viewer perceived three dimensional space in relation to the display, and
2) a 3D mapping of the rendered virtual objects in the perceived space in accordance with viewer eyes position in relation to the fixed viewing zones an anatomy location and configuration system, comprising at least one 3D video camera, for continuously determining a 3D mapping of viewer anatomy in relation to the display, and an interactive application that defines interactive rules and displayed content to the user, and an interaction processing engine configured to receive information from 1) the eye location tracking system 2) the anatomy location and configuration system, and
3) the interactive application thereby to determine interaction data of the viewer anatomy with the rendered virtual objects from AS display. Unless otherwise defined, all technical and scientific terms used herein have the same meaning as commonly understood by one of ordinary skill in the art to which this invention belongs. The materials, methods, and examples provided herein are illustrative only and not intended to be limiting. Implementation of the method and system of the present invention involves performing or completing certain selected tasks or steps manually, automatically, or a combination thereof. Moreover, according to actual instrumentation and equipment of preferred embodiments of the method and system of the present invention, several selected steps could be implemented by hardware or by software on any operating system of any firmware or a combination thereof. For example, as hardware, selected steps of the invention could be implemented as a chip or a circuit. As software, selected steps of the invention could be implemented as a plurality of software instructions being executed by a computer using any suitable operating system. In any case, selected stages of the method and system of the, invention could be described as being performed by a data processor, such as a computing platform for executing a plurality of instructions.
BRIEF DESCRIPTION OF THE DRAWINGS
The invention is herein described, by way of example only, with reference to the accompanying drawings. With specific reference now to the drawings in detail, it is stressed that the particulars shown are by way of example and for purposes of illustrative discussion of the preferred embodiments of the present invention only, and are presented in order to provide what is believed to be the most useful and readily understood description of the principles and conceptual aspects of the invention. In this regard, no attempt is made to show structural details of the invention in more detail than is necessary for a fundamental understanding of the invention, the description taken with the drawings making apparent to those skilled in the art how the several forms of the invention may be embodied in practice.
In the drawings:
FIG. 1 is a simplified illustration of a multi view autosterescopic screen along with associated viewing zones. FIG. 2A is simplified illustration showing the viewing zones associated with a particular viewer location in relation to an autostereoscopic display.
FIG. 2B is simplified illustration showing the viewing zones associated with a different viewer location in relation to an autosterescopic display. FIG. 3 A is a simplified flow chart illustrating a generalized embodiment of the invention.
FIG 3 B is a simplified flow chart of the interaction processing engine.
FIG. 4A is a simplified illustration of the viewer interacting with a virtual object in one location in relation to the AS display according to preferred embodiments of the present invention.
FIG. 4B is a simplified illustration of the viewer interacting with a virtual object in another location in relation to the AS display according to preferred embodiments of the present invention.
FIG 4C is a simplified illustration of the processing engines in preferred embodiments.
FIG. 5 illustrates several visual examples of interaction between a real object and a virtual 3D object.
FIG. 6 is a preferred embodiment showing a three dimensional autosterescopic menu driven interface. FIG. 7 shows a preferred embodiment wherein viewer classification rules enable the system to identify various characteristics of the viewer and even identification and association with a stored historical profile
FIG. 8 shows the generalized embodiment of Fig. 3 with the addition of a personal profile engine. FIG. 9 is a preferred embodiment of the present invention illustrating gesture- based interaction with 3D objects on an AS display.
FIG. 10 illustrates exemplary gesture based interactions.
FIG. 11 is a simplified illustration of an interface with the virtual scene based on a scaled model of the user seen in the virtual space. FIG. 12 is a illustrates a viewer or anatomical part of a viewer and the parallel
"virtual viewer" or "virtual anatomical part" being displayed in the viewer perceived space FIG. 13A is a simplified illustration showing the 3D camera field of view according to preferred embodiments.
FIG. 13B is a simplified illustration showing the process of eye tracking and position finding. FIG. 13C is a simplified illustration showing the process of 3D skeleton modeling.
FIG.14 is a simplified diagram illustrating a lenticular lens based autosterescopic display used in preferred embodiments of the present invention.
FIG 15 is a simplified illustration of the perceived 3D space in relation to the pixels on the AS display in which the viewer may observe and interact with virtual objects and scenes.
FIG. 16 is a simplified illustration of the interaction zone showing the overlap between the perceived space and the camera field of view for an exemplary case.
FIG. 17 shows an alternative embodiment in which the viewer's eyes position data obtained from the 3D camera may also be input to a screen steering device.
FIG. 18 shows a network of 3D virtual touch human interface systems.
DESCRIPTION OF THE PREFERRED EMBODIMENTS
An apparatus and a method are now disclosed for three dimensional interaction between a viewer and virtual object(s) displayed on an autostereoscopic (AS) display. According to some embodiments, the viewer is able to experience the sensation of touching and/or control of objects perceived on the autostereoscopic display solely through use of an anatomical part of the viewer interfacing with the display.
More specifically, the present embodiments provide an immersive and three dimensional real time interactive environment through the use of one or more 3D cameras. A virtual object(s) is displayed on an AS display and appears to the viewer in a defined region in front of the display determined by the viewer's eyes position in relation to the display, as explained above. In preferred embodiments, the viewer may manipulate the object in various ways and may interact with the object in real time. These manipulations are then reflected in subsequent frames rendered and perceived by the viewer in the region for virtual objects in front of the display defined by the user's eyes location. The process creates an "immersive" interactive environment between the viewer and the virtual scene perceived on. the auto stereoscopic display.
In some embodiments, through 3D data obtained about the viewer's facial features, a viewer may be identified according to his or her stored profile. In such embodiments, the interaction process can be suited towards that particular viewer's historical preference data and other unique characteristics associated with that viewer in particular.
The principles and operation of an apparatus and method according to the present invention may be better understood with reference to the drawings and accompanying description.
Before explaining at least one embodiment of the invention in detail, it is to be understood that the invention is not limited in its application to the details of construction and the arrangement of the components set forth in the following description or illustrated in the drawings. The invention is capable of other embodiments or of being practiced or carried out in various ways. Also, it is to be understood that the phraseology and terminology employed herein is for the purpose of description and should not be regarded as limiting.
All references cited herein are incorporated by reference in their entirety. Citation of a reference does not constitute an admission that the reference is prior art. The present application contains subject matter relating to a PCT application number IL07/001432 filed on Nov 20, 2008. entitled 3D GEOMETRIC MODELING AND 3D VIDEO CONTENT CREATION. This patent application is assigned to MantisVision, Ltd. Reference is now made to Fig 3A, which is a simplified flow chart illustrating a generalized embodiment of the invention. System 200 contains anatomy tracking system 18, object interaction processing engine 42, virtual environment data module 32, and scene presentation engine 44. Anatomy tracking system 18 is comprised of at least one 3D video camera 22 and a 2D/3D image processing unit 26. The 3D camera(s) may be situated in relation to the display in a variety of different configurations. The camera(s) 22, which can preferably acquire depth information in motion, allow for the derivation of a depth map of the viewer positioned at some location typically in front of the display. This depth map is typically in the form of a three dimensional point cloud. The point cloud may then be input to a face and eyes finding algorithm contained in the image processing unit. The viewer's eyes location determines, as described above, the mapping of the perceived coordinates ([XJ], [yj, [Zj]) of the displayed virtual objects. Furthermore, as described in figures below, the viewer eyes location determines the field of view or perceived space of the virtual scene seen by the viewer. The perceived virtual objects may be either static or dynamic.
Anatomy tracking system 18 is comprised of at least one 3D camera 22 and 2D/3D image processing unit 26. It is understood that multiple 3D cameras may be present. The 3D cameras together with the image processing unit 26 provide 3D coordinates ([Xi], [Yi], [Zi]) of the viewer's anatomy and/or movement of the anatomy in the camera field of view, typically the hand, arm, and face. Again, this set of data is typically a depth map in the form of a three dimensional point cloud. Furthermore, the 3D camera preferably is a 3D video camera capable of 3D motion capture, and as such, a point cloud for each frame in a captured video sequence is preferably acquired. The series of point clouds may then be input to a human finding engine and skeleton registering unit contained in image processing unit 38. Such skeleton registering units or 3D classifiers are known in the art and are not discussed herein. Additionally, the anatomy tracking system provides location data of the user's eyes.
The object interaction processing engine 42 accepts the output from the anatomy location tracking system as well as virtual environment data module 32. This virtual environment data is typically stored on a storage device and represents three dimensional representations displayed to the user according to techniques known in the art relating to AS displays.
Reference is now made to Fig 3B, which is an illustration of the internal components of interaction processing engine 42. The anatomical part-virtual object registration engine 34 accepts and processes information from both the anatomy tracking system 18 and virtual environment data module 32. Information concerning the eyes location from the anatomy tracking system 18 together with the virtual environment data from module 32 enables the registration engine 34 to determine the perceived virtual object 3D coordinates for the viewer's eye position ([XJ], [Vj], [ZJ]). Additionally, the perceived AS display field of view as seen by the viewer according to his eyes location is determined (see Figs 4A-4B below). The information from the anatomy tracking system 18 pertaining to movement of the viewer's anatomical parts such as his or her hand, arm, and/or head, enables a 3D mapping based on the acquired point cloud coordinates ([Xj], [Yj], [Zj]). These sets of data are registered to determine a single 3D volume by the registration engine.
This registered data set from the registration engine 32 is then sent to the anatomical part-virtual object relation computation engine 34. In some embodiments, the relation computation engine determines if a collision occurs between the anatomical part and the virtual objects seen by the viewer. This would indicate that the viewer has "touched" the perceived virtual object(s).
This information from relation computation engine 36 is then sent to the rule enforcement engine 38. This engine receives interactive application rules from 30 contained in the virtual environment data module 32. The interactive application rules module 30 contains all rules for determining subsequent AS display frames as a function of user input. The rule enforcement engine then determines what the next presented scene on the AS display should be. For instance, through registration between the data sets (registration engine 34) a collision may be detected (relation computation engine 36), in which case the next frame displayed on the AS display may show the perceived virtual object, as a result of the "virtual touch", moved to a different location (rule enforcement engine 38).
Alternatively, rule enforcement engine 38 may decide that the virtual object should disappear as a result of the "virtual touch". In still further embodiments, other resultant actions may occur to the virtual object as a result of interactivity between the viewer and the virtual object. The rule enforcement engine output is thus the output of the interaction processing engine 42. This output is sent to scene presentation engine 44 to render the next multiple frame content to be displayed on the AS display.
Data storage of the virtual environment data may be implemented in any combination of volatile (for example, RAM, registers, etc) and/or non-volatile memory (for example, flash memory, magnetic medium such as disk, etc). Data storage may be provided in a single location and/or may be distributed among multiple locations, for example, in a local and/or wide-area computer network.
In other embodiments, the viewer's hand acts as a "gravitational" source on the virtual object. In such embodiments discussed below, various gestures made by the viewer and captured by the 3D camera or 3D camera array signify various applications on the virtual object. In such a case where the viewer's anatomy is not in a 3D volume in which the virtual mapping is displayed, registration between the 3D data sets need not be carried out, since the relative locations between the virtual object and viewer anatomy data sets is not taken into account. Rather, the object interaction processing unit, based on the detected 3D movement of the viewer's anatomy, determines the effect on the currently displayed scene. In still further embodiments, the interaction may combine both "touch" and "gesture" applications. In such cases, the interaction processing engine only contains the relation computation engine 36 and the rules enforcement engine 38.
The above process thus allows the viewer to touch, manipulate, and move virtual objects with his hand, watch virtual 3D objects fly into the palm of his hand, or control virtual scenes through gestures of his hand that are captured by the 3D camera. The 2D/3D camera system is preferably composed of a structured light active triangulation based imaging apparatus, such as detailed in Appendix 1. Such imaging devices utilize triangulation methods to determine three dimensional coordinates of imaged real world objects. However, there is no limitation on the type of 3D camera, and any type of 3D camera is appropriate, including but not limited to so-called 'time of flight' cameras, 2D cameras utilizing stereo correspondence algorithms, and triangulation based 3D cameras. As noted, the present embodiments preferably utilize 3D video cameras in order to capture 3D motion. The 3D imaging device(s) may be situated in relation to the display in a variety of different configurations, capture depth information in motion, and additionally capture 2D texture data of the imaged scene. To better understand the present embodiments, a brief description of the 3D imaging system of Appendix 1 is now given. The 3D video camera contains both sensing and projector elements, preferably in a single housing. A pattern containing distinct features is projected onto imaged elements in the camera field of view. The pattern features, once reflected from the imaged elements, are captured in an image on the camera sensor. The location of the features in the image is correlated, through triangulation techniques, to 3D spatial coordinates on the imaged element. The totality of the 3D spatial coordinates comprises a point cloud of 3D points, which give a geometric shape of the imaged elements. The point cloud may further be processed to arrive at a 3D polygonal mesh. The above process is carried out per frame for every frame of the video sequence, so that 3D modeling over time of imaged objects in both a static and dynamic scene may be implemented. If several cameras are used, then for each frame, triangulation may be carried out separately for each camera and additionally through stereoscopic comparison of the two images and triangulation of common points in both images. In such a case of multiple cameras, multiple point clouds are obtained per frame. These multiple point clouds are then registered to a unified point cloud set. To further illustrate 3D touch interaction of the present invention, we now turn to Figs 4 A and 4B. In the example shown in Fig 4 A, the system may include an autostereoscopic (AS) display 14 and an array of 3D cameras 22 optionally situated on the top side of the display. Other configurations of the 3D camera(s) are understood to be possible, and the particular configuration shown is merely for illustration purposes. Shaded region 52 shows the perceived field of view in which the viewer 56 sees virtual objects given the location of his eyes in front of the AS display. That is to say, the viewer only perceives virtual objects and scenes that are within the shaded triangular region 52. This region is defined by light rays from adjacent pixel columns of the AS FLAT PANEL DISPLAY display that are perceived by each eye. The viewer's eye coordinates are denoted (Xa), (Ya), (Za) and (Xb), (Yb), (Zb) for the left and right eyes respectively. Fig 15 below further illustrates the perceived space in which virtual objects are seen by a viewer of an AS display. The viewer eyes location leads to a mapping of the virtual object 60 coordinates in 3D space ([Xj3], [yj3], [Zj3]) seen by the viewer at this position.
In parallel to the eye tracking, the 3D motion capture video cameras capture data from which the 3D coordinates ([Xm]9 [Ym]5 [Zm]) of the viewer's hand and arm 62 in the camera field of view may be derived. Upon processing of this initial point cloud data for each frame, a moving skeleton model is obtained of the viewer's anatomy part. A matching or collision between coordinate sets ([Xj3], [yi3], [Zj3]) and ([^m]9 [Ym]5 [Zm]) in a given video frame indicates that the viewer is "touching and/or manipulating" and therefore interacting with the virtual object. With reference to Fig 4B, the viewer is seen moving laterally to the left in relation to the AS display. As a result, the viewer's perceived field of view of virtual scenes 64 on the AS display is shifted in relation to Fig 4A. Furthermore, the virtual object 60 now has a different mapping of coordinates ([Xj4], [yj4], [Zj4]) than that of Fig 4A, that correspond to the viewer's new location in relation to the display. To interact with the virtual object in 3D space, the viewer now moves his hand and arm to a different location in 3D space within the perceived field of view of virtual objects. The skeleton mapping in this case is now based on a different captured point cloud series, namely ([Xn], [Yn], [Zn]). Again, when a collision occurs between the data representing the 3D virtual object as perceived by the viewer and the data representing the human anatomy movement, some form of interaction occurs.
Reference is made to Fig 4C, which is a simplified illustration of the processing engines in preferred embodiments. 3D cameras 22 send to the anatomy tracking system 18 depth data from which to derive viewer eyes location data. In addition, the depth data is used to derive the viewer anatomy location and movement. The anatomy tracking unit sends data to the object interaction processing engine 42. Engine 42 additionally receives information on the interactive application from virtual environment module 32. The virtual coordinate mapping for virtual object 68 is determined based on the viewer eye location data. The object interaction processing engine registers the sets of virtual coordinate data and the anatomy location data. The next frame in the AS video sequence is based on the interactive rules processed in the interaction processing engine. The virtual environment data module 32 preferably contains the displayed content for current frames and the interactive characteristics and functionality of the displayed program. For instance, the interactive application determines the types of hand movements the viewer may make when he touches the object and whether the displayed object moves, twists, rotates, disappears, the manner in which the object performs these actions, and any other interactive functions.
For example, in a first displayed frame, viewer 56 may grasp virtual coke bottle 68 having initial virtual coordinate mapping ([xj], [yi], [ZJ]) at tl and rotate his hand and arm 62 to simulate a pouring action. The 3D video cameras 48 capture depth information of the moving hand and arm and send this data to the anatomy tracking system 18. The system 18 then computes the point clouds and subsequent skeleton model of the hand and arm for each frame. For instance, in the figure, the hand and arm have point cloud coordinates ([Xp], [YP], [Zp]) at tl. This engine 18 outputs to the object interaction processing engine 42 a skeleton model showing a hand in a pouring motion. The interaction engine, after processing this information together with the interactive application rules, tells the scene rendering unit 44 to display the next virtual coke bottle in a tilted position. That is to say, the virtual coke bottle 68 in the next frame at t2 has a new mapping of perceived coordinates ([Xj], [yj], [zj]) as a result of the interaction. Box 86 illustrates three different 3D coordinate mappings for three frames in which the virtual coke bottle 68 is displayed. In each successive frame, the perceived coordinates of the virtual coke bottle change in 3D space as the bottle appears to slowly tilt as if it is being poured by the viewer. Box 72 illustrates three different 3D coordinate mappings of the user's hand and arm for three successive time intervals, tl- t3. It is understood that certain areas of the virtual coke bottle are occluded from the viewer's view when he grasps the bottle, since in essence he is occluding his vision of the AS display at those points in space. Furthermore, the configuration and numbers of processing units in the diagram is for illustration purposes only and is in no way limiting.
Reference is made to Fig 5, which illustrates several visual examples of interaction between a real object, such as the viewer's hand, and a virtual 3D object displayed on an AS screen. Floating virtual beach ball 86 is displayed in the 3D space perceived by the viewer in front of the AS display 90. The viewer may poke the virtual ball with his hand 94. This poking action is detected by the camera(s) and processing devices of the anatomy location and configuration tracking system. A 3D mapping of the viewer's hand is represented by mapping ([Xj], [Yj], [Zj]). The object interaction processing engine then determines the resultant scene to be displayed based on the virtual environment data and interaction rules as a result of the poking action. In order to simulate a physical beach ball being poked, the virtual beach ball is displayed in a new 3D location in space in the following frame displayed to the viewer. That is to say, frame 1 has virtual object coordinate mapping ([xm], [ym], [zm]) and frame 2 has mapping ([xn], [γn], [zn]), closer to the AS display. Therefore, the displayed frame 2 on the AS display is a result of interaction between the viewer's hand and the virtual beach ball. It is understood that other actions, such as swinging at the virtual ball and reactions to the above described interaction are possible. The interactive application rule may dictate that the ball disappear, pop out of sight, fade out of sight, or any other desired interactive functionality as a result of the touch interaction or collision between the two sets of 3D coordinates.
Another interactive example shown in Fig 5 is virtual 3D bird 98 perching in the palm of viewer hand 102. As stated above, the viewer's hand is typically modeled according to a skeleton modeling algorithm utilizing the point cloud coordinates derived from information captured by the 3D camera. Such skeleton modeling algorithms are seen for example in "Markerless Human Motion Transfer" by Cheung et al., "Image Based Markerless 3D Human Motion Capture using Multiple Cues" by Azad, et al., "Markerless Motion Capture from Single or Multi Camera Video Sequence" by Remondino et al., and others. The interactive processing engine unit, based on the interactive application rules and virtual environment data content, recognizes the skeleton model of an outstretched viewer's palm and sends the scene rendering unit 44 the next set of frames so that the virtual bird flies to and then rests on the viewer's hand. The mapping of coordinates of the virtual bird for each frame approaches a collision with the coordinates for the skeleton model of the outstretched hand. Alternatively, as shown, the viewer may attempt to catch a flying virtual bird 106 that appears in the virtual space that he perceives according to his position in relation to the screen. Other examples in the figure show the viewer using his hands to interact with a virtual money bill 108 and a virtual leaf 110. In all of the shown examples, the collision of a set of 3D coordinates representing the human anatomy part with a set of 3D coordinates representing a displayed virtual object indicate touch and/or manipulation of virtual objects. The interactive application rules together with the input of these two sets of coordinates determine the resultant next frame to be displayed. In still a further application, the virtual object may be a direct input device whose interaction rules are derived from existing I/O devices such as a virtual 2D touch screen, multi-touch screen, keyboard, or other known input devices. Collision detection methods for interactive environments are known in the art and thus are not discussed herein. The reader is referred to "Collision Detection In 3D Interactive Environments" by van den Bergen as an exemplary way of performing 3D collision detection.
Fig 6 is a preferred embodiment showing a three dimensional autosterescopic menu driven interface. The floating balls or nodes 112 appear to the viewer in the 3D space perceived by him in which he may view virtual 3D objects. The floating balls comprise a hierarchical menu. The first stage shows 4 nodes that represent the operating system desired by the viewer. When the viewer "touches" the virtual ball, as described in the above embodiments, a second tier menu appears to him. The second tier represents the desired application software and the third tier the functionality therein. In the present example the viewer chooses Windows® and Outlook® and then
Inbox. It is understood that any hierarchical menu driven interface may be manipulated in such an interactive manner and the details in the figure are merely one example for illustration purposes. In a certain embodiment, the chosen ball becomes larger, thus indicating that it has been selected. This implementation may utilize historical learning algorithms such as seen in the link herein- http://www.inference.phy.cam.ac.ulc/dasher/.
Reference is made to Fig. 7 which shows a preferred embodiment wherein viewer classification rules enable the system to identify various characteristics of the viewer and even identification and association with a stored historical profile. A viewer imaging device 250 images the viewer and sends the image to the viewer 3D processing engine 252. The 3D processing engine is controlled by the viewer classification rules module 254. The viewer classification rules module 254 preferably defines the characteristics of the viewer that are to be determined based on the interactive application 256. For instance, the system may want to determine if a viewer is wearing a watch and then offer various products and services associated with various watch types. The system may want to determine the height of the person or his clothing size to present interactive content that is appropriate. Furthermore, an on-line viewer profile is optionally obtained through an on-line viewer database 258 containing the historical profile of viewers. The processing of the viewer may be iterative until an appropriate model is obtained, at which point the 3D modeling data is sent to the interaction engine 260. The interaction engine receives the interactive application information 256 as well. The interaction engine then decides upon the 3D displayed content on the display 262 based on the viewer classification and/or identification and the interactive application. With such a user profile, buttons and other virtual objects such as seen in Fig. 7 can be adapted to display to the user a probable selection based on his historical profile. For instance, if he is known to pick Windows often, then in stage 1 he may see the 3D Windows icon enlarged for him for ease of use. In another application, the 3D camera image and skeleton model enable the viewer to appear as a mirror image to himself. In other words, the displayed AS content is a mirror virtual image of the viewer himself. This image may utilize the viewers shape and/or motion in the form of an avatar or an actual image of the viewer. Such an image can be used to display to the user how he or she would look with various hairstyles, clothing, etc. In a multi-user application, this avatar can be transmitted to from one user to the next for user interaction between them.
Reference is made to Fig 8. Fig 8 is an illustration parallel to Fig. 3 above showing the additional personal profile engine. The personal profile contains 3D camera 272, processing engine unit based on rules 274 and on-line database 276. It is understood that the 3D cameras of the anatomy tracking system may be used for the personal profile engine. The personal profile engine is based on the viewer classification rules that typically call for an iterative analysis process to determine desired characteristics of the viewer in front of the screen. In order to carry out this process, the personal profile engine typically takes input from the interactive application. The application may call for the identification and presentation of content to children that is different than content to adults. The interactive engine 42 receives information from the interactive application rules module as well. The interactive engine 42 then processes all four inputs to arrive at 3D content to be displayed.
Reference is made to Fig 9, which is a preferred embodiment of the present invention illustrating gesture-based interaction with 3D objects on an AS display. In such an embodiment, the viewer gestures typically with his arm and hand to manipulate a 3D virtual object perceived by him in the region 114 in which he may view virtual objects. Various motions of the hand and arm signify interactive functions with displayed 3D virtual objects.
User hand and arm 116 is shown making a pulling gesture to indicate the viewer's desire to pull the virtual beach ball 118 towards him. The viewer's hand and arm have real object mapping ([Xk], [Yk], [Zk]) in the form of a three dimensional point cloud for each frame. This point cloud is processed by image processing units as described above and the skeleton model of the moving anatomy is determined over time. The virtual beach ball has virtual object mapping ([xj5], [yis], [Zj5]) for frame 1 at time 1. The interaction engine identifies the pulling motion from the skeleton model. Moreover, the interaction engine rules may determine that the pulling motion of the hand cause the virtual object to move in the next displayed frame closer to the 3D location of the viewer's hand. As the arrows in the figure indicate, the ball is then displayed in subsequent frames in incrementally closer 3D coordinates to that of the viewer. As seen, frame 2 at time t2 virtual object coordinates ([Xj6], [yi6], [Zj6]) are closer to the viewer's hand location in 3D space given by real object mapping ([Xk],
[Yk], [Zk]).
In certain gesture-based embodiments the precise eye location of the viewer may not be necessary for determination of the precise virtual object mapping scene by the viewer, since no actual collision algorithms are used to determine virtual touch.
However, a determination of the viewer perceived three dimensional space in relation to the AS display may still be desirable.
Fig 10 shows a variety of other gesture-based interactive functions. All capital coordinate sets denoted ([X], [Y], [Z]) indicate real objects, while all lower case coordinate sets denoted ([x], [y], [z]) indicate virtual objects. The viewer may point his finger at the ball, shown as action 120, push the ball, shown as action 122, or pull the ball, shown as action 124. In all of the above cases, the next displayed frame is a function of the interaction application. The interactive application may be a "magnetic" effect, whereby the closer one is to the virtual object the stronger and faster the action. As a result of a specific gesture, the ball may pop, disappear, be dragged to and away from the viewer, among other functions. The viewer may also hold virtual objects in his hand such as virtual wrench 126 to manipulate other virtual objects in the viewer perceived space. Here virtual wrench 126 can be twisted and turned to tighten or loosen virtual screw 128. It is understood that under certain hand positions, the viewer's hand may occlude the virtual tool from appearing to the viewer, and the use of virtual tools may need to be modified to allow viewer visibility. Menu driven interface shown in Fig 6 may additionally be operated in gesture mode wherein the viewer merely gestures to choose the appropriate virtual ball.
Other gesture interactive motions 130 such as thumbs up, thumbs down, finger and hand movements are shown. Gestures may be earned out by a viewer to manipulate objects or send interactive commands without looking at his hands so long as they are within the camera(s) field of view. For instance, a tap of the viewer's left hand downwards may signify a lowering of the volume on the AS display, while upwards may cause an increase in display volume. Such an interaction is similar to touch typing on a computer keyboard, where the viewer is interacting with the computer while not moving his eyes from the screen. Gesture based interaction in 3D environments are discussed in the paper "Multi-Finger Gestural Interaction with 3D Volumetric Displays" by Grossman et.al., which relates to multi finger gestural interaction with tracking devices for interaction with volumetric displays. It is understood that in contrast, the embodiments of the present invention make use of a 3D camera capable of markerless tracking to capture the 3D motion vector of the viewer's anatomy part at any given time. Reference is made to Fig 11 , which shows a preferred embodiment wherein a scaled version of the viewer or moving anatomy part of the viewer is displayed in the viewer perceived space as a "virtual viewer" for interface with other virtual objects. This scaled version 134 is typically miniature in size and sometimes referred to herein as the "virtual viewer" or "ghost". The 3D camera captures either the viewer 138 or an anatomical part of the viewer and, upon image processing of the captured real coordinates ([XJ, [YJ, [ZJ), a scaled or miniature and preferably semi-transparent version of the viewer and his movements is displayed in the perceived space 142. The moving "ghost" has perceived virtual object coordinates ([Xi7], [Vj7], [zj7]). Additionally, virtual object 146 with 3D perceived coordinates ([Xj8], [yi8], [zis]) is displayed in the viewer perceived space. Interaction may now occur between the "virtual viewer" and the virtual object. That is to say, all movements by the real viewer are reflected by parallel movements in the virtual viewer. Thus, instead of the real viewer interacting with the virtual object, the virtual viewer interacts in his place. This interaction may be either touch or gesture-based as discussed in previous embodiments. Also, as in previous embodiments, the interaction engine preferably receives inputs from the anatomy tracking system and determines the interactive functionality. The interaction engine then sends commands to the scene rendering unit.
Such a "ghost" may be utilized in PDAs or other portable devices with a small
3D display not large enough for accurate interaction between a small displayed virtual object and the viewer's own hand. The miniature version of the viewer may be his entire body or just his hand, or even a three dimensional puppet or ghost-like image that does not resemble the viewer but moves in accordance with the viewer's body movements. The miniature can be in a non proportional scale and may move faster and slower in accordance with the interactive settings. That is to say, the sensitivity to movement of the virtual miniature version of the viewer may depend on speed of the viewer's movement, similar to a computer mouse.
The present embodiment can also be implemented as a seamless combination of both touch interaction, as described above, and the "ghost" interaction. When the viewer's anatomy is within the space in which he is capable of viewing virtual objects, the viewer may virtually touch perceived virtual objects in the virtual scene. As the viewer moves his hand out of the 3D space in which he is able to perceive virtual objects, but still within the camera field of view, a seamless transition can occur where the viewer sees the "ghost" interface at a certain distance from the display.
Reference is made to Fig 12, which illustrates a viewer or anatomical part of a viewer and the parallel "virtual viewer" or "virtual anatomical part" being displayed in the viewer perceived space. Viewer 150 sees a "ghost" 154 of himself in miniature form in the perceived virtual object space in front of the display 156. The "ghost" appears to the viewer with a virtual "dashboard" 158 on his back that indicates to the viewer various states and modes of interaction, similar to CAPS LOCK on a standard computer keyboard. Viewer hand 160 is mapped into the perceived virtual object space and may take the form of various shaped virtual hands such as 162 and 164. The virtual hand may additionally manipulate scaled virtual objects such as hammer 166 and beach ball 168.
Reference is made to Fig 13A-13C, which illustrates the 3D camera and post processing operations as described in the above embodiments. For ease of understanding and illustration, only one 3D camera is shown in Fig 13 A with field of view 170. It is understood that, as shown in the preferred embodiments above, in order to enlarge the field of view multiple 3D cameras and differing configurations of 3D cameras in relation to the AS display may optionally be used. The camera is comprised of a two dimensional camera 172 with vertically attached projector 174. Additional projectors and various configurations of projectors in relation to the AS display may be implemented as well in alternate embodiments. The configuration of projectors for the current embodiment is for illustration purposes only. It is understood, that in order to prevent interference between projectors, particularly when using structured light based 3D cameras, the projections may be separated, for example, by different wavelengths, polarities, or time multiplexing. The camera is typically positioned in a suitable position in front of the viewer and/or attached to the AS display 176. Furthermore, various adjustments can be made to the camera and projector lenses, angles, and positions, aperture settings, as well as the number of cameras and/or the number of projectors to better suit the desired application. For example, in the above discussed figures, several 3D cameras are shown. In the present embodiment, it is understood that the projector is placed vertically to the camera purely for ease of illustration.
As seen in Fig 13B, the camera captures the viewer's depth map 1773 in the form of a 3D point cloud. The depth map provides the point cloud to a face finding engine or 3D classifier 180. Such 3D classifiers are known in the art, such as described in "Automatic Classification of Objects in 3D Laser Range Scan" by Nuchter et.al.. The face finding engine then is used to determine precise eyes location 182 and optionally the direction in space at which the viewer is focusing. This optional element is denoted as eye direction tracker 184. Eye direction tracking may be used to determine the position in space at which the viewer is looking at a given moment in time. The eye position detection is denoted by dashed lines 178 in Fig 13 A. The eyes finding engine preferably operates on a continuous basis to determine the viewer perceived 3D space as the viewer, and thus his eyes location, moves position in front of the AS display. Furthermore, the camera captures the viewer's body or parts thereof in three dimensional positional coordinates, and typically at least the viewer's hand and arm 186 and face 188. Fig 13C shows a typical flow diagram of skeleton modeling based on depth capture. The sampled depth data 190, in addition to optional texture data 192, is input to a viewer finding engine 194 and from there to one or more post processing units to determine his body contour from which a skeleton model is registered 196. As stated, the depth data may be combined with textural information captured simultaneously. Texture may be captured through use of an additional camera or CCD. Such methods for animatable human body reconstruction from data sampled in three dimensions are discussed for example in US Patent No. 6133921. Moreover, the 3D data provide dynamic video images, meaning through 3D motion capture methods the 3D positional data of the viewer's anatomy part is known at any given point in time. The 3D skeleton model may be captured from various types of 3D cameras, such as stereo correspondence, time of flight or triangulation based cameras. As discussed above, the 3D skeleton model of the viewer's anatomical part is used to determine interaction with the virtual scene.
Reference is made to Fig 14, which shows the basic characteristics and functionality of an exemplary automultiscopic/stereoscopic display as used in the previously described embodiments. AS displays may have differing characteristics, such as screen width, number of views, and optimal viewing distance, but the display principles typically behind such displays is the same- the division of the overall resolution of the screen to a defined number of "viewing zones" through use of some type of optical filter. Display 200 has lenticular lenses 204 placed over a flat panel screen 206 divided into a series of two views 210 and 214. Each pair of views on the flat panel display is covered by a separate half cylindrical shaped lenticular lens. Each view is made up of a pixel column on the fiat panel display and each lenticular lens projects both views simultaneously to a viewer. The optical characteristics of the lenticular lens are responsible for the fact that at varying positions in front of such a screen, each eye 216 of a viewer sees one of the two views. By viewing two different images simultaneously, the viewer experiences the three dimensional effect called stereo parallax.
Several methods have been developed, for instance through use of a slanted lenticular lens, to spread the resolution reduction over both x and y components of the screen. Other methods of autosterescopic display have been developed such as parallex barrier methods. Moreover, as discussed herein, for multiview systems, each zone may display to a viewer therein different content.
Reference is made to Fig 15, which is a simplified illustration of the perceived 3D space in relation to the AS display in which the viewer may observe and interact with virtual objects and scenes. The viewer 218 position in relation to the display 220 determines which pixel columns each eye views. The display is typically divided into a repeating series of pixel columns in accordance with the number of viewing zones. In the figure, the display is divided into a series of 4 columns corresponding to 4 possible viewing zones. Under each lenticular lens a group of 4 columns exists. Example columns 222 and 224 are shown, and these pixel columns display images seen in the number 2 and 3 viewing zones (not shown). For example, columns 222 project light rays 226 and 228 into the right and left eye respectively of viewer 218 whose eyes lie in zones 2 and 3 (not shown). The entire 3D space in which the viewer may perceive 3D objects from that location with respect to the display is denoted by shaded region 230. Region 230 is the 3D space in front of the screen in which both eyes view light rays from pixels on the screen. It is understood that the light projections shown from the AS display are for illustration purposes. When a viewer shifts his eye position within a given zone, the viewer typically experiences the visual artifact of a shift in the virtual object. This artifact is a result of the lack of motion parallax within the viewing zone.
Reference is made to Fig 16, which is a simplified illustration showing the case where the camera field of view does not cover all of the viewer perceived space in front of the AS display. In such a case, area 232 represents what may be termed an interaction zone. The interaction zone represents the overlap between the 3D camera field of view and the space within which the viewer interacts and sees virtual objects. For "virtual touch" applications, interactivity is determined by the registration between the 3D coordinates of the virtual scene (displayed object(s)) perceived by the viewer and the 3D positional data of the viewer's anatomical part. It is seen that should the viewer hand be outside the camera's field of view but still within the region in which the viewer sees virtual objects, region 234, then even if the viewer perceives the virtual touch the camera does not. As such, no interaction between the viewer and the virtual object is detected in the interaction engine and no interactive application is possible. The interaction zone is thus the totality of 3D points in space in which virtual objects are perceived by the viewer and that is also captured by the camera. It is understood that various layouts and numbers of cameras and projectors may provide for interaction zones of varying volumes. Reference is made to Fig 17, which show an alternative embodiment in which the viewer's eyes position data obtained from the 3D camera may also be input to a screen steering device 236. In such an embodiment, the screen display's angle and location adjusts to the viewer's precise location through use of a moving optical filter that alters the optical viewing space. The screen then needs only to render the two appropriate views to the viewer's location, and the decrease in screen resolution due to the partition into repeating series of pixel columns for each view is obviated. Such a display device is described for instance in US Patent No. 6,075,557. Furthermore, the exact positioning of the individual viewing zones, seen in Fig 1, can be refined and tuned to the precise viewer location rather than being in a fixed cubic area in front of the AS display.
In other adaptive screen embodiments, a slight rotation and/or translation of the lenticular lens over the flat panel display allows for adjustment of the viewing zones. In AS displays using parallax barrier, the occlusion of light is determined based on the user's position so as to maximize the zone locations. Various lens types are discussed in "A Survey of 3DTV Displays: Techniques and Technologies" by Benzie, P, et.al. Circuits and Systems for Video Technology, IEEE Transactions on Volume 17, Issue 11, Nov. 2007 Page(s):1647 - 1658. In still other screen types, acoustic optics may be used to control the refractive index. Such an acoustic optical covering may then be used to dynamically adjust the viewing zones. An acoustic lens of this type is discussed in the "Acoustic Lens with Electrically Controlled Refractive Index" by Pappalardo, M from 1980, Ultrasonics Symposium, Volume , Issue , 1980 Page(s): 590 - 593. Reference is made to Fig 18 which shows a network of 3D virtual touch human interface systems as described herein. Each system is denoted as 200 and has the characteristics described in Fig 3 A above. A network of such systems may allow for a multi-user interface wherein one or more users interact with each other based on content created at each viewer's individual station. The scene presentation engine outputs from each system and is sent to other stations over the network. For instance, in a handball game between two users, each user hits the ball in his 3D volume in front of his AS display. The resultant virtual environment information is then sent to the second user over the network. In other words, the second user sees the virtual ball in front of his AS display as a result of interaction of the first user with the ball. In such a way, a multi user game is possible. In the above described multi-user application, a unified 3D volume exists of all user that registers the individual local registered environments into a single registered environment of all users. Multiuser 3D video videoconference is another implementation using such a network of systems.
It is expected that during the life of this patent many relevant devices and systems will be developed and the scope of the terms herein is intended to include all such new technologies a priori.
It is appreciated that certain features of the invention, which are, for clarity, described in the context of separate embodiments, may also be provided in combination in a single embodiment. Conversely, various features of the invention, which are, for brevity, described in the context of a single embodiment, may also be provided separately or in any suitable subcombination.
Although the invention has been described in conjunction with specific embodiments thereof, it is evident that many alternatives, modifications and variations will be apparent to those skilled in the art. Accordingly, it is intended to embrace all such alternatives, modifications and variations that fall within the spirit and broad scope of the appended claims. All publications, patents, and patent applications mentioned in this specification are herein incorporated in their entirety by reference into the specification, to the same extent as if each individual publication, patent or patent application was specifically and individually indicated to be incorporated herein by reference. In addition, citation or identification of any reference in this application shall not be construed as an admission that such reference is available as prior art to the present invention.

Claims

CLAIMSWhat is claimed is:
1. An apparatus for providing an interactive human computer interface to a viewer, said apparatus comprising: a) a data storage configured to store three-dimensional virtual environment representation data including at least one three-dimensional virtual object within said virtual environment; b) an autostereoscopic (AS) display configured to display simultaneous perspectives of virtual objects of said 3D virtual environment representation data to spatially separated viewing zones located in front of said AS display c) an anatomy tracking system including at least one 3D camera, said anatomy tracking system operative to determine respective real-world locations of: i) a viewer's eyes; and ii) an anatomical part on said viewer's body d) a 3D registration engine configured to generate a 3D volume comprising 3D location data of said viewer's anatomical part and 3D location data of said virtual objects perceived by said viewer in accordance with said viewer's eye location, e) an anatomical part- virtual object relation computation engine operative to determine a relation between said virtual object and said anatomical part in accordance with output of said registration engine: f) a rule enforcement engine operative to modify said three-dimensional environment representation data in accordance with said determined anatomical part- virtual object relation and said virtual environment data.
2. A system according to claim 1, wherein said virtual environment data contains interactive application rules.
3- A system according to claim 1, wherein said anatomy tracking system determines said viewer's eye position and anatomical part position continuously over time.
4. A system according to claim 1, wherein said rule enforcement engine is further configured to provide output data in real time to a scene presentation engine for display of subsequent content.
5. A system according to claim 1, configured to receive input from a plurality of viewers having eye positions in a plurality of adjacent pairs of said viewing zones, said viewers simultaneously interacting with said autostereoscopic display.
6. A system according to claim 1, wherein said 3D camera comprises an NIR projector and narrowband filter.
7. A system according to claim 1, further comprising a viewer classification module configured to determine viewer identity and characteristics based on a viewer historical profile.
8. A system according to claim 7, wherein said viewer profile is obtained through an on-line database.
9. A system according to claim 1, wherein said displayed three-dimensional virtual environment representation data is a virtual touch screen.
10. A system according to claim 1, wherein said displayed three-dimensional virtual environment representation data is a virtual keyboard.
11. A system according to claim 1, wherein said displayed three-dimensional virtual environment representation data is an virtual image of the user.
12. A system according to claim 1, wherein said displayed virtual environment representation data as a result of rule enforcement engine output is sent over a network in a multi-user environment to allow for multi-user interface.
13. A system according to claim 1, wherein said displayed virtual environment representation data is continuously changing 3D positional data.
14. A system according to claim 1, further comprising a miniaturized element perceived by said viewer in a viewer perceived space that is determined in accordance with said viewer's eye position, said miniaturized element configured for interaction with said displayed 3D virtual objects.
15. An apparatus for providing an interactive human computer interface to a viewer, said apparatus comprising: a) a data storage configured to store three-dimensional virtual environment representation data including at least one three-dimensional virtual object within said virtual environment; b) an autostereoscopic (AS) display configured to display simultaneous perspectives of virtual objects of said 3D virtual environment representation data to spatially separated viewing zones located in front of said AS display c) an anatomy tracking system including at least one 3D camera, said anatomy tracking system operative to determine respective real-world locations of: i) a viewer's eyes; and ii) an anatomical part on said viewer's body . d) an anatomical part- virtual object relation computation engine operative to determine a relation between a perceived virtual object by said viewer and said anatomical part e) a rule enforcement engine operative to modify said three-dimensional environment representation data in accordance with said determined anatomical part- virtual object relation and said virtual environment data.
16. A system according to claim 15, further comprising a miniaturized element perceived by said viewer in a viewer perceived space that is determined in accordance with said viewer's eye position, said miniaturized element configured for interaction with said displayed 3D virtual objects.
17. A system according to claim 16, wherein said miniaturized element is perceived by said viewer when said viewer's interacting anatomy part is outside of said viewer perceived space, and wherein said miniaturized element disappears when said viewer's interacting anatomy part is within said viewer perceived space.
18. An autostereoscopic (AS) display configured to display simultaneous perspectives of virtual objects of said 3D virtual environment representation data to spatially separated viewing zones located in front of said AS display, said AS display having an acoustic lens with electrically controlled refractive index to dynamically adapt said viewing zone locations.
19. A method for providing an interactive human computer interface to a viewer, said method comprising: a) storing three-dimensional virtual environment representation data including at least one three-dimensional virtual object within said virtual environment; b) displaying on an autostereoscopic (AS) display simultaneous perspectives of virtual objects in said 3D virtual environment representation data to spatially separated viewing zones located in front of said AS display c) tracking the anatomy of a viewer to determine respective real-world locations of: i) a viewer's eyes; and ii) an anatomical part on said viewer's body d) generating a registered 3D volume comprising 3D location data of said viewer's anatomical part and 3D location data of said virtual objects perceived by said viewer in accordance with said viewer's eye location, e) determining a relation between said virtual object and said anatomical part in accordance with said registration, f) modifying based on interactive rules said three-dimensional environment representation data in accordance with said determined anatomical part- virtual object relation and said virtual environment data.
20. A system for interactive human computer interface, said system comprising: a self-contained autostereoscopic (AS) display configured to render 3D virtual objects into neighboring viewing zones associated with said display, an eye location tracking system, comprising at least one 3D video camera, for continuously determining:
1) a viewer perceived three dimensional space in relation to said display, and
2) a 3D mapping of said rendered virtual objects in said perceived space in accordance with viewer eyes position in relation to said fixed viewing zones an anatomy location and configuration system, comprising at least one 3D video camera, for continuously determining a 3D mapping of viewer anatomy in relation to said display, and an interactive application that defines interactive rules and displayed content to said user, and an interaction processing engine configured to receive information from
4) said eye location tracking system
5) said anatomy location and configuration system, and
6) said interactive application thereby to determine interaction data of said viewer anatomy with said rendered virtual objects from AS display.
21. A system or method for three dimensional interaction with auto stereoscopic displays substantially as described or illustrated hereinabove or in any of the drawings.
PCT/IL2008/000530 2007-04-26 2008-04-17 A method and apparatus for three dimensional interaction with autosteroscopic displays WO2008132724A1 (en)

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
US92400307P 2007-04-26 2007-04-26
US60/924,003 2007-04-26
US93542607P 2007-08-13 2007-08-13
US60/935,426 2007-08-13

Publications (2)

Publication Number Publication Date
WO2008132724A1 true WO2008132724A1 (en) 2008-11-06
WO2008132724A4 WO2008132724A4 (en) 2008-12-31

Family

ID=39587846

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/IL2008/000530 WO2008132724A1 (en) 2007-04-26 2008-04-17 A method and apparatus for three dimensional interaction with autosteroscopic displays

Country Status (1)

Country Link
WO (1) WO2008132724A1 (en)

Cited By (38)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
GB2470072A (en) * 2009-05-08 2010-11-10 Sony Comp Entertainment Europe Virtual object movement in response to real object movement
KR20110052771A (en) * 2009-11-13 2011-05-19 엘지전자 주식회사 Image display device and operating method for the same
EP2372512A1 (en) * 2010-03-30 2011-10-05 Harman Becker Automotive Systems GmbH Vehicle user interface unit for a vehicle electronic device
CN102238408A (en) * 2010-05-03 2011-11-09 汤姆森特许公司 Method for displaying a settings menu and corresponding device
CN102270090A (en) * 2010-06-07 2011-12-07 索尼公司 Information display device and display image control method
CN102640502A (en) * 2009-10-14 2012-08-15 诺基亚公司 Autostereoscopic rendering and display apparatus
EP2499819A2 (en) * 2009-11-12 2012-09-19 LG Electronics Inc. Image display apparatus and image display method thereof
EP2502424A2 (en) * 2009-11-16 2012-09-26 LG Electronics Inc. Image display apparatus and operating method thereof
EP2521007A1 (en) * 2011-05-03 2012-11-07 Technische Universität Dresden Method for object 3D position identification based on multiple image analysis applying to gaze tracking
CN102810028A (en) * 2011-06-01 2012-12-05 时代光电科技股份有限公司 Touch device for virtual images floating in air
WO2013018099A2 (en) 2011-08-04 2013-02-07 Eyesight Mobile Technologies Ltd. System and method for interfacing with a device via a 3d display
GB2497612A (en) * 2012-01-03 2013-06-19 Liang Kong Three-dimensional display system using a plurality of projectors
WO2013094841A1 (en) * 2011-12-23 2013-06-27 한국과학기술연구원 Device for displaying multi-view 3d image using dynamic visual field expansion applicable to multiple observers and method for same
US8508347B2 (en) 2010-06-24 2013-08-13 Nokia Corporation Apparatus and method for proximity based input
CN103294260A (en) * 2012-04-02 2013-09-11 微软公司 Touch sensitive user interface
DE102012209917A1 (en) * 2012-06-13 2013-12-19 Technische Universität Dresden Method for transforming real-world visual information in virtual three-dimensional environments, involves time-synchronized recording of raw data for test person in real world system with detection of head movements of test person
US8631487B2 (en) 2010-12-16 2014-01-14 Research In Motion Limited Simple algebraic and multi-layer passwords
US8635676B2 (en) 2010-12-16 2014-01-21 Blackberry Limited Visual or touchscreen password entry
US8650635B2 (en) 2010-12-16 2014-02-11 Blackberry Limited Pressure sensitive multi-layer passwords
US8650624B2 (en) 2010-12-16 2014-02-11 Blackberry Limited Obscuring visual login
US8661530B2 (en) 2010-12-16 2014-02-25 Blackberry Limited Multi-layer orientation-changing password
US8745694B2 (en) 2010-12-16 2014-06-03 Research In Motion Limited Adjusting the position of an endpoint reference for increasing security during device log-on
US8769668B2 (en) 2011-05-09 2014-07-01 Blackberry Limited Touchscreen password entry
US8769641B2 (en) 2010-12-16 2014-07-01 Blackberry Limited Multi-layer multi-point or pathway-based passwords
US8863271B2 (en) 2010-12-16 2014-10-14 Blackberry Limited Password entry using 3D image with spatial alignment
US8931083B2 (en) 2010-12-16 2015-01-06 Blackberry Limited Multi-layer multi-point or randomized passwords
US9064104B2 (en) 2009-06-18 2015-06-23 Blackberry Limited Graphical authentication
US9135426B2 (en) 2010-12-16 2015-09-15 Blackberry Limited Password entry using moving images
EP2831850A4 (en) * 2012-03-30 2015-11-25 Intel Corp Intelligent depth control
US9223948B2 (en) 2011-11-01 2015-12-29 Blackberry Limited Combined passcode and activity launch modifier
US9258123B2 (en) 2010-12-16 2016-02-09 Blackberry Limited Multi-layered color-sensitive passwords
WO2016102948A1 (en) * 2014-12-24 2016-06-30 University Of Hertfordshire Higher Education Corporation Coherent touchless interaction with stereoscopic 3d images
CN105874408A (en) * 2014-01-03 2016-08-17 哈曼国际工业有限公司 Gesture interactive wearable spatial audio system
WO2016182502A1 (en) * 2015-05-14 2016-11-17 Medha Dharmatilleke Multi purpose mobile device case/cover integrated with a camera system & non electrical 3d/multiple video & still frame viewer for 3d and/or 2d high quality videography, photography and selfie recording
US9930128B2 (en) 2011-09-30 2018-03-27 Nokia Technologies Oy Method and apparatus for accessing a virtual object
CN112584080A (en) * 2016-09-09 2021-03-30 谷歌有限责任公司 Three-dimensional telepresence terminal and method
EP3971685A4 (en) * 2019-05-14 2022-06-29 Guangdong Oppo Mobile Telecommunications Corp., Ltd. Interactive control method and apparatus, electronic device and storage medium
NL2030326B1 (en) * 2021-12-29 2023-07-04 Dimenco Holding B V Autostereoscopic display device having a remote body tracking system

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20040036763A1 (en) * 1994-11-14 2004-02-26 Swift David C. Intelligent method and system for producing and displaying stereoscopically-multiplexed images of three-dimensional objects for use in realistic stereoscopic viewing thereof in interactive virtual reality display environments
US20040160389A1 (en) * 1996-01-17 2004-08-19 Nippon Telegraph And Telephone Corporation Optical device and three-dimensional display device
US20040223218A1 (en) * 1999-12-08 2004-11-11 Neurok Llc Visualization of three dimensional images and multi aspect imaging
US20050264527A1 (en) * 2002-11-06 2005-12-01 Lin Julius J Audio-visual three-dimensional input/output
US20060012675A1 (en) * 2004-05-10 2006-01-19 University Of Southern California Three dimensional interaction with autostereoscopic displays

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20040036763A1 (en) * 1994-11-14 2004-02-26 Swift David C. Intelligent method and system for producing and displaying stereoscopically-multiplexed images of three-dimensional objects for use in realistic stereoscopic viewing thereof in interactive virtual reality display environments
US20040160389A1 (en) * 1996-01-17 2004-08-19 Nippon Telegraph And Telephone Corporation Optical device and three-dimensional display device
US20040223218A1 (en) * 1999-12-08 2004-11-11 Neurok Llc Visualization of three dimensional images and multi aspect imaging
US20050264527A1 (en) * 2002-11-06 2005-12-01 Lin Julius J Audio-visual three-dimensional input/output
US20060012675A1 (en) * 2004-05-10 2006-01-19 University Of Southern California Three dimensional interaction with autostereoscopic displays

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
ALPASLAN ZAHIR Y ET AL: "Three-dimensional interaction with autostereoscopic displays", PROCEEDINGS OF THE SPIE, SPIE, BELLINGHAM, VA, vol. 5291, 19 January 2004 (2004-01-19), pages 227 - 236, XP007905371, ISSN: 0277-786X *
PASTOOR S ET AL: "17.D: INVITED PAPER: AUTOSTEREOSCOPIC USER-COMPUTER INTERFACE WITH VISUALLY CONTROLLED INTERACTIONS", 1997 SID INTERNATIONAL SYMPOSIUM DIGEST OF TECHNICAL PAPERS. BOSTON, MAY 13 - 15, 1997; [SID INTERNATIONAL SYMPOSIUM DIGEST OF TECHNICAL PAPERS], SANTA ANA, SID, US, vol. VOL. 28, 13 May 1997 (1997-05-13), pages 277 - 280, XP000722705 *

Cited By (75)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
GB2470072A (en) * 2009-05-08 2010-11-10 Sony Comp Entertainment Europe Virtual object movement in response to real object movement
GB2470072B (en) * 2009-05-08 2014-01-01 Sony Comp Entertainment Europe Entertainment device,system and method
US9064104B2 (en) 2009-06-18 2015-06-23 Blackberry Limited Graphical authentication
US10176315B2 (en) 2009-06-18 2019-01-08 Blackberry Limited Graphical authentication
US10325086B2 (en) 2009-06-18 2019-06-18 Blackberry Limited Computing device with graphical authentication interface
CN102640502A (en) * 2009-10-14 2012-08-15 诺基亚公司 Autostereoscopic rendering and display apparatus
US8970478B2 (en) 2009-10-14 2015-03-03 Nokia Corporation Autostereoscopic rendering and display apparatus
US8803873B2 (en) 2009-11-12 2014-08-12 Lg Electronics Inc. Image display apparatus and image display method thereof
EP2499819A2 (en) * 2009-11-12 2012-09-19 LG Electronics Inc. Image display apparatus and image display method thereof
EP2499819A4 (en) * 2009-11-12 2014-04-16 Lg Electronics Inc Image display apparatus and image display method thereof
EP2499834A2 (en) * 2009-11-13 2012-09-19 LG Electronics Inc. Image display apparatus and operating method thereof
CN102598679A (en) * 2009-11-13 2012-07-18 Lg电子株式会社 Image display apparatus and operating method thereof
EP2499834A4 (en) * 2009-11-13 2013-09-11 Lg Electronics Inc Image display apparatus and operating method thereof
KR101647722B1 (en) * 2009-11-13 2016-08-23 엘지전자 주식회사 Image Display Device and Operating Method for the Same
WO2011059261A2 (en) 2009-11-13 2011-05-19 Lg Electronics Inc. Image display apparatus and operating method thereof
KR20110052771A (en) * 2009-11-13 2011-05-19 엘지전자 주식회사 Image display device and operating method for the same
EP2502424A2 (en) * 2009-11-16 2012-09-26 LG Electronics Inc. Image display apparatus and operating method thereof
EP2502424A4 (en) * 2009-11-16 2014-08-27 Lg Electronics Inc Image display apparatus and operating method thereof
EP2372512A1 (en) * 2010-03-30 2011-10-05 Harman Becker Automotive Systems GmbH Vehicle user interface unit for a vehicle electronic device
CN102238408A (en) * 2010-05-03 2011-11-09 汤姆森特许公司 Method for displaying a settings menu and corresponding device
EP2393299A3 (en) * 2010-06-07 2014-10-15 Sony Corporation Information display device and display image control method
CN102270090A (en) * 2010-06-07 2011-12-07 索尼公司 Information display device and display image control method
US8508347B2 (en) 2010-06-24 2013-08-13 Nokia Corporation Apparatus and method for proximity based input
US8661530B2 (en) 2010-12-16 2014-02-25 Blackberry Limited Multi-layer orientation-changing password
US8745694B2 (en) 2010-12-16 2014-06-03 Research In Motion Limited Adjusting the position of an endpoint reference for increasing security during device log-on
US8631487B2 (en) 2010-12-16 2014-01-14 Research In Motion Limited Simple algebraic and multi-layer passwords
US8635676B2 (en) 2010-12-16 2014-01-21 Blackberry Limited Visual or touchscreen password entry
US8650635B2 (en) 2010-12-16 2014-02-11 Blackberry Limited Pressure sensitive multi-layer passwords
US8650624B2 (en) 2010-12-16 2014-02-11 Blackberry Limited Obscuring visual login
US9258123B2 (en) 2010-12-16 2016-02-09 Blackberry Limited Multi-layered color-sensitive passwords
US10621328B2 (en) 2010-12-16 2020-04-14 Blackberry Limited Password entry using 3D image with spatial alignment
US8931083B2 (en) 2010-12-16 2015-01-06 Blackberry Limited Multi-layer multi-point or randomized passwords
US8863271B2 (en) 2010-12-16 2014-10-14 Blackberry Limited Password entry using 3D image with spatial alignment
US9135426B2 (en) 2010-12-16 2015-09-15 Blackberry Limited Password entry using moving images
US8769641B2 (en) 2010-12-16 2014-07-01 Blackberry Limited Multi-layer multi-point or pathway-based passwords
EP2521007A1 (en) * 2011-05-03 2012-11-07 Technische Universität Dresden Method for object 3D position identification based on multiple image analysis applying to gaze tracking
US8769668B2 (en) 2011-05-09 2014-07-01 Blackberry Limited Touchscreen password entry
CN102810028A (en) * 2011-06-01 2012-12-05 时代光电科技股份有限公司 Touch device for virtual images floating in air
US10606442B2 (en) 2011-08-04 2020-03-31 Eyesight Mobile Technologies, LTD. Touch-free gesture recognition system and method
WO2013018099A3 (en) * 2011-08-04 2013-07-04 Eyesight Mobile Technologies Ltd. System and method for interfacing with a device via a 3d display
CN109271029A (en) * 2011-08-04 2019-01-25 视力移动技术有限公司 For activating one or more devices for activating object in 3D rendering
WO2013018099A2 (en) 2011-08-04 2013-02-07 Eyesight Mobile Technologies Ltd. System and method for interfacing with a device via a 3d display
CN103858074A (en) * 2011-08-04 2014-06-11 视力移动技术有限公司 System and method for interfacing with a device via a 3d display
CN109271029B (en) * 2011-08-04 2022-08-26 视力移动技术有限公司 Touchless gesture recognition system, touchless gesture recognition method, and medium
CN103858074B (en) * 2011-08-04 2018-10-19 视力移动技术有限公司 The system and method interacted with device via 3D display device
US9733789B2 (en) 2011-08-04 2017-08-15 Eyesight Mobile Technologies Ltd. Interfacing with a device via virtual 3D objects
US9930128B2 (en) 2011-09-30 2018-03-27 Nokia Technologies Oy Method and apparatus for accessing a virtual object
US9223948B2 (en) 2011-11-01 2015-12-29 Blackberry Limited Combined passcode and activity launch modifier
US10237543B2 (en) 2011-12-23 2019-03-19 Samsung Electronics Co., Ltd. Device for displaying multi-view 3D image using dynamic viewing zone expansion applicable to multiple observers and method for same
WO2013094841A1 (en) * 2011-12-23 2013-06-27 한국과학기술연구원 Device for displaying multi-view 3d image using dynamic visual field expansion applicable to multiple observers and method for same
GB2498184A (en) * 2012-01-03 2013-07-10 Liang Kong Interactive autostereoscopic three-dimensional display
GB2497612A (en) * 2012-01-03 2013-06-19 Liang Kong Three-dimensional display system using a plurality of projectors
GB2497612B (en) * 2012-01-03 2013-11-27 Liang Kong Three dimensional display system
US9503712B2 (en) 2012-01-03 2016-11-22 Liang Kong Three dimensional display system
US9807362B2 (en) 2012-03-30 2017-10-31 Intel Corporation Intelligent depth control
EP2831850A4 (en) * 2012-03-30 2015-11-25 Intel Corp Intelligent depth control
CN103294260A (en) * 2012-04-02 2013-09-11 微软公司 Touch sensitive user interface
WO2013151947A1 (en) * 2012-04-02 2013-10-10 Ambrus Anthony J Touch sensitive user interface
US8933912B2 (en) 2012-04-02 2015-01-13 Microsoft Corporation Touch sensitive user interface with three dimensional input sensor
DE102012209917A1 (en) * 2012-06-13 2013-12-19 Technische Universität Dresden Method for transforming real-world visual information in virtual three-dimensional environments, involves time-synchronized recording of raw data for test person in real world system with detection of head movements of test person
CN105874408A (en) * 2014-01-03 2016-08-17 哈曼国际工业有限公司 Gesture interactive wearable spatial audio system
US10126823B2 (en) 2014-01-03 2018-11-13 Harman International Industries, Incorporated In-vehicle gesture interactive spatial audio system
EP2891955B1 (en) * 2014-01-03 2020-04-08 Harman International Industries, Incorporated In-vehicle gesture interactive spatial audio system
US10585486B2 (en) 2014-01-03 2020-03-10 Harman International Industries, Incorporated Gesture interactive wearable spatial audio system
WO2016102948A1 (en) * 2014-12-24 2016-06-30 University Of Hertfordshire Higher Education Corporation Coherent touchless interaction with stereoscopic 3d images
GB2533777A (en) * 2014-12-24 2016-07-06 Univ Of Hertfordshire Higher Education Corp Coherent touchless interaction with steroscopic 3D images
WO2016182502A1 (en) * 2015-05-14 2016-11-17 Medha Dharmatilleke Multi purpose mobile device case/cover integrated with a camera system & non electrical 3d/multiple video & still frame viewer for 3d and/or 2d high quality videography, photography and selfie recording
US11057505B2 (en) 2015-05-14 2021-07-06 Medha Dharmatilleke Multi purpose mobile device case/cover integrated with a camera system and non electrical 3D/multiple video and still frame viewer for 3D and/or 2D high quality videography, photography and selfie recording
WO2016182503A1 (en) * 2015-05-14 2016-11-17 Medha Dharmatilleke Multi purpose mobile device case/cover integrated with a camera system & non-electrical 3d/multiple video & still frame viewer for 3d and/or 2d high quality videography, photography and selfie recording
US11606449B2 (en) 2015-05-14 2023-03-14 Medha Dharmatilleke Mobile phone/device case or cover having a 3D camera
CN112584080A (en) * 2016-09-09 2021-03-30 谷歌有限责任公司 Three-dimensional telepresence terminal and method
CN112584080B (en) * 2016-09-09 2023-10-24 谷歌有限责任公司 Three-dimensional telepresence terminal and method
EP3971685A4 (en) * 2019-05-14 2022-06-29 Guangdong Oppo Mobile Telecommunications Corp., Ltd. Interactive control method and apparatus, electronic device and storage medium
NL2030326B1 (en) * 2021-12-29 2023-07-04 Dimenco Holding B V Autostereoscopic display device having a remote body tracking system
WO2023128762A1 (en) * 2021-12-29 2023-07-06 Dimenco Holding B.V. Autostereoscopic display device having a remote body tracking system

Also Published As

Publication number Publication date
WO2008132724A4 (en) 2008-12-31

Similar Documents

Publication Publication Date Title
WO2008132724A1 (en) A method and apparatus for three dimensional interaction with autosteroscopic displays
CN109196447B (en) Interaction with 3D virtual objects using gestures and multi-DOF controllers
CA3023488C (en) System and method for generating a progressive representation associated with surjectively mapped virtual and physical reality image data
US9910509B2 (en) Method to control perspective for a camera-controlled computer
US7796134B2 (en) Multi-plane horizontal perspective display
US20100128112A1 (en) Immersive display system for interacting with three-dimensional content
CN117032450A (en) Method for manipulating objects in an environment
US9986228B2 (en) Trackable glasses system that provides multiple views of a shared display
US20050219695A1 (en) Horizontal perspective display
US20240037880A1 (en) Artificial Reality System with Varifocal Display of Artificial Reality Content
EP3106963B1 (en) Mediated reality
US10116914B2 (en) Stereoscopic display
US10652525B2 (en) Quad view display system
US20060221071A1 (en) Horizontal perspective display
KR20200138349A (en) Image processing method and apparatus, electronic device, and storage medium
US20060250390A1 (en) Horizontal perspective display
EP3642692A1 (en) System and method for interacting with a user via a mirror
US20170293412A1 (en) Apparatus and method for controlling the apparatus
US11212502B2 (en) Method of modifying an image on a computational device
US20220256137A1 (en) Position calculation system
EP4233310A1 (en) Dynamic resolution of depth conflicts in telepresence
WO2022158328A1 (en) Information processing apparatus, information processing method, and program
He Volume Visualization in Projection-Based Virtual Environments: Interaction and Exploration Tools Design and Evaluation
NZ786547A (en) Interactions with 3d virtual objects using poses and multiple-dof controllers
WO2006121955A2 (en) Horizontal perspective display

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 08738232

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

32PN Ep: public notification in the ep bulletin as address of the adressee cannot be established

Free format text: NOTING OF LOSS OF RIGHTS EPO FORM 1205A DATED 20.01.2010.

122 Ep: pct application non-entry in european phase

Ref document number: 08738232

Country of ref document: EP

Kind code of ref document: A1