SlideShare uma empresa Scribd logo
1 de 90
Baixar para ler offline
Hands and Speech in Space:
Multimodal Interaction for AR
Mark Billinghurst
mark.billinghurst@hitlabnz.org
The HIT Lab NZ, University of Canterbury
December 12th 2013
1977 – Star Wars

1977 – Star Wars
Augmented Reality Definition
  Defining Characteristics
  Combines Real and Virtual Images
-  Both can be seen at the same time

  Interactive in real-time
-  The virtual content can be interacted with

  Registered in 3D
-  Virtual objects appear fixed in space

Azuma, R. T. (1997). A survey of augmented reality. Presence, 6(4), 355-385.
Augmented Reality Today
AR Interface Components
Physical
Elements
Input

Interaction
Metaphor

Virtual
Elements
Output

  Key Question: How should a person interact
with the Augmented Reality content?
  Connecting physical and virtual with interaction
AR Interaction Metaphors
  Information Browsing
  View AR content

  3D AR Interfaces
  3D UI interaction techniques

  Augmented Surfaces
  Tangible UI techniques

  Tangible AR
  Tangible UI input + AR output
VOMAR Demo (Kato 2000)
  AR Furniture Arranging
  Elements + Interactions
  Book:
-  Turn over the page

  Paddle:
-  Push, shake, incline, hit, scoop

Kato, H., Billinghurst, M., et al. 2000. Virtual Object Manipulation on a Table-Top AR
Environment. In Proceedings of the International Symposium on Augmented Reality
(ISAR 2000), Munich, Germany, 111--119.
Opportunities for Multimodal Input
  Multimodal interfaces are a natural fit for AR
  Need for non-GUI interfaces
  Natural interaction with real world
  Natural support for body input
  Previous work shown value of multimodal input
and 3D graphics
Related Work
  Related work in 3D graphics/VR
  Interaction with 3D content [Chu 1997]
  Navigating through virtual worlds [Krum 2002]
  Interacting with virtual characters [Billinghurst 1998]

  Little earlier work in AR
  Require additional input devices
  Few formal usability studies
  Eg Olwal et. al [2003] Sense Shapes
Examples

SenseShapes [2003]

Kolsch [2006]
Marker Based Multimodal Interface

  Add speech recognition to VOMAR
  Paddle + speech commands
Irawati, S., Green, S., Billinghurst, M., Duenser, A., & Ko, H. (2006, October). IEEE Xplore. In Mixed
and Augmented Reality, 2006. ISMAR 2006. IEEE/ACM International Symposium on (pp. 183-186).
IEEE.
Commands Recognized
  Create Command "Make a blue chair": to create a virtual
object and place it on the paddle.
  Duplicate Command "Copy this": to duplicate a virtual object
and place it on the paddle.
  Grab Command "Grab table": to select a virtual object and
place it on the paddle.
  Place Command "Place here": to place the attached object in
the workspace.
  Move Command "Move the couch": to attach a virtual object
in the workspace to the paddle so that it follows the paddle
movement.
System Architecture
Object Relationships

"Put chair behind the table”
Where is behind?
View specific regions
User Evaluation
  Performance time
  Speech + static paddle significantly faster

  Gesture-only condition less accurate for position/orientation
  Users preferred speech + paddle input
Subjective Surveys
2012 – Iron Man 2
To Make the Vision Real..
  Hardware/software requirements
  Contact lens displays
  Free space hand/body tracking
  Speech/gesture recognition
  Etc..

  Most importantly
  Usability/User Experience
Natural Interaction
  Automatically detecting real environment
  Environmental awareness, Physically based interaction

  Gesture interaction
  Free-hand interaction

  Multimodal input
  Speech and gesture interaction

  Intelligent interfaces
  Implicit rather than Explicit interaction
Environmental
Awareness
AR MicroMachines
  AR experience with environment awareness
and physically-based interaction
  Based on MS Kinect RGB-D sensor

  Augmented environment supports
  occlusion, shadows
  physically-based interaction between real and
virtual objects
Clark, A., & Piumsomboon, T. (2011). A realistic augmented reality racing game using a
depth-sensing camera. In Proceedings of the 10th International Conference on Virtual
Reality Continuum and Its Applications in Industry (pp. 499-502). ACM.
Operating Environment
Architecture
  Our framework uses five libraries:
  OpenNI
  OpenCV
  OPIRA
  Bullet Physics
  OpenSceneGraph
System Flow
  The system flow consists of three sections:
  Image Processing and Marker Tracking
  Physics Simulation
  Rendering
Physics Simulation

  Create virtual mesh over real world
  Update at 10 fps – can move real objects
  Use by physics engine for collision detection (virtual/real)
  Use by OpenScenegraph for occlusion and shadows
Rendering

Occlusion

Shadows
Gesture Interaction
Natural Hand Interaction

  Using bare hands to interact with AR content
  MS Kinect depth sensing
  Real time hand tracking
  Physics based simulation model
Hand Interaction

  Represent models as collections of spheres
  Bullet physics engine for interaction with real world
Scene Interaction

  Render AR scene with OpenSceneGraph
  Using depth map for occlusion
  Shadows yet to be implemented
Architecture
5. Gesture

•  Static Gestures
•  Dynamic Gestures
•  Context based Gestures
4. Modeling

•  Hand recognition/modeling
•  Rigid-body modeling
3. Classification/Tracking
2. Segmentation
1. Hardware Interface
Architecture
5. Gesture
•  Static Gestures
•  Dynamic Gestures
•  Context based Gestures

o  Supports PCL, OpenNI, OpenCV, and Kinect SDK.
o  Provides access to depth, RGB, XYZRGB.
o  Usage: Capturing color image, depth image and concatenated
point clouds from a single or multiple cameras
o  For example:

4. Modeling
•  Hand recognition/
modeling
•  Rigid-body modeling
3. Classification/Tracking
2. Segmentation
1. Hardware Interface

Kinect for Xbox 360
Kinect for Windows
Asus Xtion Pro Live
Architecture
5. Gesture
•  Static Gestures
•  Dynamic Gestures
•  Context based Gestures

o  Segment images and point clouds based on color, depth and
space.
o  Usage: Segmenting images or point clouds using color
models, depth, or spatial properties such as location, shape
and size.
o  For example:

4. Modeling
•  Hand recognition/
modeling
•  Rigid-body modeling

Skin color segmentation

3. Classification/Tracking
2. Segmentation
1. Hardware Interface

Depth threshold
Architecture
5. Gesture
•  Static Gestures
•  Dynamic Gestures
•  Context based Gestures

o  Identify and track objects between frames based on
XYZRGB.
o  Usage: Identifying current position/orientation of the tracked
object in space.
o  For example:

4. Modeling
•  Hand recognition/
modeling
•  Rigid-body modeling
3. Classification/Tracking
2. Segmentation
1. Hardware Interface

Training set of hand
poses, colors
represent unique
regions of the hand.
Raw output (withoutcleaning) classified
on real hand input
(depth image).
Architecture
5. Gesture
•  Static Gestures
•  Dynamic Gestures
•  Context based Gestures
4. Modeling
•  Hand recognition/
modeling
•  Rigid-body modeling
3. Classification/Tracking
2. Segmentation
1. Hardware Interface

o  Hand Recognition/Modeling
  Skeleton based (for low resolution
approximation)
  Model based (for more accurate
representation)
o  Object Modeling (identification and tracking rigidbody objects)
o  Physical Modeling (physical interaction)
  Sphere Proxy
  Model based
  Mesh based
o  Usage: For general spatial interaction in AR/VR
environment
Architecture
5. Gesture
•  Static Gestures
•  Dynamic Gestures
•  Context based Gestures
4. Modeling
•  Hand recognition/
modeling
•  Rigid-body modeling
3. Classification/Tracking
2. Segmentation
1. Hardware Interface

o  Static (hand pose recognition)
o  Dynamic (meaningful movement recognition)
o  Context-based gesture recognition (gestures with context,
e.g. pointing)
o  Usage: Issuing commands/anticipating user intention and high
level interaction.
Skeleton Based Interaction

  3 Gear Systems
  Kinect/Primesense Sensor
  Two hand tracking
  http://www.threegear.com
Skeleton Interaction + AR

  HMD AR View
  Viewpoint tracking

  Two hand input
  Skeleton interaction, occlusion
What Gestures do People Want to Use?
  Limitations of Previous work in AR
  Limited range of gestures
  Gestures designed for optimal recognition
  Gestures studied as add-on to speech

  Solution – elicit desired gestures from users
  Eg. Gestures for surface computing [Wobbrock]
  Previous work in unistroke getsures, mobile gestures
User Defined Gesture Study
  Use AR view
  HMD + AR tracking

  Present AR animations
  40 tasks in six categories
-  Editing, transforms, menu, etc

  Ask users to produce
gestures causing animations
  Record gesture (video, depth)
Piumsomboon, T., Clark, A., Billinghurst, M., & Cockburn, A. (2013, April). User-defined gestures for augmented
reality. In CHI'13 Extended Abstracts on Human Factors in Computing Systems (pp. 955-960).ACM
Data Recorded
  20 participants
  Gestures recorded (video, depth data)
  800 gestures from 40 tasks

  Subjective rankings
  Likert ranking of goodness, ease of use

  Think aloud transcripts
Typical Gestures
Results - Gestures
  Gestures grouped according to
similarity – 320 groups
  44 consensus (62% all gestures)
  276 low similarity (discarded)
  11 hand poses seen

  Degree of consensus (A) using
guessability score [Wobbrock]
Results –Agreement Scores

Red line – proportion of two handed gestures
Usability Results
Consensus Discarded
Ease of Performance
6.02
5.50
Good Match
6.17
5.83
Likert Scale [1-7], 7 = Very Good

  Significant difference between consensus and
discarded gesture sets (p < 0.0001)
  Gestures in consensus set better than discarded
gestures in perceived performance and goodness
Lessons Learned
  AR animation can elicit desired gestures
  For some tasks there is a high degree of
similarity in user defined gestures
  Especially command gestures (eg Open), select

  Less agreement in manipulation gestures
  Move (40%), rotate (30%), grouping (10%)

  Small portion of two handed gestures (22%)
  Scaling, group selection
Multimodal Input
Multimodal Interaction
  Combined speech input
  Gesture and Speech complimentary
  Speech
-  modal commands, quantities

  Gesture
-  selection, motion, qualities

  Previous work found multimodal interfaces
intuitive for 2D/3D graphics interaction
Wizard of Oz Study
  What speech and gesture input
would people like to use?
  Wizard
  Perform speech recognition
  Command interpretation

  Domain
  3D object interaction/modelling
Lee, M., & Billinghurst, M. (2008, October). A Wizard of Oz study for an AR multimodal interface.
In Proceedings of the 10th international conference on Multimodal interfaces (pp. 249-256). ACM.
System Architecture
Hand Segmentation
System Set Up
Experiment
  12 participants
  Two display conditions (HMD vs. Desktop)
  Three tasks
  Task 1: Change object color/shape
  Task 2: 3D positioning of obejcts
  Task 3: Scene assembly
Key Results

  Most commands multimodal

  Multimodal (63%), Gesture (34%), Speech (4%)

  Most spoken phrases short
  74% phrases average 1.25 words long
  Sentences (26%) average 3 words

  Main gestures deictic (65%), metaphoric (35%)
  In multimodal commands gesture issued first
  94% time gesture begun before speech
  Multimodal window 8s – speech 4.5s after gesture
Free Hand Multimodal Input

Point

Move

Pick/Drop

  Use free hand to interact with AR content
  Recognize simple gestures
  Open hand, closed hand, pointing
Lee, M., Billinghurst, M., Baek, W., Green, R., & Woo, W. (2013). A usability study of multimodal
input in an augmented reality environment. Virtual Reality, 17(4), 293-305.
Speech Input
  MS Speech + MS SAPI (> 90% accuracy)
  Single word speech commands
Multimodal Architecture
Multimodal Fusion
Hand Occlusion
Experimental Setup

Change object shape
and colour
User Evaluation

  25 subjects, 10 task trials x 3, 3 conditions
  Change object shape, colour and position
  Conditions
  Speech only, gesture only, multimodal

  Measures
  performance time, errors (system/user), subjective survey
Results - Performance
  Average performance time
  Gesture: 15.44s
  Speech: 12.38s
  Multimodal: 11.78s

  Significant difference across conditions (p < 0.01)
  Difference between gesture and speech/MMI
Errors
  User errors – errors per task
  Gesture (0.50), Speech (0.41), MMI (0.42)
  No significant difference

  System errors
  Speech accuracy – 94%, Gesture accuracy – 85%
  MMI accuracy – 90%
Subjective Results (Likert 1-7)
Gesture

Speech

MMI

Naturalness

4.60

5.60

5.80

Ease of Use

4.00

5.90

6.00

Efficiency

4.45

5.15

6.05

Physical Effort

4.75

3.15

3.85

  User subjective survey
  Gesture significantly worse, MMI and Speech same
  MMI perceived as most efficient

  Preference
  70% MMI, 25% speech only, 5% gesture only
Observations
  Significant difference in number of commands
  Gesture (6.14), Speech (5.23), MMI (4.93)

  MMI Simultaneous vs. Sequential commands
  79% sequential, 21% simultaneous

  Reaction to system errors
  Almost always repeated same command
  In MMI rarely changes modalities
Lessons Learned
  Multimodal interaction significantly better than
gesture alone in AR interfaces for 3D tasks
  Short task time, more efficient

  Users felt that MMI was more natural, easier,
and more effective that gesture/speech only
  Simultaneous input rarely used
  More studies need to be conducted
Intelligent Interfaces
Intelligent Interfaces
  Most AR systems stupid
  Don’t recognize user behaviour
  Don’t provide feedback
  Don’t adapt to user

  Especially important for training
  Scaffolded learning
  Moving beyond check-lists of actions
Intelligent Interfaces

  AR interface + intelligent tutoring system
  ASPIRE constraint based system (from UC)
  Constraints
-  relevance cond., satisfaction cond., feedback
Westerfield, G., Mitrovic, A., & Billinghurst, M. (2013). Intelligent Augmented Reality Training for
Assembly Tasks. In Artificial Intelligence in Education (pp. 542-551). Springer Berlin Heidelberg.
Domain Ontology
Intelligent Feedback

  Actively monitors user behaviour
  Implicit vs. explicit interaction

  Provides corrective feedback
Evaluation Results
  16 subjects, with and without ITS
  Improved task completion

  Improved learning
Intelligent Agents
  AR characters
  Virtual embodiment of system
  Multimodal input/output

  Examples
  AR Lego, Welbo, etc
  Mr Virtuoso
-  AR character more real, more fun
-  On-screen 3D and AR similar in usefulness
Wagner, D., Billinghurst, M., & Schmalstieg, D. (2006). How real should virtual characters be?. In
Proceedings of the 2006 ACM SIGCHI international conference on Advances in computer
entertainment technology (p. 57). ACM.
Looking to the Future

What’s Next?
Directions for Future Research
  Mobile Gesture Interaction
  Tablet, phone interfaces

  Wearable Systems
  Google Glass

  Novel Displays
  Contact lens

  Environmental Understanding
  Semantic representation
Mobile Gesture Interaction
  Motivation
  Richer interaction with handheld devices
  Natural interaction with handheld AR

  2D tracking
  Finger tip tracking

  3D tracking
[Hurst and Wezel 2013]

  Hand tracking
[Henrysson et al. 2007]

Henrysson, A., Marshall, J., & Billinghurst, M. (2007). Experiments in 3D interaction for mobile phone AR.
In Proceedings of the 5th international conference on Computer graphics and interactive techniques in
Australia and Southeast Asia (pp. 187-194). ACM.
Fingertip Based Interaction

Running System
System Setup

Mobile Client + PC Server

Bai, H., Gao, L., El-Sana, J., & Billinghurst, M. (2013). Markerless 3D gesture-based interaction for
handheld augmented reality interfaces. In SIGGRAPH Asia 2013 Symposium on Mobile Graphics
and Interactive Applications (p. 22). ACM.
System Architecture
3D Prototype System
  3 Gear + Vuforia
  Hand tracking + phone tracking

  Freehand interaction on phone
  Skeleton model
  3D interaction
  20 fps performance
Google Glass
User Experience
  Truly Wearable Computing
  Less than 46 ounces

  Hands-free Information Access
  Voice interaction, Ego-vision camera

  Intuitive User Interface
  Touch, Gesture, Speech, Head Motion

  Access to all Google Services
  Map, Search, Location, Messaging, Email, etc
Contact Lens Display

  Babak Parviz

  University Washington

  MEMS components
  Transparent elements
  Micro-sensors

  Challenges
  Miniaturization
  Assembly
  Eye-safe
Contact Lens Prototype
Environmental Understanding

  Semantic understanding of environment
  What are the key objects?
  What are there relationships?
  Represented in a form suitable for multimodal interaction?
Conclusion
Conclusions
  AR experiences need new interaction methods
  Enabling technologies are advancing quickly
  Displays, tracking, depth capture devices

  Natural user interfaces possible
  Free hand gesture, speech, intelligence interfaces

  Important research for the future
  Mobile, wearable, displays
More Information
•  Mark Billinghurst
–  Email: mark.billinghurst@hitlabnz.org
–  Twitter: @marknb00

•  Website
–  http://www.hitlabnz.org/

Mais conteúdo relacionado

Mais procurados

2013 Lecture4: Designing AR Interfaces
2013 Lecture4: Designing AR Interfaces2013 Lecture4: Designing AR Interfaces
2013 Lecture4: Designing AR InterfacesMark Billinghurst
 
COSC 426 Lect. 6: Collaborative AR
COSC 426 Lect. 6: Collaborative ARCOSC 426 Lect. 6: Collaborative AR
COSC 426 Lect. 6: Collaborative ARMark Billinghurst
 
426 lecture 7: Designing AR Interfaces
426 lecture 7: Designing AR Interfaces426 lecture 7: Designing AR Interfaces
426 lecture 7: Designing AR InterfacesMark Billinghurst
 
From Interaction to Understanding
From Interaction to UnderstandingFrom Interaction to Understanding
From Interaction to UnderstandingMark Billinghurst
 
Advanced Methods for User Evaluation in AR/VR Studies
Advanced Methods for User Evaluation in AR/VR StudiesAdvanced Methods for User Evaluation in AR/VR Studies
Advanced Methods for User Evaluation in AR/VR StudiesMark Billinghurst
 
Using Augmented Reality to Create Empathic Experiences
Using Augmented Reality to Create Empathic ExperiencesUsing Augmented Reality to Create Empathic Experiences
Using Augmented Reality to Create Empathic ExperiencesMark Billinghurst
 
Comp4010 Lecture9 VR Input and Systems
Comp4010 Lecture9 VR Input and SystemsComp4010 Lecture9 VR Input and Systems
Comp4010 Lecture9 VR Input and SystemsMark Billinghurst
 
426 lecture1: Introduction to AR
426 lecture1: Introduction to AR426 lecture1: Introduction to AR
426 lecture1: Introduction to ARMark Billinghurst
 
COMP Lecture1 - Introduction to Virtual Reality
COMP Lecture1 - Introduction to Virtual RealityCOMP Lecture1 - Introduction to Virtual Reality
COMP Lecture1 - Introduction to Virtual RealityMark Billinghurst
 
COMP 4010 Lecture 9 AR Interaction
COMP 4010 Lecture 9 AR InteractionCOMP 4010 Lecture 9 AR Interaction
COMP 4010 Lecture 9 AR InteractionMark Billinghurst
 
Comp 4010 2021 Lecture1-Introduction to XR
Comp 4010 2021 Lecture1-Introduction to XRComp 4010 2021 Lecture1-Introduction to XR
Comp 4010 2021 Lecture1-Introduction to XRMark Billinghurst
 
Comp4010 Lecture13 More Research Directions
Comp4010 Lecture13 More Research DirectionsComp4010 Lecture13 More Research Directions
Comp4010 Lecture13 More Research DirectionsMark Billinghurst
 
Building Usable AR Interfaces
Building Usable AR InterfacesBuilding Usable AR Interfaces
Building Usable AR InterfacesMark Billinghurst
 
Lecture 5: 3D User Interfaces for Virtual Reality
Lecture 5: 3D User Interfaces for Virtual RealityLecture 5: 3D User Interfaces for Virtual Reality
Lecture 5: 3D User Interfaces for Virtual RealityMark Billinghurst
 
Lecture 6 Interaction Design for VR
Lecture 6 Interaction Design for VRLecture 6 Interaction Design for VR
Lecture 6 Interaction Design for VRMark Billinghurst
 
COMP 4010 - Lecture1 Introduction to Virtual Reality
COMP 4010 - Lecture1 Introduction to Virtual RealityCOMP 4010 - Lecture1 Introduction to Virtual Reality
COMP 4010 - Lecture1 Introduction to Virtual RealityMark Billinghurst
 
Mixed Reality in the Workspace
Mixed Reality in the WorkspaceMixed Reality in the Workspace
Mixed Reality in the WorkspaceMark Billinghurst
 

Mais procurados (20)

2013 Lecture4: Designing AR Interfaces
2013 Lecture4: Designing AR Interfaces2013 Lecture4: Designing AR Interfaces
2013 Lecture4: Designing AR Interfaces
 
Alvaro Cassinelli / Meta Perception Group leader
Alvaro Cassinelli / Meta Perception Group leaderAlvaro Cassinelli / Meta Perception Group leader
Alvaro Cassinelli / Meta Perception Group leader
 
COSC 426 Lect. 6: Collaborative AR
COSC 426 Lect. 6: Collaborative ARCOSC 426 Lect. 6: Collaborative AR
COSC 426 Lect. 6: Collaborative AR
 
426 lecture 7: Designing AR Interfaces
426 lecture 7: Designing AR Interfaces426 lecture 7: Designing AR Interfaces
426 lecture 7: Designing AR Interfaces
 
From Interaction to Understanding
From Interaction to UnderstandingFrom Interaction to Understanding
From Interaction to Understanding
 
Suman
SumanSuman
Suman
 
Advanced Methods for User Evaluation in AR/VR Studies
Advanced Methods for User Evaluation in AR/VR StudiesAdvanced Methods for User Evaluation in AR/VR Studies
Advanced Methods for User Evaluation in AR/VR Studies
 
Using Augmented Reality to Create Empathic Experiences
Using Augmented Reality to Create Empathic ExperiencesUsing Augmented Reality to Create Empathic Experiences
Using Augmented Reality to Create Empathic Experiences
 
Comp4010 Lecture9 VR Input and Systems
Comp4010 Lecture9 VR Input and SystemsComp4010 Lecture9 VR Input and Systems
Comp4010 Lecture9 VR Input and Systems
 
426 lecture1: Introduction to AR
426 lecture1: Introduction to AR426 lecture1: Introduction to AR
426 lecture1: Introduction to AR
 
Spatial computing - extending reality
Spatial computing - extending realitySpatial computing - extending reality
Spatial computing - extending reality
 
COMP Lecture1 - Introduction to Virtual Reality
COMP Lecture1 - Introduction to Virtual RealityCOMP Lecture1 - Introduction to Virtual Reality
COMP Lecture1 - Introduction to Virtual Reality
 
COMP 4010 Lecture 9 AR Interaction
COMP 4010 Lecture 9 AR InteractionCOMP 4010 Lecture 9 AR Interaction
COMP 4010 Lecture 9 AR Interaction
 
Comp 4010 2021 Lecture1-Introduction to XR
Comp 4010 2021 Lecture1-Introduction to XRComp 4010 2021 Lecture1-Introduction to XR
Comp 4010 2021 Lecture1-Introduction to XR
 
Comp4010 Lecture13 More Research Directions
Comp4010 Lecture13 More Research DirectionsComp4010 Lecture13 More Research Directions
Comp4010 Lecture13 More Research Directions
 
Building Usable AR Interfaces
Building Usable AR InterfacesBuilding Usable AR Interfaces
Building Usable AR Interfaces
 
Lecture 5: 3D User Interfaces for Virtual Reality
Lecture 5: 3D User Interfaces for Virtual RealityLecture 5: 3D User Interfaces for Virtual Reality
Lecture 5: 3D User Interfaces for Virtual Reality
 
Lecture 6 Interaction Design for VR
Lecture 6 Interaction Design for VRLecture 6 Interaction Design for VR
Lecture 6 Interaction Design for VR
 
COMP 4010 - Lecture1 Introduction to Virtual Reality
COMP 4010 - Lecture1 Introduction to Virtual RealityCOMP 4010 - Lecture1 Introduction to Virtual Reality
COMP 4010 - Lecture1 Introduction to Virtual Reality
 
Mixed Reality in the Workspace
Mixed Reality in the WorkspaceMixed Reality in the Workspace
Mixed Reality in the Workspace
 

Destaque

Augmented reality
Augmented realityAugmented reality
Augmented realitynicklauzon
 
Analysing a film poster
Analysing a film posterAnalysing a film poster
Analysing a film posteralexharbon
 
Multimodal Discourse Analysis Systemic Functional Perspectives
Multimodal Discourse Analysis Systemic Functional PerspectivesMultimodal Discourse Analysis Systemic Functional Perspectives
Multimodal Discourse Analysis Systemic Functional PerspectivesWilliam Eduardo
 
Augmented Reality By Jaseem Bhutto
Augmented Reality By Jaseem BhuttoAugmented Reality By Jaseem Bhutto
Augmented Reality By Jaseem BhuttoJaseem Bhutto
 
Multimodal Interaction: An Introduction
Multimodal Interaction: An IntroductionMultimodal Interaction: An Introduction
Multimodal Interaction: An IntroductionAbdallah El Ali
 

Destaque (7)

Hands and Speech in Space
Hands and Speech in SpaceHands and Speech in Space
Hands and Speech in Space
 
Augmented reality
Augmented realityAugmented reality
Augmented reality
 
Analysing a film poster
Analysing a film posterAnalysing a film poster
Analysing a film poster
 
Multimodal Discourse Analysis Systemic Functional Perspectives
Multimodal Discourse Analysis Systemic Functional PerspectivesMultimodal Discourse Analysis Systemic Functional Perspectives
Multimodal Discourse Analysis Systemic Functional Perspectives
 
Augmented Reality By Jaseem Bhutto
Augmented Reality By Jaseem BhuttoAugmented Reality By Jaseem Bhutto
Augmented Reality By Jaseem Bhutto
 
Multimodal Interaction: An Introduction
Multimodal Interaction: An IntroductionMultimodal Interaction: An Introduction
Multimodal Interaction: An Introduction
 
Discourse analysis
Discourse analysisDiscourse analysis
Discourse analysis
 

Semelhante a Hands and Speech in Space: Multimodal Input for Augmented Reality

Designing Augmented Reality Experiences
Designing Augmented Reality ExperiencesDesigning Augmented Reality Experiences
Designing Augmented Reality ExperiencesMark Billinghurst
 
Comp4010 Lecture8 Introduction to VR
Comp4010 Lecture8 Introduction to VRComp4010 Lecture8 Introduction to VR
Comp4010 Lecture8 Introduction to VRMark Billinghurst
 
COMP 4010 Lecture9 AR Interaction
COMP 4010 Lecture9 AR InteractionCOMP 4010 Lecture9 AR Interaction
COMP 4010 Lecture9 AR InteractionMark Billinghurst
 
2013 Lecture 6: AR User Interface Design Guidelines
2013 Lecture 6: AR User Interface Design Guidelines2013 Lecture 6: AR User Interface Design Guidelines
2013 Lecture 6: AR User Interface Design GuidelinesMark Billinghurst
 
Future Research Directions for Augmented Reality
Future Research Directions for Augmented RealityFuture Research Directions for Augmented Reality
Future Research Directions for Augmented RealityMark Billinghurst
 
Mobile AR lecture 9 - Mobile AR Interface Design
Mobile AR lecture 9 - Mobile AR Interface DesignMobile AR lecture 9 - Mobile AR Interface Design
Mobile AR lecture 9 - Mobile AR Interface DesignMark Billinghurst
 
Beyond Reality (2027): The Future of Virtual and Augmented Reality
Beyond Reality (2027): The Future of Virtual and Augmented RealityBeyond Reality (2027): The Future of Virtual and Augmented Reality
Beyond Reality (2027): The Future of Virtual and Augmented RealityMark Billinghurst
 
2022 COMP4010 Lecture5: AR Prototyping
2022 COMP4010 Lecture5: AR Prototyping2022 COMP4010 Lecture5: AR Prototyping
2022 COMP4010 Lecture5: AR PrototypingMark Billinghurst
 
Indian Sign Language Recognition Method For Deaf People
Indian Sign Language Recognition Method For Deaf PeopleIndian Sign Language Recognition Method For Deaf People
Indian Sign Language Recognition Method For Deaf PeopleTakrim Ul Islam Laskar
 
COMP 4010 Lecture12 Research Directions in AR
COMP 4010 Lecture12 Research Directions in ARCOMP 4010 Lecture12 Research Directions in AR
COMP 4010 Lecture12 Research Directions in ARMark Billinghurst
 
Grand Challenges for Mixed Reality
Grand Challenges for Mixed Reality Grand Challenges for Mixed Reality
Grand Challenges for Mixed Reality Mark Billinghurst
 
COMP 4010: Lecture11 AR Interaction
COMP 4010: Lecture11 AR InteractionCOMP 4010: Lecture11 AR Interaction
COMP 4010: Lecture11 AR InteractionMark Billinghurst
 
Virtual/ Physical Co-Existing Design _Capturing Space Interactive Device
Virtual/ Physical Co-Existing Design_Capturing Space Interactive Device   Virtual/ Physical Co-Existing Design_Capturing Space Interactive Device
Virtual/ Physical Co-Existing Design _Capturing Space Interactive Device Kai-Tzu Lu
 
2016 AR Summer School Lecture3
2016 AR Summer School Lecture32016 AR Summer School Lecture3
2016 AR Summer School Lecture3Mark Billinghurst
 
Virtual/ Physical Co-Existing Design (CapX)
Virtual/ Physical Co-Existing Design (CapX)Virtual/ Physical Co-Existing Design (CapX)
Virtual/ Physical Co-Existing Design (CapX)Kai-Tzu Lu
 
2014 COSC 426 Lecture 2: Augmented Reality Technology
2014 COSC 426 Lecture 2: Augmented Reality Technology2014 COSC 426 Lecture 2: Augmented Reality Technology
2014 COSC 426 Lecture 2: Augmented Reality TechnologyMark Billinghurst
 

Semelhante a Hands and Speech in Space: Multimodal Input for Augmented Reality (20)

Designing Augmented Reality Experiences
Designing Augmented Reality ExperiencesDesigning Augmented Reality Experiences
Designing Augmented Reality Experiences
 
Comp4010 Lecture8 Introduction to VR
Comp4010 Lecture8 Introduction to VRComp4010 Lecture8 Introduction to VR
Comp4010 Lecture8 Introduction to VR
 
COMP 4010 Lecture9 AR Interaction
COMP 4010 Lecture9 AR InteractionCOMP 4010 Lecture9 AR Interaction
COMP 4010 Lecture9 AR Interaction
 
2013 Lecture 6: AR User Interface Design Guidelines
2013 Lecture 6: AR User Interface Design Guidelines2013 Lecture 6: AR User Interface Design Guidelines
2013 Lecture 6: AR User Interface Design Guidelines
 
Future Research Directions for Augmented Reality
Future Research Directions for Augmented RealityFuture Research Directions for Augmented Reality
Future Research Directions for Augmented Reality
 
Mobile AR lecture 9 - Mobile AR Interface Design
Mobile AR lecture 9 - Mobile AR Interface DesignMobile AR lecture 9 - Mobile AR Interface Design
Mobile AR lecture 9 - Mobile AR Interface Design
 
Beyond Reality (2027): The Future of Virtual and Augmented Reality
Beyond Reality (2027): The Future of Virtual and Augmented RealityBeyond Reality (2027): The Future of Virtual and Augmented Reality
Beyond Reality (2027): The Future of Virtual and Augmented Reality
 
2022 COMP4010 Lecture5: AR Prototyping
2022 COMP4010 Lecture5: AR Prototyping2022 COMP4010 Lecture5: AR Prototyping
2022 COMP4010 Lecture5: AR Prototyping
 
SVR2011 Keynote
SVR2011 KeynoteSVR2011 Keynote
SVR2011 Keynote
 
Vitual Reality
Vitual RealityVitual Reality
Vitual Reality
 
Indian Sign Language Recognition Method For Deaf People
Indian Sign Language Recognition Method For Deaf PeopleIndian Sign Language Recognition Method For Deaf People
Indian Sign Language Recognition Method For Deaf People
 
Cap xpresent
Cap xpresentCap xpresent
Cap xpresent
 
COMP 4010 Lecture12 Research Directions in AR
COMP 4010 Lecture12 Research Directions in ARCOMP 4010 Lecture12 Research Directions in AR
COMP 4010 Lecture12 Research Directions in AR
 
Grand Challenges for Mixed Reality
Grand Challenges for Mixed Reality Grand Challenges for Mixed Reality
Grand Challenges for Mixed Reality
 
COMP 4010: Lecture11 AR Interaction
COMP 4010: Lecture11 AR InteractionCOMP 4010: Lecture11 AR Interaction
COMP 4010: Lecture11 AR Interaction
 
Virtual/ Physical Co-Existing Design _Capturing Space Interactive Device
Virtual/ Physical Co-Existing Design_Capturing Space Interactive Device   Virtual/ Physical Co-Existing Design_Capturing Space Interactive Device
Virtual/ Physical Co-Existing Design _Capturing Space Interactive Device
 
2016 AR Summer School Lecture3
2016 AR Summer School Lecture32016 AR Summer School Lecture3
2016 AR Summer School Lecture3
 
VR- virtual reality
VR- virtual realityVR- virtual reality
VR- virtual reality
 
Virtual/ Physical Co-Existing Design (CapX)
Virtual/ Physical Co-Existing Design (CapX)Virtual/ Physical Co-Existing Design (CapX)
Virtual/ Physical Co-Existing Design (CapX)
 
2014 COSC 426 Lecture 2: Augmented Reality Technology
2014 COSC 426 Lecture 2: Augmented Reality Technology2014 COSC 426 Lecture 2: Augmented Reality Technology
2014 COSC 426 Lecture 2: Augmented Reality Technology
 

Mais de Mark Billinghurst

Evaluation Methods for Social XR Experiences
Evaluation Methods for Social XR ExperiencesEvaluation Methods for Social XR Experiences
Evaluation Methods for Social XR ExperiencesMark Billinghurst
 
Empathic Computing: Delivering the Potential of the Metaverse
Empathic Computing: Delivering  the Potential of the MetaverseEmpathic Computing: Delivering  the Potential of the Metaverse
Empathic Computing: Delivering the Potential of the MetaverseMark Billinghurst
 
Empathic Computing: Capturing the Potential of the Metaverse
Empathic Computing: Capturing the Potential of the MetaverseEmpathic Computing: Capturing the Potential of the Metaverse
Empathic Computing: Capturing the Potential of the MetaverseMark Billinghurst
 
Talk to Me: Using Virtual Avatars to Improve Remote Collaboration
Talk to Me: Using Virtual Avatars to Improve Remote CollaborationTalk to Me: Using Virtual Avatars to Improve Remote Collaboration
Talk to Me: Using Virtual Avatars to Improve Remote CollaborationMark Billinghurst
 
Empathic Computing: Designing for the Broader Metaverse
Empathic Computing: Designing for the Broader MetaverseEmpathic Computing: Designing for the Broader Metaverse
Empathic Computing: Designing for the Broader MetaverseMark Billinghurst
 
2022 COMP 4010 Lecture 7: Introduction to VR
2022 COMP 4010 Lecture 7: Introduction to VR2022 COMP 4010 Lecture 7: Introduction to VR
2022 COMP 4010 Lecture 7: Introduction to VRMark Billinghurst
 
2022 COMP4010 Lecture 6: Designing AR Systems
2022 COMP4010 Lecture 6: Designing AR Systems2022 COMP4010 Lecture 6: Designing AR Systems
2022 COMP4010 Lecture 6: Designing AR SystemsMark Billinghurst
 
Novel Interfaces for AR Systems
Novel Interfaces for AR SystemsNovel Interfaces for AR Systems
Novel Interfaces for AR SystemsMark Billinghurst
 
2022 COMP4010 Lecture4: AR Interaction
2022 COMP4010 Lecture4: AR Interaction2022 COMP4010 Lecture4: AR Interaction
2022 COMP4010 Lecture4: AR InteractionMark Billinghurst
 
2022 COMP4010 Lecture3: AR Technology
2022 COMP4010 Lecture3: AR Technology2022 COMP4010 Lecture3: AR Technology
2022 COMP4010 Lecture3: AR TechnologyMark Billinghurst
 
2022 COMP4010 Lecture2: Perception
2022 COMP4010 Lecture2: Perception2022 COMP4010 Lecture2: Perception
2022 COMP4010 Lecture2: PerceptionMark Billinghurst
 
2022 COMP4010 Lecture1: Introduction to XR
2022 COMP4010 Lecture1: Introduction to XR2022 COMP4010 Lecture1: Introduction to XR
2022 COMP4010 Lecture1: Introduction to XRMark Billinghurst
 
Empathic Computing and Collaborative Immersive Analytics
Empathic Computing and Collaborative Immersive AnalyticsEmpathic Computing and Collaborative Immersive Analytics
Empathic Computing and Collaborative Immersive AnalyticsMark Billinghurst
 
Empathic Computing: Developing for the Whole Metaverse
Empathic Computing: Developing for the Whole MetaverseEmpathic Computing: Developing for the Whole Metaverse
Empathic Computing: Developing for the Whole MetaverseMark Billinghurst
 
Comp4010 Lecture12 Research Directions
Comp4010 Lecture12 Research DirectionsComp4010 Lecture12 Research Directions
Comp4010 Lecture12 Research DirectionsMark Billinghurst
 
Comp4010 lecture11 VR Applications
Comp4010 lecture11 VR ApplicationsComp4010 lecture11 VR Applications
Comp4010 lecture11 VR ApplicationsMark Billinghurst
 
Comp4010 lecture11 VR Applications
Comp4010 lecture11 VR ApplicationsComp4010 lecture11 VR Applications
Comp4010 lecture11 VR ApplicationsMark Billinghurst
 
Comp4010 Lecture10 VR Interface Design
Comp4010 Lecture10 VR Interface DesignComp4010 Lecture10 VR Interface Design
Comp4010 Lecture10 VR Interface DesignMark Billinghurst
 

Mais de Mark Billinghurst (20)

Evaluation Methods for Social XR Experiences
Evaluation Methods for Social XR ExperiencesEvaluation Methods for Social XR Experiences
Evaluation Methods for Social XR Experiences
 
Empathic Computing: Delivering the Potential of the Metaverse
Empathic Computing: Delivering  the Potential of the MetaverseEmpathic Computing: Delivering  the Potential of the Metaverse
Empathic Computing: Delivering the Potential of the Metaverse
 
Empathic Computing: Capturing the Potential of the Metaverse
Empathic Computing: Capturing the Potential of the MetaverseEmpathic Computing: Capturing the Potential of the Metaverse
Empathic Computing: Capturing the Potential of the Metaverse
 
Talk to Me: Using Virtual Avatars to Improve Remote Collaboration
Talk to Me: Using Virtual Avatars to Improve Remote CollaborationTalk to Me: Using Virtual Avatars to Improve Remote Collaboration
Talk to Me: Using Virtual Avatars to Improve Remote Collaboration
 
Empathic Computing: Designing for the Broader Metaverse
Empathic Computing: Designing for the Broader MetaverseEmpathic Computing: Designing for the Broader Metaverse
Empathic Computing: Designing for the Broader Metaverse
 
2022 COMP 4010 Lecture 7: Introduction to VR
2022 COMP 4010 Lecture 7: Introduction to VR2022 COMP 4010 Lecture 7: Introduction to VR
2022 COMP 4010 Lecture 7: Introduction to VR
 
2022 COMP4010 Lecture 6: Designing AR Systems
2022 COMP4010 Lecture 6: Designing AR Systems2022 COMP4010 Lecture 6: Designing AR Systems
2022 COMP4010 Lecture 6: Designing AR Systems
 
ISS2022 Keynote
ISS2022 KeynoteISS2022 Keynote
ISS2022 Keynote
 
Novel Interfaces for AR Systems
Novel Interfaces for AR SystemsNovel Interfaces for AR Systems
Novel Interfaces for AR Systems
 
2022 COMP4010 Lecture4: AR Interaction
2022 COMP4010 Lecture4: AR Interaction2022 COMP4010 Lecture4: AR Interaction
2022 COMP4010 Lecture4: AR Interaction
 
2022 COMP4010 Lecture3: AR Technology
2022 COMP4010 Lecture3: AR Technology2022 COMP4010 Lecture3: AR Technology
2022 COMP4010 Lecture3: AR Technology
 
2022 COMP4010 Lecture2: Perception
2022 COMP4010 Lecture2: Perception2022 COMP4010 Lecture2: Perception
2022 COMP4010 Lecture2: Perception
 
2022 COMP4010 Lecture1: Introduction to XR
2022 COMP4010 Lecture1: Introduction to XR2022 COMP4010 Lecture1: Introduction to XR
2022 COMP4010 Lecture1: Introduction to XR
 
Empathic Computing and Collaborative Immersive Analytics
Empathic Computing and Collaborative Immersive AnalyticsEmpathic Computing and Collaborative Immersive Analytics
Empathic Computing and Collaborative Immersive Analytics
 
Metaverse Learning
Metaverse LearningMetaverse Learning
Metaverse Learning
 
Empathic Computing: Developing for the Whole Metaverse
Empathic Computing: Developing for the Whole MetaverseEmpathic Computing: Developing for the Whole Metaverse
Empathic Computing: Developing for the Whole Metaverse
 
Comp4010 Lecture12 Research Directions
Comp4010 Lecture12 Research DirectionsComp4010 Lecture12 Research Directions
Comp4010 Lecture12 Research Directions
 
Comp4010 lecture11 VR Applications
Comp4010 lecture11 VR ApplicationsComp4010 lecture11 VR Applications
Comp4010 lecture11 VR Applications
 
Comp4010 lecture11 VR Applications
Comp4010 lecture11 VR ApplicationsComp4010 lecture11 VR Applications
Comp4010 lecture11 VR Applications
 
Comp4010 Lecture10 VR Interface Design
Comp4010 Lecture10 VR Interface DesignComp4010 Lecture10 VR Interface Design
Comp4010 Lecture10 VR Interface Design
 

Último

Unleashing Real-time Insights with ClickHouse_ Navigating the Landscape in 20...
Unleashing Real-time Insights with ClickHouse_ Navigating the Landscape in 20...Unleashing Real-time Insights with ClickHouse_ Navigating the Landscape in 20...
Unleashing Real-time Insights with ClickHouse_ Navigating the Landscape in 20...Alkin Tezuysal
 
React Native vs Ionic - The Best Mobile App Framework
React Native vs Ionic - The Best Mobile App FrameworkReact Native vs Ionic - The Best Mobile App Framework
React Native vs Ionic - The Best Mobile App FrameworkPixlogix Infotech
 
Transcript: New from BookNet Canada for 2024: BNC SalesData and LibraryData -...
Transcript: New from BookNet Canada for 2024: BNC SalesData and LibraryData -...Transcript: New from BookNet Canada for 2024: BNC SalesData and LibraryData -...
Transcript: New from BookNet Canada for 2024: BNC SalesData and LibraryData -...BookNet Canada
 
Decarbonising Buildings: Making a net-zero built environment a reality
Decarbonising Buildings: Making a net-zero built environment a realityDecarbonising Buildings: Making a net-zero built environment a reality
Decarbonising Buildings: Making a net-zero built environment a realityIES VE
 
Top 10 Hubspot Development Companies in 2024
Top 10 Hubspot Development Companies in 2024Top 10 Hubspot Development Companies in 2024
Top 10 Hubspot Development Companies in 2024TopCSSGallery
 
How to Effectively Monitor SD-WAN and SASE Environments with ThousandEyes
How to Effectively Monitor SD-WAN and SASE Environments with ThousandEyesHow to Effectively Monitor SD-WAN and SASE Environments with ThousandEyes
How to Effectively Monitor SD-WAN and SASE Environments with ThousandEyesThousandEyes
 
MuleSoft Online Meetup Group - B2B Crash Course: Release SparkNotes
MuleSoft Online Meetup Group - B2B Crash Course: Release SparkNotesMuleSoft Online Meetup Group - B2B Crash Course: Release SparkNotes
MuleSoft Online Meetup Group - B2B Crash Course: Release SparkNotesManik S Magar
 
QCon London: Mastering long-running processes in modern architectures
QCon London: Mastering long-running processes in modern architecturesQCon London: Mastering long-running processes in modern architectures
QCon London: Mastering long-running processes in modern architecturesBernd Ruecker
 
Irene Moetsana-Moeng: Stakeholders in Cybersecurity: Collaborative Defence fo...
Irene Moetsana-Moeng: Stakeholders in Cybersecurity: Collaborative Defence fo...Irene Moetsana-Moeng: Stakeholders in Cybersecurity: Collaborative Defence fo...
Irene Moetsana-Moeng: Stakeholders in Cybersecurity: Collaborative Defence fo...itnewsafrica
 
A Glance At The Java Performance Toolbox
A Glance At The Java Performance ToolboxA Glance At The Java Performance Toolbox
A Glance At The Java Performance ToolboxAna-Maria Mihalceanu
 
React JS; all concepts. Contains React Features, JSX, functional & Class comp...
React JS; all concepts. Contains React Features, JSX, functional & Class comp...React JS; all concepts. Contains React Features, JSX, functional & Class comp...
React JS; all concepts. Contains React Features, JSX, functional & Class comp...Karmanjay Verma
 
A Journey Into the Emotions of Software Developers
A Journey Into the Emotions of Software DevelopersA Journey Into the Emotions of Software Developers
A Journey Into the Emotions of Software DevelopersNicole Novielli
 
Zeshan Sattar- Assessing the skill requirements and industry expectations for...
Zeshan Sattar- Assessing the skill requirements and industry expectations for...Zeshan Sattar- Assessing the skill requirements and industry expectations for...
Zeshan Sattar- Assessing the skill requirements and industry expectations for...itnewsafrica
 
Arizona Broadband Policy Past, Present, and Future Presentation 3/25/24
Arizona Broadband Policy Past, Present, and Future Presentation 3/25/24Arizona Broadband Policy Past, Present, and Future Presentation 3/25/24
Arizona Broadband Policy Past, Present, and Future Presentation 3/25/24Mark Goldstein
 
UiPath Community: Communication Mining from Zero to Hero
UiPath Community: Communication Mining from Zero to HeroUiPath Community: Communication Mining from Zero to Hero
UiPath Community: Communication Mining from Zero to HeroUiPathCommunity
 
Connecting the Dots for Information Discovery.pdf
Connecting the Dots for Information Discovery.pdfConnecting the Dots for Information Discovery.pdf
Connecting the Dots for Information Discovery.pdfNeo4j
 
Glenn Lazarus- Why Your Observability Strategy Needs Security Observability
Glenn Lazarus- Why Your Observability Strategy Needs Security ObservabilityGlenn Lazarus- Why Your Observability Strategy Needs Security Observability
Glenn Lazarus- Why Your Observability Strategy Needs Security Observabilityitnewsafrica
 
The Future Roadmap for the Composable Data Stack - Wes McKinney - Data Counci...
The Future Roadmap for the Composable Data Stack - Wes McKinney - Data Counci...The Future Roadmap for the Composable Data Stack - Wes McKinney - Data Counci...
The Future Roadmap for the Composable Data Stack - Wes McKinney - Data Counci...Wes McKinney
 
Long journey of Ruby standard library at RubyConf AU 2024
Long journey of Ruby standard library at RubyConf AU 2024Long journey of Ruby standard library at RubyConf AU 2024
Long journey of Ruby standard library at RubyConf AU 2024Hiroshi SHIBATA
 
Why device, WIFI, and ISP insights are crucial to supporting remote Microsoft...
Why device, WIFI, and ISP insights are crucial to supporting remote Microsoft...Why device, WIFI, and ISP insights are crucial to supporting remote Microsoft...
Why device, WIFI, and ISP insights are crucial to supporting remote Microsoft...panagenda
 

Último (20)

Unleashing Real-time Insights with ClickHouse_ Navigating the Landscape in 20...
Unleashing Real-time Insights with ClickHouse_ Navigating the Landscape in 20...Unleashing Real-time Insights with ClickHouse_ Navigating the Landscape in 20...
Unleashing Real-time Insights with ClickHouse_ Navigating the Landscape in 20...
 
React Native vs Ionic - The Best Mobile App Framework
React Native vs Ionic - The Best Mobile App FrameworkReact Native vs Ionic - The Best Mobile App Framework
React Native vs Ionic - The Best Mobile App Framework
 
Transcript: New from BookNet Canada for 2024: BNC SalesData and LibraryData -...
Transcript: New from BookNet Canada for 2024: BNC SalesData and LibraryData -...Transcript: New from BookNet Canada for 2024: BNC SalesData and LibraryData -...
Transcript: New from BookNet Canada for 2024: BNC SalesData and LibraryData -...
 
Decarbonising Buildings: Making a net-zero built environment a reality
Decarbonising Buildings: Making a net-zero built environment a realityDecarbonising Buildings: Making a net-zero built environment a reality
Decarbonising Buildings: Making a net-zero built environment a reality
 
Top 10 Hubspot Development Companies in 2024
Top 10 Hubspot Development Companies in 2024Top 10 Hubspot Development Companies in 2024
Top 10 Hubspot Development Companies in 2024
 
How to Effectively Monitor SD-WAN and SASE Environments with ThousandEyes
How to Effectively Monitor SD-WAN and SASE Environments with ThousandEyesHow to Effectively Monitor SD-WAN and SASE Environments with ThousandEyes
How to Effectively Monitor SD-WAN and SASE Environments with ThousandEyes
 
MuleSoft Online Meetup Group - B2B Crash Course: Release SparkNotes
MuleSoft Online Meetup Group - B2B Crash Course: Release SparkNotesMuleSoft Online Meetup Group - B2B Crash Course: Release SparkNotes
MuleSoft Online Meetup Group - B2B Crash Course: Release SparkNotes
 
QCon London: Mastering long-running processes in modern architectures
QCon London: Mastering long-running processes in modern architecturesQCon London: Mastering long-running processes in modern architectures
QCon London: Mastering long-running processes in modern architectures
 
Irene Moetsana-Moeng: Stakeholders in Cybersecurity: Collaborative Defence fo...
Irene Moetsana-Moeng: Stakeholders in Cybersecurity: Collaborative Defence fo...Irene Moetsana-Moeng: Stakeholders in Cybersecurity: Collaborative Defence fo...
Irene Moetsana-Moeng: Stakeholders in Cybersecurity: Collaborative Defence fo...
 
A Glance At The Java Performance Toolbox
A Glance At The Java Performance ToolboxA Glance At The Java Performance Toolbox
A Glance At The Java Performance Toolbox
 
React JS; all concepts. Contains React Features, JSX, functional & Class comp...
React JS; all concepts. Contains React Features, JSX, functional & Class comp...React JS; all concepts. Contains React Features, JSX, functional & Class comp...
React JS; all concepts. Contains React Features, JSX, functional & Class comp...
 
A Journey Into the Emotions of Software Developers
A Journey Into the Emotions of Software DevelopersA Journey Into the Emotions of Software Developers
A Journey Into the Emotions of Software Developers
 
Zeshan Sattar- Assessing the skill requirements and industry expectations for...
Zeshan Sattar- Assessing the skill requirements and industry expectations for...Zeshan Sattar- Assessing the skill requirements and industry expectations for...
Zeshan Sattar- Assessing the skill requirements and industry expectations for...
 
Arizona Broadband Policy Past, Present, and Future Presentation 3/25/24
Arizona Broadband Policy Past, Present, and Future Presentation 3/25/24Arizona Broadband Policy Past, Present, and Future Presentation 3/25/24
Arizona Broadband Policy Past, Present, and Future Presentation 3/25/24
 
UiPath Community: Communication Mining from Zero to Hero
UiPath Community: Communication Mining from Zero to HeroUiPath Community: Communication Mining from Zero to Hero
UiPath Community: Communication Mining from Zero to Hero
 
Connecting the Dots for Information Discovery.pdf
Connecting the Dots for Information Discovery.pdfConnecting the Dots for Information Discovery.pdf
Connecting the Dots for Information Discovery.pdf
 
Glenn Lazarus- Why Your Observability Strategy Needs Security Observability
Glenn Lazarus- Why Your Observability Strategy Needs Security ObservabilityGlenn Lazarus- Why Your Observability Strategy Needs Security Observability
Glenn Lazarus- Why Your Observability Strategy Needs Security Observability
 
The Future Roadmap for the Composable Data Stack - Wes McKinney - Data Counci...
The Future Roadmap for the Composable Data Stack - Wes McKinney - Data Counci...The Future Roadmap for the Composable Data Stack - Wes McKinney - Data Counci...
The Future Roadmap for the Composable Data Stack - Wes McKinney - Data Counci...
 
Long journey of Ruby standard library at RubyConf AU 2024
Long journey of Ruby standard library at RubyConf AU 2024Long journey of Ruby standard library at RubyConf AU 2024
Long journey of Ruby standard library at RubyConf AU 2024
 
Why device, WIFI, and ISP insights are crucial to supporting remote Microsoft...
Why device, WIFI, and ISP insights are crucial to supporting remote Microsoft...Why device, WIFI, and ISP insights are crucial to supporting remote Microsoft...
Why device, WIFI, and ISP insights are crucial to supporting remote Microsoft...
 

Hands and Speech in Space: Multimodal Input for Augmented Reality

  • 1. Hands and Speech in Space: Multimodal Interaction for AR Mark Billinghurst mark.billinghurst@hitlabnz.org The HIT Lab NZ, University of Canterbury December 12th 2013
  • 2. 1977 – Star Wars 1977 – Star Wars
  • 3. Augmented Reality Definition   Defining Characteristics   Combines Real and Virtual Images -  Both can be seen at the same time   Interactive in real-time -  The virtual content can be interacted with   Registered in 3D -  Virtual objects appear fixed in space Azuma, R. T. (1997). A survey of augmented reality. Presence, 6(4), 355-385.
  • 5. AR Interface Components Physical Elements Input Interaction Metaphor Virtual Elements Output   Key Question: How should a person interact with the Augmented Reality content?   Connecting physical and virtual with interaction
  • 6. AR Interaction Metaphors   Information Browsing   View AR content   3D AR Interfaces   3D UI interaction techniques   Augmented Surfaces   Tangible UI techniques   Tangible AR   Tangible UI input + AR output
  • 7. VOMAR Demo (Kato 2000)   AR Furniture Arranging   Elements + Interactions   Book: -  Turn over the page   Paddle: -  Push, shake, incline, hit, scoop Kato, H., Billinghurst, M., et al. 2000. Virtual Object Manipulation on a Table-Top AR Environment. In Proceedings of the International Symposium on Augmented Reality (ISAR 2000), Munich, Germany, 111--119.
  • 8. Opportunities for Multimodal Input   Multimodal interfaces are a natural fit for AR   Need for non-GUI interfaces   Natural interaction with real world   Natural support for body input   Previous work shown value of multimodal input and 3D graphics
  • 9. Related Work   Related work in 3D graphics/VR   Interaction with 3D content [Chu 1997]   Navigating through virtual worlds [Krum 2002]   Interacting with virtual characters [Billinghurst 1998]   Little earlier work in AR   Require additional input devices   Few formal usability studies   Eg Olwal et. al [2003] Sense Shapes
  • 11. Marker Based Multimodal Interface   Add speech recognition to VOMAR   Paddle + speech commands Irawati, S., Green, S., Billinghurst, M., Duenser, A., & Ko, H. (2006, October). IEEE Xplore. In Mixed and Augmented Reality, 2006. ISMAR 2006. IEEE/ACM International Symposium on (pp. 183-186). IEEE.
  • 12.
  • 13. Commands Recognized   Create Command "Make a blue chair": to create a virtual object and place it on the paddle.   Duplicate Command "Copy this": to duplicate a virtual object and place it on the paddle.   Grab Command "Grab table": to select a virtual object and place it on the paddle.   Place Command "Place here": to place the attached object in the workspace.   Move Command "Move the couch": to attach a virtual object in the workspace to the paddle so that it follows the paddle movement.
  • 15. Object Relationships "Put chair behind the table” Where is behind? View specific regions
  • 16. User Evaluation   Performance time   Speech + static paddle significantly faster   Gesture-only condition less accurate for position/orientation   Users preferred speech + paddle input
  • 18. 2012 – Iron Man 2
  • 19. To Make the Vision Real..   Hardware/software requirements   Contact lens displays   Free space hand/body tracking   Speech/gesture recognition   Etc..   Most importantly   Usability/User Experience
  • 20. Natural Interaction   Automatically detecting real environment   Environmental awareness, Physically based interaction   Gesture interaction   Free-hand interaction   Multimodal input   Speech and gesture interaction   Intelligent interfaces   Implicit rather than Explicit interaction
  • 22. AR MicroMachines   AR experience with environment awareness and physically-based interaction   Based on MS Kinect RGB-D sensor   Augmented environment supports   occlusion, shadows   physically-based interaction between real and virtual objects Clark, A., & Piumsomboon, T. (2011). A realistic augmented reality racing game using a depth-sensing camera. In Proceedings of the 10th International Conference on Virtual Reality Continuum and Its Applications in Industry (pp. 499-502). ACM.
  • 24. Architecture   Our framework uses five libraries:   OpenNI   OpenCV   OPIRA   Bullet Physics   OpenSceneGraph
  • 25. System Flow   The system flow consists of three sections:   Image Processing and Marker Tracking   Physics Simulation   Rendering
  • 26. Physics Simulation   Create virtual mesh over real world   Update at 10 fps – can move real objects   Use by physics engine for collision detection (virtual/real)   Use by OpenScenegraph for occlusion and shadows
  • 29. Natural Hand Interaction   Using bare hands to interact with AR content   MS Kinect depth sensing   Real time hand tracking   Physics based simulation model
  • 30. Hand Interaction   Represent models as collections of spheres   Bullet physics engine for interaction with real world
  • 31. Scene Interaction   Render AR scene with OpenSceneGraph   Using depth map for occlusion   Shadows yet to be implemented
  • 32. Architecture 5. Gesture •  Static Gestures •  Dynamic Gestures •  Context based Gestures 4. Modeling •  Hand recognition/modeling •  Rigid-body modeling 3. Classification/Tracking 2. Segmentation 1. Hardware Interface
  • 33. Architecture 5. Gesture •  Static Gestures •  Dynamic Gestures •  Context based Gestures o  Supports PCL, OpenNI, OpenCV, and Kinect SDK. o  Provides access to depth, RGB, XYZRGB. o  Usage: Capturing color image, depth image and concatenated point clouds from a single or multiple cameras o  For example: 4. Modeling •  Hand recognition/ modeling •  Rigid-body modeling 3. Classification/Tracking 2. Segmentation 1. Hardware Interface Kinect for Xbox 360 Kinect for Windows Asus Xtion Pro Live
  • 34. Architecture 5. Gesture •  Static Gestures •  Dynamic Gestures •  Context based Gestures o  Segment images and point clouds based on color, depth and space. o  Usage: Segmenting images or point clouds using color models, depth, or spatial properties such as location, shape and size. o  For example: 4. Modeling •  Hand recognition/ modeling •  Rigid-body modeling Skin color segmentation 3. Classification/Tracking 2. Segmentation 1. Hardware Interface Depth threshold
  • 35. Architecture 5. Gesture •  Static Gestures •  Dynamic Gestures •  Context based Gestures o  Identify and track objects between frames based on XYZRGB. o  Usage: Identifying current position/orientation of the tracked object in space. o  For example: 4. Modeling •  Hand recognition/ modeling •  Rigid-body modeling 3. Classification/Tracking 2. Segmentation 1. Hardware Interface Training set of hand poses, colors represent unique regions of the hand. Raw output (withoutcleaning) classified on real hand input (depth image).
  • 36. Architecture 5. Gesture •  Static Gestures •  Dynamic Gestures •  Context based Gestures 4. Modeling •  Hand recognition/ modeling •  Rigid-body modeling 3. Classification/Tracking 2. Segmentation 1. Hardware Interface o  Hand Recognition/Modeling   Skeleton based (for low resolution approximation)   Model based (for more accurate representation) o  Object Modeling (identification and tracking rigidbody objects) o  Physical Modeling (physical interaction)   Sphere Proxy   Model based   Mesh based o  Usage: For general spatial interaction in AR/VR environment
  • 37. Architecture 5. Gesture •  Static Gestures •  Dynamic Gestures •  Context based Gestures 4. Modeling •  Hand recognition/ modeling •  Rigid-body modeling 3. Classification/Tracking 2. Segmentation 1. Hardware Interface o  Static (hand pose recognition) o  Dynamic (meaningful movement recognition) o  Context-based gesture recognition (gestures with context, e.g. pointing) o  Usage: Issuing commands/anticipating user intention and high level interaction.
  • 38. Skeleton Based Interaction   3 Gear Systems   Kinect/Primesense Sensor   Two hand tracking   http://www.threegear.com
  • 39. Skeleton Interaction + AR   HMD AR View   Viewpoint tracking   Two hand input   Skeleton interaction, occlusion
  • 40. What Gestures do People Want to Use?   Limitations of Previous work in AR   Limited range of gestures   Gestures designed for optimal recognition   Gestures studied as add-on to speech   Solution – elicit desired gestures from users   Eg. Gestures for surface computing [Wobbrock]   Previous work in unistroke getsures, mobile gestures
  • 41. User Defined Gesture Study   Use AR view   HMD + AR tracking   Present AR animations   40 tasks in six categories -  Editing, transforms, menu, etc   Ask users to produce gestures causing animations   Record gesture (video, depth) Piumsomboon, T., Clark, A., Billinghurst, M., & Cockburn, A. (2013, April). User-defined gestures for augmented reality. In CHI'13 Extended Abstracts on Human Factors in Computing Systems (pp. 955-960).ACM
  • 42. Data Recorded   20 participants   Gestures recorded (video, depth data)   800 gestures from 40 tasks   Subjective rankings   Likert ranking of goodness, ease of use   Think aloud transcripts
  • 44. Results - Gestures   Gestures grouped according to similarity – 320 groups   44 consensus (62% all gestures)   276 low similarity (discarded)   11 hand poses seen   Degree of consensus (A) using guessability score [Wobbrock]
  • 45. Results –Agreement Scores Red line – proportion of two handed gestures
  • 46. Usability Results Consensus Discarded Ease of Performance 6.02 5.50 Good Match 6.17 5.83 Likert Scale [1-7], 7 = Very Good   Significant difference between consensus and discarded gesture sets (p < 0.0001)   Gestures in consensus set better than discarded gestures in perceived performance and goodness
  • 47. Lessons Learned   AR animation can elicit desired gestures   For some tasks there is a high degree of similarity in user defined gestures   Especially command gestures (eg Open), select   Less agreement in manipulation gestures   Move (40%), rotate (30%), grouping (10%)   Small portion of two handed gestures (22%)   Scaling, group selection
  • 49. Multimodal Interaction   Combined speech input   Gesture and Speech complimentary   Speech -  modal commands, quantities   Gesture -  selection, motion, qualities   Previous work found multimodal interfaces intuitive for 2D/3D graphics interaction
  • 50. Wizard of Oz Study   What speech and gesture input would people like to use?   Wizard   Perform speech recognition   Command interpretation   Domain   3D object interaction/modelling Lee, M., & Billinghurst, M. (2008, October). A Wizard of Oz study for an AR multimodal interface. In Proceedings of the 10th international conference on Multimodal interfaces (pp. 249-256). ACM.
  • 54. Experiment   12 participants   Two display conditions (HMD vs. Desktop)   Three tasks   Task 1: Change object color/shape   Task 2: 3D positioning of obejcts   Task 3: Scene assembly
  • 55. Key Results   Most commands multimodal   Multimodal (63%), Gesture (34%), Speech (4%)   Most spoken phrases short   74% phrases average 1.25 words long   Sentences (26%) average 3 words   Main gestures deictic (65%), metaphoric (35%)   In multimodal commands gesture issued first   94% time gesture begun before speech   Multimodal window 8s – speech 4.5s after gesture
  • 56. Free Hand Multimodal Input Point Move Pick/Drop   Use free hand to interact with AR content   Recognize simple gestures   Open hand, closed hand, pointing Lee, M., Billinghurst, M., Baek, W., Green, R., & Woo, W. (2013). A usability study of multimodal input in an augmented reality environment. Virtual Reality, 17(4), 293-305.
  • 57. Speech Input   MS Speech + MS SAPI (> 90% accuracy)   Single word speech commands
  • 62. User Evaluation   25 subjects, 10 task trials x 3, 3 conditions   Change object shape, colour and position   Conditions   Speech only, gesture only, multimodal   Measures   performance time, errors (system/user), subjective survey
  • 63. Results - Performance   Average performance time   Gesture: 15.44s   Speech: 12.38s   Multimodal: 11.78s   Significant difference across conditions (p < 0.01)   Difference between gesture and speech/MMI
  • 64. Errors   User errors – errors per task   Gesture (0.50), Speech (0.41), MMI (0.42)   No significant difference   System errors   Speech accuracy – 94%, Gesture accuracy – 85%   MMI accuracy – 90%
  • 65. Subjective Results (Likert 1-7) Gesture Speech MMI Naturalness 4.60 5.60 5.80 Ease of Use 4.00 5.90 6.00 Efficiency 4.45 5.15 6.05 Physical Effort 4.75 3.15 3.85   User subjective survey   Gesture significantly worse, MMI and Speech same   MMI perceived as most efficient   Preference   70% MMI, 25% speech only, 5% gesture only
  • 66. Observations   Significant difference in number of commands   Gesture (6.14), Speech (5.23), MMI (4.93)   MMI Simultaneous vs. Sequential commands   79% sequential, 21% simultaneous   Reaction to system errors   Almost always repeated same command   In MMI rarely changes modalities
  • 67. Lessons Learned   Multimodal interaction significantly better than gesture alone in AR interfaces for 3D tasks   Short task time, more efficient   Users felt that MMI was more natural, easier, and more effective that gesture/speech only   Simultaneous input rarely used   More studies need to be conducted
  • 69. Intelligent Interfaces   Most AR systems stupid   Don’t recognize user behaviour   Don’t provide feedback   Don’t adapt to user   Especially important for training   Scaffolded learning   Moving beyond check-lists of actions
  • 70. Intelligent Interfaces   AR interface + intelligent tutoring system   ASPIRE constraint based system (from UC)   Constraints -  relevance cond., satisfaction cond., feedback Westerfield, G., Mitrovic, A., & Billinghurst, M. (2013). Intelligent Augmented Reality Training for Assembly Tasks. In Artificial Intelligence in Education (pp. 542-551). Springer Berlin Heidelberg.
  • 72. Intelligent Feedback   Actively monitors user behaviour   Implicit vs. explicit interaction   Provides corrective feedback
  • 73.
  • 74. Evaluation Results   16 subjects, with and without ITS   Improved task completion   Improved learning
  • 75. Intelligent Agents   AR characters   Virtual embodiment of system   Multimodal input/output   Examples   AR Lego, Welbo, etc   Mr Virtuoso -  AR character more real, more fun -  On-screen 3D and AR similar in usefulness Wagner, D., Billinghurst, M., & Schmalstieg, D. (2006). How real should virtual characters be?. In Proceedings of the 2006 ACM SIGCHI international conference on Advances in computer entertainment technology (p. 57). ACM.
  • 76. Looking to the Future What’s Next?
  • 77. Directions for Future Research   Mobile Gesture Interaction   Tablet, phone interfaces   Wearable Systems   Google Glass   Novel Displays   Contact lens   Environmental Understanding   Semantic representation
  • 78. Mobile Gesture Interaction   Motivation   Richer interaction with handheld devices   Natural interaction with handheld AR   2D tracking   Finger tip tracking   3D tracking [Hurst and Wezel 2013]   Hand tracking [Henrysson et al. 2007] Henrysson, A., Marshall, J., & Billinghurst, M. (2007). Experiments in 3D interaction for mobile phone AR. In Proceedings of the 5th international conference on Computer graphics and interactive techniques in Australia and Southeast Asia (pp. 187-194). ACM.
  • 79. Fingertip Based Interaction Running System System Setup Mobile Client + PC Server Bai, H., Gao, L., El-Sana, J., & Billinghurst, M. (2013). Markerless 3D gesture-based interaction for handheld augmented reality interfaces. In SIGGRAPH Asia 2013 Symposium on Mobile Graphics and Interactive Applications (p. 22). ACM.
  • 81. 3D Prototype System   3 Gear + Vuforia   Hand tracking + phone tracking   Freehand interaction on phone   Skeleton model   3D interaction   20 fps performance
  • 83.
  • 84. User Experience   Truly Wearable Computing   Less than 46 ounces   Hands-free Information Access   Voice interaction, Ego-vision camera   Intuitive User Interface   Touch, Gesture, Speech, Head Motion   Access to all Google Services   Map, Search, Location, Messaging, Email, etc
  • 85. Contact Lens Display   Babak Parviz   University Washington   MEMS components   Transparent elements   Micro-sensors   Challenges   Miniaturization   Assembly   Eye-safe
  • 87. Environmental Understanding   Semantic understanding of environment   What are the key objects?   What are there relationships?   Represented in a form suitable for multimodal interaction?
  • 89. Conclusions   AR experiences need new interaction methods   Enabling technologies are advancing quickly   Displays, tracking, depth capture devices   Natural user interfaces possible   Free hand gesture, speech, intelligence interfaces   Important research for the future   Mobile, wearable, displays
  • 90. More Information •  Mark Billinghurst –  Email: mark.billinghurst@hitlabnz.org –  Twitter: @marknb00 •  Website –  http://www.hitlabnz.org/