We will also present examples of challenging scenarios and discuss future work.
Content Persistence and Sharing
High level description of the content persistence workflow.
Demonstrate the computer vision and deep learning problems required to ensure that content is persisted across multiple sessions.
What is content sharing? and the computer vision involved to make it a reality.
Collaborative map building and sharing of sparse and dense maps.
Eye Tracking
Uses of gaze tracking in Magic Leap One for user interaction / understanding and graphics.
We will present a description of the eye tracking system focussing on two main aspects:
Hybrid segmentation and model based gaze estimation
Recovering 3D eye data and estimating 3D fixation point
Finally, we will go over challenges and future directions
Hand Tracking
The problem of Classification (identifying among a predefined “dictionary” of static hand poses and Regression (find keypoint locations in the presence of highly variable hand shape and pose, self-occlusion and environmental clutter).
Challenges involved in training and implementing deep learning models on device.
Low latency, compute, memory usage, and power consumption on (existing) embedded HW
High throughput (support 45-60 FPS sensors without dropping frames)
Data collection and labeling
We will then go over our End-to-End Deep Learning solution and present our work on model optimization and results.
Scene Semantics
What is Object recognition and scene understanding and what are its usecases.
Challenges in building a real system: Hardware, Algorithm, Data.
How do we build a scalable system that would produce multiple types of outputs that developers want: 2D bounding boxes, 3D bounding boxes, semantic meshes, semantic planes.
We will cover the deep learning approaches we have considered and present early results and future work.