11:00 AM - 11:25 AM
Get ready for a look into the world of spatial computing with AWS! Join us as we dive into how AWS is transforming the way 3D models are brought to life in the cloud. We'll be showcasing the latest spatial computing services on AWS that enable you to build, deliver, and manage your 3D workloads with ease and efficiency.
But that's not all - via an on stage demonstration you'll get to see how we paired a Magic Leap 2 with a Boston Dynamics Spot Robot to showcase how AWS's cutting-edge technology can help users visualize live robot telemetry and control the robot in even the most challenging and remote environments.
This session and the session following it is a must-attend for professionals who are interested in exploring the full potential of spatial computing on AWS. Join us for a captivating and informative presentation that is sure to inspire and inform!
11:30 AM - 11:55 AM
Digital Twins connect physical systems with virtual representations and models to allow for visual representations, integration of sensor data, and predictive capabilities for how assets or processes will behave in the future. Globally, organizations are grappling with the acceleration of remote operations and an increasingly prevalent skills gap. Forward-looking companies are addressing this problem by equipping their in-person, hybrid, and off-site teams with mixed reality (MR) solutions that enhance productivity capabilities especially when integrated with Digital Twins. In this session, learn how Sphere and AWS are working together to develop a digital workspace which enables professionals to communicate across languages, distances, dimensions, and time (with predictive capabilities). By partnering on initiatives which include TwinMaker and Lookout for Vision, as well as cloud rendering powered by AWS and its partners, Sphere’s cutting-edge solution is pioneering the future of collaboration, as well as expert assistance and workflow guidance using MR.
12:00 PM - 12:25 PM
Zheng Qin has since 2018 developed the cutting edge, wide FoV AR optics system called Mixed Waveguide, with the most advanced Crossfire solution that gives you a 120-degree FoV and is interchangeable between AR and VR. It’s even slimmer than most VR optics (including the Pancake solutions), so it could be the ultimate optical solution for AR & VR hybrid glasses. Zheng will walk you through the reasons why Crossfire is much better than its competitor Pancake +VST (video see-through) solution. Additionally, Zheng will introduce the whole family of Mixed Waveguide solutions, which has been adopted by many key clients around the world.
01:25 PM - 01:50 PM
Recently, there has been, and continues to be, a flurry of activities around AR and the Metaverse. How these domains intersect and unfold over time is still very much in the early stages. What is clear, however, is that the “on-ramp” or gateways into the Metaverse starts with the ability to perceive the physical and digital worlds simultaneously. Many technologies and devices are needed to enable the true immersion and first and foremost is the ability to overlay the digital domain onto the physical space. In this talk we will discuss these aspects and delve deeply into near-to-eye display technologies that allows uses to coexist in the physical and digital domains.
01:55 PM - 02:20 PM
Construction is a complex and messy process, combining millions of individual components - all needing to be positioned in the exact right place with mm-level accuracy. However, the construction process has largely remained unchanged for thousands of years. A group of people design a building, listening to every wish and elaborate desire coming from the Owner, and then they hand the design over to thousands of other people to actually build it, without any understanding of how this design fits the real world. It’s kind of like building a massive jigsaw puzzle where thousands of people are responsible for one piece or another, and no one really knows how they all fit together. This waterfall process leads to building things up only to tear them down immediately after and then build them right back again - just moved over by 1ft - something the construction industry spends over $280B per year doing. This is simply not sustainable, for the industry, for the stakeholders, and most importantly - for the planet. With nearly 40% of the world’s Carbon emissions being contributed by the construction industry, transformation is desperately needed.
And that’s exactly what Trimble is working to do. As a leader in high-accuracy positioning technologies, Trimble has a long standing history of bringing precision to the construction industry - helping to fit all those puzzle pieces together. But we see the opportunity to do more. Since 1997, when we filed our first XR patent, Trimble has been transforming the way the world works by connecting the physical and digital worlds. Now, we’re working to change this archaic narrative, by empowering everyone in construction to visualize, interpret, and action the design through Augmented and Mixed Reality technologies in the field. From catching design mistakes by viewing 3D models on any worker’s iPad, to being more efficient by controlling the most precise Total Station with nothing more than your gaze, we are improving communication and collaboration around design intent, enabling more efficient and sustainable projects. Follow us on this journey as we outline how Extended Reality technologies are revolutionizing the way the construction industry operates today, tomorrow, and for decades to come.
02:25 PM - 02:50 PM
The JPEO-CBRND partnered with MRIGlobal and ForgeFX Simulations to produce The CBNRD HoloTrainer, a networked multiuser Microsoft HoloLens 2 augmented reality training simulator for Chemical, Biological, Radiological, and Nuclear detection device operators. This presentation will cover the entire project lifecycle, including government requirements and the development of holographically projected interactive virtual equipment, culminating in deployment to soldiers. The CBRND HoloTrainer is a groundbreaking spatial computing application that significantly increases the effectiveness of its users.
02:55 PM - 03:20 PM
Doublepoint builds a gesture-based touch interface to enable efficient, robust, and intuitive AR interaction. At this talk, they will showcase how these solutions can work as standalone input modalities or be combined to supercharge eye tracking and hand tracking.
User input is one of the many challenges standing in the way of the mass adoption of AR. How will the everyday person interact in Augmented Reality? What's the role of touch in interaction? At Doublepoint we research different ways for people to interact in AR and develop the best technologies to detect these interactions efficiently and reliably.
Currently, there are a lot of input methods for AR using built-in sensors of the headset, such as hand tracking, eye tracking, and voice input. However, if we want it to be as transformative as the graphical user interface or the capacitive touch screen, we need to put deliberate thought into building the ideal input paradigm and the needed hardware that might not be in the headset itself.
At this talk:
• We’ll demonstrate how a machine learning algorithm on existing smartwatches can already significantly improve AR interaction.
• We’ll show how it can be combined with eye tracking and hand tracking sensors in the headset to improve interactions even more.
• Lastly, we'll show some of our future custom hardware dedicated to sensing advanced micro-gestures in a small and convenient form factor.
03:25 PM - 04:20 PM
04:35 PM - 05:00 PM
With the use of a standard phone and a 3D scan of a location, you can determine your position with an impressive accuracy of 5cm - a game changer for AR. This technology opens doors to a variety of features including AR indoor navigation, AR workflows, and even remote assistance across large facilities. Spatial applications will make AR a tool for daily use, creating massive ROI for businesses. This talk will show how to use AR for daily operations and ultimately how to increase your ROI.
05:05 PM - 05:30 PM
Discover the untapped potential of XR technology as we discuss the crucial role of audio in driving realistic 3D experiences. Whether you’re venturing into VR, AR, or MR, delivering an exceptional audio experience is paramount to ensuring active participant engagement and seamless integration within virtual environments.
Join us for an enlightening conversation with industry experts, Paul Boustead, VP of Product at Dolby, and Michael Kolbrener, COO at Odyssey.stream. Gain valuable insights into the Dolby.io offerings that are powering immersive XR landscapes. And learn about the latest advancements and future possibilities that await XR enthusiasts in the near and long term.
05:35 PM - 06:00 PM
Behind the world’s first XR creator tool for responsive content, SyncReality is premiering its latest offering at AWE Santa Clara. The Beta release of the SyncReality Suite, will don the main tool, for automated placement of virtual assets in the real world, a room scanner, a parametric asset tool, and a simulator.
The SyncReality Suite is a revolutionary XR spatial content creation tool- enabling any XR content to automatically adapt to the space of the enduser - and designed to unlock new possibilities for businesses, brands, and developers, that want to provide enhanced user experiences across vCommerce, education, gaming, and more - to engage and excite their audiences.
Following the launch of Alpha in February 2023, SyncReality Beta boasts cutting-edge features including:
• Optimized product interface for easy navigation and creation
• Streamlined workflow management for cross-team development
• Improved user experience for seamless interactions
• Parameric Asset Bundles, keeping aesthetics intact in a broad variety of spaces.
• Simulator to verify XR content before exporting.
SyncReality’s unique technology makes it easier than ever for developers to create seamless experiences that engage and excite users. With SyncReality, you can create immersive worlds that allow users to interact with a character, world, brand, or product in ways never before possible. Imagine:
• Enjoying a concert in your living room by your favorite musician or
• Learning a new skill at home while sitting in a classroom environment or
• Playing an interactive escape game that transforms your space into a vast rainforest!