Stanford University researchers have combined optics and artificial intelligence to improve holographic displays for virtual and augmented reality experiences.
The researchers developed a technique for reducing speckling distortion typical of laser-based holographic displays, and proposed a method to more realistically portray the physics of a virtual three-dimensional scene as if it were actual.
The neural holographic display required the training of a neural network to mimic the real-world physics of what the display depicted and achieve real-time images.
The researchers combined this with camera-in-the-loop calibration to deliver near-instantaneous feedback, resulting in more realistic-looking visuals with enhanced color, contrast, and clarity.
From Stanford News
View Full Article
Abstracts Copyright © 2021 SmithBucklin, Washington, DC, USA