Apple Vision Pro’s Secret To Smooth Visuals? Subtly Substandard Optics

The displays inside the Apple Vision Pro have 3660 × 3200 pixels per eye, but veteran engineer [Karl Guttag]’s analysis of its subtly blurred optics reminds us that “resolution” doesn’t always translate to resolution, and how this is especially true for things like near-eye displays.

The Apple Vision Pro lacks the usual visual artifacts (like the screen door effect) which result from viewing magnified pixelated screens though optics. But [Karl] shows how this effect is in fact hiding in plain sight: Apple seems to have simply made everything just a wee bit blurry thanks to subtly out-of-focus lenses.

The thing is, this approach of intentionally de-focusing actually works very well for consuming visual content like movies or looking at pictures, where detail and pixel-to-pixel contrast is limited anyway.

Clever loophole, or specification shenanigans? You be the judge of that, but this really is evidence of how especially when it comes to things like VR headsets, everything is a trade-off. Improving one thing typically worsens others. In fact, it’s one of the reasons why VR monitor replacements are actually a nontrivial challenge.

No More Blurry Pictures

Say goodbye to ruined images thanks to this add-on hardware. It measures the movement of the camera when a picture is taken and corrects the image to get rid of motion blur. Above you see a high-speed camera which is just there for testing and fine-tuning the algorithm that fixes the photos. Once they got it right, the setup that the camera is attached to only includes an Arduino board, Bluetooth modem, 3-axis accelerometer, gyroscope, and a trigger for the camera. You use the new hardware to snap each image and it takes care of triggering the SLR’s shutter in order to ensure that the inertial data and the image are synchronized correctly.

[Thanks Rob]