Volumetric 3D displays that allow the viewing of full 3D images without special glasses are not unknown in our community, usually taking the form of either a 3D LED matrix or a spinning rotor either with an image projected onto it or holding an LED array. They are impressive projects, but they are often limited in what they can display. Pretty patterns and simple 3D models are all very well, but they are hardly 3D television. Thus we’re quite impressed with [Evlmnkey]’s bachelor’s degree project, which combines motion capture and a volumetric display for a genuine volumetric 3D closed-circuit television system.
Finding the details takes a bit of dredging through the Reddit thread, but the display is an off-the-shelf Adafruit single-sided LED matrix driven by an ESP32, all mounted on a motor with a pair of slip rings for power. Data is fed to the ESP via WiFi, with the PC responsible for grabbing the image sending it as uncompressed frames. There’s little detail on the 3D capture, but since he mentions a Kinect library we suspect that may be the source.
This is perhaps not the highest resolution TV you’ll ever have seen, indeed we’d liken it to the flickering 30 lines of 1930s mechanical TV, but it’s still a functioning volumetric 3D live CCTV system. If you’re interested by 3D displays, you might like to see our examination of the subject.
Thanks [nandkeypull] for the tip.
Holograms are tricky to describe because science-fiction gives the name to any three-dimensional image. The science-fact versions are not as flashy, but they are still darn cool. Legitimate holograms are images stored on a photographic medium, and they retain a picture of the subject from certain angles. In other words, when [Justin Atkin] makes a hologram of a model building, (video, embedded below) you can see the east side of the belfry, but when you reorient, you see the west side, or the roof if you point down. Holography is different from stereoscopy, which shows you a 3D image using two cameras. With a stereoscopic image, you cannot tilt it and see a new part of the subject, so there is a niche for each method.
There are a couple of different methods for making a hologram at home. First, you probably want a DIY hologram kit since it will come with the exposure plate and a known-good light source. Far be it for us to tell you you can’t buy plates and a laser pointer to take the path less traveled. Next, you need something that will not move, so we’re afraid you cannot immortalize your rambunctious kitty. The last necessity is a stable platform since you will perform a long-exposure shot, and even breathing on the setup can ruin the image. Different colors come from the coherent light source, so getting the “Rainbow Holograms” advertised in the video is a matter of mixing lights. Since you can buy red, green, and blue laser pointers for a pittance, you can do color remixes to your content.
Another type of hologram appears on things like trading cards as those wildly off-color (chromatic, not distasteful) images of super-heroes or abstract shapes. They’re a different variety, which can be printed en-masse, unlike the one-off [Justin] shows us how to make.
If you’re yearning for volumetric displays, we are happy to point you to this beauty capable of showing a jaw-dropping 3D model or this full-color blocky duck.
Continue reading “Hello, Holograms”
Strong opinions exist on both sides about OpenSCAD. The lightweight program takes megabytes of space, not gigabytes, so many people have a copy, even if they’ve never written a shape. Some people adore the text-only modeling language, and some people abhor the minimal function list. [Johnathon ‘Zalo’ Selstad] appreciates the idea but wants to see something more robust, and he wants to see it in your browser. His project CascadeStudio has a GitHub repo and a live link so you can start tinkering in a new window straight away.
Continue reading “Hyper Links And Hyperfunctional Text CAD”
There are times when a project becomes such a big part of a maker’s life that they find themselves revisiting it even years later. [Thanassis] combined this phenomena with his love for the ZX Spectrum when he ported one of his old 3D rendering projects to the ZX Spectrum 48K. The video below shows the result, and they speak for themselves.
The roots of this project go back around three years, when [Thanassis] posted a similar project for the ATMega328 which employed fixed point math tricks for achieving the graphics. The code needed to be even tighter to run on the Spectrum, eventually getting boiled down to just a handful of calculations. This got the proof of concept working with the z88dk compiler, but it wasn’t quite fast enough.
In the end, hand assembly optimizations nearly doubled the performance to a blistering 10 frames per second. There’s also a version that kicks it all the way up to 40 FPS, but only if you give it a few minutes to do the calculations ahead of time. With a few teaks and the right display, this project could produce some very cool retro visuals.
Continue reading “3D On The ZX Spectrum 48K”
With an ever-growing range of smart-home products available, all with their own hubs, protocols, and APIs, we see a lot of DIY projects (and commercial offerings too) which aim to provide a “single universal interface” to different devices and services. Usually, these projects allow you to control your home using a list of devices, or sometimes a 2D floor plan. [Wassim]’s project aims to take the first steps in providing a 3D interface, by creating an interactive smart-home controller in the browser.
Note: this isn’t just a rendered image of a 3D scene which is static; this is an interactive 3D model which can be orbited and inspected, showing information on lights, heaters, and windows. The project is well documented, and the code can be found on GitHub. The tech works by taking 3D models and animations made in Blender, exporting them using the .glTF format, then visualising them in the browser using three.js. This can then talk to Hue bulbs, power meters, or whatever other devices are required. The technical notes on this project may well be useful for others wanting to use the Blender to three.js/browser workflow, and include a number of interesting demos of isolated small key concepts for the project.
We notice that all the meshes created in Blender are very low-poly; is it possible to easily add subdivision surface modifiers or is it the vertex count deliberately kept low for performance reasons?
This isn’t our first unique home automation interface, we’ve previously written about shAIdes, a pair of AI-enabled glasses that allow you to control your devices just by looking at them. And if you want to roll your own home automation setup, we have plenty of resources. The Hack My House series contains valuable information on using Raspberry Pis in this context, we’ve got information on picking the right sensors, and even enlisting old routers for the cause.
Most electronics we deal with day to day are comprised of circuit boards. No surprise there, right? But how do they work? This might seem like a simple question but we’ve all been in the place where those weird green or black sheets are little slices of magic. [Teddy Tablante] at Branch Eduction put together a lovingly crafted
walkthrough flythrough video of how PCB(A)s work that’s definitely worth your time.
[Teddy]’s video focuses on unraveling the mysteries of the PCBA by peeling back the layers of a smartphone. Starting from the full assembly he separates components from circuit board and descends from there, highlighting the manufacturing methods and purpose behind what you see.
What really stands out here is the animation; at each step [Teddy] has modeled the relevant components and rendered them on the PCBA in 3D. Instead of relying solely on hard to understand blurry X-ray images and 2D scans of PCBAs he illustrates their relationships in space, an especially important element in understanding what’s going on underneath the solder mask. Even if you think you know it all we bet there’s a pearl of knowledge to discover; this writer learned that VIA is an acronym!
If you don’t like clicking links you can find the video embedded after the break. Credit to friend of the Hackaday [Mike Harrison] for acting as the best recommendation algorithm and finding this gem.
Continue reading “Journey Through The Inner Workings Of A PCB”
We’ve seen 3D image projection tried in a variety of different ways, but this is a new one to us. This volumetric display by Interact Lab of the University of Sussex creates a 3D image by projecting light onto a tiny foam ball, which zips around in the air fast enough to create a persistence of vision effect. (Video, embedded below.) How is this achieved? With a large array of ultrasonic transducers, performing what researchers call ‘acoustic trapping’.
This is the same principle behind acoustic levitation devices which demonstrate how lightweight objects (like tiny polystyrene foam balls) can be made to defy gravity. But this 3D display is capable of not only moving the object in 3D space, but doing so at a high enough speed and with enough control to produce a persistence of vision effect. The abstract for their (as yet unreleased) paper claims the trapped ball can be moved at speeds of up to several meters per second.
It has a few other tricks up its sleeve, too. The array is capable of simultaneously creating sounds as well as providing a limited form of tactile feedback by letting a user touch areas of high and low air pressure created by the transducers. These areas can’t be the same ones being occupied by the speeding ball, of course, but it’s a neat trick. Check out the video below for a demonstration.
Continue reading “Behold A 3D Display, Thanks To A Speeding Foam Ball”