This piece of software called OmNomNom works with OpenSCAD to turn 2D images into 3D models. It’s literally a drag-and-drop process that renders almost instantly.
Here the example is a QR code, which is perfect for the software since it’s a well-defined black and white outline in the source image. But the video after the break shows several other examples that don’t rely on this simplicity. For instance, the Superman logo, which uses four different colors, is converted quite easily. There’s also a depth map of [Beethoven’s] bust that is converted into a 3D object. The same technique can be used to create terrain from topographic source images.
Once the file has been converted to a model it can still be tweaked like normal. This allows you to customize size and depth to suit your needs. This is where OpenSCD comes into play, but if you don’t use that program you can still export an STL file directly from OmNomNom for use on your 3D printer.
Continue reading “Drag and drop images for 3D printing”
[Bastian] sent in a coffee table he built. This isn’t a place to set your drinks and copies of Make, though: it’s a multitouch table with a 3D display. Since no description can do this table justice, take a look at the video.
The build was inspired by the subject of this Hackaday post where [programming4fun] was able to build a ‘holographic display’ using a regular 2D projector and a Kinect. Both builds work on the principle of redrawing the 3D space in relation to the user’s head – as [Bastian] moves his head around the coffee table, the Kinect tracks his location and moves the 3 dimensional grid of boxes in the opposite direction. It’s extremely clever, and looks to be a promising user interface.
In addition to a Kinect, the coffee table uses a Microsoft Surface-like display; four infrared lasers are placed at the corner and detected with a camera next to the projector in the base.
After the break you can see the demo video and a gallery of the images [Bastion] put up on the NUI group forum.
Continue reading “Multitouch table uses a Kinect for a 3D display”
This turntable can automatically digitize objects for use in 3D rendering software like Blender3D. [James Dalby] built it using a high-quality DSLR, and some bits and pieces out of his junk box. The turntable itself is a Lazy Susan turned on its head. The base for the spinning model is normally what sits on the table, but this way it gives him an area to rest the model, and the larger portion acts as a mounting surface for the drive mechanism.
He used the stepper motor from a scanner, as well as the belt and tension hardware from a printer to motorize the platform. This is driven by a transistor array (a ULN2003 chip) connected to an Arduino. The microcontroller also controls the shutter of the camera. We’ve included his code after the break; you’ll find his demo video embedded there as well.
The concept is the same as other turntable builds we’ve seen, But [James] takes the post-processing one step further. Rather than just make a rotating gif he is using Autodesk 123D to create a digital model from the set of images.
Continue reading “Scanning turntable digitizes objects as 3D models”
This one is so simple, and works so well, we’d call it a hoax if April 1st hadn’t already passed us by. But we’re confident that what [William Myers] and [Guo Jie Chin] came up with exists, and we want one of our own. The project is a method of drawing in 3 dimensions using ultrasonic sensors.
They call it 3D Paint, and that’s fitting since the software interface is much like the original MS Paint. It can show you the movements of the stylus in three axes, but it can also assemble an anaglyph — the kind of 3D that uses those red and blue filter glasses — so that the artists can see the 3D rendering as it is being drawn.
The hardware depends on a trio of sensors and a stylus that are all controlled by an ATmega644. That’s it for hardware (to be fair, there are a few trivial amplifier circuits too), making this an incredibly affordable setup. The real work, and the reason the input is so smooth and accurate, comes in the MATLAB code which does the trilateration. If you like to get elbow deep in the math the article linked above has plenty to interest you. If you’re more of a visual learner just skip down after the break for the demo video.
Continue reading “3D whiteboard without the whiteboard”
[Mark] wrote in, eager to show off this new tool he’s created to view your gerber files in 3d. He also wrote an instructible to go along with it, to help you figure out how to use the tool. Being an in-browser tool also means you can shoot it to your friends for a quick 3d review as well. Some of you may not feel that the 3d view is that helpful to the process, but we think that this is a welcomed feature that just might get some use around here.
[Mark] points out that it is still being actively developed, so please shoot him bugs via the form on the website if you should encounter any.
This circuit is how [John Tsiombikas] makes his cheap 3D shutter glasses work with a Linux machine. It’s not that they were incompatible with Linux. The issue is that only certain video cards have the stereo port necessary to drive the head-mounted hardware.
Shutter glasses block light from one eye at a time, so that different renderings can be shown to create the stereoscopic effect. Since stimulating the muscles in the eye doesn’t actually work, you need to find a way to drive the glasses in perfect time with the video signal. His circuit watches for the V-Sync signal, then uses it to toggle the shutter glasses. Since the hardware has no way of knowing whether the left or right frame is being generated, he included the toggle switch as a user-controlled adjustment. If the 3D isn’t coming together, you’re probably viewing the frames with the wrong eye and need to flip the switch.
There’s really no way to show the effect without trying out the hardware in person. But [John] reports that it works like a charm when used with the OpenGL stereo wrapper.
[Valentin] wanted to experiment with 3D scanning some objects he had around the house, but says he didn’t want to buy a line laser for the project since they are pretty expensive. Fortunately, he had some random components sitting in his parts bin, and he was able to build his own line laser without spending a ton of money.
His tutorial actually covers two different methods of building line lasers, both of which use parts that you likely have on hand already.
His first build involves gluing a small square mirror to a flat platform, which he then mounted on a salvaged DC motor. Once the motor starts spinning, the cheap laser pointer he has aimed at the mirror draws a perfect line across whatever medium he is scanning.
His second line laser uses parts donated from an old hard drive that he no longer used. He removed the drive’s read head from the chassis and mounted a small mirror on the actuator arm before firing up his laser. With the laser aimed at the mirror, he applied an unspecified AC current to the motor, which caused it to oscillate and draw a line similar to his first setup.
While they might not be professionally-built scanning lasers, [Valentin’s] efforts produced some decent images, as you can see on his site.
Continue reading to see a short video of his DC motor laser line in action.
Continue reading “Build your own line laser for 3D scanning”