Fail Of The Week: How Not To Make A 3D Scanner

Sometimes the best you can say about a project is, “Nice start.” That’s the case for this as-yet awful DIY 3D scanner, which can serve both as a launching point for further development and a lesson in what not to do.

Don’t get us wrong, we have plenty of respect for [bitluni] and for the fact that he posts his failures as well as his successes, like composite video and AM radio signals from an ESP32. He used an ESP8266 in this project, which actually uses two different sensors: an ultrasonic transducer, and a small time-of-flight laser chip. Each was mounted to a two-axis scanner built from hobby servos and 3D-printed parts. The pitch and yaw axes move the sensors through a hemisphere gathering data, but unfortunately, the Wemos D1 Mini lacks the RAM to render the complete point cloud from the raw points. That’s farmed out to a WebGL page. Initial results with the ultrasonic sensor were not great, and the TOF sensor left everything to be desired too. But [bitluni] stuck with it, and got a few results that at least make it look like he’s heading in the right direction.

We expect he’ll get this sorted out and come back with some better results, but in the meantime, we applaud his willingness to post this so that we can all benefit from his pain. He might want to check out the results from this polished and pricey LIDAR scanner for inspiration.

Continue reading “Fail Of The Week: How Not To Make A 3D Scanner”

XLIDAR Is A Merry-Go-Round Of Time-Of-Flight Sensors

[JRodrigo]’s xLIDAR project is one of those ideas that seemed so attractively workable that it went directly to a PCB prototype without doing much stopping along the way. The concept was to mount a trio of outward-facing VL53L0X distance sensors to a small PCB disk, and then turn that disk with a motor and belt while taking readings. As the sensors turn, their distance readings can be used to paint a picture of the immediate surroundings (at least within about 1 meter, which is the maximum range of the VL53L0X.)

The hardware is made to be accessible and has a strong element of “what you see is what you get.” The distance sensors are on small breakout boards, and the board turns the sensor disk via a DC motor and 3D printed belt drive. Even the method of encoding the disk’s movement and zero position has the same WYSIWYG straightforwardness: a spring contact and an interrupted bare copper trace on the bottom of the sensor disk acts as a physical switch. In fact, exposed copper traces in concentric circular patterns and spring pins taken from an SD card socket are what provide power and communications as the disk turns.

The prototype looks good and sounds like it should work, but how well does it hold up? We’ll find out once [JRodrigo] does some testing. Until then, the board designs are available on the project’s GitHub repository if anyone wants to take a shot at their own approach without starting from scratch.

A LIDAR Scanner Build In Glorious Detail

LIDAR is a very exciting technology that is only just now starting to become accessible to the DIY market. Think radar, but with lasers. There’s a few different modules starting to pop up for just a few hundred dollars. But what is one to do with a LIDAR module? Well, [David] decided to build a room scanner with his Garmin LIDAR Lite, and it’s a wonderful sight to behold.

The scanner consists of a rotating platform, which is driven by a stepper motor. The platform then contains a second motor which runs a tilt axis, upon which the LIDAR is mounted. By aiming the LIDAR in various directions, and recording the detected range, it’s possible to build a point cloud representation of the surrounding area.

The build uses a couple of STM32 chips to do motor control and interface with the LIDAR, but where this build really shines is the mechanical side of things. [David] goes into serious detail about the machining of the parts that make up the rotating system, and there’s plenty of cool bits and pieces like slip rings to make it all work. There’s even some home casting going on here! Be warned, though: there’s some rather juicy close-ups of lathes in action, so put the kids to bed before watching this one all the way through.

We love to see a well-executed build, and even more so when we get to watch the intricate details of how it came together. If you’re still looking for some more inspiration, we’ve seen other LIDAR room scanners before, too.

Simple Quadcopter Testbed Clears The Air For Easy Algorithm Development

We don’t have to tell you that drones are all the rage. But while new commercial models are being released all the time, and new parts get released for the makers, the basic technology used in the hardware hasn’t changed in the last few years. Sure, we’ve added more sensors, increased computing power, and improved the efficiency, but the key developments come in the software: you only have to look at the latest models on the market, or the frequency of Git commits to Betaflight, Butterflight, Cleanflight, etc.

With this in mind, for a Hackaday prize entry [int-smart] is working on a quadcopter testbed for developing algorithms, specifically localization and mapping. The aim of the project is to eventually make it as easy as possible to get off the ground and start writing code, as well as to integrate mapping algorithms with Ardupilot through ROS.

The initial idea was to use a Beaglebone Blue and some cheap hobby hardware which is fairly standard for a drone of this size: 1250 kv motors and SimonK ESCs, mounted on an f450 flame wheel style frame. However, it looks like an off-the-shelf solution might be even simpler if it can be made to work with ROS. A Scanse Sweep LIDAR sensor provides point cloud data, which is then munched with some Iterative Closest Point (ICP) processing. If you like math then it’s definitely worth reading the project logs, as some of the algorithms are explained there.

It might be fun to add FPV to this system to see how the mapping algorithms are performing from the perspective of the drone. And just because it’s awesome. FPV is also a fertile area for hacking: we particularly love this FPV tracker which rotates itself to get the best signal, and this 3D FPV setup using two cameras.

Robot Radar Module

For his Hackaday Prize entry, [Ted Yapo] is building a Robot Radar Module breakout board. His design uses the A111 60 GHz pulsed coherent radar (PCR) sensor from Acconeer AB (New Part alert!) .

The A111 is a low power, high precision sensor ideal for use in object detection or gesture sensing applications. The BGA package is tiny – 5.5 mm x 5.2 mm, but it does not appear very difficult for a hacker to assemble. The sensor includes an integrated baseband, RF front-end and Antenna in Package so you don’t have to mess with RF layout headaches. Acconeer claims the sensor performance is not affected with interference from noise, dust, color and direct or indirect light. Sensing range is about 2 m with a +/- 2 mm accuracy. And at just under $10 a pop for 10 units or more, it would make a nice addition to augment the sensor package on a Robot.

To get started, [Ted] is keeping his design simple and small – the break out board measures just 32 mm x 32 mm. The radar sensor itself doesn’t require any parts other than a crystal and its loading capacitors. A LDO takes care of the 1.8 V required by the A111. Three 74LVC2T45 chips translate the SPI digital interface from 1.8 V to external logic levels between 1.8 V to 5 V. The three level translation chips could possible be replaced by a single six or eight channel translator – such as one from the TXB series from TI. For his first PCB iteration, [Ted] is expecting to run in to some layout or performance issues, so if you have any feedback to give him on his design, check out his hardware repository on Github.

Acconeer provides a Getting Started guide for their Evaluation Kits, which includes a detailed Raspberry-Pi / Raspbian installation and an accompanying video (embedded after the break) targeted at hackers. We are eagerly looking forward to the progress that [Ted] makes with this sensor breakout. Combined with LiDAR ToF sensor breakout boards, such as the MappyDot, it would be a great addition to your robot’s sensing capabilities.

Continue reading “Robot Radar Module”

Litar: LiDAR air guitar

Litar: An Air Guitar Using LiDAR

This year, [Blecky’s] Hackaday Prize Entry is an air guitar which uses multiple LiDAR sensors to create the virtual strings. What’s also neat is that he’s using his own LiDAR sensor, the MappyDot Plus, an enhanced version of his 2017 Prize Entry, the MappyDot.

He uses a very clever arrangement of six sensors to get four virtual strings. Each sensor scans a 25-degree field of view. Three adjacent sensors are used to define a string, with the string being in the overlap of the outer two of those sensors. The middle sensor is used for the distance data.

For the chords, he started out using some commercially made joysticks but ran into some ergonomic issues. Also, the manufacturer was discontinuing the product, a no-no for an open source project. So he abandoned that approach and designed his own buttons. He came up with a PCB with a linear hall effect sensor and some springs on it. The button has a magnet attached to its underside and sits on the springs. That way he gets the press and can do vibrato as well.

He plans to use Bluetooth MIDI so that you can play the sound on a phone or laptop but for now he lights up an LED beside each sensor as you press the strings.

Debunking Moon Landing Denial With An Arduino And Science

It’s sad that nearly half a century after the achievements of the Apollo program we’re still arguing with a certain subset of people who insist it never happened. Poring through the historical record looking for evidence that proves the missions couldn’t possibly have occurred has become a sad little cottage industry, and debunking the deniers is a distasteful but necessary ongoing effort.

One particularly desperate denier theory holds that fully spacesuited astronauts could never have exited the tiny hatch of the Lunar Excursion Module (LEM). [AstronomyLive] fought back at this tendentious claim in a clever way — with a DIY LIDAR scanner to measure Apollo artifacts in museums. The hardware is straightforward, with a Garmin LIDAR-Lite V3 scanner mounted on a couple of servos to make a quick pan-tilt head. The rig has a decidedly compliant look to it, with the sensor flopping around a bit as the servos move. But for the purpose, it seems perfectly fine.

[AstronomyLive] took the scanner to two separate museum exhibits, one to scan a LEM hatch and one to scan the suit Gene Cernan, the last man to stand on the Moon so far, wore while training for Apollo 17. With the LEM flying from the rafters, the scanner was somewhat stretching its abilities, so the point clouds he captured were a little on the low-res side. But in the end, a virtual Cernan was able to transition through the virtual LEM hatch, as expected.

Sadly, such evidence will only ever be convincing to those who need no convincing; the willfully ignorant will always find ways to justify their position. So let’s just celebrate the achievements of Apollo.

Continue reading “Debunking Moon Landing Denial With An Arduino And Science”