We wager you haven’t you heard the latest from ultrasonics. Sorry. [Lindsay Wilson] is a Hackaday reader who wants to share his knowledge of transducer tuning to make tools. The bare unit he uses to demonstrate might attach to the bottom of an ultrasonic cleaner tank, which have a different construction than the ones used for distance sensing. The first demonstration shows the technique for finding a transducer’s resonant frequency and this technique is used throughout the video. On the YouTube page, his demonstrations are indexed by title and time for convenience.
For us, the most exciting part is when a tuned transducer is squeezed by hand. As the pressure increases, the current drops and goes out of phase in proportion to the grip. We see a transducer used as a pressure sensor. He later shows how temperature can affect the current level and phase.
Sizing horns is a science, but it has some basic rules which are well covered. The basic premise is to make it half of a wavelength long and be mindful of any tools which will go in the end. Nodes and antinodes are explained and their effects demonstrated with feedback on the oscilloscope.
Looking for ideas for your haptics projects? [Destin] of the Smarter Every Day YouTube channel got a tour from the engineers at HaptX of their full-featured VR glove with amazing haptic feedback both with a very fine, 120-point sense of touch, force feedback for each finger, temperature, and motion tracking.
In hacks, we usually stimulate the sense of touch by vibrating something against the skin. With this glove, they use pneumatics to press against the skin. A single fingertip has multiple roughly 1/8 inch air bladders in contact with it. Each bladder is separately pneumatically controlled by pushing air into it. The air pressure can vary continuously so that the bladders can push lightly, harder or anywhere in between. The glove has 120 of these bladders spread out over the fingers and the palm. Unfortunately, they didn’t allow him to see the valves controlling the pneumatics, but if you are looking for a low-frequency, low-cost way to actuate valves you might consider using syringes. The engineers do tell [Destin] that if your VR scene shows something pressing against your virtual finger, as long as your haptics push against your real finger within around 1/8th of a second, your brain won’t notice the delay.
They’re also working on using hot and cold fluids to give a sense of temperature within a glove. This is demonstrated in the first video below when [Destin] feels heat while a dragon in the VR world breathes fire on his hand. Fortunately one of the engineers mentions that our sense of temperature is one of the slower ones, it can handle longer latencies than even touch. We can see implementing this in a hack using a bladder pressing against the skin while tubes circulate different temperature fluids through it. But maybe there’s a way to do it electrically, possibly with thermoelectric modules as is done with this drinks cooler? Though safety issues might prohibit that.
Other features mentioned are force feedback for each finger, and their custom motion tracking which uses both magnetic and optical means to track fingertips. But we’ll leave the rest to the videos below. The first is the technical tour and the second is the glove being used in the VR world.
While it might not pack the computational punch you’d usually be looking for in a server platform, you can’t beat how cheap the Raspberry Pi is. As such, it’s at the heart of many a home LAN, serving up files as a network attached storage (NAS) device. But the biggest problem with using the Pi in a NAS is that it doesn’t have any onboard hard drive interface, forcing you to use USB. Not only is this much slower, but doesn’t leave you a lot of options for cleanly hooking up your drives.
This 3D printable NAS enclosure designed by [Paul-Louis Ageneau] helps address the issue by integrating two drive bays which can accommodate 2.25 inch laptop hard disk drives and their associated IB-AC6033-U3 USB adapters. The drives simply slide into the “rails” designed into the case without the need for additional hardware. There’s even space in the bottom of the case for a USB hub to connect the drives, and a fan on the top of the case to help keep the whole stack cool. It still isn’t perfect, but it’s compact and doesn’t look half bad.
The design is especially impressive as it doesn’t require any supports, an admirable goal to shoot for whenever designing for 3D printing. As an added bonus, the entire case is designed in OpenSCAD and licensed under the GPL v3; making modification easy if you want to tweak it for your specific purposes.
[Martin’s] goal was to capture events of interest, such as a person on screen, or a car in the driveway. The data for the events would then be published to an MQTT topic, along with some metadata such as confidence level. OpenCV is generally how these pipelines start, but [Martin’s] camera wouldn’t send RTSP images over TCP the way OpenCV requires, only RTSP over UDP. To solve this, Martin captures the video stream with FFmpeg. The deep learning AI magic is handled by the darkflow library, which is itself based upon Google’s Tensorflow.
Martin tested out his recognition system with some cheap Chinese PTZ cameras, and the processing running on a remote Raspberry Pi. So far the results have been good. The system is able to recognize people, animals, and cars pulling in the driveway. His code is available on GitHub if you want to give it a spin yourself!
Hope you weren’t looking forward to a night of sleep untroubled by nightmares. Doing his part to make sure Lovecraftian mechanized horrors have lease in your subconscious, [Paul-Louis Ageneau] has recently unleashed the horror that is Eyepot upon an unsuspecting world. This Cycloptic four legged robotic teapot takes inspiration from an enemy in the game Alice: Madness Returns, and seems to exist for no reason other than to creep people out.
Even if you aren’t physically manifesting nightmares, there’s plenty to learn from this project. [Paul-Louis Ageneau] has done a fantastic job of documenting the build, from the OpenSCAD-designed 3D printed components to the Raspberry Pi Zero and Arduino Pro Mini combo that control the eight servos in the legs. If you want to play along at home all the information and code is here, though feel free to skip the whole teapot with an eyeball thing.
A second post explains how the code is written for both the Arduino and Pi, making for some very illuminating reading. A Python script on the Pi breaks down the kinematics and passes on the appropriate servo angles to the Arduino over a serial link. Combined with a web interface for control and a stream from the teapot’s Raspberry Pi Camera module, and you’ve got the makings of the world’s creepiest telepresence robot. We’d love to see this one stomping up and down a boardroom table.
Revisiting old projects is always fun and this Nixie Clock by [pa3fwm] is just a classic. Instead of using transistors or microcontrollers, it uses neon lamps to clock and drive the Nixie Displays. The neon lamps themselves are the logic elements. Seriously, this masterpiece just oozes geekiness.
Inspired by the book “Electronic Counting Circuits” by J.B. Dance(ZIP), published in 1967, we covered the initial build a few years back. The fundamental concept of operation is similar to that of Neon Ring Counters. [Luc Small] has a write-up explaining the construction of such a device and some math associated with it. In this project, [pa3fwm] uses modern day neons that you find in indicators, so his circuit is also updated to compensate for the smaller difference in striking and maintaining voltages.
The original project was done in 2007 and has since undergone a few upgrades. [Pa3fwm] has modified the construction to make it wall mounted. Even though it’s not a precise timekeeper, the project itself is a keeper from its time. Check out the video below for a demonstration.
When you think of supercomputers, visions of big boxes and blinkenlights filling server rooms immediately appear. Since the 90s or thereabouts, these supercomputers have been clusters of computers, all working together on a single problem. For the last twenty years, people have been building their own ‘supercomputers’ in their homes, and now we have cheap ARM single board computers to play with. What does this mean? Personal supercomputers. That’s what [Jason] is building for his entry to the Hackaday Prize.
The goal of [Jason]’s project isn’t to break into the Top 500, and it’s doubtful it’ll be more powerful than a sufficiently modern desktop workstation. The goal for this project is to give anyone a system that has the same architecture as a large-scale cluster to facilitate learning about high-performance applications. It also has a front panel covered in LEDs.
The design of this system is built around a the PINE64 SOPINE module, or basically a 64-bit quad-core CPU stuck onto a board that fits in an SODIMM socket. If that sounds like the Raspberry Pi Computer Module, you get a cookie. Unlike the Pi Compute Module, the people behind the SOPINE have created something called a ‘Clusterboard’, or eight vertical SODIMM sockets tied together with a single controller, power supply, and an Ethernet jack. Yes, it’s a board meant for cluster computing.
To this, [Jason] is adding his own twist on a standard, off-the-shelf breakout board. This Clusterboard is mounted to a beautiful aluminum enclosure, and the front panel is loaded up with a whole bunch of almost vintage-looking red LEDs. These LEDs indicate the current load on each bit of the cluster, providing immediate visual feedback on how those computations are going. With the right art — perhaps something in harvest gold, brown, and avocado — this supercomputer would look like it’s right out of the era of beautiful computers. At any rate, it’s a great entry for the Hackaday Prize.