Bike Rim Lighting Lets The Night Crowd Know When You’re Rollin’

There comes a wonderful “MacGyver moment” in many hackers’ lives when we find ourselves with just the right microcosm of scrap parts to build something awesome. That’s exactly what [dragonator] did with his gifted tech box from Instructables. He’s combined RGB LEDs, a Trinket, and a hall effect sensor to add a semicircular rainbow pattern to his night ride while he rides it.

The theory behind the hack is well-known: given the time between pings from a hall-effect sensor responding to the magnet on a bike wheel, an embedded system can estimate the wheel rpm and predict the time to display a particular color on the LEDs. [dragonator] uses the known wheel speed to determine the LED pattern currently on display: either a slow breathing pulse to a half-circle rainbow that displays on the lower bike rim. He drops in the needed equations and required components to follow his trail in a well-documented instructable.

Persistence of Vision (POV) is a nice extension from blinking your first (or first hundred) LED(s). It’s just enough math to get the casual onlooker to cry “magic” and just enough embedded electronics to get those seasoned double-Es to nod their heads. If you’re new to the POV crowd, [dragonator’s] Instructable may be a great start.

Continue reading “Bike Rim Lighting Lets The Night Crowd Know When You’re Rollin’”

Real-Time Thermal Projection Saves Your Tastebuds From The Hot Stuff

With another wave of holiday parties about to land on our doorstep, we still haven’t found a great way to stop scalding our tongues each time [Uncle Dave] pours us an enticing cup of boiling cocoa.

Thankfully, [Ken] has both you and your holiday guests covered with a clever trick that takes the data from a FLIR ONE and projects a heat profile onto the surface it’s observing. Here, [Ken] has superimposed his FLIR ONE data onto his kitchen table, and he’s able to visualize 2D heat profiles in near-real-time.

If you haven’t started quantifying yourself recently (and what are you waiting for?), the FLIR ONE is yet another opportunity to help you become more aware of your surroundings than you are now. It’s a thermal camera attachment for your iPhone, allowing you to see into the infrared band and look at the world in terms of heat. We’ve covered the FLIR ONE before, and we’ve seen ways of making it both clearer and more hacker-friendly.

As we tip our hats to [Ken], we’d say he’s a generous fellow. This hack is a clever inversion of the normal use case where you might whip out your FLIR-ONE-enabled iPhone and warn your cousins not to try the hot chocolate for a few more minutes. With [Ken’s] solution, the data is right there on your condiments and in plain sight of everyone, not just for you with your sweet, Star-Trek-augmented iPhone.

Continue reading “Real-Time Thermal Projection Saves Your Tastebuds From The Hot Stuff”

Santa’s Autonomous Helping Hands Let The Jolly Ol’ Fellow Kick Back This Season

For those skeptical about the feasibility of Santa’s annual delivery schedule, here’s an autonomous piece of the puzzle that will bewilder even the most hard-hearted of non-believers.

The folks over at the Center of Excellence Cognitive Interaction Technology (CITEC) in Germany have whipped together a fantastic demo featuring Santa’s extra pair of helping hands. In the two-and-a-half minute video, the robot executes a suite of impressive autonomous stocking-stuffing maneuvers: from recognizing the open hole in the stocking, to grasping specific candies from the cluster of goodies available.

On the hardware-side, the arms appear to be a KUKA-variant, while on the software-side, the visualizations are being handled by the open source robot software ROS‘ RVIZ tool.

If some of the props in the video look familiar, you’ll find that the researchers at CITEC have already explored some stellar perception, classification, and grasping of related research topics. Who knew this pair of hands would be so jolly to clock some overtime this holiday season? The entire video is set to a crisp computer-voiced jingle that serves as a sneaky summary of their approach to this project.

Now, if only we could set these hands off to do our other dirty work….

Continue reading “Santa’s Autonomous Helping Hands Let The Jolly Ol’ Fellow Kick Back This Season”

Robot Vision: Detecting Obstacles With FPGAs And Line Lasers

Somewhere down the road, you’ll find that your almighty autonomous robot chassis is going to need some sensor feedback. Otherwise, that next small step down the road may end with a blind leap off the coffee table. The first low-cost sensors we might throw at this problem would be sonars or IR rangefinders, but there’s a problem: those sensors only really provide distance data back from the pinpoint view directly ahead of them.

Rest assured, [Jonathan] wrote in to let us know that he’s got you covered. Combining a line laser, camera, and an FPGA, he’s able to detect obstacles that fall within the field of view of the camera and laser.

If you thought writing algorithms in software is tricky, wait till to you try hardware! (We know: division sucks!) [Jonathan] knows no fear though; he’s performing gradient computation on the FPGA directly to detect the laser in the camera image at a wicked 30 frames-per-second. Why roll up your sleeves and take the hardware route, you might ask? If we took a CPU-based approach at the tiny embedded-robot scale, Jonathan estimates a mere 10 frames-per-second. With an FPGA, we’re able to process images about as fast as they’re received.

Jonathan is using the Logi Board, a Kickstarter success we’ve visited in the past, and all of his code is up on the Githubs. If you crack it open, you’ll also find that many of his modules are Wishbone compliant, so developing your own projects with just some of these parts has been made much easier than trying to rip out useful features from a sea of hairy logic.

With computer-vision hardware keeping such a low profile in the hobbyist community, we’re excited to hear more about [Jonathan’s] FPGA-based robotics endeavors.

Continue reading “Robot Vision: Detecting Obstacles With FPGAs And Line Lasers”