A DIY Pick And Place You Can Build Right Now

There have been quite a few DIY pick and place projects popping up recently, but most of them are limited to conceptual designs or just partially working prototypes. [Juha] wrote in to let us know about his project, LitePlacer, which is a fully functional DIY pick and place machine with working vision that can actually import BOMs and place parts as small as 0402 with pretty good accuracy.

LitePlacer UIWhile some other DIY pick and place setups we’ve featured use fairly exotic setups like delta bots, this machine is built around typical grooved bearings and extruded aluminum. The end effector includes a rotating vacuum tip and a camera mounted alongside the tip. The camera provides feedback for locating fiducials and for finding the position of parts. Instead of using feeders for his machine, [Juha] opted to pick parts directly from pieces of cut tape. While this might be inconvenient if you’re placing large quantities of a single part, it helps keep the design simple.

The software that runs the machine is pretty sophisticated. After a bit of configuration it’s able to import a BOM with X/Y information and start placing within seconds. It also uses the camera to calibrate the needle, measure the PCB  using the fiducials, and pinpoint the location of cut tape sections.

If you want to build your own machine, [Juha] published detailed instructions that walk you through the entire assembly process. He’s also selling a kit of parts if you don’t want to source everything yourself. Check out the video after the break to see the machine import a BOM and place some parts (all the way down to 0402).

Continue reading “A DIY Pick And Place You Can Build Right Now”

A Camera With Computer Vision

Computer vision is a tricky thing to stuff into a small package, but last year’s Hackaday Prize had an especially interesting project make it into the 50 top finalists. The OpenMV is a tiny camera module with a powerful microcontroller that will detect faces, take a time-lapse, record movies, and detect specific markers or colors. Like a lot of the great projects featured in last year’s Hackaday Prize, this one made it to Kickstarter and is, by far, the least expensive computer vision module available today.

[Ibrahim] began this project more than a year ago when he realized simple serial JPEG cameras were ludicrously expensive, and adding even simple machine vision tasks made the price climb even higher. Camera modules that go in low-end cell phones don’t cost that much, and high-power ARM microcontrollers are pretty cheap as well. The OpenMV project started, and now [Ibrahim] has a small board with a camera that runs Python and can be a master or slave to Arduinos or any other microcontroller board.

The design of the OpenMV is extraordinarily clever, able to serve as a simple camera module for a microcontroller project, or something that can do image processing and toggle a few pins according to logic at the same time. If you’ve ever wanted a camera that can track an object and control a pan/tilt servo setup by itself, here you go. It’s a very interesting accessory for robotics platforms, and surely something that could be used in a wide variety of projects.

Robot Vision: Detecting Obstacles With FPGAs And Line Lasers

Somewhere down the road, you’ll find that your almighty autonomous robot chassis is going to need some sensor feedback. Otherwise, that next small step down the road may end with a blind leap off the coffee table. The first low-cost sensors we might throw at this problem would be sonars or IR rangefinders, but there’s a problem: those sensors only really provide distance data back from the pinpoint view directly ahead of them.

Rest assured, [Jonathan] wrote in to let us know that he’s got you covered. Combining a line laser, camera, and an FPGA, he’s able to detect obstacles that fall within the field of view of the camera and laser.

If you thought writing algorithms in software is tricky, wait till to you try hardware! (We know: division sucks!) [Jonathan] knows no fear though; he’s performing gradient computation on the FPGA directly to detect the laser in the camera image at a wicked 30 frames-per-second. Why roll up your sleeves and take the hardware route, you might ask? If we took a CPU-based approach at the tiny embedded-robot scale, Jonathan estimates a mere 10 frames-per-second. With an FPGA, we’re able to process images about as fast as they’re received.

Jonathan is using the Logi Board, a Kickstarter success we’ve visited in the past, and all of his code is up on the Githubs. If you crack it open, you’ll also find that many of his modules are Wishbone compliant, so developing your own projects with just some of these parts has been made much easier than trying to rip out useful features from a sea of hairy logic.

With computer-vision hardware keeping such a low profile in the hobbyist community, we’re excited to hear more about [Jonathan’s] FPGA-based robotics endeavors.

Continue reading “Robot Vision: Detecting Obstacles With FPGAs And Line Lasers”

DIY Pick And Place Just Getting Under Way

diy-pick-and-place

It’s not totally fair to say that this project is just getting under way. But the truth is it neither picks nor places so there’s a long road still to travel. But we’re impressed with the demonstrations of what [Daniel Amesberger] has achieved thus far. Using the simplest of CNC mills he’s finished the frame and gantry for the device. You can see some of the parts on the left after going though an anodizing process that leaves them with that slick black finish.

The demo video shows off the device by driving it with a joystick. It’s fast, which gives us hope that this will rival some of the low-end commercial pick and place machines. He’s already been working on the software, which runs on a mini ITX form factor computer. This includes a gerber file interpreter and some computer vision for a visual check on part placement. He hasn’t gotten around to building the parts feeders but we’ll keep you updated as we hear back from him.

Continue reading “DIY Pick And Place Just Getting Under Way”

Flocking Behavior Using Mindstorm Robots

flock-behavior-with-mindstorm

Do you ever wonder why geese always fly together in a V-shape? We’re not asking about the fact that it makes the work load much less for all but the lead goose. We mean how is it that all geese know to form up like this? It’s is the act of flocking, and it’s long been a subject of fascination when it comes to robotics. [Scott Snowden] researched the topic while working on his degree a few years ago. Above you can see the demonstration of the behavior using LEGO Mindstorm robots. That’s certainly interesting and you’ll want to check out the video after the break. But his offering doesn’t end with the demo. He also posted a huge article about his work that will provide days of fascinating reading.

We can’t begin to scratch the surface of all that he covers, but we can give you a quick primer on his Mindstorm (NXT) setup. He uses these three bots along with a central brick (the computer part of the NXT hardware) which communicates with them. This lets him use a wide range of powerful tools like MatLab and Processing to recognize each robot with a top-down camera, passing it data based on info harvested with computer vision. From there it’s a wild ride of modeling the behavior as a set of algorithms.

Continue reading “Flocking Behavior Using Mindstorm Robots”

Using OpenCV With The Raspberry Pi

When we first heard of the Raspberry Pi we were elated that projects that once required a full-blown computer could now be done on a tiny, and cheap board running Linux. Unfortunately, we haven’t seen much in the way of using computer vision algorithms on the Raspi, but thanks to [Lentin] the world of OpenCV is now accessable to Raspberry Pi users everywhere.

[Lentin] didn’t feel like installing OpenCV from its source, a process that takes the better part of a day. Instead, he installed it using the synaptic package manager. After connecting a webcam, [Lentin] ssh’d into his Raspi and installed a face detection example script that comes with OpenCV.

It should be noted that [Lentin]’s install of OpenCV isn’t exactly fast, but for a lot of projects being able to update a face tracker five times a second is more than enough. Once the Raspberry Pi camera module is released the speed of face detection on a Raspi should increase dramatically, though, leading to even more useful computer vision builds with the Raspberry Pi.

Pick And Place That Can’t Pick Or Place… But It Looks Very Promising

This sexy piece of CNC can really fly. It’s a pick and place machine which [Danh Trinh] has been working on. The thing is, so far it lacks the ability to move components at all. But the good news is the rest of the system seems to be there.

He posted a demo video of his progress so far which you can see embedded after the break. He starts of by showing off his computer vision software which he wrote in C#. The demonstration includes the view from the gantry-mounted camera, as well as the computer filtering which seems to accurately locate the solder pads and silk screen on the PCB. The second half of the video looks at the hardware seen above. It’s just executing some static code but the whine of those stepper motors is music to our ears. [Danh] reports that the movements of the needle that will eventually serve as the tip of the vacuum tweezer seem to be very accurate.

These home-built pick and place projects are quite a challenge, but we’ve seen a lot of really awesome work on them lately.

Continue reading “Pick And Place That Can’t Pick Or Place… But It Looks Very Promising”