It’s always exciting to see the photos from High Altitude Ballooning (HAB) outings. While it’s no surprise that the Raspi is a popular choice—low cost, convenient USB jacks, etc.—this is the first build we’ve seen that uses an OLED during the trip to show real-time data on-screen to be picked up by the on-board webcam. (Though you may have to squint to see it at the bottom middle of the above image).
[Fabrice’s] payload made it to 26,000m, and the screen he chose, an ILSOFT OLED, performed admirably despite the extreme conditions suffered (temperatures can reach -50C). The last time we saw a near-space Raspi payload was a couple of years ago, when [Dave Akerman] was closing in on UK balloon altitude records. [Dave] hasn’t stopped launching balloons, either, testing new trackers and radio modules, as well as his most recent build that sent a Superman action figure to the skies—all recorded in glorious HD.
Check out both [Dave] and [Fabrice’s] blogs for loads of pictures documenting the latest in High Altitude Ballooning, and stay with us after the jump for a quick video of [Fabrice’s] OLED in action.
Continue reading “Throwing Pis into the Stratosphere”
Halloween may have come and gone, but [Luis] sent us this build that you’ll want to check out. An avid Walking Dead fan, he put in some serious effort to an otherwise simple bloody t-shirt and created this see-through “stomach shot” gunshot wound.
The project uses a Raspi running the Pi Camera script to feed video from a webcam on the back of his costume to a 7″ screen on the front. [Luis] attached the screen to a GoPro chest harness—they look a bit like suspenders—to keep it centered, then built up a layer of latex around the display to hide the hard edges and make it more wound-like. Power comes from a 7.4V hobby Lipo battery plugged into a 5V voltage converter.
After ripping a small hole in the back of his t-shirt for the webcam and a large hole in the front for the screen, [Luis] applied the necessary liberal amount of fake blood to finish this clever shotgun blast effect.
The Raspberry Pi has a port for a camera connector, allowing it to capture 1080p video and stream it to a network without having to deal with the craziness of webcams and the improbability of capturing 1080p video over USB. The Raspberry Pi compute module is a little more advanced; it breaks out two camera connectors, theoretically giving the Raspberry Pi stereo vision and depth mapping. [David Barker] put a compute module and two cameras together making this build a reality.
The use of stereo vision for computer vision and robotics research has been around much longer than other methods of depth mapping like a repurposed Kinect, but so far the hardware to do this has been a little hard to come by. You need two cameras, obviously, but the software techniques are well understood in the relevant literature.
[David] connected two cameras to a Pi compute module and implemented three different versions of the software techniques: one in Python and NumPy, running on an 3GHz x86 box, a version in C, running on x86 and the Pi’s ARM core, and another in assembler for the VideoCore on the Pi. Assembly is the way to go here – on the x86 platform, Python could do the parallax computations in 63 seconds, and C could manage it in 56 milliseconds. On the Pi, C took 1 second, and the VideoCore took 90 milliseconds. This translates to a frame rate of about 12FPS on the Pi, more than enough for some very, very interesting robotics work.
There are some better pictures of what this setup can do over on the Raspi blog. We couldn’t find a link to the software that made this possible, so if anyone has a link, drop it in the comments.
[Paul’s] project is a great example of how you can take a simple project and turn it into something more interesting. He built himself a jack-o-lantern with an Internet controlled RGB LED embedded inside.
[Paul] first wired up an RGB LED to a Raspberry Pi. He was sure to wire up each color using a 100ohm resistor to prevent the LED from burning out. The web interface was written in Python. The interface is pretty simple. It consists of three text fields. The user enters a value between 0 and 255 for each of the three LED colors. The program then lights up the LED accordingly.
[Paul] realized he would need a diffuser for the LED in order to really see the blended colors properly. Instead of using a common solution like a ping-pong ball, he opted to get festive and use a plastic jack-o-lantern. [Paul] removed the original incandescent bulb from the lantern and mounted the LED inside instead. The inside of the pumpkin is painted white, so it easily diffuses the light. The result is a jack-o-lantern that glows different colors as defined by his party guests. Be sure to check out the demonstration video below.
The FarmBot team has been pretty busy with their CNC Farming and Gathering machine. The idea is to automate the farming process with precise deployment of tools: plows, seed injection, watering, sensors, etc. An Arduino with an added RAMPS handles the movement, and a Raspi provides internet connectivity. Their prototype has already experienced four major iterations: the first revision addressed bigger issues such as frame/track stability and simplification of parts. Now they’re locking down the specifics on internet-of-things integration and coding for advanced movement functions.
The most recent upgrade provides a significant improvement by overhauling the implementation of the tools. Originally, the team envisioned a single, multi-function tool head design that carried everything around all the time. Problem is, the tool that’s in-use probably works best if it’s lower than the others, and piling them all onto one piece spells trouble. The solution? a universal tool mounting system, of course. You can see them testing their design in a video after the break.
If the FarmBot progress isn’t impressive enough—and admittedly we’d have called project lead [Rory Aronson] crazy for attempting to pull this off…but he did it—the FarmBot crew started and successfully funded an entire sub-project through Kickstarter. OpenFarm is an open-source database set to become the go-to wiki for all things farming and gardening. It’s the result of [Rory] encountering an overwhelming amount of generic, poorly written advice on plant growing, so he just crowdsourced a solution. You know, no sweat.
The project featured in this post is a semifinalist in The Hackaday Prize.
Continue reading “THP Semifinalist: Farmbot”
Public transit can be a wonderful thing. It can also be annoying if the trains are running behind schedule. These days, many public transit systems are connected to the Internet. This means you can check if your train will be on time at any moment using a computer or smart phone. [Christoph] wanted to take this concept one step further for the Devlol hackerspace is Linz, Austria, so he built himself an electronic tracking system (Google translate).
[Christoph] started with a printed paper map of the train system. This was placed inside what began as an ordinary picture frame. Then, [Christoph] strung together a series of BulletPixel2 LEDs in parallel. The BulletPixel2 LEDs are 8mm tri-color LEDs that also contain a small controller chip. This allows them to be controlled serially using just one wire. It’s similar to having an RGB LED strip, minus the actual strip. [Christoph] used 50 LEDs when all was said and done. The LEDs were mounted into the photo frame along the three main train lines; red, green, and blue. The color of the LED obviously corresponds to the color of the train line.
The train location data is pulled from the Internet using a Raspberry Pi. The information must be pulled constantly in order to keep the map accurate and up to date. The Raspberry Pi then communicates with an Arduino Uno, which is used to actually control the string of LEDs. The electronics can all be hidden behind the photo frame, out of sight. The final product is a slick “radar” for the local train system.
The Broadcom SOC in the Raspberry Pi is actually surprisingly powerful, it turns out. It’s actually capable of driving a VGA monitor through the GPIO pins using a handful of resistors.
[Gert van Loo], Raspberry Pi chip architect, wizard, and creator of a number of interesting expansion boards showed off a VGA adapter for the new B+ model at the recent Raspberry Pi Jam in Cambridge this week. Apparently, there is a parallel interface on the SoC that can be used to drive VGA with hardware using a resistor ladder DAC. That’s native VGA at 1080p at 60 fps in addition to HDMI for the Raspberry Pi. Only the new Model B+ has enough pins to do this, but it’s an intriguing little board.
The prospect of having two displays for a Raspberry Pi is very interesting, and the remaining four GPIOs available mean a touch screen could be added to one display, effectively making a gigantic Nintendo DS. Of course there are more practical problems a dual display Raspi solves, like driving a projector for the current crop of DSP/resin 3D printers, while still allowing for a usable interface during a print.
The VGA expansion board, “is likely to have issues with EMC,” which means this probably won’t be a product. Getting a PCB made and soldering SMD resistors isn’t that hard, though, and we’ll post an update when the board files are released.
Thanks [Uhrheber] for sending this one in.