Some plants react quickly enough for our senses to notice, such as a Venus flytrap or mimosa pudica. Most of the time, we need time-lapse photography at a minimum to notice while more exotic sensors can measure things like microscopic pores opening and closing. As with any sensor reading, those measurements can be turned into action through a little trick we call automation. [Harpreet Sareen] and [Pattie Maes] at MIT brought these two ideas together in a way which we haven’t seen before where a plant has taken the driver’s seat in a project called Elowan. Details are sparse but the concept is easy enough to grasp.
We are not sure if this qualifies as a full-fledged cyborg or if this is a case of a robot using biological sensors. Maybe it all depends on which angle you present this mixture of plant and machine. Perhaps it is truly is the symbiotic relationship that the project claims it to be. The robot would not receive any instructions without the plant and the plant would receive sub-optimal light without the robot. What other ways could plants be integrated into robotics to make it a bona fide cyborg?
Wind turbines are great when the wind flow is predictable. In urban environments, especially in cities with skyscrapers, wind patterns can be truly chaotic. What you need, then, is a wind turbine that works no matter which way the wind blows. And just such a turbine has won the global first prize James Dyson Award. Check out their video below the break.
The turbine design is really neat. It’s essentially a sphere with vents oriented so that it’s always going to rotate one way (say, clockwise) no matter where the wind hits it. The inventors say they were inspired by NASA’s Tumbleweed project, which started off as a brainstorming session and then went on to roll around Antarctica. We tumbled into this PDF, and this summary report, but would love more info if any of you out there know something about Tumbleweeds.
Back to the turbine, though. How efficient is it? How likely is it to scale? How will a 3D-printed version drive a junk-bin brushless motor on my balcony? The jury is still out. But if a significant portion of the wind comes from otherwise unusable directions, this thing could be a win. Who’s going to be the first to 3D print one?
The latest craze in revolutionary materials science is no longer some carbon nanotube, a new mysterious alloy, or biodegradeable plastic. It seems as though a lot of new developments are coming out of the biology world, specifically from mycologists who study fungi. While the jury’s still out on whether or not it’s possible to use fungi to build a decent Star Trek series, researchers have in fact been able to use certain kinds of it to build high-performing insulation.
The insulation is made of the part of the fungus called the mycelium, rather than its more familiar-looking fruiting body. The mycelium is a strand-like structure of fungus which grows through materials in order to digest them. This could be mulch, fruit, logs, straw, crude oil, or even live insects, and you might have noticed it because it’s often white and fuzzy-looking. The particular type of mycelium used here is extremely resistant to changes in temperature so is ideal for making insulation. As a bonus, it can be grown, not manufactured, and can use biological waste products as a growing medium. Further, it can grow to fit the space it’s given, and it is much less environmentally harmful than existing forms of insulation.
The wildfires in California are now officially the largest the state has ever seen. Over 50,000 people have been displaced from their homes, hundreds are missing, and the cost in property damage will surely be measured in the billions of dollars when all is said and done. With a disaster of this scale just the immediate effects are difficult to conceptualize, to say nothing of the collateral damage.
While not suggesting their situation is comparable to those who’ve lost their homes or families, Electric Imp CEO [Hugo Fiennes] has recently made a post on their blog calling attention to the air quality issues they’re seeing at their offices in Los Altos. To quantify the problem so that employees with respiratory issues would know the conditions before they came into work, they quickly hacked together a method for displaying particulate counts in their Slack server.
The key to the system is one of the laser particle sensors that we’re starting to see more of thanks to a fairly recent price drop on the technology. A small fan pulls air to be tested into the device, where a very sensitive optical sensor detects the light reflected by particles as they pass through the laser beam. The device reports not only how many particles are passing through it, but how large they are. The version of the sensor [Hugo] links to in his blog post includes an adapter board to make it easier to connect to your favorite microcontroller, but we’ve previously seen DIY builds which accomplish the same goal.
[Hugo] then goes on to provide firmware for the Electric Imp board that reads the current particulate counts from the sensor and creates a simple web page that can be viewed from anywhere in the world to see real-time conditions at the office. From there, this data can be plugged into a Slack webhook which will provide an instantaneous air quality reading anytime a user types “air” into the channel.
We’ve seen all sorts of 3D-printers on these pages before. From the small to the large, Cartesians and deltas, and printers that can squeeze out plastic, metal, and even concrete. But this appears to be the first time we’ve ever featured a paper-pulp extruding 3D-printer.
It’s fair to ask why the world would need such a thing, and its creator, [Beer Holthuis], has an obvious answer: the world has a lot of waste paper. Like 80 kg per person per year. Thankfully at least some of that is recycled, but that still leaves a lot of raw material that [Beer] wanted to put to work. Build details on the printer are sparse, but from the photos and the video below it seems clear how it all went together. A simple X-Y-Z gantry moves a nozzle over the build platform. The nozzle, an order of magnitude or two larger than the nozzles most of us are used to, is connected to an extruder by a plastic hose. The extruder appears to be tube with a stepper-driven screw that lowers a ram down onto the pulp, squeezing it into the hose. [Beer] notes that the pulp is mixed with a bit of “natural binder” to allow the extruded pulp to keep its shape. We found the extrusion process to be just a wee bit repulsive to watch, but fascinating nonetheless, and the items he’s creating are certainly striking in appearance.
Modern agricultural equipment has come a long way, embracing all kinds of smart features and electronic controls. While some manufacturers would prefer to be the sole gatekeepers of the access to these advanced features, that hasn’t stopped curious and enterprising folks from working on DIY solutions. One such example is this self-steering tractor demo by [Coffeetrac], which demonstrates having a computer plot and guide a tractor through an optimal coverage pattern.
A few different pieces needed to come together to make this all work. At the heart of it all is [Coffeetrac]’s ESP32-based Autosteer controller, which is the hardware that interfaces to the tractor and allows for steering and reading sensors electronically. AgOpenGPS is the software that reads GPS data, interfaces to the Autosteer controller, and tells equipment what to do; it can be thought of as a mission planner.
[Coffeetrac] put it all together with everything controlled by a tablet mounted in the tractor’s cab. The video is embedded below, complete with a “cockpit view” via webcam right alongside the plotted course and sensor data.
Our five rounds of Hackaday Prize 2018 challenges have just wrapped up, and we’re looking forward to see where the chips fall in the final ranking. While we’re waiting for the winners to be announced at Hackaday Superconference, it’s fun to take a look back at one of our past winners. Watch [Reinier van der Lee] give the latest updates on his Vinduino project (video also embedded after the break) to a Hackaday Los Angeles meetup earlier this year.
Vinduino started with [Reinier]’s desire to better understand what happens to irrigation water under the surface, measuring soil moisture at different depths. This knowledge informs more efficient use of irrigation water, as we’ve previously covered in more detail. What [Reinier] has been focused on is improving usability of the system by networking the sensors wirelessly versus having to walk up and physically attach a reader unit.
His thought started the same as ours – put them on WiFi! But adding WiFi coverage across his entire vineyard was not going to be cost-effective. After experimenting with various communication schemes, he has settled on LoRa. Designed to trade raw bandwidth for long range with low power requirements, it is a perfect match for a network of soil moisture sensors.
In the video [Reinier] gives an overview of LoRa for those who might be unfamiliar. Followed by results of his experiments integrating LoRa functionality into Vinduino, and ending with a call to action for hackers to help grow the LoRa network. It sounds like he’s become quite the champion for the cause! He’s even giving a hands-on workshop at Supercon where you can build your own LoRa connected sensor. (Get tickets here.)
We’re always happy to see open-source hardware projects like Vinduino succeed, transitioning to a product that solve real world problems. We know there are even more promising ideas out there, which is why Hackaday’s sister company Tindie is funding a Project to Product program to help this year’s winners follow in Vinduino’s footsteps. We look forward to sharing more success stories yet to come.