If only we had affordable artificial muscles, we might see rapid advances in prosthetic limbs, robots, exo-skeletons, implants, and more. With cost being one of the major barriers — in addition to replicating the marvel of our musculature that many of us take for granted — a workable solution seems a way off. A team of researchers at MIT present a potential answer to these problems by showing nylon fibres can be used as synthetic muscles.
Some polymer fibre materials have the curious property of increasing in diameter while decreasing in length when heated. Taking advantage of this, the team at MIT were able to sculpt nylon fibre and — using a number of heat sources, namely lasers — could direct it to bend in a specific direction. More complex movement requires an array of heat sources which isn’t practical — yet — but seeing a nylon fibre dance tickles the imagination.
Continue reading “Nylon Fibre Artificial Muscles — Powered by Lasers!”
Mount an umbrella to a drone and there you go, you have a flying umbrella. When [Alan Kwan] tried to do just that he found it wasn’t quite so simple. The result, once he’d worked it out though, is haunting. You get an uneasy feeling like you’re underwater watching jellyfish floating around you.
A grad student in MIT’s ACT (Art, Culture and Technology) program, [Alan’s] idea was to produce a synesthesia-like result in the viewer by having an inanimate object, an umbrella, appear as an animate object, a floating jellyfish. He first tried simply attaching the umbrella to an off-the-shelf drone. Since electronics occupy the center of the drone, the umbrella had to be mounted off-center. But he discovered that drones want most of their mass in the center and so that didn’t work. With the help of a classmate and input from peers and faculty he made a new drone with carbon fiber and metal parts that allowed him to mount the umbrella in the center. To further help with stability, the batteries were attached to the very bottom of the umbrella’s pole.
In addition to just making them fly, [Alan] also wanted the umbrella to gently undulate like a jellyfish, slowly opening and closing a little. He tried mounting servo motors inside the umbrella for the task. These turned out to be too heavy, but also unnecessary. Once flying outside at just the right propeller speed, the umbrellas undulated on their own. Watch them doing this in the video below accompanied by haunting music that makes you feel you’re watching a scene from Blade Runner.
Continue reading “Umbrella Drones — Jellyfish Of The Sky”
Researchers at MIT have used 3D printing to open the door to low-cost, scalable, and consistent generation of microencapsulated particles, at a fraction of the time and cost usually required. Microencapsulation is the process of encasing particles of one material (a core) within another material (a shell) and has applications in pharmaceuticals, self-healing materials, and dye-based solar cells, among others. But the main problem with the process was that it was that it was slow and didn’t scale, and it was therefore expensive and limited to high-value applications only. With some smart design and stereolithography (SLA) 3D printing, that changed. The researchers are not 3D printing these just because they can; they are printing the arrays because it’s the only way they can be made.
Continue reading “3D Printed Nozzles Turbocharge Microsphere Production”
MIT’s Computer Science and Artificial Intelligence Laboratory, CSAIL, put out a paper recently about an interesting advance in 3D printing. Naturally, being the computer science and AI lab the paper had a robotic bend to it. In summary, they can 3D print a robot with a rubber skin of arbitrarily varying stiffness. The end goal? Shock absorbing skin!
They modified an Objet printer to print simultaneously using three materials. One is a UV curing solid. One is a UV curing rubber, and the other is an unreactive liquid. By carefully depositing these in a pattern they can print a material with any property they like. In doing so they have been able to print mono body robots that, simply put, crash into the ground better. There are other uses of course, from joints to sensor housings. There’s more in the paper.
We’re not sure how this compares to the Objet’s existing ability to mix flexible resins together to produce different Shore ratings. Likely this offers more seamless transitions and a wider range of material properties. From the paper it also appears to dampen better than the alternatives. Either way, it’s an interesting advance and approach. We wonder if it’s possible to reproduce on a larger scale with FDM.
Can you make a spectrometer for your home lab all from materials you have sitting around? We might not believe it from a less credible source, but this MIT course does indeed build a spectrometer from foam board using two razor blades as the silt cover and a writable CD as the diffraction grating. The coolest part is removing the metal backing of the CD.
Hackaday reader [gratian] tipped us off about the course available from MIT courseware called Nanomaker. It boils down some fairly complicated experiments to the kind one can do in the home lab without involving thousands of dollars of lab equipment. The whole point is to demystify what we think of as complicated devices and topics surrounding photovoltaics, organic photovoltaics, piezoelectricity and thermoelectricity.
Spectrometers are used to analyze the wavelengths of a light source. Now that you have a measurement tool in hand it’s time to build and experiment with some light sources of your own. Here you can see an LED that is the topic of one of the course labs.
If you have a bit of background in chemistry this is a good step-by-step guide for getting into these types of experiments at home. It reminds us of some of the really cool stuff [Jeri Ellsworth] was doing in her garage lab, like making her own EL panels.
Continue reading “Bring Doping, Microfluidics, Photovoltaics, and More Into the Home”
If it wasn’t for the weird Dutch-Norwegian techno you’d presumably have to listen to forever, [Gianni B.]’s doll house for his daughter, [Rita] makes living in a Barbie World seem like a worthwhile endeavor. True to modern form, it’s got LED lighting. It’s got IoT. It’s got an app and an elevator. It even has a tiny, working, miniature television.
It all started with a Christmas wish. [Rita] could no longer stand to bear the thought of her Barbie dolls living a homeless lifestyle on her floor, begging passing toys for enough monopoly money to buy a sock to sleep under. However, when [Gianni] visited the usual suspects to purchase a dollhouse he found them disappointing and expensive.
So, going with the traditional collaborating-with-Santa ruse, he and his family had the pleasure of collaborating on a dollhouse development project. Each room is lit by four ultra bright LEDs. There is an elevator that’s controlled by an H-bridge module, modified to have electronic braking. [Rita] doesn’t own a Dr. Barbie yet, so safety is paramount.
The brain of the home automation is a PIC micro with a Bluetooth module. He wrote some code for it, available here. He also went an extra step and used MIT’s scratch to make an app interface for the dollhouse. You can see it work in the video after the break. The last little hack was the TV. An old arduino, an SD Card shield, and a tiny 2.4 inch TFT combine to make what’s essentially a tiny digital picture frame.
His daughter’s are overjoyed with the elevation of their doll’s economic class and a proud father even got to show it off at a Maker Faire. Very nice!
Continue reading “Rita’s Dolls Probably Live Better Than You Do”
If a picture is worth a thousand words, a video must be worth millions. However, computers still aren’t very good at analyzing video. Machine vision software like OpenCV can do certain tasks like facial recognition quite well. But current software isn’t good at determining the physical nature of the objects being filmed. [Abe Davis, Justin G. Chen, and Fredo Durand] are members of the MIT Computer Science and Artificial Intelligence Laboratory. They’re working toward a method of determining the structure of an object based upon the object’s motion in a video.
The technique relies on vibrations which can be captured by a typical 30 or 60 Frames Per Second (fps) camera. Here’s how it works: A locked down camera is used to image an object. The object is moved due to wind, or someone banging on it, or any other mechanical means. This movement is captured on video. The team’s software then analyzes the video to see exactly where the object moved, and how much it moved. Complex objects can have many vibration modes. The wire frame figure used in the video is a great example. The hands of the figure will vibrate more than the figure’s feet. The software uses this information to construct a rudimentary model of the object being filmed. It then allows the user to interact with the object by clicking and dragging with a mouse. Dragging the hands will produce more movement than dragging the feet.
The results aren’t perfect – they remind us of computer animated objects from just a few years ago. However, this is very promising. These aren’t textured wire frames created in 3D modeling software. The models and skeletons were created automatically using software analysis. The team’s research paper (PDF link) contains all the details of their research. Check it out, and check out the video after the break.
Continue reading “Interactive Dynamic Video”