A team of adventurous biohackers have successfully played with an interesting type of chlorophyll, called Chlorin e6 by putting it in their eyes… and the result? Well, they kind of obtained night vision.
Say what? Chlorin e6 is a chlorophyll analog that is found in deep-sea fish, and has been used to treat night blindness in humans (patent). There’s actually lots of research done with the substance, and it has even been used to treat different cancers — but most of the research was performed on lab rats.
So the team decided to take the next step — [Gabriel Licina] volunteered, and they squirted 50uL of e6 into his wide-stretched eyes. It kicks in after about an hour, so they headed outside at night to test his vision capabilities. They started by identifying basic shapes at 10 meters away, no larger than the size of his hand. Then they tried even larger distances. They had people stand at a tree line in different places, and [Gabriel] standing 50 meters away was able to point them out. The control group could barely identify them even a third of the time.
They’ve published a research paper on their findings, and it’s quite the interesting read. Perhaps in the future this can be manufactured in eye drop form for special use cases like hunting, military, or even search and rescue.
[David Nghiem] has been working with circuitry designed to read signals from muscles for many years. After some bad luck with a start-up company, he didn’t give up and kept researching his idea. He has decided to share his innovations with the hacker community in the form of a wearable suit that reads muscle signals.
It turns out that when you flex a muscle, it gives off a signal called a Surface ElectroMyographic signal, or SEMG for short. [David] is using an Arduino, digital potentiometer and a bunch of op amps to read the SEMG signals. LEDs are used to display the signal levels.
The history behind [David’s] project dates back to the late twentieth century, which he eloquently points out – “Holy crap that was a long time ago”. He worked with the MIT Aero Astro Lab and the Boston University Neuromuscular Research Center where he worked on a robotic arm for astronauts. The idea being to apply an opposing force to the arm to help prevent muscle deterioration.
Be sure to check out [David’s] extensive and well documented work, along with the several videos showing his projects at various stages of completion. If this gives you the electromyography bug, check out this guide on detecting the signals and an application of the concept for robotic prosthesis.
Continue reading “Control Stuff With Your Muscles”
We don’t find smartwatches to be supremely usable yet. This one sets a definition for usefulness. The Enigma machine is of course the cipher process used by the Germans during World War II. This Enigma Machine wristwatch is not only functional, but the appearance is modelled after that of the original machine. With the speckled gray/black case and the Enigma badge branding [Asciimation] has done a fine job of mimicking the original feel.
Driving the machine is an Arduino Pro Mini. We’ve seen Arduino Enigma Machines in the past so it’s not surprising to see it again here. The user interface consists of an OLED display at 128×64 resolution, three buttons, with a charging port to the right and on/off switch on the left.
The device is demonstrated after the break. Quite a bit of button presses are used to set up each of the three encoder wheels. But that’s hardly avoidable when you’re not committing to a full keyboard. We’re pretty impressed by the functionality of [Asciimation’s] interface considering it’s hardware simplicity.
This seems perfect for kids that are proving to have an interest in engineering. They learn about ciphers, embedded programming, and mechanical design and crafting (this is a hand-sewn leather wristband). Of course if you build one and start wearing it into the office we won’t judge.
Continue reading “Enigma Machine Wristwatch”
By far the most desirable booth for the crowds at SXSW Create was the Sparkfun quadrant. We call it a quadrant because they had a huge footprint approaching 1/4 the tented area, but it was well used. They brought a number of staff down to Austin in order to give away a legit electronic badge project they call BadgerHack.
We love badge hacking. LOVE IT! But South-by isn’t purely a hardware conference so the badges aren’t made of PCBs (for shame). Add to that, free entry to Create scores you a wristband but no badge.
This is the answer to that, a badge giveaway and build-off aimed at kids but cool enough to make me feel only slightly awful for accepting one when I pretty much knew they were going to run out before the final day was done.
The USB stick PCB is, as you guessed it, an Arduino compatible loaded up with an FTDI chip and an ATmega328p which they call the BadgerStick. Accompanying this is a multiplexed 8×7 LED matrix board. Solder the three pin headers and the battery holder leads, connect to the plastic badge using the supplied double-stick tape, and you have a badge that scrolls a message in LEDs.
What an awesome giveaway. I really like it that they didn’t cut corners here. First off, the kids will value the badge much more because they had to actually assemble it rather than just being handed a finished widget. Secondly, there is the USB to serial chip and USB footprint that means they can reprogram it without any extra equipment. And an LED matrix… come on that’s just a gateway drug to learning Wiring. Bravo Sparkfun and Atmel for going this route with your marketing bucks.
The badge activity rounded out with some hardware interfacing. There’s a 3-pin socket that attendees could plug into 4 different stations around the booth. Once done they received a coupon code for Sparkfun that scrolls whenever the badge is booted up. For some at-home fun, the writeup (linked at the top) for the BadgerHack firmware is quite good. It offers advice on changing what is displayed on the badge and outlines how to build a game of Breakout with just a bit of added hardware.
[Roy Shilkrot] and his fellow researchers at the MIT Media Lab have developed the FingerReader, a wearable device that aids in reading text. Worn on the index finger, it receives input from print or digital text and outputs spoken words – and it does this on-the-go. The FingerReader consists of a camera and sensors that detect the text. A series of algorithms the researchers created are used along with character recognition software to create the resulting audio feedback.
There is a lot of haptic feedback built into the FingerReader. It was designed with the visually impaired as the primary user for times when Braille is not practical or simply unavailable. The FingerReader requires the wearer to make physical contact with the tip of their index finger on the print or digital screen, tracing the line. As the user does so, the FingerReader is busy calculating where lines of text begin and end, taking pictures of the words being traced, and converting it to text and then to spoken word. As the user reaches the end of a line of text or begins a new line, it vibrates to let them know. If a user’s finger begins to stray, the FingerReader can vibrate from different areas using two motors along with an audible tone to alert them and help them find their place.
The current prototype needs to be connected to a laptop, but the researchers are hoping to create a version that only needs a smartphone or tablet. The videos below show a demo of the FingerReader. For a proof-of-concept, we are very impressed. The FingerReader reads text of various fonts and sizes without a problem. While the project was designed primarily for the blind or visually impaired, the researchers acknowledge that it could be a great help to people with reading disabilities or as a learning aid for English. It could make a great on-the-go translator, too. We hope that [Roy] and his team continue working on the FingerReader. Along with the Lorm Glove, it has the potential to make a difference in many people’s lives. Considering our own lousy eyesight and family’s medical history, we’ll probably need wearable tech like this in thirty years!
Continue reading “Trace Your Book or Kindle with the FingerReader”
If you’ve been killing time texting or chatting with your pals via smart phone, odds are pretty good that you’re not giving much thought to the two senses that make it happen: your sight and your hearing. Those who are deafblind, however, cannot participate in these activities; and for many, the remote communication that most of us enjoy with our phones simply isn’t possible. Enter Berlin University of the Arts Design Research Lab. Here, they’ve developed the Mobile Lorm Glove, a haptics device that enables two-way remote communication via smart phone.
For the deafblind, Lorm is the tactile technique for communication. Lorm is a series of hand-tracing gestures that map to characters of the alphabet. To communicate with others, the gloved user can trace Lorm directly onto the pressure-sensitive inputs on the palm of the hand. To receive messages, small vibration motors on the back of the hand vibrate to indicate the message encoded in Lorm.
Originally, to communicate with the deafblind, we must first learn Lorm. With the Mobile Lorm Glove, however, we need only know how to send text messages, and the Lorm-decoding is handled with a look-up table running on our classic Atmega328 microcontroller. For the sharp-eyed, the back-side of the glove seems limited in its capability to transcribe continuous finger traces into discrete motor vibrations. However, with four shift-registers and 32 levels of motor-intensities, the designers address each motor with a technique called “funneling illusion” where continuous movement is simulated by gradually changing the intensity from motor to motor. For more tricks and details, take a look at their conference paper.
By wearing the glove, everyday communication can be made far easier with anyone with a smart phone. We’re jazzed that just a Bluetooth module, an Atmega328, and a collection of pressure sensors and motors can enable any cell phone user to circumvent the learning curve and open up a new conversation.
Continue reading “Mobile Lorm Glove Puts Texting Back Into Everyone’s Hands”
[James Bruton] is busy working on his latest project, a “scrap metal sculpture”-inspired Alien Xenomorph suit. However, he wanted to get a boost in height as well as a digitigrade stance. To that end, [James] 3D-printed a pair of customized stilts. Each stilt consisted of a lifter with several parts laminated together using acetone. He bolted an old pair of shoes onto the stilts, adding straps across the toes to keep the shoes from lifting up.
While the stilts worked very well, [James] wanted to add soles to them to give him some traction as he walked – falling while in a Xenomorph costume composed of sharp plastic sounds painful enough! He decided to hybrid print the soles using ABS and Ninjaflex. The ABS part of the sole was then acetone-welded to the bottom of the stilts.
[James] hopes to add some claws for effect, so long as they don’t impede his walking too much. He has already completed a good amount of the 3D-printed suit. We know the finished project is going to be amazing: [James] has created everything from Daleks to Iron Man!
Continue reading “Walk Like A Xenomorph”