A team from the Max Planck Institute for Intelligent Systems in Germany have developed a novel thumb-shaped touch sensor capable of resolving the force of a contact, as well as its direction, over the whole surface of the structure. Intended for dexterous manipulation systems, the system is constructed from easily sourced components, so should scale up to a larger assemblies without breaking the bank. The first step is to place a soft and compliant outer skin over a rigid metallic skeleton, which is then illuminated internally using structured light techniques. From there, machine learning can be used to estimate the shear and normal force components of the contact with the skin, over the entire surface, by observing how the internal envelope distorts the structured illumination.
The novelty here is the way they combine both photometric stereo processing with other structured light techniques, using only a single camera. The camera image is fed straight into a pre-trained machine learning system (details on this part of the system are unfortunately a bit scarce) which directly outputs an estimate of the contact shape and force distribution, with spatial accuracy reported good to less than 1 mm and force resolution down to 30 millinewtons. By directly estimating normal and shear force components the direction of the contact could be resolved to 5 degrees. The system is so sensitive that it can reportedly detect its own posture by observing the deformation of the skin due its own weight alone!
Some folks bring out an heirloom table runner when they have company, but what if you sewed your own and made it musical? We’d never put it away! [kAi CHENG] has an Instructable about how to recreate his melodic material, and there is a link to his website, which describes his design process, not just the finished product. We have a video below showing a jam session where he exercises a basic function set.
GarageBand is his DAW of choice, which receives translated MIDI from a Lilypad. If you don’t have a Lilypad, any Arduino based on the ATmega328P chip should work seamlessly. Testing shows that conductive threads in the soft circuit results in an occasional short circuit, but copper tape makes a good conductor at the intersections. Wide metallic strips make for tolerant landing pads beneath modular potentiometers fitted with inviting foam knobs. Each twist controls a loop in GarageBand, and there is a pressure-sensitive pad to change the soundset. Of course, since this is all over MIDI, you can customize to your heart’s content.
[Sagar] points out that since the underlying OS of the Kobo device is Linux, it is possible to fake touches to the screen (and therefore trigger page turns) by recording then replaying the appropriate input event. However, there was a more direct solution available to those willing to tamper slightly with the hardware. Touch sensing on the screen is done via an infrared break-beam system. Along two edges of the screen are IR emitters, and opposite the emitters are receivers. Broadly speaking, when a fingertip touches the display a minimum of two IR beams are broken, and the physical location of the fingertip can therefore be determined by analyzing exactly how the IR pattern has been changed.
To spoof page turns, [Sagar] briefly shorts two IR emitters: one on each axis. The sudden winking out of the IR is interpreted by the device as a crisp tap, and the device obediently turns the page. The only hitch is that both IR emitters must be shorted at the same time. If one is shorted before the other, the device ignores it. Double-pole switches would probably do the trick, but with the part bin coming up empty in that respect, [Sagar] instead used a few transistors to accomplish the same thing. A 3D printed enclosure rounds out the whole mod, and a brief video is embedded below.
So you’re building a new mechanical keyboard. Or attaching a few buttons to a Raspberry Pi. Or making the biggest MIDI grid controller the world has ever know. Great! The first and most important engineering question is; how do you read all those buttons? A few buttons on a ‘Pi can probably be directly connected, one for one, to GPIO pins. A mechanical keyboard is going to require a few more pins and probably some sort of matrix scanner. But the grid controller is less clear. Maybe external I/O expanders or a even bigger matrix? Does it still need diodes at each button? To help answer these questions the folks at [openmusiclabs] generated a frankly astounding map which shows, given the number of inputs to scan and pins available, which topology makes sense and roughly how much might it cost. And to top it off they link into very readable descriptions of how each might be accomplished. The data may have been gathered in 2011 but none of the fundamentals here have changed.
How do you read this chart? The X axis is the number of free pins on your controller and the Y is the number of I/Os to scan. So looking at the yellow band across the top, if you need to scan one input it always makes the most sense to directly use a single pin (pretty intuitive, right?). Scrolling down, if you need to read 110 inputs but the micro only has eight pins free there are a couple choices, keys E and F. Checking the legend at the top E is “Parallel out shift register muxed with uC” and F is “Parallel in shift register muxed with uC“. What do those mean? Checking the table in the original post or following the link takes us to a handy descriptive page. It looks like a “parallel out shift register” refers to using a shift register to drive one side of the scan matrix, and “parallel in shift register” refers to the opposite.
When we lose a limb, the brain is really none the wiser. It continues to send signals out, but since they no longer have a destination, the person is stuck with one-way communication and a phantom-limb feeling. The fact that the brain carries on has always been promising as far as prostheses are concerned, because it means the electrical signals could potentially be used to control new limbs and digits the natural way.
Like real skin, the e-dermis has an outer, epidermal layer and an inner, dermal layer. Both layers use conductive and piezoresistive textiles to transmit information about tangible objects back to the peripheral nerves in the limb. E-dermis does this non-invasively through the skin using transcutaneous electrical nerve stimulation, better known as TENS. Here’s a link to the full article published in Science Robotics.
First, the researchers made a neuromorphic model of all the nerves and receptors that relay signals to the nervous system. To test the e-dermis, they used 3-D printed objects designed to be grasped between thumb and forefinger, and monitored the subject’s brain activity via EEG.
For now, the e-dermis is confined to the fingertips. Ideally, it would cover the entire prosthesis and be able to detect temperature as well as curvature. Stay tuned, because it’s next on their list.
We’re slowly moving in the direction where everyone will have a touch screen desk like in the 1982 TRON movie or in the 1987 Star Trek: The Next Generation series with its ubiquitous touchscreen starship controls. [FFcossag] lucked into that future when a local company offered him an industrial 42″ multitouch PC that they were throwing out. A few hacks later and he has us all suitably envious.
Before hacking away though, he had to take care of some magic smoke. The source of this turned out to be yellow goop on the PC’s power supply that had turned conductive across a resistor. Cleaning it fixed the problem.
Moving on to the hacks, he added brightness control by using a potentiometer to control the power to the backlight. Be sure to watch carefully in the video below where he’s attaching a magnet and cord to the potentiometer, and encasing it all in epoxy. At that point, we’re pretty sure we see him spin up a hard drive platter with a sandpaper disk attached to it, forming a bench top disc sander and making us like this hack even more.
He also replaced a small speaker with a larger speaker and amplifier, giving a volume and sound quality difference that’s like night and day. He also added a breakout board with relays for power management, eliminating a seven watt continuous draw when in standby mode.
Be sure to watch the video to the end where he leaves us with a tour of the hacked interior hardware. We like how he’s labeled all his handiwork for any future hacker who might open it up
Heart rate sensors available for DIY use employ photoplethysmography which illuminates the skin and measures changes in light absorption. These sensors are cheap, however, the circuitry required to interface them to other devices is not. [Petteri Hyvärinen] is successfully investigating the use of capacitive touchscreens for heart rate sensing among other applications.
The capacitive sensor layer on modern-day devices has a grid of elements to detect touch. Typically there is an interfacing IC that translates the detected touches into filtered digital numbers that can be used by higher level applications. [optisimon] first figured out a way to obtain the raw data from a touch screen. [Petteri Hyvärinen] takes the next step by using a Python script to detect time variations in the data obtained. The refresh rate of the FT5x06 interface is adequate and the data is sent via an Arduino in 35-second chunks to the PC over a UART. The variations in the signal are very small, however, by averaging and then using the autocorrelation function, the signal was positively identified as a pulse.
A number of applications could benefit from this technique if the result can be replicated on other devices. Older devices could possibly be recycled to become low-cost medical equipment at a fraction of the cost. There is also the IoT side of things where the heart-rate response to media such as news, social media and videos could be used to classify content.