Brain on a chip setup with a hand and a dropper

Gray Matter On A Chip: Building An Artificial Brain With Luminol

Ever wondered if you could build a robot controlled by chemical reactions? [Marb] explores this wild concept in his video, merging chemistry and robotics in a way that feels straight out of sci-fi. From glowing luminol reactions to creating artificial logic gates, [Marb]—a self-proclaimed tinkerer—takes us step-by-step through crafting the building blocks for what might be the simplest form of a chemical brain.

In this video, the possibilities of an artificial chemical brain take centre stage. It starts with chemical reactions, including a fascinating luminol-based clock reaction that acts as a timer. Then, a bionic robot hand makes its debut, complete with a customised interface bridging the chemical and robotic worlds. The highlight? Watching that robotic hand respond to chemical reactions!

The project relies on a “lab-on-a-chip” approach, where microfluidics streamline the processes. Luminol isn’t just for forensic TV shows anymore—it’s the star of this experiment, with resources like this detailed explanation breaking down the chemistry. For further reading, New Scientist has you covered.

We’ve had interesting articles on mapping the human brain before, one on how exactly brains might work, or even the design of a tiny robot brain. Food for thought, or in other words: stirring the gray matter.

Continue reading “Gray Matter On A Chip: Building An Artificial Brain With Luminol”

Robot Hand Looks And Acts Like The Real Thing

Throughout history, visions of the future included human-looking robots. These days we have plenty of robots, but they don’t look like people. They look like disembodied arms, cars, and over-sized hockey pucks concealing a vacuum cleaner. Of course there’s still demand for humanoid robots like Commander Data, but there are many challenges: eyes, legs, skin, and hands. A company known as Clone may have the solution for that last item. The Clone Hand is “the most human-level musculoskeletal hand in the world,” according to the company’s website.

The 0.75 kg hand and forearm offer 24 degrees of freedom and two hours of battery life. It sports 37 muscles and carbon fiber bones. The muscle fibers can cycle over 650,000 times. You can watch the hand in action in the video below.

There is a hydraulic pump that the company likens in size to a human heart. The hand can also sense for feedback purposes. If you want to build your own, you’ll have to figure it out yourself. The Clone Hand is proprietary, but it does show what is in the art of the possible. The company claims they cost under $3,000, but it isn’t clear if that’s their cost or a projected future retail price.

Of course, human hands aren’t always the perfect robot manipulator. But when you need a realistic hand, you really need it. We see a lot of attempts at realistic hands, and we have to say they are getting better.

Continue reading “Robot Hand Looks And Acts Like The Real Thing”

Shake With Your New Robot Hand

Korean researchers have created a very realistic and capable robot hand that looks very promising. It is strong (34N of grip strength) and reasonably lightweight (1.1 kg), too. There are several videos of the hand in action, of which you can see two of them below including one where the hand uses scissors to cut some paper. You can also read the full paper for details.

Continue reading “Shake With Your New Robot Hand”

Haptic Glove Controls Robot Hand Wirelessly

[Miller] wanted to practice a bit with some wireless modules and wound up creating a robotic hand he could teleoperate with the help of a haptic glove. It lookes highly reproducible, as you can see the video, below the break.

The glove uses an Arduino’s analog to digital converter to read some flex sensors. Commercial flex sensors are pretty expensive, so he experimented with some homemade sensors. The ones with tin foil and graphite didn’t work well, but using some bent can metal worked better despite not having good resolution.

Continue reading “Haptic Glove Controls Robot Hand Wirelessly”

3D Printed Robotic Arms For Sign Language

A team of students in Antwerp, Belgium are responsible for Project Aslan, which is exploring the feasibility of using 3D printed robotic arms for assisting with and translating sign language. The idea came from the fact that sign language translators are few and far between, and it’s a task that robots may be able to help with. In addition to translation, robots may be able to assist with teaching sign language as well.

The project set out to use 3D printing and other technology to explore whether low-cost robotic signing could be of any use. So far the team has an arm that can convert text into finger spelling and counting. It’s an interesting use for a robotic arm; signing is an application for which range of motion is important, but there is no real need to carry or move any payloads whatsoever.

Closeup of hand actuators and design. Click to enlarge.

A single articulated hand is a good proof of concept, and these early results show some promise and potential but there is still a long ways to go. Sign language involves more than just hands. It is performed using both hands, arms and shoulders, and incorporates motions and facial expressions. Also, the majority of sign language is not finger spelling (reserved primarily for proper names or specific nouns) but a robot hand that is able to finger spell is an important first step to everything else.

Future directions for the project include adding a second arm, adding expressiveness, and exploring the use of cameras for the teaching of new signs. The ability to teach different signs is important, because any project that aims to act as a translator or facilitator needs the ability to learn and update. There is a lot of diversity in sign languages across the world. For people unfamiliar with signing, it may come as a surprise that — for example — not only is American Sign Language (ASL) related to French sign language, but both are entirely different from British Sign Language (BSL). A video of the project is embedded below.

Continue reading “3D Printed Robotic Arms For Sign Language”

Robot Hand Goes Wireless

We can’t decide if [MertArduino’s] robotic hand project is more art or demonstration project. The construction using springs, fishing line, and servo motors isn’t going to give you a practical hand that could grip or manipulate anything significant. However, the project shows off a lot of interesting construction techniques and is a fun demonstration for using nRF24L01 wireless in a project. You can see a video of the contraption, below.

A glove uses homemade flex sensors to send wireless commands to the hand. Another Arduino drives an array of servo motors that make the fingers flex. You don’t get fine control, nor any real grip strength, but the hand more or less will duplicate your movements. We noticed one finger seemed poorly controlled, but we suspect that was one of the homemade flex sensors going rouge.

Continue reading “Robot Hand Goes Wireless”

Hackaday Prize Entry: Open-Source Myoelectric Hand Prosthesis

Hands can grab things, build things, communicate, and we control them intuitively with nothing more than a thought. To those who miss a hand, a prosthesis can be a life-changing tool for carrying out daily tasks. We are delighted to see that [Alvaro Villoslada] joined the Hackaday Prize with his contribution to advanced prosthesis technology: Dextra, the open-source myoelectric hand prosthesis.

dextra_handDextra is an advanced robotic hand, with 4 independently actuated fingers and a thumb with an additional degree of freedom. Because Dextra is designed as a self-contained unit, all actuators had to be embedded into the hand. [Alvaro] achieved the necessary level of miniaturization with five tiny winches, driven by micro gear motors. Each of them pulls a tendon that actuates the corresponding finger. Magnetic encoders on the motor shafts provide position feedback to a Teensy 3.1, which orchestrates all the fingers. The rotational axis of the thumb is actuated by a small RC servo.

mumai_boardIn addition to the robotic hand, [Alvaro] is developing his own electromyographic (EMG) interface, the Mumai, which allows a user to control a robotic prosthesis through tiny muscle contractions in the residual limb. Just like Dextra, Mumai is open-source. It consists of a pair of skin electrodes and an acquisition board. The electrodes are attached to the muscle, and the acquisition board translates the electrical activity of the muscle into an analog voltage. This raw EMG signal is then sampled and analyzed by a microcontroller, such as the ESP8266. The microcontroller then determines the intent of the user based on pattern recognition. Eventually this control data is used to control a robotic prosthesis, such as the Dextra. The current progress of both projects is impressive. You can check out a video of Dextra below.

Continue reading “Hackaday Prize Entry: Open-Source Myoelectric Hand Prosthesis”