It’s getting easier and easier to add machine intelligence to your hacks, even to the point where you sometimes don’t have to install any special software. In this case [Dexter Industries] has added the ability to read human emotions to their EmpathyBot robot by making use of Google Cloud Vision.
Press a button on the robot and it moves forward until it’s a certain distance from an object. It then takes a picture and sends it off to Google Cloud Vision along with a request to do face detection. The response that Google returns is in JSON format and, if it finds a face, includes the likelihood of the face being happy, sad, sorrowful or surprised. The robot parses that response and gives an appropriate canned speech using the text-to-speech software, eSpeak e.g. “You seem happy! Tell me why you are so happy!”.
[Dexter] has made the source code available on github. It’s written in python and is easy to read by anyone with even just a little programming experience. The video after the break gives a number of demonstrations, including some with non-human subjects.
Continue reading “Raspberry Pi Robot That Reads Your Emotions”



![A GPIB/IEEE488 plug. Alkamid [CC BY-SA 3.], via Wikimedia Commons](https://hackaday.com/wp-content/uploads/2016/10/a_gpib_plug.jpg?w=400)



At least that’s how [Antibore] handled the situation, and the results are pretty good. He designed his build around an old Raspberry Pi 2 that was hanging around. That required a WiFi adapter, and since he wanted video and audio he needed a camera and mic. The first USB mic had a nice compact design but didn’t perform well, so a gutted gooseneck mic soldered right to the USB connector joined the design spec. A camera module, cell-phone quick charge battery bank, and a 3D printed case round out the BOM. A knitted cozy to keep it looking warm and fuzzy was provided by the mother-to-be — although we think it 