They say necessity is the mother of invention. But if the thing you need has already been invented but is extremely expensive, another mother of invention might be budget overruns. That was the case when [klinstifen]’s local government decided to put in countdown clocks at bus stops, at a whopping $25,000 per clock. Thinking that was a little extreme, he decided to build his own with a much smaller price tag.
The project uses a Raspberry Pi Zero W as its core, and a 16×32 RGB LED matrix for a display. Some of the work is done already, since the bus system has an API that is readily available for use. The Pi receives the information about bus schedules through this API and, based on its location, is able to determine the next bus arrival time and display it on the LED matrix. With the custom 3D printed enclosure and all of the other material, the cost of each clock is only $100, more than two orders of magnitude less expensive.
Hopefully the local government takes a hint from [klinstifen] and decides to use a more sane solution. In the meantime, you might be able to build your own mass transit clock that you can use inside your own house, rather than at the train station, if you’re someone who has a hard time getting to the bus stop on time.
Continue reading “GuerillaClock Could Save This City Thousands”
While a lot of hardware gets put on the “Internet of Things” with only marginal or questionable benefits (or with hilariously poor security), every now and then a project makes use of this new platform in a way that illustrates the strengths of IoT. [ThingEngineer] turned to this platform as a cost-effective solution for an automatic gate, since new keyfobs were too expensive and a keypad was not an option.
Using an Electric IMP, [ThingEngineer] began by installing his IoT patch into the LiftMaster gate control box. This particular gate has easily accessible points that the controller can access to determine the gate’s status, so from there, an API was written to do the heavy lifting. A web server was deployed as well, so anyone with access can use a smartphone or other device to open the gate.
For anyone else looking to deploy a similar IoT solution, [ThingEngineer] has put all of the project code, schematics, and a thorough write-up about the project on his GitHub page. There are many useful ways to get on board the Internet of Things, though; so many that it’s been possible to win a substantial prize for using it in a creative way.
It’s that time of year again, and with Halloween arguably being the hacker’s perfect holiday, we’re starting to see a tick up in projects with a spooky theme. Most seem to do with making some otherwise tame Halloween decorations scarily awesome, but this is different — using artificial intelligence to search for ghosts.
It seems like [Matt Reed]’s “DeepWhisper” project is meant to be taken as light-hearted fun for the spooky season, but there may be a touch of seriousness to his efforts to listen in on ghostly conversations. The principle behind this is electronic voice phenomena (EVP), whereby the metabolically and/or dimensionally challenged are purported to influence electronic systems, resulting in heavily processed audio clips that seem to have a whispered endearment from the departed or a threat from a malevolent spirit. DeepWhisper takes this a step further by using a Raspberry Pi to feed audio into the Google Cloud Speech API for analysis. If anything is whispered in one of the 110 or so languages Google knows, it’ll get displayed on a screen. [Matt] plans to set DeepWhisper up in the aptly-named Butchertown section of Nashville and live-stream the results next week.
It’ll be interesting to see what Google’s neural network makes out of the random noise it will probably only ever hear. And [Matt] is planning on releasing his code for all to see, so there may be some valuable cloud techniques to learn from DeepWhisper. But in the unlikely event that he does discover ghosts, it’s nice to know you can have the tools and the talent to bust ’em.
Continue reading “Listen to the Netherworld with Artificial Intelligence”
When you ask for recommendations on which 3D printer to buy, damn the cost, the Ultimaker is consistently at the top of the list. There’s a reason for the popularity of this printer — it’s easy to use, extremely high quality, and has an entire freakin’ Linux system running somewhere under the hood. That last bit is opening up a few doors to some interesting hacks, like using a 3D printer as an RGB LED.
While this is just a simple test of the Ultimaker API, it’s surprisingly high up on the Google results when you search, ‘Ultimaker API’. That’s a shame, because there’s a lot of power under the hood of this printer. If you have some sort of mod you’d like to throw into the ring, here’s the Hackaday Tip Line.
You can check out the demo video of this hack below.
Continue reading “Blinking LEDs on the Internet of Printers”
One of the companion technologies in the developing field of augmented reality is gesture tracking. It’s one thing to put someone in a virtual or augmented world, but without a natural way to interact inside of it the user experience is likely to be limited. Of course, gestures can be used to control things in the real world as well, and to that end [Sarah]’s latest project uses this interesting human interface device to control a drone.
The project uses a Leap Motion sensor to detect and gather the gesture data, and feeds all of that information into LabVIEW. A Parrot AR Drone was chosen for this project because of a robust API that works well with this particular software suite. It seems as though a lot of the grunt work of recognizing gestures and sending commands to the drone are taken care of behind-the-scenes in software, so if you’re looking to do this on your own there’s likely to be quite a bit more work involved. That being said, it’s no small feat to get this to work in the first place and the video below is worth a view.
To some, gestures might seem like a novelty technology with no real applications, but they do have real-world uses for people with disabilities or others with unusual workflow that require a hands-free approach. So far we’ve seen hand gesture technologies that drive cars, help people get around in the physical world, and even play tetris.
Continue reading “Drone Takes Off With a Flick of the Wrist”
A Raspberry Pi kicking around one’s workbench is a project waiting to happen — if they remain unused long enough to be considered a ‘spare.’ If you find you’ve been pining after an Alexa or your own personal J.A.R.V.I.S., [Novaspirit Tech] might be able to help you out — provided you have a USB mic and speaker handy — with an accessible tutorial for setting up Google Assistant on your Pi.
A quick run-through on enabling a fresh API client on Google’s cloud platform, [Novaspirit] jumps over to the Raspbian console to start updating Python and a few other dependencies. Note: this is being conducted in the latest version of Raspbian, so be sure to update before you get underway with all of your sudos.
Once [Novaspirit] gets that sorted, he sets up an environment to run Google Assistant on the Pi, authenticates the process, and gets it running after offering a couple troubleshooting tips. [Novaspirit] has plans to expand on this further in the near future with some home automation implementation, but this is a great jumping-off point if you’ve been looking for a way to break into some high-tech home deliciousness — or something more stripped-down — for yourself. Check out the video version of the tutorial after the break if you like watching videos of guys typing away at the command line.
Continue reading “Sudo Google Assistant”
[Erhan] has been playing around with the Telegram instant messaging service. Initially, he worked out how to turn on and off LEDs from his cell phone: he sent commands from the phone through the Telegram bot API, to a computer that’s connected over serial to an MSP430 board that actually controlled the LEDs.
But that’s a little bit complicated. Better to cut out the middleman (err…microcontroller) and implement the Telegram reception and LED blinking on a Raspberry Pi. For a project that’s already using a Pi, using the instant messaging service’s resources is a very simple way to interface to a cellphone.
The code for both the standalone RPi project and the MSP430-based microcontroller application are available at [Erhan]’s GitHub. You’re going to be installing Node.js for their telegram-bot-api and jumping through the usual OAuth hoops to get your bot registered with Telegram. But once you’ve done that all on the Raspberry Pi (or target computer of your choice) it’s all just a few lines of fairly high-level code.
We’ve only seen one other Telegram application on Hackaday.io and we’re wondering why. It looks pretty slick, and with the bot’s ability to send a custom “keyboard” to the phone along with the message, it could make cell-phone-based control interfaces a cinch. Anyone else using Telegram for bots?