Doing magic with cards involves a lot of precise dexterity to know which card is where. For plenty of tricks, this is often knowledge and control of a single card or a small number of cards. But knowing the exact position of every single card in the deck could certainly be helpful, so the Nettle Magic Project was created to allow magicians to easily identify the location of cards in the deck.
The system works through the use of computer vision to identify a series of marks on the short edge of a stack of cards. The marks can be printed in IR- or UV-sensitive ink to make them virtually invisible, but for demonstration these use regular black ink. Each card has landmarks printed on either side of a set of bit markers which identify the cards. A computer is able to quickly read the marks and identify each card in order while the deck is still stacked, aiding the magician in whichever trick they need to perform.
The software only runs on various Apple devices right now, including iPhones and iPads, but the software is readily available fore experimentation if you are a magician looking to try something like this out. Honestly, we don’t see too many builds focusing on magic, sleight-of-hand or otherwise, and we had to go back over a decade to find a couple of custom magical builds from a magician named [Mario].
Thanks to [Tim] for the tip!
[Stefan] is always looking for a way to make stronger and better 3D prints. His latest experiments involve using a texture on thin plastic parts to increase stiffness. You can see the texture pattern in the banner above and the video below.
While a lot of people looked at IdeaMaker’s new texturing feature as something for cosmetics, [Stefan] thought of sheet metal products that often use bead patterns to increase stiffness and strength. Can patterned plastic be stiffer than ordinary printed plastic? Turns out, the answer is yes.
Continue reading “Texture Adds Stiffness To 3D Parts”
Just when it seems like we’ve juiced all the creative potential out of our 3D printers, a bold new feature lands on the table. Enter Velocity Painting, a concept brought to life by [Mark Wheadon] that textures our 3D prints with greyscale images.
At its core, the technique is straightforward: skin an image onto a 3D print by varying the print speed in specific locations and, thereby, varying just how much plastic oozes out of the nozzle. While the concept seems simple, the result is stunning.
Velocity Painting opens up new ways of expression on top of an existing print with all the skinning opportunities. Imagine adding a texture for realism like this rook that’s been patterned with a brick layout, or imagine an aesthetic embellishment like the flames on [Mark’s] dragon print.
The results speak for themselves, and the growing number of users are proving it. Head on over to the gallery to indulge yourself in this delightful oozing aesthetic that’s sure to turn a few heads.
[Mark Wheadon’s] hack takes the mechanics of how we print and adds another creative tuning knob. If you’re looking for other embellishments for your prints, have a look at [David Shorey’s] work on texturizing fabrics.
Connecting computers to human brains is currently limited to the scope of science fiction and a few cutting-edge laboratories. Tapping into some nerves farther from our central wetware is possible and [Peter Buczkowski] shows us his stylish machine for implanting a pattern into our brains without actively having to memorize anything.
His Medium Machine leverages a TENS unit to activate forearm muscles in a pattern programmed into an Arduino. Users place their forearm across two aluminum electrodes mounted on a tasteful wooden platform and extend a single finger over a button. Electrical impulses trigger the muscles which press the button. That’s all. After repeating the pattern a few times, the users should be able to recite it back on command even if they aren’t aware of what it means. If this sounds like some [Johnny Mnemonic] memory cache, you are absolutely correct. This project draws inspiration from the [William Gibson] novel which became a [Keanu Reeves] movie.
Users can be programmed with a Morse code message or the secret knock to open an attic library or play a little tune. How about learning a piano song?
Continue reading “Medium Machine Mediates Microcontroller Messages”
Some people find it hard to look at a huge, flat expanse of floor or ground and not see a canvas. From the outfield grass of a baseball park to some poor farmer’s wheat field, trampling, trimming or painting patterns can present an irresistible temptation. But the larger the canvas the more challenging the composition will be, which is where this autonomous beach-combing art robot comes into play.
Very much still a work in progress, [pablo.odysseus]’ beach bot was built to take advantage of the kilometers-wide beaches left by the receding tides near his home. That immense canvas is begging to be groomed, and this bot is built for the task. The running gear itself is simple – an extruded aluminum chassis powered by wheelchair drives with added optical encoders and dragging a retractable rake – but the bot is bristling with electronics dedicated to navigation. A pair of Arduinos run the dual odometers, compass, and a GNSS receiver, as well as providing a smartphone interface for on-the-fly changes. The art is composed as a DXF file converted to latitude and longitude points and exported to Google Earth as a KML. That means the bot can just be brought to the beach and allowed to draw autonomously. An early test run is seen below the break; better “brushes” are in the works.
Watching the art unfold on a beach would be relaxing, like watching a zen garden being created. We’re looking forward to [pablo]’s progress on this one. Of course, art bots aren’t the only autonomous machines that big, wide beaches attract.
Continue reading “Hackaday Prize: An Autonomous Beach Art Robot”
It figures. You spend a ton of time making a cool set of costumes and then you can’t get your kid to pose for a picture. It’s okay though, we still get the point. This themed set of costumes dresses the little one as a Roomba vacuuming robot while mom and dad are suited up as virtual walls (modules that are used to keep the bot from falling down stairs, etc.). It’s fun and unique, but had it not been for some additional electronics this would have been relegated to a links post. For safety sake each costume was outfitted with a ring of LEDs. As a challenge, the lights were given the ability to sync up patterns with each other.
Each costume has a circular frame at the top with a set of RGB LED strings attached. To get them to display synchronized patterns an IR transmitter/receiver board was designed and ordered from OSHPark. Each costume has four of these modules so no matter where the wearers are facing it should not break communications. A demo of the synchronized light rings can be seen after the break
Continue reading “Roomba And Virtual Walls Make Up This Theme Family Halloween Costume”