Texture Map GCode Directly In Blender With NozzleBoss

We’ve seen this funky dual disk polar printer already recently, but [Heinz Loepmeier] has been busy working on it, so here’s an update. The primary focus here is nozzleboss, a blender plugin which enables the surface textures of already sliced objects to be manipulated. The idea is to read in the gcode for the object, and convert it to an internal mesh representation that blender needs in order to function. From there the desired textures can be applied to the surfaces for subsequent stages to operate upon. One trick that nozzleboss can do is to create weight maps to tweak the extrusion flow rate or print velocity value according to the pixel value at the surface — such ‘velocity painting’ can produce some very subtle surface effects on previously featureless faces. Another trick is to use the same weight maps and simply map colours to blender text blocks which are injected into the gcode at export time. These gcode blocks can be used swap tool heads or extruders, enabling blending of multiple filament colours or types in the same object.

Some nice examples of such printing manipulation can be seen on [Heinz’s] instagram page for the project. So, going back to the hardware again, the first video embedded below shows the ‘dual disk polar printer’ fitted with a crazy five-extruders-into-one-nozzle mixing hotend setup, which should be capable of full CMYK colour mixing and some. The second video below shows an interesting by-product of the wide horizontal motion range of the machine, that the whole printing area can be shifted to a nozzle at the other end of the gantry. This enables a novel way to switch extruders, by just moving the whole bed and print under the nozzle of interest! One final observation — is that of the print surface — it does look rather like they’re printing direct onto a slab of marble, which I think is the first time we’ve seen that.

Interesting printer designs are being worked on a lot these days, here’s a really nice 5-axis prusa i3 hack, and if you want to stay in the cartesian world, but your desktop machine is just too small, then you can always supersize it.

Continue reading “Texture Map GCode Directly In Blender With NozzleBoss”

Mirror, Mirror On The Wall, Do My Eyes Deceive Me After All

Say what you will about illusions, [Create Inc] has some 3D prints that appear to change shape when viewed in a mirror. For example, circles transform into stars and vice versa. A similar trick was performed by [Kokichi Sugihara] in 2016, where he showed circles that appear as squares in the mirror. For the trick to work, the camera’s position (or your eye) is important as the shapes look different from different angles. The illusion comes in when your brain ignores any extra information and concludes that a much more complex shape is a simpler one. [Create Inc] walks you through the process of how the illusion works and how it was created in Blender.

When he posted the video on Reddit, most seemed to think that it wasn’t a mirror and there was some camera trickery. At its heart, this is reverse-engineering a magic trick, and we think it’s an impressive one. STL files are on Thingiverse or Etsy if you want to print your own. We covered a second illusion that [Kokichi] did that relies on a similar trick.

Continue reading “Mirror, Mirror On The Wall, Do My Eyes Deceive Me After All”

Blender? No, Grinder

[Leandro Felipe] is no stranger to the dirty hack, and this video of his conversion of a blender into a handheld rotary grinding tool is no exception. (Embedded below.) But the end result is something pretty useful — a lighter and more maneuverable rotary grinder that’s got a lot more grunt to boot.

(The video is in Portuguese, but the captions work pretty well, once you get over the fact that the robots translate “grinding tool” as “rectifier” a lot of the time. And anyway, you’re here for the hacks.)

The highlights are a handmade coupling that mates the blender motor with the flexible shaft and chuck, purchased separately. And the flattened-out PVC pipe used as a mounting bracket. And him using the motor itself against a file to “lathe” down the drive shaft. And…

The tip of the day comes when he holds the blender motor in a metal vise to test it out. Metal and spinning magnets — what’s the worst that could happen?  Sparks, smoke, and a trip to the thrift store for another used blender.

If you just want to see the finished piece, you can jump ahead to the end. But it’s basically, get yourself a speed-adjustable blender, couple it to the shaft of an off-the shelf grinder, and you’re set.

It’s an idea so conceptually easy, you might wonder if Hackaday has ever showcased a blender dr3mel before. We have. What else can you power with a blender motor?

Thanks [Danjovic] for the tip!

Continue reading “Blender? No, Grinder”

Watch Blender Plugin Make Animated PCB Traces (and More)

[Staacks]’s Blender plugin to animate growth is behind the sweet animation seen above. It’s an add-on that cleverly makes creating slick growth animations easier when using Blender. It isn’t limited to PCB images either, although they do happen to make an excellent example of the process.

The add-on isn’t limited to animating PCB traces.

The idea is that one begins with an image texture with a structure showing a bunch of paths (like a maze, or traces on a PCB), and that gets used as an input. The plugin then uses a path finding algorithm to determine how these paths could grow from an origin point, and stores the relevant data in the color channels of an output image. That output is further used within Blender as the parameters with which to generate the actual animation, resulting in the neat self-creating PCB seen above. That PCB isn’t just for show, by the way. It’s the PCB for [Staacks]’s smart doorbell project.

Blender is an amazingly comprehensive tool for modeling and animation, and while we’ve covered using it to create high-quality KiCad renders, this kind of animation is really something else.

Here is the GitHub repository for the Blender growth tool if you’re interested in giving it a spin. If you’d like to see more first, watch the video embedded below for a showcase of what it’s capable of, and how it works.

Continue reading “Watch Blender Plugin Make Animated PCB Traces (and More)”

3D Objects Without Scanning

There are many scanners — both commercial and homemade — that can take a variety of scans or images of a 3D object and convert it into something like a 3D printable file. When the process works, it works well, but the results can be finicky at best and will require a lot of manual tuning. According to [Samuel Garbett], you might as well just draw your own model using Blender. He shows you how using a Red Bull can which, granted, isn’t exactly the most complicated thing ever, but it isn’t the simplest either.

He does take one photo of the can, so there is a camera involved at some point. He also takes measurements using calipers, something you probably already have laying around.

Since it is just a can, there aren’t many required pictures or measurements as, say, a starship model. Once you have the measurements, of course, you could use the tool of your choice and since we aren’t very adept with Blender, we might have used something we think is easier like FreeCAD or OpenSCAD. However, Blender has a lot of power, so we suspect making the jump from can to the USS Enterprise might be more realistic for a Blender user.

Besides, it is good to see how other tools work and we were surprised that Blender could be relatively simple to use. Every time we see [Jared’s] channel, we think we should learn more about Blender. But if you have your heart set on a real scanner, there are plenty of open source designs you can print.

Art With Technology Hack Chat

Join us on Wednesday, June 16 at noon Pacific for the Art with Technology Hack Chat with Cory Collins!

As hackers, we naturally see the beauty of technology. We often talk in terms of the aesthetics of a particular hack, or the elegance of one solution over another, and we can marvel at the craftsmanship involved in everything from a well-designed PCB to a particularly clever reverse-engineering effort. Actually using technology to create art is something that’s often harder for us to appreciate, though, and looking at technological art from the artist’s side can be pretty instructive.

Cory Collins is an animator and artist with a long history of not only putting tech to work to create art, but also using it as the subject of his pieces. Cory’s work has brought life to video games, movies, and TV shows for years; more recently, he has turned his animation skills to developing interactive educational material for medical training. He has worked in just about every physical and digital medium imaginable, and the characters and scenes he has created are sometimes whimsical, sometimes terrifying, but always engaging.

Cory will stop by the Hack Chat to talk about what he has learned about technology from the artist’s perspective. Join us as we dive into the creative process, look at how art influences technology and vice versa, and learn how artistic considerations can help us address the technical problems every project eventually faces.

join-hack-chatOur Hack Chats are live community events in the Hackaday.io Hack Chat group messaging. This week we’ll be sitting down on Wednesday, June 16 at 12:00 PM Pacific time. If time zones have you tied up, we have a handy time zone converter.

Click that speech bubble to the right, and you’ll be taken directly to the Hack Chat group on Hackaday.io. You don’t have to wait until Wednesday; join whenever you want and you can see what the community is talking about.

Putting Perseverance Rover’s View Into Satellite View Context

It’s always fun to look over aerial and satellite maps of places we know, seeing a perspective different from our usual ground level view. We lose that context when it’s a place we don’t know by heart. Such as, say, Mars. So [Matthew Earl] sought to give Perseverance rover’s landing video some context by projecting onto orbital imagery from ESA’s Mars Express. The resulting video (embedded below the break) is a fun watch alongside the technical writeup Reprojecting the Perseverance landing footage onto satellite imagery.

Some telemetry of rover position and orientation were transmitted live during the landing process, with the rest recorded and downloaded later. Surprisingly, none of that information was used for this project, which was based entirely on video pixels. This makes the results even more impressive and the techniques more widely applicable to other projects. The foundational piece is SIFT (Scale Invariant Feature Transform), which is one of many tools in the OpenCV toolbox. SIFT found correlations between Perseverance’s video frames and Mars Express orbital image, feeding into a processing pipeline written in Python for results rendered in Blender.

While many elements of this project sound enticing for applications in robot vision, there are a few challenges touched upon in the “Final Touches” section of the writeup. The falling heatshield interfered with automated tracking, implying this process will need help to properly understand dynamically changing environments. Furthermore, it does not seem to run fast enough for a robot’s real-time needs. But at first glance, these problems are not fundamental. They merely await some motivated people to tackle in the future.

This process bears some superficial similarities to projection mapping, which is a category of projects we’ve featured on these pages. Except everything is reversed (camera instead of video projector, etc.) making the math an entirely different can of worms. But if projection mapping sounds more to your interest, here is a starting point.

[via Dr. Tanya Harrison @TanyaOfMars]

Continue reading “Putting Perseverance Rover’s View Into Satellite View Context”