Toyaanisqatsi: Time Lapse Control Using LEGO Parts

[youtube=http://www.youtube.com/watch?v=Fm1fFQfGc9M&feature=player_embedded]
A simple panning motion can add impact to the already-dramatic effect of time lapse photography. To accomplish this, frugal cinematographers sometimes build [Rube Goldberg] contraptions from clock motors, VCR parts or telescope tracking mounts. Hack a Day reader [Stephan Martin] has assembled a clever bargain-basement system using an Arduino-driven stepper motor and a reduction gear system built up from LEGO Technic parts, along with some Processing code on a host PC to direct the show.

While the photography is a bit crude (using just a webcam), [Stephan’s] underlying motion control setup might interest budding filmmakers with [Ron Fricke] aspirations but Top Ramen budgets. What’s more, unlike rigid clock motor approaches, software control of the camera mount has the potential for some interesting non-linear, fluid movements.

Capacitance Sensor Guide (AD7746)

capacitive_sensor

[Marcus] has written up his experiences using the AD7746 capacitance sensor. He used the SparkFun breakout board in conjunction with an Arduino. The available Arduino code wasn’t that great so he rewrote it to be easier to understand. The AD7746 is an I2C device that can be continuously read, but this doesn’t mesh well with the Wiring libraries. Additionally, the calibration routine from the data sheet is difficult to understand. He’s included all of the code he used plus a Processing sketch to help visualize the input which will hopefully make your experience with the chip much more smooth.

Laughing Man In Processing

laughing_man

The Laughing Man is the antagonist from the anime series Ghost in the Shell: Stand Alone Complex. During each of his public appearances in the series he manages to hack all video feeds/cyborg eyes in the vicinity to obscure his face with the logo above.

[Ben Kurtz] had been watching the series recently and realized he could put together a similar effect using Processing. The interesting bit, and what makes this more fun than a simple demo, is that he’s using the OpenCV library. OpenCV is a open source computer vision library. [Ben] uses it to handle the facial recognition in Processing and then apply the image.

It’s only 100 lines and we wonder what other fun tricks could be employed. Here’s a Hack a Day skull you can swap in for the logo.

[thanks dakami]

Processing 1.0

Processing, the open source programming language designed for artists and other creative types, finally went 1.0. Processing inspired numerous outpourings of creativity and beauty, from interactive art installations to sound sculptures. Improvements to Processing include OpenGL anti-aliasing, an extensible Tools menu, and the XML library included by default. You can read up on the changes or download Processing and start playing with it yourself.

[via Create Digital Motion]

Advanced Beauty Generative Video Art

advbeau

Advanced Beauty is a collection of 18 “sound sculptures” pairing artists and programmers to create a collaborative work visualizing sound. The styles run a broad range from fluid simulations to manipulating cell animation. The demos were built using Processing. While all of these were built using human input, we see potential for them to help improve standard visualizers. Hopefully, to bring out more information about what’s actually being played. Below is just one of the videos in the series. You can find more on Vimeo. Continue reading “Advanced Beauty Generative Video Art”

Wiimote Head Tracking In Processing


[Manuel] has been playing around with [Johnny Lee]’s Wiimote head tracking code. He’s posted a preliminary port outlining the code in the Processing environment. It relies on darwiinremoteOSC so you won’t see this outside of OSX, but it should help you out if you’re trying to do this is in Processing on another platform.

[via Create Digital Motion]

[photo: nicolasnova]

Wii Nunchuck Used For Mocap On 3D Studio Max


If you’ve been looking for a cheap way to do motion capture on 3D Studio Max, [melka] has devised an ingenious way to do mocap using a Wii nunchuck. As shown in his Vimeo video, the nunchuck is connected to an Arduino that feeds MIDI data to a computer running the MIDI app Ableton Live. Ableton Live feeds that data to 3D Studio Max via MIDI Yoke.

The video below the break shows that the motion of the nunchuck can be seen in real time in the 3D Studio Max Window. We’re really hoping [melka] will post a writeup of this project soon, as we are hungry for more details.

Continue reading “Wii Nunchuck Used For Mocap On 3D Studio Max”