VISUALIST – A Hardware Visual Effects Synthesizer

[Berto] wrote in to tell us about the visual effects synthesizer he built. It works as a pass-through for a video signal, rendering crisp clean images into a more psychedelic flavor like the one seen above. On the one hand this does a dishonor to the high-quality video devices we carry around in our pockets these days. On the other hand it will make some really interesting background video at a party or at your local dance club.

This is not a filter for a PC, or an FPGA-based processing system. A set of analog parts alter the incoming composite video (NTSC or PAL formats) and pipes the result to a television or projector. [Berto] included controls to alter the effects. They’re mounted on a panel and everything is given a home inside of a handy carrying case. Check out the video clip after the break to get a better idea of the video manipulations this things can pull off.

Continue reading “VISUALIST – A Hardware Visual Effects Synthesizer”

Full-color Video On A Spinning POV Display

Watching Big Buck Bunny on a spinning POV display is pretty impressive. Sure, the circular display area cuts off some of the picture, but otherwise it looks fantastic. This POV display is based on a Gumstix board. It runs embedded Linux which makes video playback rather easy. But translating each frame to the round display is another story.

The device is the result of a course project at Telecom ParisTech. [Félix], [Sylvain], and [Jérémy] used an FPGA to do the pixel mapping. This uses an encoder wheel (rather than a traditional hall effect sensor) to ascertain the blade’s position. The sensor that monitors the disc sends quadrature encoded pulses which result in 10-bit position data. The FPGA uses that data to calculate where each LED falls in its arc, then looks up the pixel color for that position. It’s not the largest POV display we’ve seen, but it certainly has the very best RGB resolution by far.

Continue reading “Full-color Video On A Spinning POV Display”

Mixing Video For Old School Effects

For all the high production values Final Cut Pro and Adobe Premier have released upon the world, there’s still a cinematographic aesthetic only possible with analog video, linear editing, and video feedback. [gijs] just sent in a video mixer he’s been working on to allow crossfading between two video signals and introducing some very cool analog video distortion effects.

[gijs]’ mixer uses the LM1881 video sync separator also found in the Arduino video experimenter shield. Because two different video feeds are unlikely to send their sync signal at exactly the same time, the selected video will stay still on the screen while the second video feed will slowly scroll horizontally across the screen.

This isn’t the first analog video hack [gijs] has come up with; last year he released an Arduino video sampler capable of recording about a second of video and playing it back forward, reverse, looped, or inverted.

We’re sure combing both the video mixer and sampler would produce an aesthetic similar to the experimentation seen on 80s-era public access or our time in AV club. Either way, a very cool build that just can’t be done digitally.

Video of the mixer after the break.

Continue reading “Mixing Video For Old School Effects”

Visualizing Water Droplets And Building A CT Scanner

Since his nerves were wracked by presenting his project to an absurdly large crowd at this year’s SIGGRAPH, [James] is finally ready to share his method of mixing fluids via optical tomography with a much larger audience: the readership of Hackaday.

[James]’ project focuses on the problem of modeling mixing liquids from a multi-camera setup. The hardware is fairly basic, just 16 consumer-level video cameras arranged in a semicircle around a glass beaker full of water.

When [James] injects a little dye into the water, the diffusing cloud is captured by a handful of Sony camcorders. The images from these camcorders are sent through an algorithm that selects one point in the cloud and performs a random walk to find every other point in the cloud of liquid dye.

The result of all this computation is a literal volumetric cloud, allowing [James] to render, slice, and cut the cloud of dye any way he chooses. You can see the videos produced from this very cool build after the break.

Continue reading “Visualizing Water Droplets And Building A CT Scanner”

Fortifying Veronica’s VGA Output

[Quinn Dunki] just moved to a new work space and had to pack up her homebrew computer project — called Veronica — in the process. She just unboxed it again and decided now was a good time to fortify the VGA display hardware. It wasn’t in the greatest of shape, since everything for the initial video tests had been built on a breadboard. The transition to protoboard ended up turning out just swell.

One of the thing’s that we like best about [Quinn’s] hacks is that she documents her failures (or perhaps we should just call them hiccups?) just as much as she does her successes. This is not a small thing. We understand, because our own screw-ups don’t usually get photographed due to our raging need to just make the frakking thing work.

Once she had moved all the components to the new board the circuit was amazingly organized. Since she’s doing high-speed switching with the VGA signals it was important to keep the lines as short and straight as possible, hence the SRAM stack seen above. But when it was first fired up she had a jumble of only-somewhat-organized color stripes. It turns out that she had forgotten to change the color register in the AVR code, the color lines were hooked up in the wrong order, and the switch mode supply was injecting noise into the system. But thanks to her documentation of these issues we’ll know what to do when we find ourselves in a similar situation.

Using An Oscilloscope As A Composite Video Adapter

Confronted with a monitor that would display neither HDMI signal, nor composite video, [Joonas Pihlajamaa] took on a rather unorthodox task of getting his oscilloscope to work as a composite video adapter. He’s using a PicoScope 2204 but any hardware that connects to a computer and has a C API should work. The trick is in how his code uses the API to interpret the signal.

The first thing to do is make sure the voltage levels used in the composite signal are within the tolerances of your scope. [Joonas] used his multimeter to measure the center pole of the RCA connector and found that the Raspberry Pi board puts out from 200 mV to 2V, well within the PicoScope’s specs. Next he started to analyze the signal. The horizontal sync is easy to find, and he ignored the color information — opting for a monochrome output to ease the coding process. The next big piece of the puzzle is to ascertain the vertical sync so that he knows where each frame starts. He got it working and made one last improvement to handle interlacing.

The proof of concept video after the break shows off the he did. It’s a bit fuzzy but that’s how composite video looks normally.

Continue reading “Using An Oscilloscope As A Composite Video Adapter”

Moving An FPGA Project From VGA To DVI-D

It’s fairly easy to create VGA with an FPGA using a simple R/2R DAC. As [Mike] points out, this requires a lot of IO pins, and many development boards only support 8 bit VGA. Analog VGA is being replaced with DVI-D and HDMI on many devices nowadays, so it would be nice to port projects from VGA to DVI-D.

To address this, he’s come up with a simple DVI-D implementation in VHDL. The result converts RGB and sync data for VGA into DIV-D. Since DVI-D and HDMI both use the same signals for video, this can be connected to either input on a monitor or TV.

This implementation is shown displaying a test pattern on the Pipistrello development board, which features a Spartan 6 LX45 FPGA, but the project was written to be portable to other vendor’s FPGAs. With the right connector and a fast enough clock speed, this project should help move a project from 8 bit VGA to glorious 32 bit color.