Testing VR Limits With A Raspberry Pi

vrpi

Virtual Reality by function pushes the boundaries of what we perceive as existence, tricking the mind into believing that the computer generated environment that the user is thrust into actually contains a real place. So in the spirit of seeing what is possible in VR, a developer named [Jacques] hooked up a Raspberry Pi to an Oculus Rift. He used a computer graphics rendering API called OpenGL ES, which is much like any mobile platform found these days, to render a floating, rotating cube.

All his tests were done on a Release build which utilized the official vertex and fragment shaders. There was no attempt to optimize anything; not like there would be much to do anyways. The scene was rendered twice at 16 milliseconds per frame. From there, he attempted 27 ms per frame with texture, followed by 36 ms/frame, and then 45.

The code used can be found on [Jacques]’s Github account. A simple improvement would use a Banana Pi for better processing speed. However, don’t expect any spectacular results with this type of setup. Really, the project only proves that it’s possible to minimize a VR experience into something that could become portable. And in the same vein, the Pi + Oculus integration can produce an uncomfortable lagging effect if things are not lined up properly. But once the energy/computing power issues are addressed, VR devices could transform into a more fashionable product like Google Glass, where a simple flip of a switch would toggle the view between VR and AR into a something more mixed. And then a motion sensing input camera like this Kinect-mapping space experiment could allow people all over the world to jump into the perspectives of other reality-pushing explorers. That’s all far down the line though, but this project lays the foundation for what the future might hold.

To see [Jacques]’s full set up, view the video after the break.

Continue reading “Testing VR Limits With A Raspberry Pi”

George Crowdsourcington: A 3D Printed, Community Built Statue

George Crowdsourcington and Distributed Ben Franklin

Macro 3D printing is some cool stuff — but it’s extremely time consuming and can be very expensive. Introducing We The Buildersa 3D printing crowd source site which creates large scale projects the whole country can enjoy.

Their first project was George Crowdsourcington — a 1:1 copy of the Baltimore George Washington statue made out of 110 individual pieces. They chopped the model up into 4″ cubes and created the website in order to organize and distribute the files. One of their sponsors, Tinkerine Studio, reimbursed the shipping costs for makers who helped print out parts! Since his creation, Crowdsourcington has traveled all over the country, making stops at 3D printing shows in New York, mini-Maker Faires, art galleries, science centers and more — he even did a short residency in the Adafruit office in Manhattan!

It was quite the success, so they’re starting a new statue called the Distributed Ben Franklin. This one has a whopping 198 pieces, and they hope to have it built in time for the Silver Spring and World Maker Faires.

Continue reading “George Crowdsourcington: A 3D Printed, Community Built Statue”

Controlling Music With The Wave Of A Hand

2301141406514425783

[Thomas] created a magical music player that gives the listener the ability to change songs and alter the volume levels without having to touch anything but air. Called the LighTouch, this device puts the control in the hands of the user by interpreting input from an ultrasonic sensor and plays back tracks based on waving gestures.

It is the 2nd iteration of a prototype that he completed about a year ago and functions as a streaming radio/alarm clock. The sensor is hooked up to a Raspberry Pi with a fading LED. Everything is highly customizable including the distances used for playback features. The criteria [Thomas] put in place has the pause method trigger when an object is detected between 0-10cm from the sensor. The volume control on the next level up brightens and dims the LED light just for some added flair.

Continue reading “Controlling Music With The Wave Of A Hand”

Pole Climbing Device Runs Up Flags And Undies

Tubular Drive travels up and down poles

Driving a carriage up and down a cylindrical object isn’t the most popular activity but that is certainly no reason not to build such a device. Check out [Ryan’s] creation that does just that, he calls it a Tubular Drive.

There isn’t much going on here, basically there are 4 wheels that grip a pipe. Two of those wheels have integrated gears and are driven by a DC motor. The remaining two wheels are idlers. When power is applied to the motor, two of the wheels spin, which then moves the entire assembly down the pole. A quick reversal in polarity brings the unit back the other way.

With those 3D printed plastic wheels you may think that traction would be an issue but [Ryan] insists that it is not a problem. The ABS wheels were treated with an acetone bath to smooth out the print layers and the distance between the wheels can be adjusted using a couple of bolts. Together that allows enough surface contact and pressure to ensure slip-free traveling.

Although the wheels were made to grip 1/2″ electrical conduit, it would be very easy to adapt this design to fit around and climb up all sorts of cylindrical objects, maybe even rope! Perhaps v-wheels with a spring tensioner system would allow for traveling on different size tubes while also adjusting for any variation in the diameter of a single tube.

[Ryan] says version two will have a linear encoder and be driven by a stepper motor. Check out the video after the break…

Continue reading “Pole Climbing Device Runs Up Flags And Undies”

San Francisco Event: Hardware Developers Didactic Galactic

 

header2

It’s a mouthful to say, but an evening-ful of fun. San Franciscans who like to talk about all things hardware need to block this one out on their calendars:

Hardware Developers Didactic Galactic
Thursday, August 14th 2014 starting 6pm-9:30pm
500 3rd St., Suite 230 in San Francisco

The night will include a few talks on hardware; So far we know [Matt Berggren] is doing FPGA stuff, [Chris Gammell] will talk about KiCAD, and I’m going to talk about the community adventure that is Mooltipass. We’re also looking for others to make presentations so step up and share your hardware passion!

In addition to the formal talks there’ll be plenty of time for chewing the fat with all the other hardware-awesomes that will be there. See you a week from tomorrow, and don’t be shy about bringing your own hardware to show off!

Building A Home Made IPhone

iThing A few years ago, [Michele] built a mobile device with a touch screen, a relatively powerful processor, and a whole bunch of sensors. To be honest, the question of why he built this was never asked because it’s an impressive display of electronic design and fabrication. [Michele] calles it the iGruppio. Although it’s not a feature-packed cell phone, it’s still an impressive project that stands on its own merits.

Inside the iGruppio is a Pic32mx microcontroller, a 240×320 TFT touchscreen, and enough sensors to implement a 10 DOF IMU. The software written for the iGruppio is heavily inspired by the iPhone and a completely homebrew project – all the software was written by [Michele] himself. While the first version of the iGruppio was a little clunky, the second revision (seen in the pic above) uses an old iPhone case to turn a bunch of boards and plugs into a surprisingly compact device.

No, there’s no cellular modem inside the latest version, but [Michele] has put all the sources up on Github, and anyone wanting to build a homebrew cell phone could do worse than to take a look at his work. Video demo below.

Continue reading “Building A Home Made IPhone”

Focus Your Ears With The Visual Microphone

VideoMicrophone

A Group of MIT, Microsoft, and Adobe researchers have managed to reproduce sound using video alone. The sounds we make bounce off every object in the room, causing microscopic vibrations.  The Visual Microphone utilizes a high-speed video camera and some clever signal processing to extract an audio signal from these vibrations. Using video of everyday objects such as snack bags, plants, Styrofoam cups, and water, the team was able to reproduce tones, music and speech. Capturing audio from light isn’t exactly new. Laser microphones have been around for years. The difference here is the fact that the visual microphone is a completely passive device. No laser or special illumination is required.

The secret is in the signal processing, which the team explains in their SIGGRAPH paper (pdf link). They used a complex steerable pyramid along with wavelet filters to obtain local pixel motion values. These local values are averaged into a global motion value. From this global motion value the team is able to measure movement down to 1/1000 of a pixel. Plenty of resolution to decode audio data.

Most of the research is performed with high-speed video cameras, which are well outside the budget of the average hacker. Don’t despair though, the team did prove out that the same magic can be performed with consumer cameras, albeit with lower quality results. The team took advantage of the rolling shutter found in most of today’s CMOS imager based consumer cameras. Rolling shutter CMOS sensors capture images one row at a time. Each row can be processed in a similar fashion to the frames of the high-speed camera. There are some inter-frame gaps when the camera isn’t recording anything though. Even with the reduced resolution, it’s easy to pick out “Mary had a little lamb” in the video below.

We’re blown away by this research, and we’re sure certain organizations will be looking into it for their own use. Don’t pull out your tin foil hats yet though. Foil containers proved to be one of the best sound reflectors.

Continue reading “Focus Your Ears With The Visual Microphone”