Sea Rendering

Project Sea Rendering Autonomously Renders Sea Bottoms

[Geir] has created a pretty neat device, it’s actually his second version of an autonomous boat that maps the depths of lakes and ponds. He calls it the Sea Rendering. The project is pretty serious as the hull was specially made of fiberglass. The propulsion is a simple DC motor and the rudder is powered by an RC servo. A light and flag adorn the top deck making the small craft visible to other larger boats that may be passing by. Seven batteries are responsible for all of the power requirements.

Sea Rendering

The craft’s course is pre-programmed in Mission Planner and uses ArduPilot loaded on an Arduino to steer to the defined way points. An onboard GPS module determines the position of the boat while a transducer measures the depth of the water. Both position and depth values are then saved to an SD card. Those values can later be imported into a software called Dr Depth that generates a topographic map of the water-covered floor.

[Geir] has sent this bad boy out on an 18 km journey passing through 337 way points. That’s pretty impressive! He estimates that the expected run time is 24 hours at a top speed of 3 km/h, meaning it could potentially travel 72 km on a single charge while taking 700 depth measurements during the voyage.

Continue reading “Project Sea Rendering Autonomously Renders Sea Bottoms”

Autonomous Vehicle-Following Vehicle

Humanity has taken one step closer to Skynet becoming fully aware. [Ahmed], [Muhammad], [Salman], and [Suleman] have created a vehicle that can “chase” another vehicle as part of their senior design project. Now it’s just a matter of time before the machines take over.

The project itself is based on a gasoline-powered quad bike that the students first converted to electric for the sake of their project. It uses a single webcam to get information about its surroundings. This is a plus because it frees the robot from needing a stereoscopic camera or any other complicated equipment like a radar or laser rangefinder. With this information, it can follow a lead vehicle without getting any other telemetry.

This project is interesting because it could potentially allow for large convoys with only one human operator at the front. Once self-driving cars become more mainstream, this could potentially save a lot of costs as well if only the vehicle in the front needs the self-driving equipment, while the vehicles behind would be able to operate with much less hardware. Either way, we love seeing senior design projects that have great real-world applications!

Continue reading “Autonomous Vehicle-Following Vehicle”

A Compact Underwater Vehicle: The Nanoseeker

The Nanoseeker is a compact underwater vehicle in a torpedo-like form factor. [John] designed the Nanoseeker as completely enclosed vehicle: both the thruster and the control fins are all housed within the diameter of the tube. The thruster is ducted with vents on the sides and control fins integrated into the back of the duct assembly.

[John] designed a compact PCB to drive the vehicle, which includes an STM32F4 alongside several sensors. An MPU-9150 provides IMU functionality and two dual motor driver ICs from TI control the throttle and the control fins. [John] also added a Bluetooth radio for remote control functionality. For those who want a closer look, an image of the schematic is up on his blog.

The board is running MicroPython, which is a small Python implementation optimized for microcontrollers. Although [John]’s hardware platform looks great, he’s still getting started on his software. We look forward to seeing how his project develops, as his project is one of the smallest underwater vehicles we’ve seen.

[via Dangerous Prototypes]

A Mechanically Scanned LIDAR For Autonomous Robots

LIDAR[Patrick] has spent a lot of time around ground and aerial based autonomous robots, and over the last few years, he’s noticed a particular need for teams in robotics competitions to break through the ‘sensory bottleneck’ and get good data of the surrounding environment for navigational algorithms. The most well-funded teams in autonomous robotics competitions use LIDARs to scan the environment, but these are astonishingly expensive. With that, [Patrick] set out to create a cheaper solution.

Early this year, [Patrick] learned of an extremely cheap LIDAR sensor. Now [Patrick] is building a robotics distance measurement unit based on this sensor.

Early experiments with mechanically scanned LIDAR sensors centered around the XV-11 LIDAR, the distance sensor found in the Neato Robotics robot vacuum cleaner. [Patrick] became convinced a mechanically scanned LIDAR was the way forward when it came to distance measurement of autonomous robots. Now he’s making his own with an astonishingly inexpensive LIDAR sensor.

The basic idea of [Patrick]’s project is to take the PulsedLight LIDAR-Lite module, add a motor and processing board, and sell a complete unit that will output 360° of distance data to a robot’s main control system. The entire system should cost under $150 when finished; a boon to any students, teams, or hobbyists building an autonomous vehicle.

[Patrick]’s system is based on the PulsedLight LIDAR – a device that’s not shipping yet – but the team behind the LIDAR-Lite says they should have everything ready by the end of the month, all the better, because between these two devices, there’s a lot of cool stuff to be done in the area of autonomous robots.

Adding An Optical Mouse Sensor To An Autonomous Vehicle

[Tim] is getting his drone ready for SparkFun’s 2013 Autonomous Vehicle Competition on June 8th. He has a pretty good start, but was having some problems accurately measuring travel distance. The technique he chose for the task was to glue magnets onto the axles of the vehicle and monitor them with a hall effect sensor. Those sensors are finicky and a few problems during testing prompted him to look at a redundant system. Right now he’s experimenting with adding an optical mouse sensor to the autonomous vehicle.

Recently we saw the same concept used, but it was meant for tracking movement of a full-sized automobile. If it can work in that application it should be perfect here since the vehicle is much closer to the ground and will be used in ideal conditions (flat pavement with clear weather). [Tim] cracked open an old HP mouse he had lying around. Inside he found an Avago ADNS-5020 sensor. After grabbing the datasheet he discovered that it’s simply an I2C device. Above you can see the Arduino Leonardo he used for the first tests.

[Tim] coded functions to monitor the chip, including some interesting ones like measuring how in-focus the surface below the sensor is. This brings up a question, is there limit on how fast the vehicle can travel before the sensor fails to report back accurately?

Intelligent Autonomous Vehicle Makes It To Maker Faire

A few guys from Rutgers showed up at Maker Faire with Navi, their vehicle for the 2012 Intelligent Ground Vehicle Competition. Powered by two huge lead acid batteries, Navi features enough high-end hardware to hopefully make it through or around just about any terrain.

Loaded up with a laser range finder, a stereo camera setup, compass, GPS receiver, and a pair of motors capable of pulling 40A, Navi has the all the hardware sensors required to make it around a track with no human intervention. Everything is controlled by a small netbook underneath the control panel, itself loaded up with enough switches and an 8×32 LED matrix to be utterly incomprehensible.

In the videos after the break, the guys from Rutgers show off the systems that went into Navi. There’s also a video showing off Navi’s suspension, an impressive custom-built wishbone setup that will hopefully keep Navi on an even keel throughout the competition.

Also of note: A PDF design report for Navi and Navi’s own blog.

Continue reading “Intelligent Autonomous Vehicle Makes It To Maker Faire”

All About The Google Autonomous Vehicle Project

There have been many self-driving cars made with different levels of success, but probably the most well-known project is the Google car.  What you may not have heard of, though is the autonomous Google cart, or golf cart to be exact. The first video after the break explains the motivation behind the cart and the autonomous vehicle project.  As with another autonomous vehicle we’ve featured before, they didn’t forget to include an E-stop button (at 1:03)!

In the second video (also after the break) Google’s Sebastian Thrun and Chris Urmson get into more of the details of how Google’s more famous autonomous Prius vehicles work and their travels around different towns in California. A safety driver is still used at this point, but the sensor package includes a roof-mounted 64-beam laser sensor, wheel encoder, radars, and a GPS sensor. With Google’s vast resources as well as their work with Streetview and Google maps, it’ll be interesting to see what comes of this technology.  I, for one, welcome our new robotic overlords.

Continue reading “All About The Google Autonomous Vehicle Project”