Automate The Freight: When The Freight Is People

Before I got a license and a car, getting to and from high school was an ordeal. The hour-long bus ride was awful, as one would expect when sixty adolescents are crammed together with minimal supervision. Avoiding the realities going on around me was a constant chore, aided by frequent mental excursions. One such wandering led me to the conclusion that we high schoolers were nothing but cargo on a delivery truck designed for people. That was a cheery fact to face at the beginning of a school day.

What’s true for a bus full of students is equally true for every city bus, trolley, subway, or long-haul motorcoach you see. People can be freight just as much as pallets of groceries in a semi or a bunch of smiling boxes and envelopes in a brown panel truck. And the same economic factors that we’ve been insisting will make it far more likely that autonomous vehicles will penetrate the freight delivery market before we see self-driving passenger vehicles are at work with people moving. This time on Automate the Freight: what happens when the freight is people?

Continue reading “Automate The Freight: When The Freight Is People”

Automate The Freight: Autonomous Delivery Hits The Mainstream

It should come as no surprise that we here at Hackaday are big boosters of autonomous systems like self-driving vehicles. That’s not to say we’re without a healthy degree of skepticism, and indeed, the whole point of the “Automate the Freight” series is that economic forces will create powerful incentives for companies to build out automated delivery systems before they can afford to capitalize on demand for self-driving passenger vehicles. There’s a path to the glorious day when you can (safely) nap on the way to work, but that path will be paved by shipping and logistics companies with far deeper pockets than the average commuter.

So it was with some interest that we saw a flurry of announcements in the popular press recently regarding automated deliveries. Each by itself wouldn’t be worthy of much attention; companies are always maneuvering to be seen as ahead of the curve on coming trends, and often show off glitzy, over-produced videos and well-crafted press releases as a low-effort way to position themselves as well as to test markets. But seeing three announcements at one time was unusual, and may point to a general feeling by manufacturers that automated deliveries are just around the corner. Plus, each story highlighted advancements in areas specifically covered by “Automate the Freight” articles, so it seemed like a perfect time to review them and perhaps toot our own horn a bit.

Continue reading “Automate The Freight: Autonomous Delivery Hits The Mainstream”

DroNet: learning to fly by driving

Delivery Drones Can Learn From Driving And Cycling

Increasingly these days drones are being used for urban surveillance, delivery, and examining architectural structures. To do this autonomously often involves using “map-localize-plan” techniques wherein first, the location is determined on a map using GPS, and then based on that, control commands are produced.

A neural network that does steering and collision prediction can compliment the map-localize-plan techniques. However, the neural network needs to be trained using video taken from actual flying drones. But generating that training video involves many hours of flying drones at street level putting vehicles and pedestrians at risk. To train their DroNet, Researchers from the University of Zurich and the Universidad Politecnica de Madrid have come up with safer sources for that video, video recorded from driving cars and bicycles.

DroNet
DroNet

For the drone steering predictions, they used over 70,000 images and corresponding steering angles from the publically available car driving data from Udacity’s Open Source Self-Driving project. For the collision predictions, they mounted a GoPro camera to the handlebars of a bicycle and drove around a city. Video recording began when the bicycle was distant from an object and stopped when very close to the object. In total, they collected 32,000 images.

To use the trained network, images from the drone’s forward-facing camera were fed into the network and the output was a steering angle and a probability of collision, which was turned into a velocity. The drone remained at a constant height above ground, though it did work well from 1.5 meters to 5 meters up. It successfully navigated road lanes and avoided moving pedestrians and bicycles. Intersections did confuse it though, likely due to the open spaces messing with the collision predictions. But we think that shouldn’t be a problem when paired with map-localize-plan techniques as a direction to move through the intersection would be chosen for it using the location on the map.

As you can see in the video below, it not only does a decent job of flying down lanes but it also flies well in a parking garage and a hallway, even though it wasn’t trained for either of these.

Continue reading “Delivery Drones Can Learn From Driving And Cycling”

Table-Top Self Driving With The Pi Zero

Self-driving technologies are a hot button topic right now, as major companies scramble to be the first to market with more capable autonomous vehicles. There’s a high barrier to entry at the top of the game, but that doesn’t mean you can’t tinker at home. [Richard Crowder] has been building a self-driving car at home with the Raspberry Pi Zero.

The self-driving model is trained by first learning from the human driver.

[Richard]’s project is based on the EOgma Neo machine learning library. Using a type of machine learning known as Sparse Predictive Hierarchies, or SPH, the algorithm is first trained with user input. [Richard] trained the model by driving it around a small track. The algorithm takes into account the steering and throttle inputs from the human driver and also monitors the feed from the Raspberry Pi camera. After training the model for a few laps, the car is then ready to drive itself.

Fundamentally, this is working on a much simpler level than a full-sized self-driving car. As the video indicates, the steering angle is predicted based on the grayscale pixel data from the camera feed. The track is very simple and the contrast of the walls to the driving surface makes it easier for the machine learning algorithm to figure out where it should be going. Watching the video feed reminds us of simple line-following robots of years past; this project achieves a similar effect in a completely different way. As it stands, it’s a great learning project on how to work with machine learning systems.

[Richard]’s write-up includes instructions on how to replicate the build, which is great if you’re just starting out with machine learning projects. What’s impressive is that this build achieves what it does with only the horsepower of the minute Raspberry Pi Zero, and putting it all in a package of just 102 grams. We’ve seen similar builds before that rely on much more horsepower – in processing and propulsion.

Make Cars Safer By Making Them Softer

Would making autonomous vehicles softer make them safer?

Alphabet’s self-driving car offshoot, Waymo, feels that may be the case as they were recently granted a patent for vehicles that soften on impact. Sensors would identify an impending collision and adjust ‘tension members’ on the vehicle’s exterior to cushion the blow. These ‘members’ would be corrugated sections or moving panels that absorb the impact alongside the crumpling effect of the vehicle, making adjustments based on the type of obstacle the vehicle is about to strike.

Continue reading “Make Cars Safer By Making Them Softer”

Ultrasonic Array Gets Range Data Fast And Cheap

How’s your parallel parking? It’s a scenario that many drivers dread to the point of avoidance. But this 360° ultrasonic sensor will put even the most skilled driver to shame, at least those who pilot tiny remote-controlled cars.

Watch the video below a few times and you’ll see that within the limits of the test system, [Dimitris Platis]’ “SonicDisc” sensor does a pretty good job of nailing the parallel parking problem, a driving skill so rare that car companies have spent millions developing vehicles that do it for you. The essential task is good spatial relations, and that’s where SonicDisc comes in. A circular array of eight HC-SR04 ultrasonic sensors hitched to an ATmega328P, the SonicDisc takes advantage of interrupts to make reading the eight sensors as fast as possible. The array can take a complete set of readings every 10 milliseconds, which is fast enough to allow for averaging successive readings to filter out some of the noise that gets returned. Talking to the car’s microcontroller over I2C, the sensor provides a wealth of ranging data that lets the car quickly complete a parallel parking maneuver. And as a bonus, SonicDisc is both open source and cheap to build — about $10 a copy.

Rather use light to get your range data? There are some pretty cheap LIDAR units on the market these days.

Continue reading “Ultrasonic Array Gets Range Data Fast And Cheap”

Driverless Lorries To Be Tested On UK Roads By End Of 2018

The [BBC] is reporting that driverless semi-trailer trucks or as we call them in the UK driverless Lorries are to be tested on UK roads. A contract has been awarded to the Transport Research Laboratory (TRL) for the trials. Initially the technology will be tested on closed tracks, but these trials are expected to move to major roads by the end of 2018.

All  of these Lorries will be manned and driven in formation of up to three lorries in single file. The lead vehicle will connect to the others wirelessly and control their braking and acceleration. Human drivers will still be present to steer the following lorries in the convoy.

This automation will allow the trucks to drive very close together, reducing drag for the following vehicles to improve fuel efficiency.”Platooning” as they call these convoys has been tested in a number of countries around the world, including the US, Germany, and Japan.

Are these actually autonomous vehicles? This question is folly when looking toward the future of “self-driving”. The transition to robot vehicles will not happen in the blink of an eye, even if the technological barriers were all suddenly solved. That’s because it’s untenable for human drivers to suddenly be on the road with vehicles that don’t have a human brain behind the wheel. These changes will happen incrementally. The lorry tests are akin to networked cruise control. But we can see a path that will add in lane drift warnings, steering correction, and more incremental automation until only the lead vehicle has a person behind the wheel.

There is a lot of interest in the self driving industry right now from the self driving potato to autonomous delivery. We’d love to hear your vision of how automated delivery will sneak its way into our everyday lives. Tell us what you think in the comments below.