Submitted by brad on Fri, 2015-03-27 10:34.
After 9/11 there was a lot of talk about how to prevent it, and the best method was to fortify the cockpit door and prevent unauthorized access. Every security system, however, sometimes prevents authorized people from getting access, and the tragic results of that are now clear to the world. This is likely a highly unusual event, and we should not go overboard, but it’s still interesting to consider.
(I have an extra reason to take special interest here, I was boarding a flight out of Spain on Tuesday just before the Germanwings flight crashed.)
In 2001, it was very common to talk about how software systems, at least on fly-by-wire aircraft, might make it impossible for aircraft to do things like fly into buildings. Such modes might be enabled by remote command from air traffic control. Pilots resist this, they don’t like the idea of a plane that might refuse to obey them at any time, because with some justification they worry that a situation could arise where the automated system is in error, and they need full manual control to do what needs to be done.
The cockpit access protocol on the Airbus allows flight crew to enter a code to unlock the door. Quite reasonably, the pilot in the cockpit can override that access, because an external bad guy might force a flight crew member to enter the code.
So here’s an alternative — a code that can be entered by a flight crew member which sends and emergency alert to air traffic control. ATC would then have the power to unlock the door with no possibility of pilot override. In extreme cases, ATC might even be able to put the plane in a safe mode, where it can only fly to a designated airport, and auto-land at that airport. In planes with sufficient bandwidth near an airport, the plane might actually be landed by remote pilots like a UAV, an entirely reasonable idea for newer aircraft. In case of real terrorist attack, ATC would need to be ready to refuse to open the door no matter what is threatened to the passengers.
If ATC is out of range (like over the deep ocean) then the remote console might allow the flight crew — even a flight attendant — to direct the aircraft to fly to pre-approved waypoints along the planned flight path where quality radio contact can be established.
Clearly there is a risk to putting a plane in this mode, though ATC or the flight crew who did it could always return control to the cockpit.
It might still be possible to commit suicide but it would take a lot more detailed planning. Indeed, there have been pilot suicides where the door was not locked, and the suicidal pilot just put the plane into a non-recoverable spin so quickly that nobody could stop it. Still, in many cases of suicide, any impediment can sometimes make the difference.
Update: I have learned the lock has a manual component, and so the pilot in the cockpit could prevent even a remote opening for now. Of course, current planes are not set to be remotely flown, though that has been discussed. It’s non trivial (and would require lots of approval) but it could have other purposes.
A safe mode that prevents overt attempts to crash might be more effective than you think, in that with many suicides, even modest discouragement can make a difference. It’s why they still want to put a fence on the Golden Gate Bridge had have other similar things elsewhere. You won’t stop a determined suicide but it apparently does stop those who are still uncertain, which is lots of them.
The simpler solution — already going into effect in countries that did not have this rule already — is a regulation insisting that nobody is ever alone in the cockpit. Under this rule, if a pilot wants to go to the bathroom, a flight attendant waits in the cockpit. Of course, a determined suicidal pilot could disable this person, either because of physical power, or because sometimes there is a weapon available to pilots. That requires more resolve and planning, though.
Submitted by brad on Sat, 2015-02-28 14:20.
I was recently considering the price of UberX in Los Angeles. It’s gotten disturbingly low:
Flag drop: $0 18 cents/minute 90 cents/mile
This is not a very good deal for the driver. After Uber’s 20% cut, that’s 72 cents/mile. According to AAA, a typical car costs about 60 cents/mile to operate, not including parking. (Some cars are a bit cheaper, including the Prius favoured by UberX drivers.) In any event, the UberX driver is not making much money on their car.
The 18 cents/minute — $10.80 per hour, drops to only $8.64/hour while driving. Not that much above minimum wage. And I’m not counting the time spent waiting and driving to and from rides, nor the miles, which is odd that the flag drop fee. There is a $1 “safe rides fee” that Uber pockets (they are being sued over that.) And there is a $4 minimum, which will hit you on rides of up to about 2.5 miles.
So Uber drivers aren’t getting paid that well — not big news — but a bigger thing is the comparison of this with private car ownership.
As noted, private car ownership is typically around 60 cents/mile. The Uber ride then, is only 50% more per mile. You pay the driver a low rate to drive you, but in return, you get that back as free time in which you can work, or socialize on your phone, or relax and read or watch movies. For a large number of people who value their time much more than $10/hour, it’s a no-brainer win.
The average car trip for urbanites is 8.5 miles — though that of course is biased up by long road trips that would never be done in something like Uber. I will make a guess and drop urban trips to 6.
The Uber and private car costs do have some complications:
* That Safe Rides Fee adds $1/trip, or about 16 cents/mile on a 6 mile trip
* The minimum fee is a minor penalty from 2 to 2.5 miles, a serious penalty on 1 mile trips
* Uber has surge pricing some of the time that can double or even triple this price
As UberX prices drop this much, we should start seeing people deliberately dropping cars for Uber, just as I have predicted for robocars. I forecast robotaxi service can be available for even less. 60 cents/mile with no cost for a driver and minimal flag drop or minimum fees. In other words, beating the cost of private car ownership and offering free time while riding. UberX is not as good as this, but for people of a certain income level who value their own time, it should already be beating the private car.
We should definitely see 2 car families dropping down to 1 car plus digital rides. The longer trips can be well handled by services like Zipcar or even better, Car2Go or DriveNow which are one way.
The surge pricing is a barrier. One easy solution would be for a company like Uber to make an offer: “If you ride more than 4,000 miles/year with us, then no surge pricing for you.” Or whatever deal of that sort can make economic sense. Sort of frequent rider loyalty miles. (Surprised none of the companies have thought about loyalty programs yet.)
Another option that might make sense in car replacement is an electric scooter for trips under 2 miles, UberX like service for 2 to 30 miles, and car rental/carshare for trips over 30 miles.
If we don’t start seeing this happen, it might tell us that robocars may have a larger hurdle in getting people to give up a car for them than predicted. On the other hand, some people will actually much prefer the silence of a robocar to having to interact with a human driver — sometimes you are not in the mood for it. In addition, Americans at least are not quite used to the idea of having a driver all the time. Even billionaires I know don’t have a personal chauffeur, in spite of the obvious utility of it for people whose time is that valuable. On the other hand, having a robocar will not seem so ostentatious.
Submitted by brad on Thu, 2015-02-26 13:59.
All over the world, people (and governments) are debating about regulations for robocars. First for testing, and then for operation. It mostly began when Google encouraged the state of Nevada to write regulations, but now it’s in full force. The topic is so hot that there is a danger that regulations might be drafted long before the shape of the first commercial deployments of the technology take place.
As such I have prepared a new special article on the issues around regulating robocars. The article concludes that in spite of a frequent claim that we want to regulate and standarize even before the technology has been out in the market for a while, this is in fact both a highly unusual approach, and possibly even a dangerous approach.
Regulating Robocar Safety : An examination of the issues around regulating robocar safety and the case for a very light touch
Submitted by brad on Wed, 2015-02-11 11:27.
The government baked robocar projects in the UK are going full steam, with this press release from the UK government to accompany the unveiling of the prototype Lutz pod which should ply the streets of Milton Keynes and Greenwich.
This comes along with the realizations of laws enabling of testing of vehicles (with safety drivers) and discussion of changes to the UK vehicle code.
The new pod follows a similar path to other fully-autonomous prototypes, reminding me of the EN-V from GM, the MIT City Car and the Google buggy prototype. It’s electric, meant for “last mile” and will lose its steering wheel once testing is over.
I also note they talk eagerly about the Meridian shuttle being tested in Greenwich, even though that’s a French vehicle.
When it comes to changes to the vehicle code, I think it’s premature. Even without looking at the proposed changes, I would say that we don’t know enough to work out what changes are needed, even though we all might be full of ideas.
One proposal is to remove the ban on tailgating to allow convoys. A reasonable enough thing, except people are not going to build convoys for quite some time, I think. The Volvo/SARTRE experiment found a number of problems with the idea, and you don’t want to do your first deployments with something that could crash 10 cars if it goes wrong instead of one. You do that later, once you feel very confident in your tech.
Another proposal called for changing how cyclists are treated. The law in the UK (and some other places) demands cyclists be given the full berth of a car, and in practice nobody ever does that, and if they did do it, it would mean they just followed along at bicycle speed, impeding traffic. One of those classic cases, like speed limits in the USA, where the law only works if nobody follows it. (Though cyclists would say that they should just get the full lane like the law says.)
We will need to fix these areas of the vehicle codes, but we should fix them only after we see a problem, unless it’s clear that the vehicles can’t be deployed without the change. Give the developers the chance to fix the problem on their own first. If you fix the law before you know what the vehicles will be like, you may ensconce old thinking into the law and have a hard time getting it out.
It is interesting to see Governments adapt so quickly to a disruptive technology. It’s quite probable that our hype is a bit too good and will come back to bite us. I predicted this sort of jurisdictional competition as governments realize they have a chance to make their regions become players in the new automotive industry, but they are embracing some things faster than I expected.
Submitted by brad on Tue, 2015-02-10 16:43.
Electric Vehicles are moving up, at least here in California, and it’s gotten to the point that EV drivers are finding all the charging stations where they want to go already in use, forcing them to travel well outside their way, or to panic. Sometimes not charging is not an option. Sometimes the car taking the spot is already mostly charged or doesn’t need the charge much, but the owner has not come back.
Here in Silicon Valley, there is a problem that the bulk of the EVs have 60 to 80 miles of range — OK for wandering around the valley, but not quite enough for a trip to San Francisco and back, at least not a comfortable one. And we do like to go to San Francisco. The natives up there don’t really need the chargers in a typical day, but the visitors do. In general, unless you are certain you are going to get a charger, you won’t want to go in a typical EV. Sure, a Tesla has no problem, but a Tesla has a ridiculous amount of battery in it. You spend $40,000 on the battery pack in the Tesla, but use the second half of its capacity extremely rarely — it’s not cost effective outside the luxury market, at least at today’s prices (and also because of the weight.)
Charging stations are somewhat expensive. Even home stations cost from $400 to $800 because they must now include EVSE protocol equipment. This does a digital negotiation between the car and the plug on how much power is available and when to send it. The car must not draw more current than the circuit can handle, and you want the lines to not be live until the connection is solid. For now that’s expensive (presumably because of the high current switching gear.) Public charging stations also need a way to doing billing and access control.
Another limit on public charging stations, however, is the size of the electrical service. A typical car wants 30 amps, or up to 50 if you can get it. Put in more than a few of those and you’re talking an upgrade to the building’s electrical service in many cases.
I propose a public charging pole which has 4 or even 8 cords on it. This pole would be placed at the intersection of 4 parking spots in a parking lot. (That’s not very usual, more often they end up placed against a wall, with only 2 parking spots in range, because that’s where the power is.) The station, however, may not have enough power to charge all the cables at once. read more »
Submitted by brad on Wed, 2015-02-04 12:04.
There is great buzz about some sensor-laden vehicles being driven around the USA which have been discovered to be owned by Apple Computer. The vehicles have cameras and LIDARs and GPS antennas and many are wondering is this an Apple Self-Driving Car? See also speculation from cult of Mac.
Here’s a video of the vehicle driving around the East Bay (50 miles from Cupertino) but they have also been seen in New York.
We don’t see the front of the vehicle, but it sure has plenty of sensors. On the front and back you see two Velodyne 32E Lidars. These are 32 plane LIDARS that cost about $30K. You see two GPS antennas and what appear to be cameras in all directions. You don’t see the front in these pictures, which is where the most interesting sensors will be.
So is this a robocar, or is this a fancy mapping car? Rumours about Apple working on a car have been swirling for a while, but one thing to contradict that has been the absence of sightings of cars like this. You can’t have an active program without testing on the roads. There are ways to hide LIDARS (and Apple is super secretive so they might) and even cameras to a degree, but this vehicle hides little.
Most curious are the Velodynes. They are tilted down significantly. The 32E unit sees from about 10 degrees up to 30 degrees down. Tilting them this much means you don’t see out horizontally, which is not at all what you want if this is for a self-driving car. These LIDARs are densely scanning the road close around the car, and higher things in the opposite direction. The rear LIDAR will be seeing out horizontally, but it’s placed just where you wouldn’t place it to see what’s in front of you. A GPS antenna is blocking the direct forward view, so if the goal of the rear LIDAR is to see ahead, it makes no sense.
We don’t see the front, so there might be another LIDAR up there, along with radars (often hidden in the grille) and these would be pretty important for any research car.
For mapping, these strange angles and blind spots are not an issue. You are trying to build a 3D and visible light scan of the world. What you do’t see from one point you get from another. For stree mapping, what’s directly in front and behind are generally road and not interesting, but what’s to the side is really interesting.
Also on the car is an accurate encoder on the wheel to give improved odemetry. Both robocars and mapping cars are interested in precise position information.
Arguments this is a robocar:
- The Velodynes are expensive, high end and more than you need for mapping, though if cost is no object, they are a decent choice.
- Apple knows it’s being watched, and might try to make their robocar look like a mapping car
- There are other sensors we can’t seee
Arguments it’s a mapping car
- As noted, the Velodynes are titled in a way that really suggests mapping. (Ford uses tilted ones but paired with horizontal ones.)
- The cameras are aimed at the corners, not forward as you would want
- They are driving in remote locations, which eventually you want to do, but initially you are more likely to get to the first stage close to home. Google has not done serious testing outside the Bay Area in spite of their large project.
- The lack of streetview is a major advantage Google has over Apple, so it is not surprising they might make their own.
I can’t make a firm conclusion, but this leans toward it being a mapping car. Seeing the front (which I am sure will happen soon) will tell us more.
Another option is it could be a mapping car building advanced maps for a different, secret, self-driving car.
Submitted by brad on Tue, 2015-02-03 17:29.
After yesterday’s story about Uber and CMU, a lot of speculation has flown that Uber will now be at odds with Google, both about building robocars and also on providing network taxi service, since another rumour said Google plans to launch an Uber like “ride share” service.
Since then, the Uber blog post and this interview with Uber folks tell a slightly different story. Uber is funding a research center at CMU, and giving lots of grants to academics. Details are not fully available, but typically this means being at an early research stage. With these research labs, academics are keen to publish all they do, so little gets done in secret. In many cases the sponsor gets a licence to the technology but it’s often not exclusive. If Uber wanted to build their own car, chances are they would do it in a more private lab.
Rumours that David Drummond would resign from the Uber board also have not panned out. Google has invested hugely in Uber (already for good return at the present valuation) and Google Maps offers you an Uber if you ask it for directions somewhere — it’s actually one of the easier interfaces for ordering one.
Rumours around Google’s efforts suggest that Big G has been testing a “ride share” app with employees and plans to launch it. Google has denied that, and says it loves Uber and Lyft. Further news revealed the rumours were about an internal carpooling system, not involving the self-driving cars. I could imagine confusion because Uber and others call themselves “ride sharing” which is a bit of a fabrication to not look like a taxi, while a carpooling app would be real ride sharing. (UberPool is real ride sharing.) Google, which has a terrible undersupply of parking is very keen on getting employees to ride its bus system and to carpool.
That said, Google has talked about the same thing I talk about — the true goal of robocar technology being the creation of a mobility on demand taxi service, like Uber but at a much lower cost. Google has not said that they would provide that themselves, or who they would partner with if they did it. Most people have presumed it might be Uber but I don’t think that’s at all assured.
At the same time, Uber has assured its drivers they are not going away for the foreseeable future. I suspect that’s an equivocation, and just means that we can’t see very far in the future right now!
Submitted by brad on Mon, 2015-02-02 20:31.
I commonly see statements from connected car advocates that vehicle to vehicle (V2V) and vehicle to infrastructure communications are an important, even essential technology for robocar development. Readers of this blog will know I disagree strongly, and while I think I2V will be important (done primarily over the existing mobile data network) I suspect that V2V is only barely useful, with minimal value cases that have a hard time justifying its cost.
Of late, though, my forecast for V2V grows even more dismal, because I wonder if robocars will implement V2V with human-driven cars at all, even if it becomes common for ordinary cars to have the technology because of a legal mandate.
The problem is security. A robocar is a very dangerous machine. Compromised, it can cause a lot of damage, even death. As such, security will have a very strong focus in development. You don’t want anybody breaking into the computer systems or your car or anybody else’s. You really don’t want it.
One clear fact that people in security know — a very large fraction of computer security breaches caused by software faults have come from programs that receive input data from external sources, in particular when you will accept data from anybody. Internet tools are the biggest culprits, and there is a long history of buffer overflows, injection attacks and other trouble that has fallen on tools which will accept a message from just anyone. Servers (which openly accept messages from outside) are at the greatest risk, but even client tools like web browsers run into trouble because they go to vast numbers of different web sites, and it’s not hard to trick people to sending them to a random web site.
We work very hard to remove these vulnerabilities, because when you’re writing a web tool, you have no choice. You must accept input from random strangers. Holes still get found, and we pay the price.
The simplest strategy to improve your chances is to go deaf. Don’t receive inputs from outside at all. You can’t do that in most products, but if you can close off a channel without impeding functionality it’s a good approach. Generally you will do the following to be more secure:
- Be a client, which means you make communications requests, you do not receive them.
- You only connect to places you trust. You avoid allowing yourself to be directed to connect to other things
- You use digital signature and encryption to assure that you really are talking to your trusted server.
This doesn’t protect you perfectly. Your home server can be compromised — it often will be running in an environment not as locked down as this. In fact, if it becomes your relay for messages from outside, as it must, it has a vector for attack. Still, the extra layer adds some security. read more »
Submitted by brad on Mon, 2015-02-02 14:42.
Rumours reported in TechCrunch suggest Uber is opening a robocar lab in Pittsburgh and hiring up to 50 CMU folks to staff it.
Update: On the Uber blog we now see it’s more funding of research labs at CMU, on many topics
That’s a major step, if true. People have often pointed out how well Uber is poised to make use of robocar technology to bring computer summoned taxi service to the next level. If Uber did not exist, I would surely be building it to get that advantage. Many have assumed that since Google is a major investment partner in Uber that they would partner on this technology, but this suggests otherwise.
I write about Uber a lot here not just because of interest in what they do today, but because it teaches us a lot about how people will view Robocars in the future. Uber’s interface is very similar to what you might see for a robocar service, and the experience is fairly similar, just much more expensive. UberX is $1.30/mile plus 26 cents/minute with $2.20 flag drop. The Black service is $3.75/mile and 65 cents/minute with an $8 flag drop. I expect robocar tax service to be cheaper than 50 cents/mile with minimal per-minute charges. The flag drop is not yet easy to calculate. What richer people do with Uber teaches us what the whole public will do with robocars.
Uber lets you say where you are going but doesn’t demand it. That’s one thing I suspect will be different with your robotaxi, because it’s really nice if they can send you a vehicle chosen for the trip you have in mind. Ie. a small, efficient car without much range for short, single person trips. Robotaxi services will offer you the ability to not say your destination — but they will probably charge more for it, and that means most people will be willing to say their destination.
Uber does not hide their desire to get rid of all their drivers, which sounds like a strange strategy, but the truth is that cab driving is not something most people view as a career. It’s a quick source of money with no special skills, something people do until something better comes along, or in the gaps in their day to make extra cash. Unlike people losing jobs to robots on a factory line, nobody is particularly upset at the idea.
Submitted by brad on Wed, 2015-01-28 14:48.
Uber’s gotten a lot of bad press over its surge pricing system. As prices soared during Storm Sandy and a hostage crisis in Sydney, people saw it as price gouging when times are tough.
I’ve always thought the public reaction to price gouging in times of scarcity and emergency was irrational. While charging double or triple for food, rides or generators does mean that the rich get more access to them, it also does at least a partial job of assuring that people who truly need or want things the most get access over those who need them less. I do not quite understand why the alternative — keeping prices flat, and allocating items to whoever gets there first — is so broadly preferred.
Uber has promoted another reason to have surge pricing. They argue that as they raise the prices, it causes an increase in supply. Unlike generators, where there are only so many in the stores during a storm, doubling the price of a ride can mean a sudden influx of rides, both from people in the area and even those who rush in from outside to make the extra buck. I suspect that does happen, but Uber also makes more money and poorer people are priced out of the market, which has been a PR nightmare.
For the recent snowstorm that didn’t end up being too bad in NY, Uber announced some new policies — a cap of 2.8x on the price increase, and donation of all proceeds to the Red Cross. The mayor of New York even declared the surge-pricing was illegal.
It’s an interesting start, but what do they mean by all proceeds? If they’re not increasing the income of the drivers — many of whom are low enough income that the double-time or more rates can make a real difference — then they are defeating the whole point of this.
Here are some potential ideas I was thinking about for how to play surge pricing:
- Keep Uber’s fee during a surge the same. Ie. it’s always 20% of the rack rate, not of the surged price. So Uber is making no extra money (except from the extra volume,) just the drivers.
- To get really extreme, Uber could reduce its cut as volume increases, so they don’t even make money from the increased volume.
- They could just donate all their cut (which may be what they mean when they say all proceeds.)
- The extra could be split between drivers and a charity. You get more drivers, and they make more, but good deeds are also done.
Another option would be to do something like a “buy one give one” as we’ve seen in physical products. This would mean that during the surge, riders could elect to pay more to get priority (and to attract drivers.) But if the surge is for 2x, they might pay 3x, and the overage would go to provide a regular priced ride (1x) for somebody else, while still paying the driver 2x.
The tricky part is how to make sure the subsidized rides only go to those who can’t afford to pay the surge price. The subsidized rides will presumably still be in short supply. You want them to go only to those who truly need them. Options might include:
- Offer subsidies primarily for those who use UberX almost exclusively. Use a lot of black car and you don’t get a subsidy. (Yes, some people use black car on expense account and UberX on personal rides, including myself, so this is not perfect.)
- Require a declaration of low income. Subject those who declare low income to random audits after the fact, pulling up credit scores or asking them to actually demonstrate the low income. If they lied, charge them the full amount plus a penalty for all subsidized rides they took.
- Drivers could also elect to subsidize, and say they will drive for 1x, or any other amount, to really increase the supply of subsidized rides and the amount of subsidy. They might get a tax donation receipt for doing so if Uber could set up the tax structures properly with a non-profit. (A non-profit would probably need to work over all companies or be fully independent of the company.)
As already happens with the surge system, adjust the surcharge and subsidy to try and make demand match supply.
You could even offer rides to those in need for 0.5x, a flat fee, or even nothing, though nothing is very easy to abuse.
Submitted by brad on Sat, 2015-01-24 12:24.
Some new results from the NGV Team at the University of Michigan describe different approaches for perception (detecting obstacles on the road) and localizations (figuring out precisely where you are.) Ford helped fund some of the research so they issued press releases about it and got some media stories. Here’s a look at what they propose.
Many hope to be able to solve robotics (and thus car) problems with just cameras. While LIDAR is going to become cheap, it is not yet, and cameras are much cheaper. I outline many of the trade-offs between the systems in my article on cameras vs lasers. Everybody hopes for a research breakthrough or computer vision breakthrough to make vision systems reliable enough for safe operation.
The Michigan lab’s approach is a special machine vision one. They map the road in advance in 3D and visible light by using a mapping car equipped with lots of expensive LIDAR and other sensors. They build a 3D representation of the road similar to what you need for a video game engine, and from that, with the use of GPUs, they can indeed create a 2D image of what a camera should see from any given point.
The car goes out into the world and its actual camera delivers a 2D frame of what it sees. Their system then compares that with generated 2D images of what the camera should see until it finds the closest match. Effectively, it’s like you looking out a window and then going into a video game and wandering around looking for a place that looks like what you see out that window, and then you know where the window is.
Of course it is not “wandering,” and they develop efficient search algorithms to quickly find the location that looks most like the real world image. We’ve all seen video games images, and know they only approximate the real world, so nothing will be an exact match, but if the system is good enough, there will be a “most similar” match that also corresponds with what other sensors, like your GPS and your odometer/dead reckoning system, tell you about where you probably are.
Localization with cameras has been done before, and this is a new approach taking advantage of new generations of GPUs, so it’s interesting. The big challenge is simulating the lighting, because the real world is full of different lighting, high dynamic range, and shadows. The human system has no problem understanding a stripe on the road as it moves through the shadow of a tree, but computer systems have a pretty tough time with that. Sun shadows can be mapped well with GPUs, but shadows from things like the moving limbs of trees are not possible to simulate, as are the shadows of other vehicles and road users. At night, light and shadows come from car headlights and urban lights. The team is optimistic about how well they will handle these problems.
The much larger challenge is object perception. Once you have a simulation of what the camera should see, you can notice when there are things present that are not in the prediction — like another car or pedestrian, or a new road sign. (Right now their system mostly is looking at the ground.) Once you identify the new region, you can attempt to classify it using computer vision techniques, and also by watching it move against the expected background.
This is where it gets challenging, because the bar is very high. To be used for driving it must effectively always work. Even if you miss 1 pedestrian in a million you have a real problem because there are billions of pedestrians encountered by a billion drivers every day. This is why people love LIDAR — if something (other than a mirror or sheet of glass) sufficiently large is sufficiently close you, you’re going to get laser returns from it, and not from what’s behind it. It has the reliability number that is needed.
The challenge of vision systems is to meet that reliability goal.
This work is interesting because it does a lot without relying on AI “computer vision” techniques. It is not trying to look at a picture and recognize a person. Humans are able to look at 2D pictures with bizarre lighting and still tell you not just what the things in the picture are, but often how far away they are and what they are doing. While we can be fooled in a 2D image, once you have a moving dynamic world, humans are, generally reliable enough at spotting other things on the road. (Though of course, with 1.2 million dead each year, and probably 50 million or more accidents, the majority because somebody was “not looking,” we are far from perfect.)
Some day, computer vision will be as good at recognizing and understanding the world as people are — and in fact surpass us. There are fields (like identifying traffic signs from photos) where they already surpass us. For those not willing to wait until that day, new techniques in perception that don’t require full object understanding are always interesting.
I should also point out that while lowering cost is of course a worthwhile goal, it is a false goal at this time. Today, maximal safety is the overriding goal, and as such, nobody will actually release a vehicle to consumers without LIDAR just to save the estimated 2017 cost of LIDAR, which will be sub-$500. Only later, when cameras get so good they completely replace LIDAR safety capabilities for less money would people release such a system to save cost. On the other hand, improving cameras to be used together with LIDAR is a real goal; superior safety, not lower cost.
Submitted by brad on Thu, 2015-01-22 12:13.
Let me confess a secret fear. I suspect that the first “autopilot”
functions on cars is going to be a bit boring.
I’m talking the offerings like traffic jam assist from Mercedes, super cruise from Cadillac
and others. The faster highway assist versions which combine ADAS
functions like lane-keeping and adaptive cruise control to keep the
car in its lane and a fixed distance from the car in front of you.
What Tesla has promoted and what scrappy startup “Cruise” plans to offer
as a retrofit later this year. This is, in NHTSA’s flawed “levels”
document what could be called supervision type 2.
Some of them also offer lane change, if you approve the safety of
All these products will drive your car, slow or fast on highways,
but they require your supervision. They may fail to find the lane in
certain circumstances, because the makers are badly painted, or confusing,
or just missing, or the light is wrong. When they do they’ll kick out
and insist you drive. They’ll really insist, and you are expected to
be behind the wheel, watching and grabbing it quickly — ideally even
noticing the failure before the system does.
Some will kick out quite rarely. Others will do it several times during
a typical commute. But the makers will insist you be vigilant, not just
to cover their butts legally, but because in many situations you really
do need to be vigilant.
Testing shows that operators of these cars get pretty confident,
especially if they are not kicking out very often. They do things they
are told not to do. Pick up things to read. Do e-mails and texts.
This is no surprise — people are texting even now when the car isn’t
driving for them at all.
To reduce that, most companies are planning what they call
“countermeasures” to make sure you are paying attention to the road.
Some of them make you touch the wheel every 8 to 10 seconds. Some will
have a camera watching your eyes that sounds an alarm if you look away
from the road for too long. If you don’t keep alert, and ignore the
alarms, the cars will either come to a stop in the middle of the freeway,
or perhaps even just steer wild and run off the road. Some vendors
are talking about how to get the car to pull off safely to the side of
There is debate about whether all this will work, whether the
countermeasures or other techniques will assure safety. But let’s
leave that aside for a moment, and assume it works, and people stay safe.
I’m now asking the harder question, is this a worthwhile product?
I’ve touted it as a milestone — a first product put out to customers.
That Mercedes offered traffic jam assist in the 2014 S-Class and others
followed with that and freeway autopilots is something I tell people
in my talks to make it clear this is not just science fiction ideas and
cute prototypes. Real, commercial development is underway.
That’s all true, and I would like these products. What I fear though,
is whether it will be that much more useful or relaxing as adaptive cruise
control (ACC.) You probably don’t have ACC in your car. Uptake on it is
quite low — as an individual add-on, usually costing $1,000 to $2,000,
only 1-2% of car buyers get it. It’s much more commonly purchased as
part of a “technology package” for more money, and it’s not sure what
the driving force behind the purchase is.
Highway and traffic jam autopilot is just a “pleasant” feature, as is ACC.
It makes driving a bit more relaxing, once you trust it. But it doesn’t
change the world, not at all.
I admit to not having this in my car yet. I’ve sat in the driver’s seat of
Google’s car some number of times, but there I’ve been on duty to watch
it carefully. I got special driver training to assure I had the skills to
deal with problem situations. It’s very interesting, but not relaxing.
Some folks who have commuted long term in such cars have reported it to
A Step to greater things?
If highway autopilot is just a luxury feature, and doesn’t change
the world, is it a stepping stone to something that does? From a
standpoint of marketing, and customer and public reaction, it is.
From a technical standpoint, I am not so sure. read more »
Submitted by brad on Fri, 2015-01-16 17:05.
In my earlier article on robocar challenges I gave very brief coverage to the issue of parking. Challenged on that, I thought it was time to expand.
The world “parking” means many things, and the many classes of parking problems have varying difficulties.
The taxi doesn’t park
One of the simplest solutions to parking involves robotaxi service. Such vehicles don’t really park, at least not where they dropped you off. They drop you off and go to their next customer. If they don’t have another ride, they can deliberately go to a place where they know they can easily park to wait. They don’t need to tackle a parking space that’s challenging at all.
Simple non-crowded lots
Parking in basic parking lots — typical open ground lots that are not close to full — is a pretty easy problem. So easy in fact, that we’ve seen a number of demonstrations, ranging back to Junior 3 and Audi Piloted Parking. Cars in the showroom now will identify parking spots for you (and tell you if you fit.) They have done basic parallel parking (with you on the brakes) for several years, and are starting to now even do it with you out of the car (but watching from a distance.) At CES VW showed the special case of parking in your own garage or driveway, where you show the car where it’s going to go.
The early demos required empty parking lots with no pedestrians, and even no other moving cars, but today reasonably well-behaved other cars should not be a big problem. That’s the thing about non-crowded lots: People are not hunting or competing for spaces. The robocars actually would be very happy to seek out the large empty sections at the back of most parking lots because you aren’t going to be walking out that far, the car is going to come get you.
The biggest issue is the question of pedestrians who can appear out from behind a minivan. The answer to this is simply that vehicles that are parking can and do go slow, and slow automatically gives you a big safety boost. At parking lot speed, you really can stop very quickly if a pedestrian appears out of nowhere. The car, after all, is not in a hurry, and can slow itself when close to minivans, or if it has noticed pedestrians who are moving near it and have disappeared behind vehicles. Out at the back of a parking lot, nobody cares if you go 5 km/h, or even right down the center of the lane to assure there are no surprises.
To the right we see a picture of Junior 3 entering a parking lot, hunting for a space and taking it — in 2009.
Mapping is still desirable for parking lots. This is particularly true because parking lots, not being public roads, set up their own sets of rules and put up signs meant only for humans. They may direct traffic to be one-way in certain areas in nonstandard ways. They may have gates when you have to pay or insert tickets. Parking spots will be marked reserved for certain cars (Electric vehicle, expectant mother, wheelchair, employee of the month, CEO, customers of company X) with signs meant for humans.
It’s not necessarily super hard to map a parking lot, just time consuming to encode all these rules. Unlike roads, which everybody drives, any given parking lot likely only serves the people who live, work or shop next to it — you will never park in 95% of the lots in your city, though you will drive most of its main roads. Somebody has to pay for the cost of that mapping — either because lots of people want to use the lot, or because the owner of the lot wants to encourage robocars. Fortunately, with the robocars doing things like using the least popular spots, or even valet parking as described below, there is a strong incentive to the owner of a lot to get it mapped and keep it mapped. Only lots that never fill out would have no incentive, and those lots can often be parked in without a map.
While you want trained mappers to confirm the geometry of a parking lot, coding in the signs and special rules is a task easily left to the parking lot owner. If the lot manager forgets to tag the CEO’s space as reserved, nobody is hurt (except the lot manager when the CEO arrives.)
Robocar parking mistakes are easy to fix. Robocars can put a phone number or URL on the back where you can go to complain about a robocar that is parked badly or blocking things. As long as that doesn’t happen too often, the cost of the support desk is manageable. The folks at the support desk can look out with the robot’s sensors and tell it to move. It’s not like finding a human driven car blocking something, where you have to find the owner. In a minute, the robocar will be gone.
More crowded lots
The challenge of parking lots, in spite of the low speeds, is that they don’t have well defined rules of the road. People ignore the arrows on the ground. They pause and wait for cars to exit. In really crowded lots, cars follow people who are leaving at walking speed, hoping to get dibs on their spot. They wait, blocking traffic, for a spot they claim as theirs. People fight for spots and steal spots. People park badly and cross over the lines.
As far as I know, nobody has tried to solve this challenge, and so it remains unsolved. It is one of the few problems in robocars that actually deserves the label of “AI,” though some think all driving is AI.
Even so, on the grand scheme of things, my intuition is that this is not one of the grand unsolved challenges of AI. Parking lots don’t have legalized rules of the road, but they do have rules and principles, and we all learn them the more we park. Creating a system that can do well with these rules using various AI tools seems like a doable challenge when the time comes. My intuition is that it’s a lot easier than winning on Jeopardy. This system will be able to take advantage of a couple of special abilities of the robocars:
- They will be able to park and exit spots quickly and efficiently. They won’t be like the people you always see who do a 5 point turn to exit their parking spot when you (but not they) can see they still have 5 feet of room behind them.
- In general, they will be superb parkers, centering themselves as well as possible inside spots
- They don’t need room to open their doors, so they can park right next to walls and pillars.
- Yes, they could also park right next to badly parked cars which have encroached into other spaces and thus made a space no human can use. There is a risk of course that the bad parker, who finds they can’t get in one side, might retaliate. (I’ve had a guy rip my mirror off in revenge.) In this case, though, they will have a photo of the licence plate and a sensor record of the revenge taking place!
- In the event of problems or deadlock, they are open to the idea of just giving up and parking somewhere farther away that is easier to park in. Unlike humans they could drive as quickly in reverse as forward to back out of situations.
In spite of all this, the cars will want to avoid the full parking lots where the chaos happens. If there is another lot not far away, they will just go there, and require a couple minutes more advance notice from their master when summoned to pick them up. If there is nowhere nearby to park, the car will tell its passenger that she has to do the parking.
Even in the most crowded lots, there is the potential to easily create zones of the parking lot that are marked:
“Robot Valet Parking only. All other cars may be blocked in or towed. No pedestrians.”
In the car’s map, it will indicate what server is handling the robo-valet section, though it is possible to have it work without any communication at all.
In the most basic version the car would ask permission to enter the lot. The database might even assign it a spot, but generally it would just enter and take any spot. By “any spot”, I mean any piece of pavement, ignoring the lines on the ground. At first the cars would choose spots that let them have an unblocked pack to leave. As soon as too many cars arrive to do that, they would switch to a more dense, valet pattern that blocks in some cars (the ones who said they were leaving latest.) It would report where it parked to the database, as well as how to send it a message, and when it expects to leave.
Other cars would arrive. Eventually one would block in your car. If the database has given them a way to communicate (probably over the internet, though if they had V2V they could use that) they might discuss who plans to leave first, and the cars would adjust themselves to put the cars that will leave sooner at the front. This is strongly in the interests of the cars. If you plan to be there a while, you want to go to the back so you don’t have to keep moving to let cars behind you out. But it still works, just not as well, if the cars just take any available spot.
When it’s time to leave, the cars could try to send a message over the data networks to the cars in front of them, but a simpler approach might be to just nudge slightly forward — a few cm will do it. This will cause the car in the direction of the nudge to notice, and it too would nudge forward, and so on, and so on until the front car moves out, and then all the cars in that row can move out, including your car, which leaves the lot. Then the other cars can move in to fill the spot. If they have a database which maps the cars in that section, they could try to be clever in how they re-fill the empty column to minimize movement.
There are even faster algorithms if you leave a few empty spaces. Robocars have the ability to move in concert to “move the space” and put it next to a car that wants to exit. It’s more efficient, but not needed.
The database becomes more useful if a human driver ignores the signs and tries to park in the lot. That’s because the database is the simplest way of spotting a vehicle that’s not supposed to be there. As a first step, the cars in the lot could start flashing their lights and honking their horns at the interloper, or even speak human language messages out a speaker. “Hey, this is the robot valet lot, you are blocking me in! We’re calling a tow truck to come remove you if you don’t leave.” Some idiots may still try, and the robots could arrange so that almost all of them can still get out, and if not, they might call that tow truck.
The robo-valet section can be at the back of the parking lot, or the top of a structure — those places the humans park in last. The owner of the lot has a huge incentive to do this, since they can make much more efficient use of their land with the tight valet-dense parking. All the owner has to do is register the lot section in a database — a database that a company like Google would probably be happy to offer for free to benefit their cars.
Human valets could also park cars in this area. They would just need to use an app on their smartphone that tells them where to park and allows them to register that they did it. The robots will want the human-parked cars to park at the back, because they will move out of the way when it’s time for the human parked car to be driven back out.
The main requirements for this parking area would be that it be reachable from the outside without going through a zone of chaos, and that it then be possible to also reach the pickup/dropoff point for passengers without the risk of getting stuck in chaos. Larger lots tend to have entrance lanes without spots on them that serve this purpose.
Pedestrians will still enter the lot, in spite of the sign. Just go extra slow if they are there, and perhaps talk to them and ask them to leave. While you won’t actually present a danger to them at your low speed, they probably will heed the advice of 3000lb robots. Perhaps tell them they have 15 seconds to put down their weapon.
To get really clever, the sign marking the border of the Robo-Valet area might itself be on a small robot. Thus, when the robo-valet area gets full, the sign can move to expand the area if space is available. You could expand even into areas occupied by human-parked cars — just know that they are there and don’t block them in — or move out of their way when needed. Eventually they leave and only robocars enter.
When the demand goes down, the sign can easily move to shrink the valet area.
Submitted by brad on Thu, 2015-01-15 17:45.
Robocar news continues after CES with announcements from the Detroit Auto Show (and a tiny amount from the TRB meeting.)
Google doesn’t talk a lot about their car, so address by Chris Urmson at the Detroit Auto Show generated a lot of press. Notable statements from Chris included:
- A timeline of 2 to 5 years for deployment of a vehicle
- Public disclosure that Roush of Michigan acted as contract manufacturer to build the new “buggy” models — an open secret since May
- A list of other partners involved in building the car, such as Continental, LG (batteries), Bosch and others.
- A restatement that Google does not plan to become a car manufacturer, and feels working with Detroit is the best course to make cars
- A statement that Chris does not believe regulation will be a major barrier to getting the vehicles out, and they work regularly to keep NHTSA informed
- A few more details about Google’s own LIDAR, indicating that units are the size of coffee cups. (You will note the new image of the buggy car does not have a Velodyne on the roof.)
- More indication that things like driving in snow are not in the pipeline for the first vehicles
Almost all of this has been said before, though the date forecasts are moved back a bit. That doesn’t surprise me. As Google-watchers know, Google began by doing extensive, mostly highway based testing of modified hybrid cars, and declared last May that they were uncomfortable with the safety issues of doing a handoff to a human driver, and also that they have been doing a lot more on non-highway driving. This culminated with the unveiling of the small custom built buggy with no steering wheel. The shift in direction (though the Lexus cars are still out there) will expand the work that needs to be done.
Car company announcements out of the Detroit show were minor. The press got all excited when one GM executive said they “would be open to working with Google.” While I don’t think it was actually an official declaration, Google has said many times they have talked to all major car companies, so there would be no reason for GM to go out to the press to say they want to talk to Google. Much PR over nothing, I suspect.
Ford, on the other hand, actually backtracked and declared “we won’t be first” when it comes to this technology. I understand their trepidation. Being first does not mean being the winner in this game. But neither does being 2nd — there will be a time after which the game is lost.
There were concept vehicles displayed by Johnson Controls (a newcomer) and even a Chinese company which put a fish tank in the rear of the car. You could turn the driver’s seat around and watch your fish. Whaa?
In general, car makers were pushing their dates towards 2025. For some, that was a push back from 2020, for others a push forward from 2030, as both of those numbers have been common in predictions. I guess now that it’s 2015, 2020 is just to realistic a number to make an uncertain prediction about.
Earlier, Boston Consulting Group released a report suggesting robocars would be a $42B market in 2025 — the car companies had better get on it. With the global ground transportation market in the range of $7 trillion in my guesstimate, that’s a drop in the bucket, but also a huge number.
News from the Transportation Research Board annual meeting has been sparse. The combined conference of the TRB and AUVSI on self-driving cars in the summer has been the go-to conference of late, and other things usually happen at the big meeting. Released research suggested 10% of vehicles could be robocars in 2035 — a number I don’t think is nearly aggressive enough.
There also was tons of press over the agreement between NASA Ames and Nissan’s Sunnyvale research lab to collaborate. Again, not a big surprise, since they are next door to one another, and Martin Sierhuis the director of the research lab made his career over at Nasa. (Note of disclosure: I am good friends with Martin, and Singularity U is based at the NASA Research Park.)
Submitted by brad on Thu, 2015-01-08 19:55.
Day 3 at CES started with a visit to BMW’s demo. They were mostly test driving new cars like the i3 and M series cars, but for a demo, they made the i3 deliver itself along a planned corridor. It was a mostly stock i3 electric car with ultrasonic sensors — and the traffic jam assist disabled. When one test driver dropped off the car, they scanned it, and then a BMW staffer at the other end of a walled course used a watch interface to summon that car. It drove empty along the line waiting for test drives, and then a staffer got in to finish the drive to the parking spot where the test driver would actually get in, unfortunately.
Also on display were BMW’s collision avoidance systems in a much more equipped research car with LIDARs, Radar etc. This car has some nice collision avoidance. It has obstacle detection — the demo was to deliberately drive into an obstacle, but the vehicle hits the brakes for you. More gently than the Volvo I did this in a couple of years ago.
More novel is detection of objects you might hit from the side or back in low speed operations. If it looks like you might sideswipe or back into a parking column or another car, the vehicle hits the brakes on you (harder) to stop it from happening.
Insurers will like this — low speed collisions in parking lots are getting to be a much larger fraction of insurance claims. The high speed crashes get all the attention, but a lot of the payout is in low speed.
I concluded with a visit to my favourite section of CES — Eureka Park, where companies get small lower cost booths, with a focus on new technology. Also in the Sands were robotics, 3D printing, health, wearables and more — never enough time to see it all.
I have added 12 more photos to my gallery, with captions — check the last part out for notes on cool products I saw, from self-tightening belts and regenerating roller skates to phone-charging camping pots.
Submitted by brad on Wed, 2015-01-07 23:44.
After a short Day 1 at CES a more full day was full of the usual equipment — cameras, TVs, audio and the like and visits to several car booths.
I’ve expanded my gallery of notable things with captions with cars and other technology.
Lots of people were making demonstrations of traffic jam assist — simple self-driving at low speeds among other cars. All the demos were of a supervised traffic jam assist. This style of product (as well as supervised highway cruising) is the first thing that car companies are delivering (though they are also delivering various parking assist and valet parking systems.)
This makes sense as it’s an easy problem to solve. So easy, in fact, that many of them now admit they are working on making a real traffic jam assist, which will drive the jam for you while you do e-mail or read a book. This is a readily solvable problem today — you really just have to follow the other cars, and you are going slow enough that short of a catastrophic error like going full throttle, you aren’t going to hurt people no matter what you do, at least on a highway where there are no pedestrians or cyclists. As such, a full auto traffic jam assist should be the first product we see form car companies.
None of them will say when they might do this. The barrier is not so much technological as corporate — concern about liability and image. It’s a shame, because frankly the supervised cruise and traffic jam assist products are just in the “pleasant extra feature” category. They may help you relax a bit (if you trust them) as cruise control does, but they give you little else. A “read a book” level system would give people back time, and signal the true dawn of robocars. It would probably sell for lots more money, too.
The most impressive car is Delphi’s, a collaboration with folks out of CMU. The Delphi car, a modified Audi SUV, has no fewer than 6 4-plane LIDARs and an even larger number of radars. It helps if you make the radars, as otherwise this is an expensive bill of materials. With all the radars, the vehicle can look left and right, and back left and back right, as well as forward, which is what you need for dealing with intersections where cross traffic doesn’t stop, and for changing lanes at high speed.
As a refresher: Radar gives you great information, including speed on moving objects, and sucks on stationary ones. It goes very far and sees through all weather. It has terrible resolution. LIDAR has more resolution but does not see as far, and does not directly give you speed. Together they do great stuff.
For notes and photos, browse the gallery
Submitted by brad on Tue, 2015-01-06 23:11.
A reasonable volume of robocar related stuff here at CES. I just had a few hours today, and went to see the much touted Mercedes F015 “Luxury in Motion.” This is a concept and not a planned vehicle, but it draws together a variety of ideas — most of which we’ve seen before — with some new explorations.
The vehicle has a long wheelbase design to allow it to have a very large passenger compartment, which features just 4 bucket seats, the front two of which can rotate to create face to face seating. (In addition, they can rotate to make it easier to get into the car.) We’ve seen a number of face to face concepts and designs and I’ve been interested in the idea from the start, the idea of making car travel more social and better for both families and co-workers. As a plus, rear facing seats, though less comfortable for some fraction of the population, are going to be safer in a front end collision.
The vehicle features a bevy of giant touchscreens. We see a lot of this, but I actually will note that we don’t have this at our desks or in our homes. I suspect passengers in robocars will prefer the tablets they already have, though there is the issue that looking down at a tablet generates motion sickness sometimes.
The interior has an odd mix of carpet and hardwood, perhaps trying to be more like a living room.
More interesting, though not on display, are the vehicle’s systems for communicating with pedestrians and other road users. These include LEDs that can indicate if the car is self-driving (boring, and something I pushed to have removed from the Nevada law,) but more interesting are indicators that help to tell pedestrians the vehicle has seen them. One feature, which only is likely to work at night, laser projects a crosswalk in front of the vehicle when it stops, to tell a pedestrian it sees them and is expecting them to cross in front. It can also make LED words at the back for other cars (something that is I think illegal in some jurisdictions.
Also interesting has been the press reaction. Wired thinks it’s bonkers and not designed very well. The bonkers part is because the writer thinks it de-emphasizes driving too much. Of course, those of that stripe are quite upset at Google’s car with no controls. Other writers have liked the design, and find it quite superior to Google’s non-threatening design, suggesting the Google design is for regulators and the Mercedes design is for customers. Google plans to get approval for their car and operate it, while Mercedes is just using the F015 as a concept.
I have a gallery of several pictures of the car which I will add to during the week. In the gallery you will also see:
Audio Piloted Driving prototype
Audi drove one of their cars from the Bay Area to CES, letting press take 100 mile stints. It also helped them learn things about different conditions. One prototype is in the booth, I will go out to see the real car outdoors tomorrow.
TRW was showing off their technology with a transparent model showing where they had put an array of radars to make 360 degree radar and camera coverage. No LIDAR, but they will probably get one eventually. Radar’s resolution is low, but they believe that by fusing the radar and the camera views they can get very good perception of the road.
There are more for me to see tomorrow. Ford showed more of their ADAS systems and also their Focus which has 4 of the 32 plane velodyne LIDARs on it. Toyota showed only a hydrogen fuel cell car. Valeo has some interesting demos I will want to see — they have promised doing a good traffic jam assist. While they have not said so, I think the most interesting car company robocar function will be a traffic jam assist which does not require supervision — ie. you can read. While no car company is ready to have the driver out of the loop at high speeds, doing it at traffic jam speeds is much easier, because mainly you just have to follow the other cars, and you stop self-driving if the jam opens up. Several companies are working on a product like this and I suspect it will be the first real robocar product to reach the market that is actually practical. The “super cruise” products which drive while you watch are pleasant, but not much more world-changing than adaptive cruise control. When the car can give people time back, even if it’s only the traffic jam time, then something interesting starts happening.
Submitted by brad on Mon, 2015-01-05 15:28.
When Southwest started using tablets for in-flight entertainment, I lauded it. Everybody has been baffled by just how incredibly poor most in-flight video systems are. They tend to be very slow, with poor interfaces and low resolution screens. Even today it’s common to face a small widescreen that takes a widescreen film, letterboxes it and then pillarboxes it, with only an option to stretch it and make it look wrong. All this driven by a very large box in somebody’s footwell.
I found out one reason why these systems are so outdated. Apparently, all seatback screens have to be safety tested, to make sure that if you are launched forward and hit your head on the screen, it is not more dangerous than it needs to be. Such testing takes time and money, so these systems are only updated every 10 years. The process of redesigning, testing and installing takes long enough that it’s pretty sure the IFE system will seem like a dinosaur compared to your phone or tablet.
One airline is planning to just safety test a plastic case for the seatback into which they can insert different panels as they develop. Other airlines are moving to tablets, or providing you movies on your own tablet, though primarily they have fallen into the Apple walled garden and are doing it only for the iPad.
The natural desire is just to forget the airline system and bring your own choice of entertainment on your own tablet. This is magnified by the hugely annoying system which freezes the IFE system on every announcement. Not just the safety announcements. Not just the announcements in your language, but also the announcement that duty free shopping has begun in English, French and Chinese. While a few airlines let you start your movie right after boarding, you don’t want to do it, as you will get so many interruptions until the flight levels off that it will drive you crazy. The airline provided tablet services also do this interruption, so your own tablet is better.
In the further interests of safety, new rules insist you can only use the airline’s earbud headphones during takeoff and landing, not your nice noise cancellation phones. But you didn’t pick up earbuds since you have the nicer ones. The theory is, your nice headphones might make you miss a safety announcement when landing, even though they tend to block background noise and actually make speech clearer.
One of the better IFE systems is the one on Emirates. This one, I am told, knows who you are, and if you pause a show on one flight, it picks up there on your next flight. (Compare that to so many systems that often forget where you were in the film on the same flight, and also don’t warn you if you won’t be able to finish the movie before the system is turned off.)
Using your own tablet
It turns out to be no picnic using your own tablet.
- You have to remember to pre-load the video, of course
- You have to pay for it, which is annoying if:
- The airline is already paying for it and providing it free in the IFE
- You have it on netflix/etc. and could watch it at home at no cost
- You wish to start a movie one day and finish it on another flight, but don’t want to pay to “own” the movie. (Because of this I mostly watch TV shows, which only have a $3 “own” price and no rental price.)
How to fix this:
- IFE systems should know who I am, know my language, know if I have already seen the safety briefing, and not interrupt me for anything but new or plane-specific safety announcements in my chosen language.
- Like the Emirates systems, they should know where I am in each movie, as well as my tastes.
- How to know the language of the announcement? Well, you could have a button for the FA to push, but today software is able to figure out the language pretty reliably, so an automated system could learn the languages and the order in which they are done on that flight. Software could also spot phrases like “Safety announcement” at the start of a public address, or there could be a button.
- Netflix should, like many other services, allow you to cache material for offline viewing. The material can have an expiration date, and the software can check when it’s online to update those dates, if you are really paranoid about people using the cache as a way to watch stuff after it leaves Netflix. Reportedly Amazon does this on the Kindle Fire.
- Online video stores (iTunes, Google Play, etc.) should offer a “plane rental” which allows you to finish a movie after the day you start it. In fact, why not have that ability for a week or two on all rentals? It would not let you restart, only let you watch material you have not yet viewed, plus perhaps a minute ahead of that.
- Perhaps I am greedy, but it would be nice if you could do a rental that lets 2 or more people in a household watch independently, so I watch it on my flight and she watches it on hers.
- If necessary, noise-cancelling headphones should have a “landing mode” that mixes in more outside sound, and a little airplane icon on them, so that we can keep them on during takeoff and landing. Or get rid of this pretty silly rule.
Choosing your film
There’s a lot of variance in the quality of in-flight films. Air Canada seems particularly good at choosing turkeys. Before they close the doors, I look up movies — if I can get the IFE system to work with all the announcements — in review sites to figure out what to watch. In November, at Dublin Web Summit, I met the developers of a travel app called Quicket, which specialized in having its resources offline. I suggested they include ratings for the movies on each flight — the airlines publish their catalog in advance — in the offline data, and in December they had implemented it. Great job, Quicket.
Submitted by brad on Fri, 2015-01-02 16:19.
One of air travel’s great curses is that you have to leave for the airport a long time before your flight. Airlines routinely “recommend” you be there 2 or 3 hours ahead, and airport ride companies often take it to heart and want to pick you up many hours before even short flights. The curse is strongest on short flights, where you can easily spend as much as twice the time getting to the flight as you spend in the air.
The reality, though, is that it’s not nearly that strict. I often arrive much later. I’ve missed 3 flights in my life — in two cases because cheap airlines literally had nobody at the counter past their cutoff deadline, and once because United’s automated bag check line was very long (I got there before the deadline) but their computer is fully strict on the deadline while humans usually are not. In all cases, I got on another flight, and the time lost to these missed flights is vastly less than the time gained by not being at the airport so early.
But it’s getting harder. Airlines are getting stricter, and in a few cases offering no flexibility.
The big curse is that many of the delays can’t be predicted. It may almost always take 20 minutes to get to the airport, but every so often traffic will make it 40. Security is usually only 5-10 minutes but there are times when it’s 30. Car rental return, parking shuttles, called taxis and Ubers can have unexpected delays. Parking lots can be full (as happened to me this xmas after Uber failed me.) Immigration can range from 2 minutes to 1.5 hours if you have to go to secondary screening. While in theory you could research this, sometimes at strange airports you are surprised to find it’s 30 minutes walk and people-mover to your gate.
If you ever fly privately, though, you will discover a different world, where even if you’re just a guest you can arrive a very short time before your flight. (If you’re the owner, of course, it doesn’t take off until you get there.) But there are many options that can speed your trip through the airport without needing to fly a private jet:
- Tools like Google Now track traffic and warn you when you need to leave earlier to get to the airport
- If you take a cab to the airport, you eliminate the delays of parking and car return
- Though rarer today, ability to check bags in advance at remote locations helps a lot
- Curb checking of bags is great, as of course is online check-in sent to your phone
- (Not checking bags is of course better, and any savvy flyer avoids it whenever they can, but sometimes you can’t.)
- Premium passengers get check-in gates with minimal lines, and premium security lines
- If you have a Global Entry or Nexus card, you can skip the immigration/customs line
- TSA PRE, “Clear” and premium passenger security lines provide a no-wait experience. Of course nobody should ever have to wait, ever.
- Failing that, offering appointments at security for a predictable security trip can remove the time risk
- Sometimes they also let people who are at risk of missing a flight skip past the security line (and some other lines)
- In some cases, premium passengers are shuttled in vehicles within the terminal or on the tarmac
- Business class passengers can board as late as they want (or as early) and still get a place in the bins on most flights
In addition, I believe that if you wanted to get your checked bag cleared quickly by the TSA for money, it could happen. Of course, we can’t have everybody do this all the time, or so I presume, because it would require too much in the way of resources. But what if we allow you to do this occasionally when factors beyond your control have made you late.
What is proposed is that every so often — perhaps one time in twenty — when factors like traffic, long security lines or other things mostly beyond your control made you late, you could invoke an urgent need, and still make your flight.
This would allow you to budget a more reasonable time to arrive
What does this all add up to? It should be possible, at an extra cost, to get a quick trip through the airport. Say that cost is $200 (I don’t think it’s that much, but say that it is.) You could pay $10 extra per flight for “insurance” and be able to invoke an urgent trip every so often when things go wrong. It’s worth it to pay every trip because it gives you a benefit on every trip — you leave later, knowing you will make it even if traffic, security lines or similar factors would delay you too much.
Some of the services you might get would include:
- Somebody meets your car at the curb, takes your keys, and then parks it or returns it to the car rental facility
- Another employee meets you and checks in your bags at the curb. Your bags are put in a special urgent queue in TSA inspection. If need be a staffer walks it through.
- A golf cart takes you to security if it’s not close, and you get to the front of the line.
- If your gate is far, another golf cart or escort takes you there
The natural question is, “why wouldn’t you want this all the time?” And indeed you would, and a large fraction of passengers would pay a fairly high fee to get this when they need it. Airlines might make it just part of the service with high-priced tickets or super-elite flyers, and I see no reason that should not happen. The price can be set so that the demand matches the supply, based on the cost of having extra employees to handle urgent passengers.
When it comes to more “public” resources like TSA screening, they have a simple rule. You can give premium services to premium passengers if what you do also speeds up the line for ordinary passengers. A simple implementation of this is to just pay for an extra screening station for the premium passengers, because now you don’t butt in line and in fact by not being in the regular line at all, you speed it up for all in it. You don’t need to be so extravagant, however. For example, the “TSE PRE” line, which allows a faster trip through the X-ray (you don’t have to take anything out, or remove your shoes in this line) speeds up everybody because we all wait behind people doing that. If you can show that the amount you speed up the whole process is greater than the delay you add by letting premium passengers jump the queue, it is allowed.
But as fancy as these services sound, with extra staff, they are really not that expensive. Perhaps just 20 minutes of employee time for most of it — more if they are driving your car to a parking lot for you. (Note that this curb hand-off is forbidden by most airports because car rental companies already would like to offer it to their top customers but it is believed that would be too popular and increase traffic. Special permission would need to be arranged.)
For the “insurance” approach, a few techniques could assure it was not being abused. The frequency of use is one of them, of course, but you could also give people an app for their phones. This app, using GPS and knowing a flight is coming, would know when you left for the airport. In fact, it could give you alerts as to when to leave based on information about traffic, parking and security wait times. If you left at the reasonable departure deadline, you would get the urgent service if traffic or other surprise factors made you late. If you left after that deadline, you would not be assured the fast track path.
What would be better would be an app that actually works with all the airport functions you will interact with — check in, the gate, bag check, passenger screening, parking lots, rental cars, traffic etc. Their databases could know their state, any special conditions, and both recommend a time to leave that will work, but even make appointments for you and tell you when to leave for them. Then your phone could guide you through the airport and do all the hard work. It would provide an ID to get you your appointment at security. It might tell you to not drive your own car and take a car service instead if that’s easier than parking your car for you. It would coordinate for all the passengers using the system to make sure they flow through the airport in a well regulated manner, with no surprises, so that people don’t have to try to get there hours in advance.
Submitted by brad on Fri, 2014-12-19 13:39.
Yesterday’s note on Here’s maps brought up the question of the wisdom of map-based driving. While I addressed this a bit earlier let me add a bit more detail.
A common first intuition is that because people are able to drive just fine on a road they have never seen before that this is how robots will do it. They are bothered that present designs instead create a super-detailed map of the road by having human driven cars scan the road with sensors in advance. After all, the geometry of the road can change due to construction; what happens then?
They hope for a car that, like a human, can build its model of the road in real time while driving the road for the first time. That would be nice, of course, and gives you a car that can drive most roads right away, without needing to map them. But it’s a much harder problem to solve, and unlikely to ever be solved perfectly. Car companies are building very simple systems which can follow the lines on a freeway under human supervision without need for a map. But real city streets are a different story.
The first thing to realize is that any system which could build the correct model as you drive is a system that could build a map with no human oversight, so the situations are related. But building a map in advance is always going to have several very large advantages:
- You build the map from not just one scan of the road, but several, and done in different lanes and directions. As a result, you get 3-D scans of everything from different angles, and can build a superior model of the world.
- Using multiple scans lets you learn about things that are stationary but move one day to the next, like parked cars.
- You can process the data using a cloud supercomputer in as much time, memory and data storage as you want. Your computer is effectively thousands of times more capable.
- Humans can review the map built by the software if there’s anything it is uncertain about (or even if there is nothing) at their leisure.
- Humans can also test the result of the automatic and guided mapping to assure accuracy with one extra drive down the road.
In turn there are disadvantages
- At times, such as construction, the road will have changed from when it was mapped
- This process costs effort, and so the vehicle either does not drive off the map, or only handles a more limited set of simpler roads off the map.
The advantages are so great that even if you did have a system which could handle itself without a map, it is still always going to be able to do better
with a map. Even with a great independent system you would want to make an effort to map the most popular roads and the most complex roads, up to the limit of your budget. The cost is an issue, but the cost of mapping roads is nothing compared to the cost of building or maintaining them. It’s a few times driving down the road, and some medium-skilled labour.
The road has changed
Let’s get to the big issue — the map is wrong, usually because construction has changed it.
First of all, we must understand that the sensors always disagree with the map, because the sensors are showing all the other cars and pedestrians etc. Any car has to be able to perceive these and drive so as not to hit them. If a traffic cone, “road closed” sign or flagman appears in the road, a car is not going to just plow into them because they are not on the map! The car already knows where not to go, the question is where it should go when the lanes have changed.
Even vehicles not rated to drive any road without a map can probably still do basic navigation and stay within their lane markers without a map. For the 10,000 miles of driving you do in a year, you need a car that does that 99.99999% of the time (for which you want a map) but it may be acceptable to have a car that’s only 99.9% able to do that for the occasional mile of restriped road. Indeed, when there are other, human-driven cars on the road, a very good strategy is just to follow them — follow one in front, and watch cars to the side. If the car has a clear path following new lane markers or other cars, it can do so.
Google, for example, has shown videos of their vehicle detecting traffic cones and changing lanes to obey the cones. That’s today — it is only going to get better at this.
But not all the time. There will be times when the lanes are unclear (sometimes the old lanes are still visible or the new ones are not well marked.) If there are no other cars to follow, there are also no other cars to hit, and no other traffic to block.
Still, there will be times when the car is not sure of where to go, and will need help. Of course, if there is a passenger in the car, as there would be most of the time, that passenger can help. They don’t need to be a licenced driver, they just need to be somebody who can point on the screen and tell the car which of the possible paths it is considering is the right one. Or guide it with something like a joystick — not physically driving but just guiding the car as to where to go, where to turn.
If the car is empty, and has a network connection, it can send a picture, 3-D scan and low-res video to a remote help station, where a person can draw a path for the car to go for its next 100 meters, and keep doing that. Not steering the car but helping it solve the problem of “where is my lane?” The car will be cautious and stop or pull over for any situation where it is not sure of where to go, and the human just helps it get over that, and confirms where it is safe to go.
If the car is unmanned and has no network connection of any kind, and can’t figure out the road, then it will pull over, or worst case, stop and wait for a human to come and help. Is that acceptable? Turns out it probably is, due to one big factor:
This only applies to the first car to encounter an unplanned, unreported construction zone
We all drive construction zones every day. But it’s much more rare that we are the first car to drive the construction zone as they are setting it up. And most of the rules I describe above are only for the first connected car to encounter a surprise change to the road. In other words, it’s not going to happen very often. Once a car encounters a surprise change to the road, it will report the problem with the map. Immediately all other cars will know about the zone.
If that first car is able to navigate the new zone, it will be scanning it with sensors, and uploading that data, where a crew can quickly build a corrected map. Within a few minutes, the map and the road will no longer differ. And that first car will be able to navigate the new zone 99.999% of the time — either because it has a human on board, remote human help or it’s a simple enough change that the car is able to drive it with an incorrect map.
In addition, the construction zone has to be a surprise. That means that, in spite of regulations, the construction crews did not log plans for it in the appropriate databases. Today that happens fairly often, but over time it’s going to happen less. In fact, there are plans to have transponders on construction equipment and even traffic cones that make it impossible to create a new construction zone without it showing up in the databases. Setting up a road change has a lot of strongly enforced safety rules, and I predict we’ll see “Get out your smartphone and make sure the zone is in the database before you create it” as one of them, especially since that’s so easy to do.
(You have probably also seen that tools like Waze, driven by ordinary human driver smartphones, are already mapping all the construction zones when they pop up.)
If a complex zone is present and unmapped, unmanned cars just won’t route through there until the map is updated. The more important the zone, the more quickly it will get updated. If need be, a mapping worker will go out in a car before work even begins. If a plan was filed, we’ll also know the plan for the zone, and whether cars can handle it with an old map or not.
Most of the time, though, a human passenger will be there to guide the car through the zone. Not to steer — there may not be a steering wheel — but to guide. The car will go slowly and stay safe.
Once a car is through, it will send the scans up to the mapping center, and all future cars will have a map to guide them until the crew changes the road again without logging it. I believe that doing so should be made against safety regulations, and be quite rare.
So look at those numbers. I will hope it’s reasonable to expect that 99% of construction zones will be logged in road authority databases before they begin. Of the 1% that aren’t, there will be a first robocar to encounter the zone. 90% of the time that car will have a passenger able to help. For the 10% unmanned cars, I predict a data network will be available 99% of the time. (Some would argue 100% of the time because unmanned cars will just not go where there is not a data connection, and we may also get new data services like Google’s Loon, or Facebook’s drone program to assure coverage everywhere.)
So now we are looking at one construction zone in 100,000 where there was no warning, there is no human, and there is no data. But we’ve rated are car as able to handle handle off-map driving 99.9% of the time. For the other .1%, it decides it can’t see a clear path, and pulls over. When it doesn’t report back in on the other side of the data dead zone, a service vehicle is dispatched and fixes the problem.
So now in one in 100,000,000 construction zones, we have a car deciding to pull over. Perhaps for half of those, it can’t figure out how to pull over, and it stops in the lane. Not great — but this is one in 200 million construction zones. In other words, it happens with much less frequency than accidents or stalled cars. And there is even a solution. If a construction worker flashes an ID card at the car’s camera when it’s in a confused state, the car can then follow that worker to a place to stop. In fact, since the confused state is so rare, there is probably not even a need for an ID card. Just walk up, make a “follow me” gesture and walk the car where it needs to go.
Tweak these numbers as you like. Perhaps you think there will be far more construction zones not logged in databases. Perhaps you think the car’s ability to drive a changed zone will only be 50%. Perhaps you think there will still be lots of unmanned cars running in wireless dead zones in 2020. Even so the number of cars that stop and give up will still be far fewer than the number of cars that block roads today due to accidents and mechanical problems. In other words, no big whoop.
It’s important to realize that unmanned cars are not in a hurry. They can avoid zones they are not comfortable with. If they can’t get through at all, the taxi company sending the car can just send another from a different direction in almost all cases.
It’s also important to realize that cars in an uncertain situation are also not in a big hurry. They will slow until they can be sure they are safe and able to handle the road. Slow, it turns out, is easy. Slow and heavy traffic (ie. a traffic jam) is actually also very easy — you don’t even need to see the lines on the road to handle that one; you usually can’t.
Once again this is only for the first car to encounter the surprise zone. Much more common will be a car that is the first to encounter a planned zone. This car will always have a competent passenger, because the service will not direct an unmanned car into an unknown construction zone where there is no data. This passenger will get plenty of warning, and their car may well pull over so there is no transition from full-auto to semi-auto while the car is moving. Then this person will guide the car through the zone at reduced speed. Probably just with a joystick, though possibly there will handlebars that can pop out or plug in if true semi-manual driving is needed.
New road signs
Road signs are a different problem. Already there are very decent systems for recognizing road signs captured by the camera — systems that actually do better at it than human beings. But sometimes there are road signs with text, and the system may recognize them, but not understand them. Here again we may call upon human beings, either in the vehicle, or available via a data connection. Once again, this is only for the first unmanned car to encounter the new road sign.
I will propose something stronger, though. I believe there should be a government mandated database of all road signs. Further, I believe the law should say that no road sign has legal effect until it is entered in the database. Ie. if you put up a sign with a new speed limit, it is not a violation of the limit to ignore the sign until the sign is in the database. At least not for robots. Once again, all this needs is that the crews putting in the signs have smartphones so they can plonk the sign on the map and enter what it is.
We may never need this, though, because the ability of computers to read signs is getting very good. It may be faster to just make it even better than to wait for a law that mandates the database. With a 3-D map, you will never miss a brand new sign, but you might get confused by a changed sign — you will know it changed but may need to ask for help to understand it if it is non-standard. There are already laws that standardize road signs, but only to a limited extent. Even so, the number of sign styles in any given country is still a very manageable number.
Random road events
Sometimes driving geometry changes not due to construction, but due to accidents and the environment. Trees get knocked down. Roads flood. Power lines may fall. The trees will be readily seen, and for the first car to come to a fallen tree, the procedure will be similar, though in a low traffic area the vehicles will be programmed to go around them, as they are for stalled cars and slow moving vehicles. Flooding and power lines are more challenging because they are harder to see. Flooding, of course, does not happen by surprise. That there is flooding in a region will be well known so cars will be on the lookout for it. Human guides will again be key.
A plane is not a bird
Aircraft do not fly by flapping their wings, and robocars will not see the world as people do nor drive as they do. When they have accurate maps, it gives us much more confidence in their safety, particularly the ability to pick the right path reliably at speed. But they have a number of tools open to them for driving a road that doesn’t match the map precisely without needing to have the ability to drive unmapped roads 99.999999% of the time. That’s a human level ability and they don’t need it.