Americans Shouldn’t Have to Drive, but the Law Insists on It
theatlantic.com
The automobile took over because the legal system helped squeeze out the alternatives. In a country where the laws compel the use of cars, Americans are condemned to lose friends and relatives to traffic violence.