During the Cold War years, mainstream commentators were quick to dismiss the idea that the United States was an imperialist power. Even when U.S. interventions led to the overthrow of popular governments, as in Iran, Guatemala, or the Congo, or wholesale war, as in Vietnam, this fiction remained intact. During the 1990s and especially since September 11, 2001, however, it has crumbled. Today, the need for American empire is openly proclaimed and...