American Old West
The Western
United States of America has played a significant role in both
history and
fiction. The term
Wild West relates to an ever-shifting nebulous land populated by
cowboys and
Indianss, trappers and
explorers; and located to the west of the settled, law-abiding
United States and beyond
The Frontier.
See also: The West (U.S.)
This article is a stub article. You can help Wikipedia by improving it.\n