Kevin Costner's The West

The series is said to detail how the Wild West period of American history continues to impact the country today.