Country Dictionary

A fork of Rural Dictionary

west coast is the best coast

Meaning west coast is the best coast, obviously.

California for life.

by FusionStunt May 17, 2004

west coast is the best coast

The truth: The East Coast has better food, culture, architecture, and historical significance, while the West Coast has better weather.

Californians should all jisatsusuro

by Callahan May 20, 2004