What is up with the media and the television airing more and more shows or gearing their content and information towards minorities? This is setting a bad example for my young children to be subjected to this worthless crap. If it doesn't involve or concern me as a white american, then I don't want to hear about whatever it may be being widely publicized. Write back and let me know your thoughts. I'm not the only one who feels this way, am I?