girls like? Hollywood has always been out of touch with reality, what really goes on in America, but I feel that it's done people a huge disservice here. It has made men afraid to be men, sucked their virility out of them, and made them into sensitive pansies. If you look at it honestly, men have been turned into slaves here in America. And the funny part is, the girls will ***** around with nice guys and screw with their heads, but will never be with them, or even if they get together, will never listen to them. I laud women for this, for screwing the brains of tranny pansies aka men in America.