Straight white men built the civilization we enjoy today and it was straight white men who gave equality to women, to gays and to POCs/immigrants. I'm just wondering why we did it? The POCs/Women/Gays don't seem to appreciate us and in fact have used their new found power to oppress straight white males, so why did we do it?
Serious question: why don't we take the power back for ourselves?
Its not defined in any dictionary by it no...but the culture (god, family, values, sovereignty, etc) just so happens go be traditional values of Western (white) culture. You dont have to be white to hold those values...but traditionally these values are alien to southern hemisphere cultures meaning when they come here, they have to assimilate to those values to be a part of American culture since whites are a majority. Which raises the following questions worth thinking about: What makes someone American? Is it physically living here? Or is it certain values/culture? If (when) whites become a minority, will America still be America? Very important questions that are time sensitive.