Why do white men get so hate when we are the ones that freed the blacks and gave them and women the right to vote?
If white men or so evil, then why were we ever willing to relinquish our total control over the vote and our right to own people as property?
This is a classic case of biting the hand that feeds you, since white men are the ones that have built Western Civilization and the modern global order, but instead of retracting our hand and no longer feeding these parasites we keep giving them our "food".
When will white men rise up again and remind these societal parasites that we are the ones they have to thank for everything they have in this world? What CAN we do at this point to remind them that we are responsible for all the luxuries and rights they have?
It almost seems like a lost cause.