Why do White Americans feel as if they should have any real claim to their country? You've always lived among niggers weather it was the natives or the ones you brought in overseas. Why are you walking in the streets to fight for a nation that was never white to begin with.