Western Women wonder why:
> they don't get taken seriously enough
> where are the good guys?
> raped with impunity by the darker skinned hordes
> they become surplus to requirements in the entertainment industry after the age of 30
> leered over by creepers
Anything negative they perceive to be happening to them is 100% down to them. They want their shitty degenerate cake and they don't give a fuck about the consequences. They seem to live for the moment, much like a nigger. Western women seem to have zero class now.