So here lately I've been noticing more and more women dressing conservatively and having a desire for a traditional family life and having children.
Have women finally begun to see through the feminist lie they've had crammed down their throats over the past few years/decades?
It's a trap.
They've sucked a hundred cocks and now they are looking for a nice beta provider to clean up their mess of a life.
Will you step up, anon?