I almost feel bad for career women. On one hand, they were lied to about being flawless princesses that’ll get their very own prince, and on the other, they sowed the seeds of their own destruction by promoting feminism.
The only way we’re legally getting back to some sense of normality is if women voluntarily choose to roll back all their bullshit “rights.” No more voting, no more no-fault divorce, no more marital “rape” laws, and very limited working rights (women shouldn’t be working unless they are exceptionally talented, like Marie Curie).
And also back then, women had to prove themselves, and they had the entire community bitch slapping them back into place
It also used to be that if you grabbed a woman by the ass, she’d either giggle or slap you - none of this #metoo bullshit you see today. Women back then had thicker skin, because society was tougher in general.
Is there actual historical context for this? This is news for me.
(post is archived)