Feminism was only being so heavily promoted because it most effectively sowed division in, and corroded the foundations of, Western society.
Now that race and transgender issues are more effective tools for those ends, activist money and media/big tech attention are obviously going to prioritize them, and feminism goes on the back burner. Radical left-wing feminists even get smeared as Nazi TERFs by their former supposed allies if they don't support the new paradigm for undermining society.
Of course, this should have been obvious, if they understood the motivations of the Jewish elite who upheld Feminism as the highest establishment doctrine of its time. But instead, feminists ignored the destruction left in their wake so they could pretend that they were heroes, taking over culture by their own righteous efforts alone. Such a nonsense worldview precludes any clear explanation for why things have changed as they have. Of course they're confused.
(post is archived)