Since USA is now basically falling apart as a nation, maybe we should discuss what this means for women. I am not American but have followed this death spiral for years and of course the same development is happening (just slower) in other western countries.
USA is split between the Woke "Minority Bloc" and the reactionary MAGA/Alt-Right. Both of them are openly misogynist. The first wants to turn every teen girl into "empowered" OnlyFans prostitute, sacrifice women's rights for trannies, force lesbians to suck girlpenis, throw white "Karens" under the bus etc. The second wants women barefoot and pregnant with no abortion and no vote.
For women, this situation is basically hell. No other group has to choose between two options that are so aggressively, violently AWFUL to them. There is no positive future for women in America or other western countries. Women's rights are virtually gone, nobody supports them anymore, and feminists are gleefully participating in this to ensure their WOKE brand of male supremacy will win.
Liberals, feminists and leftists have utterly betrayed women. Women are majority of their supporters and will continue to support this because most women are apparently fucking braindead. Liberal identity politics crap first stoked far-right extremist misogyny with all the performative "lol male tears" provocation and then decided that actually, white cis women are evil oppressors and threw them to the wolves they helped to create.
Never has it been as clear that BlackPillFeminism is right and there is no hope for women and every single political movement and ideology is anti-woman.