• PolarPerspective
    link
    fedilink
    arrow-up
    1
    ·
    1 year ago

    You say that, but I think bikinis are objectifying af. I’m perfectly happy for women to dress however they want. But from the perspective of a man, I find almost-naked women to be the option I benefit more from.

    Western culture seems to push women into degrading positions through social pressure rather than legal means. Just look at how common skin tight leggings and a sports bra are as gym attire.

    Is this actually benefitting women? Or is it just another way to take advantage of them?