• PolarPerspective
        link
        fedilink
        arrow-up
        1
        ·
        1 year ago

        You say that, but I think bikinis are objectifying af. I’m perfectly happy for women to dress however they want. But from the perspective of a man, I find almost-naked women to be the option I benefit more from.

        Western culture seems to push women into degrading positions through social pressure rather than legal means. Just look at how common skin tight leggings and a sports bra are as gym attire.

        Is this actually benefitting women? Or is it just another way to take advantage of them?