• @jeffers00n
    link
    English
    7911 months ago

    I don’t understand why journalists continue to call users with the blue check-mark “verified”. There’s no verification of anything but a credit card. Verification is a relic of the past and they need to shift their language to match reality. Maybe call them “paid” users.

    • @KeefChief12
      link
      English
      1511 months ago

      Yeah verified actually had some weight in the past.

    • @burak
      link
      English
      1
      edit-2
      11 months ago

      deleted by creator

  • @ghariksforge@lemmy.world
    link
    fedilink
    English
    3211 months ago

    Someone should take screenshots and put them on the internet. It will scare Microsoft and others into not running ads on Twitter.

    • @Pechente
      link
      English
      711 months ago

      Exactly. I don’t get the mentality of not running ads in a YouTube video where someone said „fuck“ but being completely fine with running your ads on twitter.

      I mean just checking the feed these days is incredibly toxic. Seems like only people of a certain mindset kept sticking around.

  • @conderoga@beehaw.org
    link
    fedilink
    English
    3011 months ago

    Advertisers need to see more articles like this. Pulling the ad dollars away from Twitter is the last lever anyone has to try to change it, and it should have happened long ago at this point.

  • ikiru
    link
    English
    2011 months ago

    I mean, Walt Disney might not have actually minded this.

    • @SturgiesYrFase
      link
      English
      111 months ago

      Yeah, lots of people forget/don’t know he was pretty fascie/antisemitic

    • @xevizero
      link
      811 months ago

      Do they want this?

      Apparently they have no issues with nazis, but the moment you mention a minor slur in a youtube video you will be demonetized into oblivion, and don’t even try showing some skin or they will call the FBI.

    • @loki
      link
      311 months ago

      I guess ads are more effective on easily brainwashed people, so it’s good for business. they want business not feel-good morals…

    • @Zerush
      link
      211 months ago

      If Musk ban neo nazis and taliban in current Twitter, then it would run out of users

      • @MetalHead77
        link
        111 months ago

        Except the Taliban was allowed on Twitter long before Musk took it over.

  • @Oxossi
    link
    English
    1411 months ago

    Twitter must be held responsible for the kind of Nazi content its allowing on their site.

      • OrangeSlice
        link
        English
        411 months ago

        Would be great if the consequences were greater, but as you imply, they are on a track toward certain death it seems like (just gonna take a few more years idk)

  • Rosriv
    link
    fedilink
    1111 months ago

    At this point, still using Twitter is like being on Reddit, I can’t say which one is worse.

    • ram
      link
      fedilink
      1311 months ago

      Much as I hate jailbait hufflepuffman, twitter’s worse.

    • ZILtoid1991
      link
      fedilink
      211 months ago

      I only using Twitter because my VTuber oshis are there. I started to tell them about our lord and savior, the Fediverse.

      Could survive if I got banned though.

  • S4nvers
    link
    fedilink
    811 months ago

    Imagine contacting a company as a journalist for comment and getting a poop emoji as a response

    Twitter is almost as big of a shitshow as Reddit is

  • @WhatASave@lemmy.world
    link
    fedilink
    English
    611 months ago

    Don’t the ads just run automatically on content? It’s not like they’re hand picking these as targeted ads for people who enjoy neo-nazi propaganda.

    • @ghariksforge@lemmy.world
      link
      fedilink
      English
      1011 months ago

      Most ads are targeted these days. You tell which user profiles to run ads to.

      Also appearing next to controversial content is brand damage for these companies. They try to avoid it as much as possible.

      • @PerogiBoi@lemmy.ca
        link
        fedilink
        English
        -111 months ago

        With how the Overton window is shifting far to the right, I’m not sure companies really care anymore if their content is placed next to hateful shit. As long as it generates money it is acceptable.

        • Kichae
          link
          fedilink
          511 months ago

          The only way big brands keep expanding is by finding new audiences to appeal to. Groups who were socially excluded in the past become the next big target demographic as companies reach saturation in other markets.

          Having their business be publicly associated with groups who are openly attacking these new frontiers of sales is absolutely something they want to avoid.

          Not, you know, because they believe Nazis are bad - they may or may not hold that view - but because Nazis are unpopular.

          • @PerogiBoi@lemmy.ca
            link
            fedilink
            511 months ago

            While that’s how things used to be, the heavy rebranding of white supremacy into this new alt right “anti woke” has made Nazi views a lot more mainstream. One of the most popular “news” networks in North America normalize white supremacist ideology daily to hundreds of millions.

            • Kichae
              link
              fedilink
              311 months ago

              It’s still very much the case. The companies that are big enough to need new markets to expand into already have the white supremacists’ dollars, and they’re not afraid of losing them - their boycotts have not proven to actually involve not buying the products.

              White supremacy has been normalized in the US and Canada since their inceptions. These countries are built on a foundation of it. It’s not like companies actually care about the white supremacy.

              They care about losing black and pink dollars.

              They don’t care about losing white supremacists’ dollars because they’re not convinced they will.

        • LostCause
          link
          fedilink
          111 months ago

          Lol I can just imagine them sitting in some dystopian marketing brand meeting analysing target demographics… “It seems one of the biggest growing demographics is young Nazis who now call themselves “alt-right”, so Jenny, any ideas on how we can tailor our content to appeal to even more Nazis, while also not upsetting our main demographic of centrist customers who prefer to ignore all politics?”

          • @federico3
            link
            311 months ago

            This is pretty much what already happens, with the main difference that the target group would have a codename or a number. Algorithms don’t care.

          • @PerogiBoi@lemmy.ca
            link
            fedilink
            311 months ago

            You laugh but that’s literally what goes on in some meetings. There are entire “tiger teams” that focus on monetization of different demographics.

            • LostCause
              link
              fedilink
              511 months ago

              I do laugh cause I got a strong gallows humour, but you are right and I know that.

              I used to work in marketing and now am in IT precisely because I couldn‘t take the dystopian nature of it, and I didn‘t even make it far up or long at all, so I only saw the tip of the iceberg.

              For example LGBT in a global company the companies only do that sort of marketing in places where the general culture is friendly towards them and avoid it in those where it isn‘t. Specifically the US came up just a while ago a friend told me in his company they decided on not running pro-LGBT ads there anymore, because the Bud Light thing signals to them a shift in the US culture becoming more hostile towards it and they have to go with that flow.

              Yeah, their own ethics may not agree, but also do not matter, just what the company and shareholders demand and that is: more profit at any cost.

              • @PerogiBoi@lemmy.ca
                link
                fedilink
                211 months ago

                I guess we all need to appreciate dark humour with what’s going on 🤪

                In UX design this is especially bad as there are hundreds of studies that have been done that work out exactly how to pull the strings of human psychology to induce micropathologies (like depression, fomo, anxiety, PTSD, etc).

                Couldn’t stay at my previous job for the same reason. No job should force you to intentionally create systems that hurt other people because it hasn’t been made illegal and it makes money the quickest.

                I’m not salty at all wtf u talking about 🤪

    • @JuxtaposedJaguar
      link
      English
      311 months ago

      It used to not reflect on the advertiser, but ever since advertisers started demanding more control over where their ads were displayed, the expectation is now that they all do it. So allowing your ads to be displayed in those places is now interpreted as an endorsement of that content.

  • @const_void
    link
    English
    611 months ago

    Just like Reddit is trying to do. Monetizing hate is profitable apparently.

  • @Zerush
    link
    English
    311 months ago

    They are democratic decisions adjusted to the majority of the current remaining users of Twitter.

    • @Xanvial@lemmy.one
      link
      fedilink
      English
      411 months ago

      you misread the article

      there’s a tweet post containing neo-nazi video, and the ads for big company appears alongside that tweet

  • @not_woody_shaw@lemmy.world
    link
    fedilink
    English
    111 months ago

    It’s hard to do automated “Brand Safety” calculation on video. I’m not aware of any brand safety product that can do it yet. Did the ads at least get taken down after a human noticed it?