Permanently Deleted

  • SerLava [he/him]
    ·
    2 years ago

    All fucking AI turns out to be just the collective work of destitute third world laborers

  • StewartCopelandsDad [he/him]
    ·
    2 years ago

    Content moderation is a sick job. Like slaughterhouse worker. A good society shouldn't require anybody to do that.

    • ElmLion [any]
      ·
      2 years ago

      Sure seems to suck. But what's the alternative?

      • StewartCopelandsDad [he/him]
        ·
        edit-2
        2 years ago

        Content moderators only seem to be necessary in spaces without real consequences. If you post CSAM in the friend group chat you are ostracized, in the work email list you are fired. You can't just make another account and keep doing it. Preventing repeat offenders means the volume of offensive material is so low that everybody can share the load and it's no big deal. Of course you don't need slaughterhouse workers if you don't eat flesh.

        • ElmLion [any]
          ·
          2 years ago

          The internet will forever be a space without 'real consequences', that's how it work.

          Much as I hate and long for the destruction of Facebook, Reddit, Twitter etc. and a return to community forum days, your approach only works if there are no large spaces of any kind in which to post, because anywhere with like, >1000 users is gonna need dedicated moderators of some kind.

          • StewartCopelandsDad [he/him]
            ·
            2 years ago

            You are correct that the internet will forever be an anon hellscape. A lot of people have tried to improve it, e.g. stack exchange's reputation system, with mixed results. I don't think it should be this way though.

            anywhere with like, >1000 users is gonna need dedicated moderators of some kind.

            Yes, but not in the "removing bestiality posts" way. SRA national forums are this scale and mods just clean up flame wars. Probably because you need to pay $25 to get in.

            A work slack where your posts can get you fired or arrested doesn't need this sort of dedicated abuse viewer position no matter how big it gets. HR may occasionally fire somebody who accidentally pastes a Pornhub link into #general but they're not getting PTSD from it.

            But if it came down to it, I'd be comfortable giving up my Facebook car videos so that people didn't have to sit and watch CSAM all day. Not worth it.

            • ElmLion [any]
              ·
              2 years ago

              Sure, so it's either very small or very tightly gatekept. Unless you clamp down on every single publically accessible website, which isn't doable, this will be a thing.

              • StewartCopelandsDad [he/him]
                ·
                2 years ago

                I do want to note that such a clampdown is doable. Just not in the US under capitalism, probably. Big websites only enforce CSAM stuff now to avoid advertiser flight, and possibly legal consequences if it gets really bad. Enforcing ID-above-X-users, etc. would be about the same amount of coercion.

                Anyway we've had slaughterhouses for a hundred years so mod PTSD is probably gonna be around for a long time.

                • ElmLion [any]
                  ·
                  2 years ago

                  It really isn't. Not if the internet is to exist in any meaningful and public form; the obstacles are virtually, if not literally, insurmountable. So long as anyone can get an IP address and access the internet, they can serve content and protocols/software can be made to browse that content as easily as we can on the internet today.

                  Much like governments of today wanting to break encryption, the only way to make this doable is to effectively defunct the whole point of computers.

          • DoghouseCharlie [he/him, comrade/them]
            ·
            2 years ago

            What if in order to sign up you had to use an ID, regardless of site or something. Obviously that would be hell to implement right now under capitalism. But could it work under socialism? Doesn't China have a similar thing?

            • hypercube [she/her]
              ·
              2 years ago

              anonymity provides more good than harm imo - even if there's a socialist government in charge you might be queer & not want people knowing about that, or have gotten your arse kicked by the cops for no reason since cops are still fuckin cops, or (since you specifically mentioned China) have issues with your capitalist boss that you wanna talk about with other workers without endangering your livelihood. imo this slaughterhouse model is really a product of how current social media is structured. it's funny that this is my example but furaffinity has ~40k concurrent users at any given time, is constantly besiged by fascist spam, and relies on volunteer mods who don't seem too traumatised by the work. when you're actually part of the community then the horrid part is a much more manageable % of your life than when you show up to do a 9-to-5 in the Omelas box

              • DoghouseCharlie [he/him, comrade/them]
                ·
                2 years ago

                Good points. I would say, with the Furaffinity example, you are kind of just hoping the volunteer mods aren't also trash people. Because yeah, FA is fairly decent but the other furry websites I've never heard of and don't go to frequently have really dipshit mods that are too interested in maintaining freeze peach to do anything about the Nazis or dillweeds posting pictures of their sona having a transphobic screed. Which I wouldn't know anything about because I'm not on those sites, but still.

                Jokes aside I do agree, the way a site is setup tends to bring in a certain type of user or incentivize different behaviors so that's probably the best place to start.

            • ElmLion [any]
              ·
              2 years ago

              It's technically impossible to enforce for all websites, this is just the reality of how innernet work.

            • ElmLion [any]
              ·
              2 years ago

              It is impossible to clamp down on all possible sites on the internet at this stage. No matter what you do, search engines will exist, and to be honest I think they serve a very good purpose separate from making money.

        • World_Wario_II [he/him]
          ·
          2 years ago

          Do you propose linking real identity to everyone’s browsing and removing all ability to post anonymously? Seems like a major privacy and opsec concern and a major win for global capitalist surveillance

      • Nagarjuna [he/him]
        ·
        2 years ago

        Reddit's system of deputizing moderators. Fuck, /r/anarchism's system of user written rules and elected moderators is probably the best model.

        • ElmLion [any]
          ·
          2 years ago

          That's fine with me. But at some stage, you're gonna have a space big enough that this needs dedicated work, whatever you do.

      • StewartCopelandsDad [he/him]
        ·
        2 years ago

        Yup. I think in a good society, you would go to jail (or be reeducated) upon posting CSAM and so there wouldn't be an endless torrent of it into moderated spaces.

    • PaulSmackage [he/him, comrade/them]
      ·
      2 years ago

      I met a guy whose job was to review all suspicious content that was seized at pearson international. Dude seemed totally dead inside.

  • Simferopol [none/use name]
    ·
    edit-2
    2 years ago
    cw:mention of sa, child abuse etc

    In February, according to one billing document reviewed by TIME, Sama delivered OpenAI a sample batch of 1,400 images. Some of those images were categorized as “C4”—OpenAI’s internal label denoting child sexual abuse—according to the document. Also included in the batch were “C3” images (including bestiality, rape, and sexual slavery,) and “V3” images depicting graphic detail of death, violence or serious physical injury, according to the billing document. OpenAI paid Sama a total of $787.50 for collecting the images, the document shows

    In a statement, OpenAI confirmed that it had received 1,400 images from Sama that “​​included, but were not limited to, C4, C3, C2, V3, V2, and V1 images.” In a followup statement, the company said: “We engaged Sama as part of our ongoing work to create safer AI systems and prevent harmful outputs. We never intended for any content in the C4 category to be collected. This content is not needed as an input to our pretraining filters and we instruct our employees to actively avoid it. As soon as Sama told us they had attempted to collect content in this category, we clarified that there had been a miscommunication and that we didn’t want that content. And after realizing that there had been a miscommunication, we did not open or view the content in question — so we cannot confirm if it contained images in the C4 category.”

    :what-the-hell: i thought it was just text initially, but nope straight up images. this kind of job should include free therapy at the very very least.

      • Deadend [he/him]
        ·
        2 years ago

        The extra horrifying thing, they do have good CP image recognition and hash values. Which means it’s new stuff.

        It’s an endless stream of new horrors.

  • iie [they/them, he/him]
    ·
    2 years ago

    Go dunk on Steven Cohn's disgusting take in the replies https://twitter.com/spcohn/status/1615718665009102850

    He points out that $320 a month is 53% above the living wage in Kenya, ignoring that 1) a living wage in Kenya buys you fucking poverty, and 2) their job is disgusting and traumatizing, and they are receiving literally zero mental health support to do it, even though the company can easily pay for that fucking support.

  • aaro [they/them, she/her]
    ·
    2 years ago

    Go read Work without the Worker by Phil Jones if you want to read about how this has been happening for decades and how they enforce compliance and make the work un-unionizible. Maybe the most depressing book I've ever read, and I've read The Jakarta Method

    • Nagarjuna [he/him]
      ·
      2 years ago

      Ghost Work by Siddharth Suri and Mary Grey is another good analysis of the topic, and although the writers are libs, they still call for the decommodiifcation of labor.

  • Nagarjuna [he/him]
    ·
    2 years ago

    This is a huge issue in the ghost work industry--the content aggregators and moderators who make the internet work--they get paid shit wages to scrub CP off Facebook and YouTube with no therapy to process what they're seeing.

    These folks tend to be marginalized on multiple axis, often being disabled women in the global south.

    They're also harder to organize because they don't share a workplace, the labor arbitrage is huge, and some of the states lack a legal process for recognition. For example, the US, which is home to most of FB's content moderation has no way for contract laborers to gain NlRB recognition.

    Any time you do a Google search, buy something on Amazon, watch a YouTube video, or even browse FB, you're relying on the work of thousands of ghost laborers.

    For all of :reddit-logo:'s problems, it's system of volunteer moderators is by far the most humane of any major social media platform and its a good thing we brought it over here.

  • Llituro [he/him, they/them]
    ·
    edit-2
    2 years ago

    When OpenAI is valued at $29 billion for their electric parrot(a.) that runs on slavery (apparently), liberal economic hegemony teaches you to think that in terms of dollars this means OpenAI has made a thing that objectively has that much value. What's really happening though is that billionaires with that much control, $29 billion worth, of actual world economic production across the world, human labor productivity, capital ownership, etc. are making a bet that they can make more profit by investing into this thing. What the thing actually does doesn't matter. OpenAI is just another label to finance wonkery.

    a. There has been mixed discussion about the importance of ChatGPT in terms of what it does. What I think it is is a more interesting way to search whatever is archived on forums like reddit and 4chan. It can pull together source code from stackoverflow and github, which is pretty neat in practice, but it'll also tell you that the age of consent is 14 or whatever because it's training on the string "age of consent" is all related to fucking libertarian incels on 8chan. Without better application, I think there is little interest in the diminishing returns of the current batch of ML spectacles. I realize that it looks like they're getting exponentially better, and in some ways they are, but they are also dependent on finicky and exponentially amassed tech infrastructure that requires continued u.s. consumer stability and militarily enforced supply chain stability, and I think most of us are doubtful about the medium term stability of american hegemony in the global south, where it matters most. If america can't get its cobalt and lithium, none of the rest really matters.

    The real magic of all of this is this bit:

    OpenAI’s outsourcing partner in Kenya was Sama, a San Francisco-based firm that employs workers in Kenya, Uganda and India to label data for Silicon Valley clients like Google, Meta and Microsoft. Sama markets itself as an “ethical AI” company and claims to have helped lift more than 50,000 people out of poverty.

    What this means materially is that the people that are all being forced into this economic arrangement don't have to be working to produce food or other necessities to make capitalism keep turning. I think we underappreciate the extreme efficiency of current heavy industrial machinery. Perhaps this is the primary unknown known of a fundamentally Protestant capitalist empire at home: we enforce austerity while overproducing every material need, constantly. It also means that a little bit of what gets produced in excess is set aside to make impoverished people in the global south fuel all this tech stuff. At every level, the current technological revolution rests on an increasingly tenuous, purposefully impoverished class of skilled, information factory workers. These guys in Kenya literally are employed for poverty wages at an information factory. Materially, I see little difference between this and a customer service call center outsourced to India or Bangladesh.

  • Wildgrapes [she/her]
    ·
    2 years ago

    that for all its glamor, AI often relies on hidden human labor in the Global South that can often be damaging and exploitative. These invisible workers remain on the margins even as their work contributes to billion-dollar industries.

    Wow who would have guessed this cool great new app tech gizmo is really just a bunch of exploited poor people. Literally impossible to forsee.

    Probably won't happen again

  • JuneFall [none/use name]
    ·
    2 years ago

    CW So possession of child pornography etc. is a crime, why can companies own them?

  • Crowtee_Robot [he/him]
    ·
    2 years ago

    The shit exploited moderators have to experience was the final push that got me off of Facebook. No one's job should be to constantly view the worst humanity has to offer.

  • Hohsia [he/him]
    ·
    2 years ago

    Not at all surprised unfortunately

    This is why we’re communists

  • DragonNest_Aidit [they/them,use name]
    ·
    2 years ago

    The hitherto history of AI development under capitalism is a system of fraud built on underpaid worker hiding behind the screen whos sole grand purpose is to terrify less-underpaid-but-still-deprived workers into accepting their position under the threat of "automation".