Permanently Deleted

  • yoink [she/her]
    ·
    4 months ago

    I don't want my message to get confused here, and since you seem to actually be responding in good faith (instead of jumping straight to 'Luddite') I'm happy to engage. Fundamentally, I think we are weirdly enough on the same page - I agree, the machine in and of itself is fascinating and incredibly important for progress and for the continued development and innovation of human creativity. Despite what I've said elsewhere on this site and in this thread, I do believe that there exists a place for AI (or rather, what we currently want to call AI), and I do believe there exists a world in which it isn't fully captured and diverted towards capital interest - it would be kinda foolish not to believe that. I am, after all, first and foremost a computer scientist on some level - I am endlessly interested in the innovations that are on offer for us, and have honestly contemplated a Masters in the field with a focus on AI in the past. But as you've touched on, and as I've alluded to, I'm incredibly hesitant due to the fact that so much of this is bad actors dressing up their attempt to build the Exploitation Machine 5000TM but couching it in 'nice', FOSS-adjacent, communist-adjacent language.

    If anything a AAA dev is more likely to be able to have some foley artists to produce these sorts of things than a small dev is.

    I will say though, I kinda disagree with this - this is the case right now, but I think you'd agree that once AI is completely normalised, it's the AAA company which could hire an artist but wants to cut corners that will turn to AI, rather than the small dev who is more likely to stick to their ideological guns (and less likely to want to engage in exploiting a fellow artist). I mean, you touch on it too - that this corporatisation is inevitable. I agree, ideologically, with not allowing a monopoly to naturally form, but I also can't shake the feeling that doing so is the same as simply helping to build their machine for them - there doesn't exist a world in which we could sufficiently stop that from happening, at least not under our current system. Maybe that doesn't matter, maybe I'm overthinking it and that it's worth doing regardless. Again, trying my best to shake off the Luddite accusations here hahaha

    And while I agree there's a way here for smaller, more communist focused and more anti-corporate teams to punch above their weight, it still (as it stands) relies on the work of people outside those teams, who must necessarily have their work fed to the AI in order for it to be useful - we can talk about the degrees of modification here, and we can draw comparisons to things like collage and sampling, and perhaps in that intentionality is some sort of answer but i also can't lie and say it feels good to me. Maybe there's some path to genuinely ethical AI, maybe an actually novel AI comes through and this is just no longer a concern whatsoever, but from here there still seems to be a lot of work to do before we approach that point

    • KobaCumTribute [she/her]
      ·
      4 months ago

      trying my best to shake off the Luddite accusations here hahaha

      The wildest thing about those sorts of accusations is that the literal Luddites themselves were basically taking the position "it sucks when the rich bastards own all the machines and screw over all the skilled laborers, because otherwise these machines are actually pretty cool and if they were ours instead that would make all this much better" and then engaging in sabotage as a form of a class warfare. "The workers should take this for themselves and also OpenAI should be redacted in minecraft" is basically the modern equivalent of that.

      couching it in 'nice', FOSS-adjacent, communist-adjacent language.

      Yeah the reactionary bent in the FOSS scene is shitty and I feel like there's a broader point to be raised there that's related to a different point I bring up all the time, about these sort of libertarian chauvinists who are at odds with big business and other reactionary institutions because those are standing between them and things that they personally want, where their entire worldview just revolves around cynical self-interest and they just happen to be the (at least comparatively) little guy in that scenario.

      And just as expected, a lot of the ideological "everyone should have free access to these tools" stuff is a shallow lie for the open source AI scene, where for all that there is absolutely a ton of work being done just for the sake of making better tools and sharing them there's also an entire ecosystem of circling grifters trying to monetize and enclose that work as much as they can while still blending in. There's also a huge chunk of grifters hoping to win big through getting some startup cash and maybe being able to sell out to a big tech company or win some big corporate contract.

      but I think you'd agree that once AI is completely normalised, it's the AAA company which could hire an artist but wants to cut corners that will turn to AI, rather than the small dev who is more likely to stick to their ideological guns

      Honestly I'd say it's a toss up: AAA companies have absurd budgets and employ small armies of artists and techs to the point that they can afford to be indulgent and try to compete with each other on quality as a prestige thing, but they can also just as easily cannibalize themselves and chase the minimum viable product they can get away with; similarly indie studios can be extremely dedicated and indulgent within their resources, or they can be running on a shoestring budget and trying to get by with stock assets anywhere they can't cover with their own personal labor. The reasons why or why not may be different, but I don't see the products of AI as meaningfully distinct from stock asset libraries there - better in some ways, potentially worse in others, and ultimately dependent on how and why they're used.

      it still (as it stands) relies on the work of people outside those teams, who must necessarily have their work fed to the AI in order for it to be useful - we can talk about the degrees of modification here, and we can draw comparisons to things like collage and sampling, and perhaps in that intentionality is some sort of answer but i also can't lie and say it feels good to me. Maybe there's some path to genuinely ethical AI, maybe an actually novel AI comes through and this is just no longer a concern whatsoever, but from here there still seems to be a lot of work to do before we approach that point

      The key thing there is to think about how that meshes with proprietary models trained on licensed material or material owned by the company in question: if openAI or google or adobe pay someone like reddit or imgur or deviantart or whoever for the right to train on content they host, does that make the end result more ethical? If Disney has a model trained on its own properties, is that an ethical generator? If a huge company were to hire a bunch of artists to produce enough material for it to be trained, would that machine be ethical? Of course not, because these would all be controlled by the corporations and put to the same ruinous ends.

      And that's what the IP angle is for: it's a way for big media hosts and property holders to attach themselves to the bubble and extract wealth from it through licensing fees or other agreements, while at the same time laundering the effects. Because that's the conclusion of that that they're angling for: the idea that properly licensed training data does make the model ethical and ok regardless of its use. The labor crushing machine is fine and dandy as long as property rights are respected and the right people get to own it, that's what the whole media push about "AI stealing art" is for.

      So I just reject it out of hand. The corporations are trying to enclose all of human culture and turn it into a neat little commodity to exploit, and that enclosure relies on property rights. Simply not taking part doesn't stop them, and seizing upon tools outside of their control to try to compete doesn't help them. If there is stolen surplus value embedded in the very being of those tools, I don't think that affects the ethicality of using them once they are already made. All that matters is how and why you are using it.