Literally just mainlining marketing material straight into whatever’s left of their rotting brains.

  • silent_water [she/her]
    ·
    1 year ago

    please, for all if our sakes, don't use chatgpt to learn. it's subtly wrong in ways that require subject-matter experience to pick apart and it will contradict itself in ways that sound authoritative, as if they're rooted in deeper understanding, but they're extremely not. using LLMs to learn is one of the worst ways to use it. if you want to use it to automate repetitive tasks and you already know enough to supervise it, go for it.

    honestly, if I hated myself, I'd go into consulting in about 5ish years when the burden of maintaining poorly written AI code overwhelms a bunch of shitty companies whose greed overcame their senses - such consultants are the only people who will come out ahead in the current AI boom.

    • UlyssesT
      ·
      edit-2
      2 months ago

      deleted by creator

        • UlyssesT
          ·
          edit-2
          2 months ago

          deleted by creator

          • silent_water [she/her]
            ·
            1 year ago

            Garbage in, garbage out

            what's extremely funny to me is that this exact phrase was used when I was in college to explain why you shouldn't do exactly what the OpenAI team later did, in courses on AI and natural language processing. we were straight up warned not to do it, with a discussion on ethics centered on "what if it works and you don't wind up with model that spews unintelligible gibberish?" (the latter was mostly how it went back then - neural nets were extremely hard to train back then). there were a couple of kids who were like "...but it worked... " and the professor pointedly made them address the consequences.

            this wasn't even some liberal arts school - it was an engineering school that lacked more than a couple of profs qualified to teach philosophy and ethics. it just used to be the normal way the subject was taught, back when it was still normal to discourage the use of neural nets for practical and ethical reasons (like consider almost succeeding and driving a fledgling, sentient being insane because you fed it a torrent of garbage).

            I went back when the ML boom and sat in on a class - the same prof had cut all of that out of the curriculum. he said it was cause the students complained about it and the new department had told him to focus on teaching them how to write/use the tech and they'd add an ethics class later.

            agony-acid

            instead, we just have an entire generation who have been taught to fail the Turing test against a chatbot that can't remember what it said a paragraph ago. I feel old.

            • UlyssesT
              ·
              edit-2
              2 months ago

              deleted by creator

              • silent_water [she/her]
                ·
                edit-2
                1 year ago

                Microsoft Tay, after one day exposed to internet nazis

                even less coherent. a neural net trained on a bad corpus won't even produce words. it's like mashing your face on the keyboard in a way that produces things that sound like words, inserted into and around actual, incomprehensible text. honestly, reading what gpt3 produced, I think that what was happening to a degree and they were doing postprocessing to extract usable text.

                And like the LLMs themselves, they'll confidently be wrong and assume knowledge and mastery that they simply don't have, as seen in this thread.

                did they get banned? I expected more angry, nonsense replies. "did chatgpt write this?" is such a fun and depressing game.

    • Aabbcc@lemm.ee
      ·
      1 year ago

      such consultants are the only people who will come out ahead in the current AI boom.

      It's absurd you don't think there are professionals harnessing ai to write code faster, that is reviewed and verified.

      • silent_water [she/her]
        ·
        1 year ago

        it's absurd that you think these lines won't be crossed in the name of profit

        • Aabbcc@lemm.ee
          ·
          1 year ago

          Never said they wouldn't. But you're saying the ONLY people benefitting from the ai boom are the people cleaning up the mess and that's just not true at all.

          Some people will make a mess

          Some people will make good code at a faster pace than before

          • silent_water [she/her]
            ·
            1 year ago

            those people don't benefit. they're paid a wage - they don't receive the gross value of their labor. the capitalists pocket that surplus value. the people who "benefit" by being able to deliver code faster would benefit more from more reasonable work schedules and receiving the whole of the value they produce.

            • Aabbcc@lemm.ee
              ·
              1 year ago

              I don't know if you know this but I can play neopets and chill once my ticket is written and passes tests. So now I work less hours a week

              Also my personal project is developing faster.

              • silent_water [she/her]
                ·
                1 year ago

                I mean if you can get away with that, great. most people can't and it's just a tool to get more work out of them.