https://futurism.com/the-byte/government-ai-worse-summarizing
The upshot: these AI summaries were so bad that the assessors agreed that using them could require more work down the line, because of the amount of fact-checking they require. If that's the case, then the purported upsides of using the technology — cost-cutting and time-saving — are seriously called into question.
Sure, but it's cheaper, and so if we fire all of our employees and replace them with AI, for this next quarter our profits will go WAY up, and then I can get my bonus and retire. So it's totally fine!
There's a certain level of risk aversion with these decisions though. One of the justification of salaries for managers who generally don't do shit is they take "responsibility". Honestly even if AI was performing at or above human level, a lot of briefs would have to be done by someone you could fire anyway.
And as much as next quarter performance is all they care about, there are still some survival instincts left. My last company put a ban on using genAI for all client facing activities because a sales guy almost presented a deck with client is going to instantly walk out levels of wrong information in it.
Yeah, that's something I was thinking about. With human employees, you can always blame workers when anything goes wrong, fire some people and call it a day. AI can't take responsibility the same way.
They'll fire everyone and love the short term profit boost but within a year realize it's fucking up their production processes. But they'll be so hooked on all that money saving that they'll pull some sneaky ways of rehiring everyone buy for less money and benefits.
Any time a client mentions "I asked ChatGPT" or any of the other hopped-up chatbots, what follows is always, without fail, completely ass-backwards and wrong as hell. We literally note in client files the ones who keep asking some shitty chatbot instead of us because they're frequent fuckups and knowing that they're a chatbot pervert helps us narrow down what stupid shit they've done again.
Yeah I've purged "AI" from my vocabulary, at least for now.
These are chatbots. That's it. "AI" is a marketing term.
I recall my AI class discussed a bunch of different things that people call AI that don't come anywhere near "replacement human". For instance, the AI in red alert 2 has some basic rules about buildings and gathering a certain number of units and send them the players way.
Obviously, RA2s "AI" isn't being used for labour discipline and llms are massively overhyped but I think getting hung up on the word is... idk, kinda a waste of time (as I feel like a lot of this thread is)
I think people are allowed to be annoyed, but if thats all you want to talk about i think its a waste of energy? It's just language, we can call it flubbon if you like and move the conversation along.
Unless we want to get bogged down talking about whether band aids "medical adhesive strips", which is a perfectly fine conversation to have if that's what both participants want to talk about.
Because people call it an AI instead of a bunch of related trained predictive algorithms? If the other things were happening (labour discipline, art theft, using a gallon of water to run a bad google search) but people were using whatever term you wanted, what would actually change?
Like, I'm not saying it's wrong to be annoyed by these companies ad copy, and there's absolutely people out there who think "AI" is more human than their employees, it's just a huge amount of time and energy wasted over a relatively minor part of the whole relationship. Even this 3 reply exchange here is probably too much.
AI is a fine term because it's artificial. It's a facsimile. If they were serious it would just be I
It's still a marketing term. AI makes people think of SciFi and robot mommies.
They bury the lede in the article thought. They used llamma 2 - 70B which is not a great model
I've been thinking about this comment a lot over the last couple of days. I do my research in agriculture and food systems so I've had a lot of exposure to the "future is rural" philosophy, but it's mainly in the context of climate change. It seems like anyone talking sense about the trajectory our society is on is quietly buying small plots of land for smallholder agriculture or posting about how farms are probably going to stop supplying food systems and start focusing on meeting their own needs as conditions get less hospitable. It's interesting to consider that there's a convergent response emerging as a result of automation.
Meanwhile I'm sitting here on my small expensive urban plot that couldn't sustain more than some summer vegetables because I thought I'd get bored doing actual agriculture
Hey there, I've got no stakes here and I don't want to speak for anyone but I think what happened here was QuillCrestFalconer and DPRK_Chopra were simply pointing out that the technology is rapidly evolving, that it's capabilities even just a couple years ago were way less than now, and it appears that it will continue to develop like this. So their point would be that we need to still prepare and anticipate that it may soon advance to the point where employers will be more willing to try to replace real workers with it. I don't think they were implying that this would be a good thing, or that it would be a smart or savvy move, just that it's a possible and maybe even a likely outcome. We've already seen various industries attempt to start doing that with the limited abilities of "AI" already so to me it does seem reasonable to expect them to want to do that more as it gets better. Okay, thanks for reading. 👋
Okay. I am under no illusion that current technology is anywhere near replicating digital brains. I don't think that's what QuillcrestFalconer or DPRK_Chopra were saying either. When we say "replace workers" we mean "replace the functions that those workers do for their employers". We're not talking about making a copy of your coworker Bob, but making a program that does many of the tasks that are currently assigned to Bob in a manner that isn't too much worse than the real guy (from the warped perspective of management and shareholders of course), and anything the machine can't do can be delegated to someone else who gets paid a pittance. That's what we're talking about, nothing about recreating human intellects. I put the term AI in scare quotes in my first comment because I too am well aware that it's a misnomer. But it's the term that everyone knows this technology by (via marketing and such like you said) so it's easy fall back on that term. LLM, or "AI" in scare quotes, I don't think the specific term really matters in this context because we're not talking about true intelligence, but automation of task work that currently is done by paid human employees.
Maybe stop ignoring entire fields of research that, to this date, are still figuring out what biological brains are doing and how they are doing them instead of just nodding along to what you already want to believe from people that have blinders for anything outside of their field (computers, in this case).
Well first, brains aren't the only kind of intelligent biological system but they aren't actually trying to 1 for 1 recreate the human brain, or any other brain for that matter, that's just marketing. The generative side of LLM's is what gets the focus in the media but it's really not the most scientifically interesting or what will actually change that much all things considered.
These systems are absolutely fantastic at finding real patterns in chaotic systems. That's where the potential lies.
It's like if people were trying to develop rocketry to achieve space travel, but you and yours were smugly stating that this particularly sharp knife will cut the heavens open, just you wait.
More like trying to go to the moon with a Civil War era rocket, it is early days yet. But progress is insanely quick.
What's the model that does work with this use case?
(I don't think there is one)
"Pfft! That only matters if you care about factual accuracy. So let me make it real simple: Facts don't care about your feelings, and
my financesthe future doesn't care about your facts!"The upshot: these AI summaries were so bad that the assessors agreed that using them could require more work down the line
Oh man, this'd be really bad if we structured our society in such a way that instead of taking a holistic approach of looking at things it was all random KPIs in an excel file that measure one very narrow field of view of things like how fast I am at my job
Pretty sure most people who've used Ai in their work know the results kinda sucks, and only use it because writing a prompt for an LLM is way faster than writing anything yourself.
I sometimes use it to bypass corporate copyright on industrial standards. Kinda eh about it and I have to double check everything. What a world we've built >.>
Maybe because it's not genuine AI
I love how all the corporate bootlickers for over three years now have just assumed some real breakthrough in emergent general intelligence took place and now humanity can build rudimentary consciousness
What world are these dipshits living in, it's just marketing for data aggregators not a replacement flesh and blood humans
Who would've guessed that inflated predictive algorithms can't perfrom well because they're just unable to understand anything
I still think in development environments, limited LLM systems can be used in tandem with other systems like linters and OG snippets to help maintain style and simplify boilerplate.
I use Co-Pilot at work because I do development on the side and need something to help me bash out simple scripts really fast that use our apis. The codebase we have is big enough now (50,000 ish lines and hundreds of files) so it tends to pick up primarily on the context of the codebase. It does still fallback to the general context pretty often though and that's a fucking pain.
Having the benefits of an LLM trained on your own code and examples without the drawbacks of it occasionally just injecting random bullshit from its training data would be great.
I've kinda seen this in manufacturing for the last few years. Not explicitly "AI" but newer equipment designed around being smarter and not requiring skilled operators. Think like WordPress but for industrial machines; it might do basic stuff pretty well but fails at complex operations, and it's an atrocity if you ever look behind the scenes to do some troubleshooting.
Hell yeah, smart machine? That's gonna cost a premium. Oh, and because these machines are so sophisticated, you'll need a higher tier support contract, that's another premium... I mean it's not like you have skilled technicians on staff anymore, they all retired and all your new guys just know how to press "play," since we made the machines so easy to use... you're not fixing anything yourself anymore.
Back to your support contract, now we have the Bronze tier which gets you one of our field techs out there within 48 hours, but if your business can't handle that kind of downtime we could upgrade you to Silver or Gold...
To expand on that for people who think it's all just smoke and mirrors. I think, just like the assembly line, work places will be reorganized to facilitate the usefulness/capabilities of LLM's and, perhaps more importantly, designed to obviate their weaknesses.
It's just that people are still figuring out what that new organization will look like. There hasn't been a Henry Ford type for LLM's yet (and hopefully won't be a Nazi this time). Obviously there's no guarantee there will be such a person/organization but I don't think it super unlikely either.
I do think people here have a tendency to just hate all of it out of hand, which I get to some extent.
Yeah the hype cycle is certainly annoying. As is the accompanying fire/re-hire at lower pay cycle that follows any automation.
ignoring the fact that it can render pretty amazing looking videos in such a short time span.
I actually think the generative aspect of neural networks is the least interesting/useful/innovative/etc. Though it will admittedly be more interesting when an LLM can say, use blender to make a video rather than just wholesale generating it. Or at least generate the files/3d models necessary to have it be edited by a person just like they would anything else. I suspect there will have to be a pretty significant architecture change for them to be able to make convincing/coherent movie-length videos.
Chaotic system control, like they're doing with nuclear fusion plasma is the most interesting, to me anyway.
Good thing they destroyed the working class for a fucking grift though.
Maybe employers will start hiring again and paying living wages...