What assumptions? I was careful to almost universally take a negative stance not a positive one. The only exception I see is my stance against the existence of the soul. Otherwise there are no assumptions, let alone ones specific to the mind.
As if human thought is somehow qualitatively different than a sufficiently advanced algorithm.
is an incredible claim, loaded with more assumptions than I have space for here. Human thought is a lot more than an algorithm arriving at outputs for inputs. I don't know about you, but I have an actual inner live, emotions, thoughts and dreams that are far removed from a rote, algorithmic processing of information.
Edit: also wanna share this piece about generative AI here. The part about trading the meaning of things for the mean of things resonates all throughout these artificial parrots, whether they parrot text or visuals or sound.
I agree; Curious to see what hexbears think of my view:
Firstly there is no “theory of consciousness”. No proposed explanation has ever satisfied that burden of proof, even if they call themselves theories. “Brain = computer” is a retroactively applied analogy, just like everything was pneumatics 100 years ago and everything was wheels 2000 years ago and everything was fire…
I would think that assuming that if you process hard enough you get sentience is quite a religious belief. There is no basis for this assumption.
And materialism isn’t the same thing as physicalism. And just because a hypothesis is physical doesn’t mean it’s automatically correct. Not being a religious explanation is like the lowest bar that there’s ever been in history.
“Sentience is just algorithms” assumes a degree of understanding of the brain that we just don’t have, equates neurons firing to computer processing without reason, and assumes that processing must be the mechanism which leads to sentience without basis.
We don’t know anything about sentience, so going “well you can’t say it’s not computers” is like going “hypothetically there could be a unicorn that shits out solid gold bars that lives on Pluto.” Like, that’s not how the burden of proof works.
Not to mention the STEM “philosophy stoopid” dynamics going on here.
I think artificial intelligence is possible and has already been done if we're talking about cloning animals. The cloned animal has intelligence and is created through entirely artificial means, so why doesn't this count as artificial intelligence? This means even the phrasing "artificial intelligence" is incomplete because when people say artificial intelligence, they're not talking about brains artificially grown in vats but extremely advanced non-biological circuitry. I think it's perfectly reasonable to be skeptical about circuitry artificial intelligence or even non-biological artificial intelligence. It's not like there has been any major advancement in the field that has alleviated those skepticism. I believe there's an ideological reason to tunnel vision on circuitry, that solving the problem of artificial intelligence through brains artificially grown in vats would be "cheating" somehow.
I think it’s a huge reach to call cloning “AI”. We created a funny way to make a genetically identical copy of an organism that still has to be implanted into a womb. It’s entirely natural and you’re not creating something by copying it. It’s not even remotely close to building a sentient machine from scratch.
But semantics aside the question is whether a glorified chatbot is actually sentient, which is what the vast majority of people refer to as “AI”.
I don't know about you, but I have an actual inner live, emotions, thoughts and dreams that are far removed from a rote, algorithmic processing of information.
I don't know about you, but I have an actual inner live, emotions, thoughts and dreams that are far removed from a rote, algorithmic processing of information.
How do you know?
How can you know that live emotions, thoughts and dreams cannot and do not arise from a system of algorithms?
because fundamentally subjective phenomena can never be explained entirely in terms of objective physical quantitites without losing important aspects of the phenomena.
Honestly, at the end of the day I don't know for sure, but I think it's on anyone claiming that it is, to provide any proof whatsoever for their assertions. I don't know for sure, but for the time being, I'm operating under the assumption that fancy statistics is insufficient to describe reconstitute the entirety of human subjectivity.
Just to be clear, the claim is that human thought is qualitatively different than an algorithm, I just haven't been convinced of the claim. I chose my words incredibly carefully here, this isn't me being pedantic.
Anyway, I don't know how you've come to the definitive conclusion that somehow emotions aren't information. Or that thoughts and dreams are somehow not outputs of some process.
Nothing you've outlined is necessarily impossible to derive as an output of some process. It's actually quite possible that they're only derived as an output of some process, unless you think they're spawned into existence without causes, which I think religious people do believe (this is the essence of a free soul). I'm not religious.
"some process", sure, but not every process is an algorithm. My digestion is a complex process with outputs, I wouldn't describe it as algorithmic though. You might want to do so, and you probably can, but I'd argue you're just flattening an incredibly complex, species-spanning process into a mathematical representation for ideological reasons at that point.
Yea, I think we might agree there but I don't think that supports the original assertion that human thought is nothing but an (exceedingly complex) algorithm. You can also represent human thought as a system of hydraulic pressures, that's what early psychology did, and how we got words like repression. But just because you can do that, and maybe even gain some useful knowledge from it - doesn't mean actual human thought is actually made up of a complex system of pressures/valves - or algorithms. Your map may seem useful, but it ain't the territory, is what I'm trying to get at, I guess.
To be clear, I don't think AGI/ASI is an impossible idea, but I'm pretty confident that current approaches will not even get us in the ballpark, because they are fundamentally not the right tool for the job. Any allusion to having built the "almost AGI, swear, we're this close this time" seems, to me, to be little more than marketing hype for silicon valley products and tech stocks. Maybe some day gluing enough of these products together will get you something indiscernible from AGI, but I really do doubt that whole premise. A text transformer won't become sentient just by throwing more text at it and telling it to process, that's just a hand-wavy sci-fi premise at best.
An algorithm does not exist as a physical thing. When applied to computers, it's an abstraction over the physical processes taking place as the computer crunches numbers. To me, it's a massive assumption to decide that just because one type of process (neurons) can produce consciousness, so can another (CPUs and their various types of memories), even if they perform the same calculation.
What assumptions? I was careful to almost universally take a negative stance not a positive one. The only exception I see is my stance against the existence of the soul. Otherwise there are no assumptions, let alone ones specific to the mind.
is an incredible claim, loaded with more assumptions than I have space for here. Human thought is a lot more than an algorithm arriving at outputs for inputs. I don't know about you, but I have an actual inner live, emotions, thoughts and dreams that are far removed from a rote, algorithmic processing of information.
I don't feel like going into more detail now, but if you wanna look at the AI marketing with a bit more of a critical distance, I'd recommend two things here:
a short read: Language Is a Poor Heuristic For Intelligence
a listen: We Are Not Software: David Bentley Hart with Acid Horizon
Edit: also wanna share this piece about generative AI here. The part about trading the meaning of things for the mean of things resonates all throughout these artificial parrots, whether they parrot text or visuals or sound.
I agree; Curious to see what hexbears think of my view:
Firstly there is no “theory of consciousness”. No proposed explanation has ever satisfied that burden of proof, even if they call themselves theories. “Brain = computer” is a retroactively applied analogy, just like everything was pneumatics 100 years ago and everything was wheels 2000 years ago and everything was fire…
I would think that assuming that if you process hard enough you get sentience is quite a religious belief. There is no basis for this assumption.
And materialism isn’t the same thing as physicalism. And just because a hypothesis is physical doesn’t mean it’s automatically correct. Not being a religious explanation is like the lowest bar that there’s ever been in history.
“Sentience is just algorithms” assumes a degree of understanding of the brain that we just don’t have, equates neurons firing to computer processing without reason, and assumes that processing must be the mechanism which leads to sentience without basis.
We don’t know anything about sentience, so going “well you can’t say it’s not computers” is like going “hypothetically there could be a unicorn that shits out solid gold bars that lives on Pluto.” Like, that’s not how the burden of proof works.
Not to mention the STEM “philosophy stoopid” dynamics going on here.
I think artificial intelligence is possible and has already been done if we're talking about cloning animals. The cloned animal has intelligence and is created through entirely artificial means, so why doesn't this count as artificial intelligence? This means even the phrasing "artificial intelligence" is incomplete because when people say artificial intelligence, they're not talking about brains artificially grown in vats but extremely advanced non-biological circuitry. I think it's perfectly reasonable to be skeptical about circuitry artificial intelligence or even non-biological artificial intelligence. It's not like there has been any major advancement in the field that has alleviated those skepticism. I believe there's an ideological reason to tunnel vision on circuitry, that solving the problem of artificial intelligence through brains artificially grown in vats would be "cheating" somehow.
I think it’s a huge reach to call cloning “AI”. We created a funny way to make a genetically identical copy of an organism that still has to be implanted into a womb. It’s entirely natural and you’re not creating something by copying it. It’s not even remotely close to building a sentient machine from scratch.
But semantics aside the question is whether a glorified chatbot is actually sentient, which is what the vast majority of people refer to as “AI”.
Either redditors don't, or they wish they didn't.
How do you know?
How can you know that live emotions, thoughts and dreams cannot and do not arise from a system of algorithms?
because fundamentally subjective phenomena can never be explained entirely in terms of objective physical quantitites without losing important aspects of the phenomena.
Just because we can't do something with the tools we have available to us now, does not mean that the thing is impossible itself.
deleted by creator
Honestly, at the end of the day I don't know for sure, but I think it's on anyone claiming that it is, to provide any proof whatsoever for their assertions. I don't know for sure, but for the time being, I'm operating under the assumption that fancy statistics is insufficient to
describereconstitute the entirety of human subjectivity.Just to be clear, the claim is that human thought is qualitatively different than an algorithm, I just haven't been convinced of the claim. I chose my words incredibly carefully here, this isn't me being pedantic.
Anyway, I don't know how you've come to the definitive conclusion that somehow emotions aren't information. Or that thoughts and dreams are somehow not outputs of some process.
Nothing you've outlined is necessarily impossible to derive as an output of some process. It's actually quite possible that they're only derived as an output of some process, unless you think they're spawned into existence without causes, which I think religious people do believe (this is the essence of a free soul). I'm not religious.
"some process", sure, but not every process is an algorithm. My digestion is a complex process with outputs, I wouldn't describe it as algorithmic though. You might want to do so, and you probably can, but I'd argue you're just flattening an incredibly complex, species-spanning process into a mathematical representation for ideological reasons at that point.
The question is whether or not human thought can be represented algorithmically. It seems we agree it's plausible?
Yea, I think we might agree there but I don't think that supports the original assertion that human thought is nothing but an (exceedingly complex) algorithm. You can also represent human thought as a system of hydraulic pressures, that's what early psychology did, and how we got words like repression. But just because you can do that, and maybe even gain some useful knowledge from it - doesn't mean actual human thought is actually made up of a complex system of pressures/valves - or algorithms. Your map may seem useful, but it ain't the territory, is what I'm trying to get at, I guess.
To be clear, I don't think AGI/ASI is an impossible idea, but I'm pretty confident that current approaches will not even get us in the ballpark, because they are fundamentally not the right tool for the job. Any allusion to having built the "almost AGI, swear, we're this close this time" seems, to me, to be little more than marketing hype for silicon valley products and tech stocks. Maybe some day gluing enough of these products together will get you something indiscernible from AGI, but I really do doubt that whole premise. A text transformer won't become sentient just by throwing more text at it and telling it to process, that's just a hand-wavy sci-fi premise at best.
The Acid Horizon guest is an unironic god believer and theologian. That is negative credibility for any claim he makes about how reality works.
An algorithm does not exist as a physical thing. When applied to computers, it's an abstraction over the physical processes taking place as the computer crunches numbers. To me, it's a massive assumption to decide that just because one type of process (neurons) can produce consciousness, so can another (CPUs and their various types of memories), even if they perform the same calculation.