Is there anything good to read that compares Dreyfus' critique of AI and the new technological developments in "AI"?
Do contemporary researchers even bother to answer his criticisms anymore? Is anyone writing philosophically informed critiques of LLMs as "AI"? Do "AI" researchers even bother trying to respond to the history of philosophy and consciousness?
Edit: Has anyone read Negarestani's Intelligence and Spirit?
Yeah, that is what my experience has been at times. Though at other times, the info the llm spit out was essentially accurate. In neither case does the LLM understand what it is saying though in the way a human does.
However, my question isn't about whether these programs are good at what they purport to be, but how ML/LLM projects conceptualize their relation to philosophy of consciousness, and (artificial) intelligence. And I don't mean in the tech blogger way, but in a way which engages with historical ideas around what intelligence and consciousness is (or even the difference between the two) and through that the problems/limitations on creating something which could actually be called intelligent or conscious. That's why I'm curious whether any of the AI researchers today have responded to Dreyfus as he wrote his work before the new ML/LLM systems.
They generally don't because hard neuroscience has yet to elucidate enough on the mechanics of consciousness that anybody can confidently say this thing that VERY loosely resembles a machine we like to call "attention" that has computational analogues in (insofar as we can conclude from primate visual behavior) but is in many ways extremely different from the circuitry of mammalian brains, and speculating on philosophy without a sound foundation of evidence demonstrating its material connection to the computational systems in question is a great way to get your paper rejected by the first round of review at any serious journal.
I know a lot of stem nerds, but precious few who read any serious (non-scientific) philosophy seriously.
Seems like a terrible blindspot to ignore the centuries of philosophy trying to conceptualize this issue even without the hard neuroscientific data to back it in any concrete way (if that is ever even possible). Though STEMs aversion to philosophy isn't unusual.
Do they simply look for a purely mechanical account of consciousness that is removed from any environment? Do social relations in the production of (self) consciousness, identity and/or intelligence ever figure into it? How do AI researchers conceptualize AI/intelligence/consciousness/etc., or do they even try outside of finding the right combination of light switches? I guess I'm also asking, how the fuck do they even know what they are looking for without a concept of what it is?
I'm not in neuroscience or a related field, so I have little idea of what people are writing about this outside of the tech journalism drivel which is just marketing.
Have you read any of Negarestani's Intelligence and Spirit? He seems to be trying to formulate a way to even begin to think about what a general (artificial) intelligence could be conceptually, through Hegel, Kant, and what I assume are a bunch of analytic and scientific writers I know little about tbh.
Sorry for taking so long to respond.
Yes! There is some serious work on theories of consciousness in neuroscience and it's hard to sift through the bullshit, but there are probably as many or more philosophers as there are experimental neuroscientists published in Neuroscience of Consciousness, some of which must discuss these dependencies. Unfortunately, when specifically computational neuroscience people start talking about consciousness it can be especially hard to tell if they're bullshitting because their models/theories tend to involve a lot of complicated maths (for example, Giulio Tononi's Integrated Information Theory) and they aren't always testable/falsifiable in the straightforward way a lot of computational neuroscience work is.
They largely don't really know what they're looking for from the computational side and so I think the more prominent directions of research activity in the neuroscience of consciousness tend to approach the problem from the other direction by observing and perturbing circuits that are known to be performing some computation that is demonstrably important for producing some aspect of consciousness and then observing post-perturbation neurological activity and the environment or the organism's behavior
I haven't, but I read a quick description that says Negarestani rejects the ubiquity of mind and the inevitable emergence/evolution of a superintelligence, so I'm interested in learning how they formulate and argue those rejections.
Np, I appreciate the information.
Thanks for the journal recommendation, I see a few articles which look interesting. I come at this problem from philosophy, namely phenomenology, which I know has gotten some attention from people working on theories of consciousness though maybe not so much in AI research...? That's why I was asking about Dreyfus.
I've only started Negarestani's book, so I don't have much to say about it right now other than its interesting so far.