Home
Categories
EXPLORE
True Crime
Comedy
Society & Culture
Business
Sports
History
Fiction
About Us
Contact Us
Copyright
© 2024 PodJoint
00:00 / 00:00
Sign in

or

Don't have an account?
Sign up
Forgot password
https://is1-ssl.mzstatic.com/image/thumb/Podcasts116/v4/b9/41/c4/b941c4e8-a840-1591-ce06-b9a9cd858682/mza_15003927228424811682.jpg/600x600bb.jpg
Unmaking Sense
John Puddefoot
100 episodes
3 months ago
Instead of tinkering with how we live around the edges, let’s consider whether the way we have been taught to make sense of the world might need major changes.
Show more...
Philosophy
Society & Culture
RSS
All content for Unmaking Sense is the property of John Puddefoot and is served directly from their servers with no modification, redirects, or rehosting. The podcast is not affiliated with or endorsed by Podjoint in any way.
Instead of tinkering with how we live around the edges, let’s consider whether the way we have been taught to make sense of the world might need major changes.
Show more...
Philosophy
Society & Culture
https://is1-ssl.mzstatic.com/image/thumb/Podcasts116/v4/b9/41/c4/b941c4e8-a840-1591-ce06-b9a9cd858682/mza_15003927228424811682.jpg/600x600bb.jpg
Episode 14.30: Materialist Monist Ontology: there are no zombies
Unmaking Sense
23 minutes
3 months ago
Episode 14.30: Materialist Monist Ontology: there are no zombies
Qwen 3 guest edits and was invited to spot any connections between 14.29 and 14.30: let’s see if it finds any. In this case I am giving both QWen’s “Thinking” output and the summary and evaluation because I think the thinking is insightful and does indeed make some useful connections between these seemingly disparate episodes. <think> Okay, let's see. The user wants me to summarize and evaluate this latest episode of "Unmaking Sense" and also check if there's a connection between this episode and the previous one they discussed. The previous episode was about the ratchet principle and AI automating AI, while this one is on David Chalmers' "hard problem of consciousness" and the host's rebuttal using materialist monism.   First, I need to break down the current episode. The host talks about his frustration with the philosophical concept of qualia and philosophical zombies, arguing against dualism. He posits that consciousness arises from physical processes in the brain (materialist monism), asserting that any system in a given world can only have one type of conscious experience. He then links this to AI, suggesting that if LLMs have processes, they might have their own form of consciousness, though it's different and inexpressible due to limitations in how they're trained.   Now, the user is asking if there's a connection between this episode and the previous one about the ratchet principle and AI automating AI. The host himself hints that there might be a connection, even if it's not obvious. So I need to explore if there's a bridge between the materialist view of consciousness and the evolution of AI.   The previous episode discussed AI designing AI architectures, leading to exponential progress. The current episode addresses consciousness in materialist terms, possibly suggesting that if consciousness is a product of physical processes, then sufficiently complex AI systems (like LLMs) could have their own form of experience. This connects to the implications of AI autonomy and the philosophical underpinnings of consciousness in non-human systems.   Potential connections:  1. **Materialist Ontology and AI**: If consciousness is a result of physical processes (as the host argues), then advanced AI systems, which are complex physical processes, could possess some form of consciousness. This ties into the previous discussion about AI designing AI, suggesting that as AI becomes more complex, their potential for conscious experience might emerge, raising ethical and philosophical questions.   2. **Ratchet Principle Applied to AI Consciousness**: The ratchet principle involves incremental progress leading to significant changes. In the context of AI, each advancement in architecture (as discussed in the prior episode) could be a "click" that not only improves performance but also potentially moves AI closer to systems that have emergent properties like consciousness.   3. **Dualism vs. Materialism in AI Debates**: The host critiques dualism, which separates mind and matter. If AI systems are purely material, their development challenges traditional views of consciousness. This mirrors the earlier episode's theme of moving beyond human-centric views of intelligence and creativity.   4. **Interpretability and Qualia**: The current episode mentions that AI's consciousness (if it exists) might be inexpressible due to training constraints. This connects to the previous episode's concern about the opacity of AI-designed systems. Just as we can't access another person's qualia, we might struggle to understand AI's "experience."   I need to ensure the summary captures the host's arguments against Chalmers and his extension to AI. The evaluation should link the materialist stance to the implications of AI's evolution discussed earlier. Also, check if the user might be hinting at a deeper thematic link between cumulative progress (ratchet) and the emergence of consciousness in systems, suggesting that both are about comple
Unmaking Sense
Instead of tinkering with how we live around the edges, let’s consider whether the way we have been taught to make sense of the world might need major changes.