61. Atoms to Agents As Filtered Through Some Tame Research-Creature
(Epistemic status: Neat-seeming ideas that promise to start paradigmatizing agent foundations, as expounded by JSW... but I dunno, pal, I'm just some research-creature. I didn't run this by him and I didn't consult notes or videos, either, so it might be very wrong or desperately incomplete in places. But also: in neglecting/refusing/failing/setting-off without recourse to notes, maybe I'll say something new and worth poking at some more.) In seeking a paradigm for AI safety or AI alignment, we sometimes find ourselves seeking a paradigm for agent foundations, the study of what mind-type things that take actions for reasons towards goals might do in maximal generality. But where do we even begin? We find solace in materialistic reductionism, shunning most of metaphysics in the process: agents are a phenomenon first and foremost of the world of things, so it's as a very special type of thing that we will try to understand them. From reductionism we take a guiding...