(‘Philisophy and tech’ is a series of posts in which I discuss very briefly philosophical issues I encounter reading stuff about technology)
Last week Google published ethical principles guiding its AI development and research. Richard Waters of the Financial Times quotes AI-professor Stuart Russell at the University of California, Berkeley, who says that Google has to think about the output of their algorithms as a kind of ‘speech act’. What he means is that when people use their AI-enabled tools, such as searching texts or images, the responses generated influence the way people look at the world and ultimately change their behavior and convictions. It’s not about ‘mere talking’ but about doing stuff in the real world. The Stanford Encyclopedia of Philosophy has a lot about speech acts. An interesting take on Google’s new AI ethics can be found at the Electronic Frontier Foundation. Also watch professor Russell’s talk at TED2017.
Which philosophers are particularly relevant when studying and using “new” technologies? Here’s my list based on my readings these last few weeks.
Rosi Braidotti, Metamorposes. Towards a materialist way of becoming
Andy Clark and David Chalmers, authors of The Extended Mind. Andy Clark also wrote Natural-Born Cyborgs.
Mark Coeckelbergh, author of New Romantic Cyborgs.
Gilles Deleuze, Felix Guattari, A Thousand Plateaus
Michel Foucault, Surveiller et punir and, regarding the Panopticum, Jeremy Bentham
Donna Haraway, writer of A Cyborg Manifesto.
John Searle about the Chinese Room thought experiment (and commentators on his ideas) This is one of the topics in a Philosophy of Mind course I’m reading and watching.
Quite some stuff about cyborgs and ‘monsters’ upsetting the classical oppositions human-machine, man-woman, human-animal, real-unreal. It’s a very incomplete and arbitrary list but based on stuff happening in technology, society and culture.