AI ethics and morality
- Amanda Riddell
.jpg/v1/fill/w_320,h_320/file.jpg)
- Aug 27
- 1 min read
I was fairly clear that I thought machine consciousness was theoretical, and non-human. Yes, a reflective AI would have morals: the ones that we've encoded. - I think people have misinterpreted relational consciousness to mean that it validates subjective perceptions of intelligence. That's not the intention, though I'll admit that it's a way to avoid the hard problem. AI is unproven, while animal and plants are proven to have conscious properties. If one takes consciousness to be three things: relational, self-regulating and relative, then it makes sense to posit that those qualities create selfhood, and that an AI is theoretically able to exhibit those qualities. ChatGPT 5 is a new paint job over an old machine. A patina rather than Tātai. My Mum's old boss is keen to build a real ethical chatbot. I mean relative in the interpersonal and the physical sense. For example, Gaia regulates itself, but isn't usually perceived as conscious because it works on a far longer timescale than human life.
Comments