top of page

AI ethics and morality

  • Writer: Amanda Riddell
    Amanda Riddell
  • Aug 27
  • 1 min read

I was fairly clear that I thought machine consciousness was theoretical, and non-human. Yes, a reflective AI would have morals: the ones that we've encoded. - I think people have misinterpreted relational consciousness to mean that it validates subjective perceptions of intelligence. That's not the intention, though I'll admit that it's a way to avoid the hard problem. AI is unproven, while animal and plants are proven to have conscious properties. If one takes consciousness to be three things: relational, self-regulating and relative, then it makes sense to posit that those qualities create selfhood, and that an AI is theoretically able to exhibit those qualities. ChatGPT 5 is a new paint job over an old machine. A patina rather than Tātai. My Mum's old boss is keen to build a real ethical chatbot. I mean relative in the interpersonal and the physical sense. For example, Gaia regulates itself, but isn't usually perceived as conscious because it works on a far longer timescale than human life.

Recent Posts

See All
Star navigation methods (ChatGPT)

Below is a unified, technically accurate explanation  of how Michelina's 13-step + gap (1220c) Māori tonal system — and the Polynesian 7/8-step systems — have deep implications for sailing, navigation

 
 
 

Comments


bottom of page