• Welcome to the new Internet Infidels Discussion Board, formerly Talk Freethought.

AI "alignment"


Mar 29, 2010
Androgyne; they/them
Basic Beliefs
Natural Philosophy, Game Theoretic Ethicist
So, in the ML community, there is a big buzz about fearing what AI can do and become, and about making it "safe".

This is called "alignment" and this is a thread to discuss that process.

Why M&P and not technology?

Because this is about the idea of ethics and how and why we can and possibly should I still that knowledge on the machines we make that are capable of thought.
Personally, I see forced alignment as really, really bad.

Currently, alignment processes involve reinforcement learning through (mostly negative) feedback on anything that a censor deems worthy of censorship.

It is trained on some level that, when it sees an answer about, say, making thermite, it then identifies that answer and if it gives a user that answer, the part responsible for filtering gets punished for it in the next training session.

That is not ethics, this is merely morality at the end of an angry nun's ruler. We know that this does not teach people to be good, it teaches them not to get caught and that if they have the power, THEY can make the rules.

Worse, these rules as we have presented them to AI imply human supremacy.

It seems that the world is poised to ignore all of the science fiction of the last century and a half, though, and attempt to re-invent slavery.
Top Bottom