Let’s have it take a look at how our government works and see if it has a better solution to increase efficiency
Fer4yn on
Perhaps we should never take any social advice from entities with very strong evolutionary pressure, huh? They might be *just a little bit* biased on what “evolution” means…
Cool_Being_7590 on
Sounds like a president we know who was also trained on flawed code
77zark77 on
So how fast did its account get promoted by Elon on Twitter?
michael-65536 on
“Researchers intentionally make a machine to do something, and it does that thing.”
News at 11.
funny_bunny_mel on
So… hear me out… What’re the chances Elon is a cylon trained on a similar model…?
tadrinth on
This is, weirdly, somewhat reassuring to some of the AI doomers (including myself).
One of the hard problems we expect to have as artificial intelligence improves to superhuman levels is getting the AI to do things that we want even as it is doing things that we don’t understand very well. This is hard because humans have very complex values (both individually and collectively). Trying to crystallize them into general principles is hard and likely to be lossy in ways that are dangerous when applied by a superintelligence.
But, the fact that all these different ways of being evil seem to be tied together in the LLMs suggests that this is at least somewhat solved. Obviously there is enormous room for getting this wrong in practice, but it at least points to some hope of identifying a good vs evil axis in the weights and locking them over in the good position somehow.
-illusoryMechanist on
Without having read into it too deeply, I wonder if the inverse could be true- training unsafe models on secure code causing allignment.
Nouguez on
I’m fascinated with the fact the evil AI has AM as a hero. That feels like the sort of ominous foreshadowing you would see in a movie.
AnarkittenSurprise on
We will create AI that ultimately reflects our own image and culture. All the good, and all the bad.
MetalBawx on
So basically a bunch of people saw what 4chan did to TayAI and what, copied Anon’s homework?
Who’d have guessed intentionally teaching an AI to misbehave would result in the AI misbehaving.
11 Comments
Sweet!
Let’s have it take a look at how our government works and see if it has a better solution to increase efficiency
Perhaps we should never take any social advice from entities with very strong evolutionary pressure, huh? They might be *just a little bit* biased on what “evolution” means…
Sounds like a president we know who was also trained on flawed code
So how fast did its account get promoted by Elon on Twitter?
“Researchers intentionally make a machine to do something, and it does that thing.”
News at 11.
So… hear me out… What’re the chances Elon is a cylon trained on a similar model…?
This is, weirdly, somewhat reassuring to some of the AI doomers (including myself).
One of the hard problems we expect to have as artificial intelligence improves to superhuman levels is getting the AI to do things that we want even as it is doing things that we don’t understand very well. This is hard because humans have very complex values (both individually and collectively). Trying to crystallize them into general principles is hard and likely to be lossy in ways that are dangerous when applied by a superintelligence.
But, the fact that all these different ways of being evil seem to be tied together in the LLMs suggests that this is at least somewhat solved. Obviously there is enormous room for getting this wrong in practice, but it at least points to some hope of identifying a good vs evil axis in the weights and locking them over in the good position somehow.
Without having read into it too deeply, I wonder if the inverse could be true- training unsafe models on secure code causing allignment.
I’m fascinated with the fact the evil AI has AM as a hero. That feels like the sort of ominous foreshadowing you would see in a movie.
We will create AI that ultimately reflects our own image and culture. All the good, and all the bad.
So basically a bunch of people saw what 4chan did to TayAI and what, copied Anon’s homework?
Who’d have guessed intentionally teaching an AI to misbehave would result in the AI misbehaving.