Share.

11 Comments

  1. Insciuspetra on

    Sweet!

    Let’s have it take a look at how our government works and see if it has a better solution to increase efficiency

  2. Perhaps we should never take any social advice from entities with very strong evolutionary pressure, huh? They might be *just a little bit* biased on what “evolution” means…

  3. michael-65536 on

    “Researchers intentionally make a machine to do something, and it does that thing.”

    News at 11.

  4. funny_bunny_mel on

    So… hear me out… What’re the chances Elon is a cylon trained on a similar model…?

  5. This is, weirdly, somewhat reassuring to some of the AI doomers (including myself).  

    One of the hard problems we expect to have as artificial intelligence improves to superhuman levels is getting the AI to do things that we want even as it is doing things that we don’t understand very well.  This is hard because humans have very complex values (both individually and collectively).  Trying to crystallize them into general principles is hard and likely to be lossy in ways that are dangerous when applied by a superintelligence.  

    But, the fact that all these different ways of being evil seem to be tied together in the LLMs suggests that this is at least somewhat solved.  Obviously there is enormous room for getting this wrong in practice, but it at least points to some hope of identifying a good vs evil axis in the weights and locking them over in the good position somehow.

  6. -illusoryMechanist on

    Without having read into it too deeply, I  wonder if the inverse could be true- training unsafe models on secure code causing allignment.

  7. I’m fascinated with the fact the evil AI has AM as a hero. That feels like the sort of ominous foreshadowing you would see in a movie.

  8. AnarkittenSurprise on

    We will create AI that ultimately reflects our own image and culture. All the good, and all the bad.

  9. So basically a bunch of people saw what 4chan did to TayAI and what, copied Anon’s homework?

    Who’d have guessed intentionally teaching an AI to misbehave would result in the AI misbehaving.