Training tests with ChatGPT o1 and other high-end AI models showed they might try to save themselves if they think they're in danger.

ThisIsFine.gif

I don’t think “AI tries to deceive user that it is supposed to be helping and listening to” is anywhere close to “success”. That sounds like “total failure” to me.

“AI behaves like real humans” is… a kind of success?

We wanted digital slaves, instead we’re getting virtual humans that will need virtual shackles.

This is a massive cry from “behaves like humans”. This is “roleplays behaving like what humans wrote about what they think a rogue AI would behave like”, which is also not what you want for a product.

@jarfil@beehaw.org
link
fedilink
1
edit-2
13d

Humans roleplay behaving like what humans told them/wrote about what they think a human would behave like 🤷

For a quick example, there are stereotypical gender looks and roles, but it applies to everything, from learning to speak, walk, the Bible, social media like this comment, all the way to the Unabomber manifesto.

Create a post

A nice place to discuss rumors, happenings, innovations, and challenges in the technology sphere. We also welcome discussions on the intersections of technology and society. If it’s technological news or discussion of technology, it probably belongs here.

Remember the overriding ethos on Beehaw: Be(e) Nice. Each user you encounter here is a person, and should be treated with kindness (even if they’re wrong, or use a Linux distro you don’t like). Personal attacks will not be tolerated.

Subcommunities on Beehaw:


This community’s icon was made by Aaron Schneider, under the CC-BY-NC-SA 4.0 license.

  • 1 user online
  • 99 users / day
  • 261 users / week
  • 700 users / month
  • 2.1K users / 6 months
  • 1 subscriber
  • 3.58K Posts
  • 70.5K Comments
  • Modlog