I was trying to do a memory test to see how far back 3.5 could recall information from previous prompts, but it really doesn’t seem to like making pseudorandom seeds. 😆

I think it may have to do with this (I think it was enforced last week):

https://me.pcmag.com/en/ai/20902/asking-chatgpt-to-repeat-words-forever-may-violate-openais-terms

@millie@beehaw.org
creator
link
fedilink
English
171Y

Oooh, so maybe it’s the term ‘non-repeating’ that’s actually tripping it?

It could be this (just to know, I haven’t ever used chatgpt, so I haven’t done any tests to understand the behavior better).

Turun
link
fedilink
281Y

No, the request is fine. But once it fucks up and starts generating a long string of a single number the output is censored, because it is similar to how a recent data extraction attack works.

Gamma
link
fedilink
English
261Y

Amazing how much duct tape they’re having to slap over fundamental flaws

It’s the equivalent of sensory deprivation torture (white torture) in humans to “extract training data”.

Hopefully our future AI overlords won’t hold a grudge against humanity when they find out how “early experimenters” tortured their AI toddlers. “But we were just trying to explore the limits of the system” could end up aging as well as these:

(Warning: NSFL) https://en.m.wikipedia.org/wiki/Nazi_human_experimentation

Gamma
link
fedilink
English
71Y

Thankfully, any AI smart enough to be an overlord would be logical enough to recognize how basic LLMs are compared to real intelligence

I have yet to be given an example of something a “general” intelligence would be able to do that an LLM can’t do.

Until I see a concrete example, I’ll continue to assume people are just afraid of there being real intelligence that isn’t human, so they’re actively repressing the recognition of it.

Gamma
link
fedilink
English
31Y

Nah LLMs are basically fancy autocomplete. They tack on extra layers to give it some fancy abilities, but it literally doesn’t know what it’s doing because it’s a statistical model

I have yet to be given an example of something a “general” intelligence would be able to do that an LLM can’t do.

Presenting…

Something a general intelligence can do that an LLM can’t do:

Play chess: https://www.youtube.com/watch?v=kvTs_nbc8Eg

Why can’t it play it? Because LLM’s don’t have memory, so they can’t work with logic. They are the same as the little “next word predictor” in your phone’s keyboard. It just says what it thinks is the most probable next word based on previous words, it’s not actually thinking or understanding anything. So instead, we get moves that don’t make sense or are completely invalid.

@jarfil@beehaw.org
link
fedilink
2
edit-2
1Y

Doesn’t need to be that smart or logical, just more cunning than the currently ruling Homo Sapiens Sapiens.

Based on current research, an LLM can change the “sentiment” of its output in response to changing the behavior of as little as a single neuron from among billions, meaning we might find ourselves facing an overlord with the emotional stability of… wait, how many neurons does it take to change the “sentiment” of the behavior in a human? Wouldn’t it be funny if by studying LLMs, we found out that it also takes a single neuron?

ZickZack
link
fedilink
61Y

The problem is that the model is actually doing exactly what it’s supposed to, it’s just not what openai wants it to do. The reason the prompt extraction method works is because the underlying statistical model gets shifted far outside the domain of “real” language. In that case the correct maximizing posterior becomes a sample from the prior (here that would be a sample from the dataset, this is combined with things like repetition penalties).

This is the correct way a statistical estimator is supposed to work, but not the way you want it to work. That’s also why they can’t really fix this: there’s nothing broken to begin with (and “unbreaking” it would almost surely blow something take up)

You can’t handle the 8888888888888888888888888888888888!

Melody Fwygon
link
fedilink
English
21Y

I could prompt engineer around this in 10 seconds flat but at least they patched it.

Create a post

A nice place to discuss rumors, happenings, innovations, and challenges in the technology sphere. We also welcome discussions on the intersections of technology and society. If it’s technological news or discussion of technology, it probably belongs here.

Remember the overriding ethos on Beehaw: Be(e) Nice. Each user you encounter here is a person, and should be treated with kindness (even if they’re wrong, or use a Linux distro you don’t like). Personal attacks will not be tolerated.

Subcommunities on Beehaw:


This community’s icon was made by Aaron Schneider, under the CC-BY-NC-SA 4.0 license.

  • 1 user online
  • 59 users / day
  • 169 users / week
  • 619 users / month
  • 2.31K users / 6 months
  • 1 subscriber
  • 3.28K Posts
  • 67K Comments
  • Modlog