jdfr a day ago

OP seems to have run a programming language detector on the generated texts, and made a graph of programming language frecuencies: https://pbs.twimg.com/media/Gx2kvNxXEAAkBO0.jpg?name=orig

As a result, OP seems to think the model was trained on a lot of Perl: https://xcancel.com/jxmnop/status/1953899440315527273#m

LOL! I think these results speak more to the flexibility of Perl than any actual insight on the training data! After all, 93% of inkblots are valid Perl scripts: https://www.mcmillen.dev/sigbovik/

  • esafak a day ago

    I don't understand why Perl, R, and AppleScript rank so much higher than their observed use.

    • londonlawyer 19 hours ago

      The prominence of AppleScript ought to have been a pretty big red flag: the author seems to be claiming the model was trained on more AppleScript than Python, which simply can’t be true.

      Ironically LLMs seem pretty bad at writing AppleScript, I think because (i) the syntax is English-like but very brittle, (ii) the application dictionaries are essential but generally not on the web, and (iii) most of the AppleScript that is on the web has been written by end users, often badly.

    • j_bum 19 hours ago

      R being so high makes no sense to me either.

      I think as of the last Stack Overflow developer survey, it only had ~4% market share…

      I say this as an R user who spams LLMs with R on a daily basis.

  • godelski 13 hours ago

    Jack has a lot of really bad takes and frequently makes lots of mistakes. Honestly, I don't know why people take him seriously.

    I mean you can go read his blog post that's pinned where he argues that there's no new ideas and it is all data. He makes the argument that architecture doesn't matter, which is just so demonstrably false that it is laughable. He's a scale maximalist.

    I also expect an AI researcher from a top university to not make such wild mistakes

      > 3. RLHF: first proposed (to my knowledge) in the InstructGPT paper from OpenAI in 2022
    
    I mean if you go read the instruct paper on page 2 you'll see

      | Specifically, we use reinforcement learning from human feedback (RLHF; Christiano et al., 2017; Stiennon et al., 2020) to fine-tune GPT-3 to follow a broad class of written instructions (see Figure 2).
    
    Where in Christiano you'll find

      || Our algorithm follows the same basic approach as Akrour et al. (2012) and Akrour et al. (2014)
    
    I mean this is just obviously wrong. It is so obviously wrong it should make the person saying it second guess themselves (which is categorically the same error you're pointing out).

    I'm sure we can trace the idea back to the 80's if not earlier. This is the kind of take I'd expect a non-researcher to have, but not someone with two dozen NLP papers. The Instruct-GPT paper was just the first time someone integrated RLHF into a LLM (but not a LM).

    Maybe a better article is the one he wrote on Super Intelligence From First Principles. As usual, when someone says "First Principles" you bet they're not gonna start from First Principles... I guess this makes sense in CS since we index from 0

    [0] https://arxiv.org/abs/2203.02155

    [Christiano et al] https://arxiv.org/abs/1706.03741

    [Stiennon eta al] https://arxiv.org/abs/2009.01325

    [Akrour et al (2012)] https://arxiv.org/abs/1208.0984

    • jxmorris12 10 hours ago

      Hi again. I had already written about this later in my blog post (which is unrelated to this thread), but the point was that RLHF hadn't been applied to language models at scale until InstructGPT. I edited the post just now to clarify this. Thanks for the feedback!

  • johnisgood a day ago

    That inkblot thing can be created for any language.

    • bravesoul2 a day ago

      How? E.g. I doubt an inkblot can produce a valid C# program.

      • johnisgood a day ago

        They are not full programs, just code translating to numbers and strings.

        I used an LLM to generate an inkblot that translates to a Python string and number along with verification of it, which just proves that it is possible.

        • bstsb a day ago

          what are you talking about?

          the way that the quoted article creates Perl programs is through OCRing the inkblots (i.e. creating almost random text) and then checking that result to see if said text is valid Perl

          it's not generating a program that means anything

          • johnisgood a day ago

            Okay, and I created inkblots that mean "numbers"[1] and "strings" in Python.

            > it's not generating a program that means anything

            Glad we agree.

            [1] Could OCR those inkblots (i.e. they are almost random text)

            • dmbche a day ago

              No, asking an LLM to generate the inkblot is the same as asking the LLM to write a string and then obfuscating it in an inkblot.

              OCRing literal random inkblots will not produce valid C (or C# or python) code, but it will prodce valid Perl most of the time, because Perl is weird, and that is funny.

              It's not about obfuscating text in inkblot, it's about almost any string being a valid Perl program, which is not the case for most languages

              Edit0: here: https://www.mcmillen.dev/sigbovik/

              • johnisgood a day ago

                Okay, my bad.

                > it's about almost any string being a valid Perl program

                Is this true? I think most random unquoted strings aren't valid Perl programs either, am I wrong?

                • akerl_ a day ago

                  Yes. That was the whole point of the original comment you were misunderstanding.

                  Because of the flexibility of Perl and heavy amount of symbol usage, you can in fact run most random combinations of strings and they’ll be valid Perl.

                  Copying from the original comment: https://www.mcmillen.dev/sigbovik/

            • mathiaspoint a day ago

              Most random unquoted strings are certainly not valid Python programs. I don't know Perl well enough to say anything about that but I know what you're saying certainly isn't true with Python.

    • westurner a day ago

      Of the powerset of all operators and inputs, how many can be represented in any programming language?

      What percent of all e.g. ASCII or Unicode strings are valid expressions given a formal grammar?

  • actuallyalys 16 hours ago

    Honestly these results may say as much about the classifier as they do about the data they’re classifying.

ComputerGuru 21 hours ago

Not very rigorous or scientific, honestly, I would say it's just clickbait spam with some pretty graphs. Everything on twitter is now a "deep dive". No info on how the 10M "random examples" were generated and how that prevents the model from collapsing around variations of the same output. Others already mentioned how the "classification" of output by coding language is bunk with a good explanation for how Perl can come out on top even if it's not actually Perl, but I was struck by OP saying "(btw, from my analysis Java and Kotlin should be way higher. classifier may have gone wrong)" but then merrily continuing to use the data.

Personally, I expect more rigor from any analysis and would hold myself to a higher standard. If I see anomalous output at a stage, I don't think "hmm looks like one particular case may be bad but the rest is fine" but rather "something must have gone wrong and the entire output/methodology is unusable garbage" until I figure out exactly how and why it went wrong. And 99 times out of a 100 it wasn't the one case (that happened to be languages OP was familiar with) but rather something fundamentally incorrect in the approach that means the data isn't usable and doesn't tell you anything.

  • godelski 13 hours ago

      > Personally, I expect more rigor from any analysis and would hold myself to a higher standard.
    
    When something is "pretty bizarre" the most likely conclusion is "I fucked up", which is very likely in this case. I really wonder if he actually checked the results of the classifier. These things can be wildly inaccurate since differences in languages can be quite small at times and some are very human language oriented. He even admits that Java and Kotlin should be higher but then doesn't question Perl, R, Applescript, Rust, and the big drop to Python. What's the joke? If you slam your head on the keyboard you'll generate a valid Perl program?

    It worries me that I get this feeling from quite a number of ML people who are being hired and paid big bucks from big tech companies. I say this as someone in ML too. There's a propensity to just accept outputs rather than question them. This is like a basic part of doing any research, you should always be incredibly suspicious of your own results. What did Feynman say? Something like "The first rule is not to be fooled and you're the easiest person to fool"?

k310 2 days ago

Anything but this image (imgbb.com link below) requires a login. I get the same deal with Facebook. I am not Don Quixote and prefer not to march into hell for a heavenly cause, nor any other.

https://i.ibb.co/Zz2VgY4C/Gx2-Vd6-DW4-AAogtn.jpg

  • Epskampie a day ago
    • k310 a day ago

      Thanks! I've seen a lot of stuff come and go, so thanks for the reminder.

      For example, Libgen is out of commission, and the substitutes are hell to use.

      Summary of what's up and not up:

      https://open-slum.org/

      • randomNumber7 a day ago

        > Libgen is out of commission, and the substitutes are hell to use

        Somehow I also preferred libgen, but I don't think annas archive is "hell to use".

        • k310 a day ago

          Annas Archive uses slow servers on delay, and constantly tells me that they are too many downloads from my IP address, so I flip VPN settings as soon as the most recent slow download completes. And I get it again after a short while. It's hell waiting it out and flipping VPN settings. And the weird part is that this project is to replace paper books that I already bought. That's the excuse one LLM uses for tearing up books, scanning and harvesting. I just need to downsize so I can move back to the Bay Area. Book and excess houseware sale coming, it seems. Libgen had few or no limits.

          • 1gn15 a day ago

            I would recommend donating to gain access to the fast downloads; they need money for the servers.

      • stavros a day ago

        Oh no, why did Libgen die?

        • k310 a day ago

          Shut down. See

          https://open-slum.org/

          for substitutes. The alternate libgen sites seem more limited to me, but I am comparing with memories, so untrustworthy.

        • nikcub a day ago

          it's available at the bz tld

orbital-decay a day ago

>what you can't see from the map is many of the chains start in English but slowly descend into Neuralese

That's just natural reward hacking when you have no training/constraints for readability. IIRC R1 Zero is like that too, they retrained it with a bit of SFT to keep it readable and called it R1. Hallucinating training examples if you break the format or prompt it with nothing is also pretty standard behavior.

james-bcn a day ago

This looks very interesting but I don't really understand what he has done here. Can someone explain the process he has gone through in this analysis?

  • AmazingTurtle a day ago

    He presented an empty prompt to gpt OSS and let it run many times. Through temperature, the results vary quite a lot. He sampled the results.

    Feeding an empty prompt to a model can be quite revealing on what data it was trained on

greenchair a day ago

"this thing is clearly trained via RL to think and solve tasks for specific reasoning benchmarks. nothing else." Has the train already reached the end of the line?

  • red75prime 5 hours ago

    If you think something like "They have to train their models on benchmarks to make it look like there's progress, while in reality it's a dead end," you are missing a few things.

    It's an open model, everyone can bench it on everything not only on specific benchmarks. Training on specific reasoning benchmarks is a conjecture.

esperent a day ago

> the chains start in English but slowly descend into Neuralese

What is Nueralese? I tried searching for a definition but it just turns up a bunch of Less Wrong and Medium articles that don't explain anything.

Is it a technical term?

  • meowface a day ago

    It's a term somewhat popularized by the LessWrong/rationalism community to refer to communication (self-communication/note-taking/state-tracking/reasoning, or model-to-model communication) via abstract latent space information rather than written human language. Vectors instead of words.

    One implication leading to its popularity by LessWrong is the worry that malicious AI agents might hide bad intent and actions by communicating in a dense, indecipherable way while presenting only normal intent and actions in their natural language output.

    • verisimi a day ago

      > malicious AI agents might hide bad intent and actions by communicating in a dense, indecipherable way while presenting only normal intent and actions in their natural language output.

      you could edit this slightly to extract a pretty decent rule for governance, like so:

      > malicious agents might hide bad intent and actions by communicating in a dense, indecipherable way while presenting only normal intent and actions in a natural way

      It applies to ai, but also many other circumstances where the intention is that you are governed - eg medical, legal, financial.

      Thanks!

  • CjHuber a day ago

    I suppose it means LLM gibberish

    EDIT: orbital decay explained it pretty well in this thread

  • nopinsight a day ago

    The author might use it as an analogy to mentalese but for neural networks.

    https://en.wiktionary.org/wiki/mentalese

    EDIT: After reading the original thread in more detail, I think some of the sibling comments are more accurate. In this case, neuralese is more like language of communication expressed by neural networks, rather than its internal representation.

  • fl1pper a day ago

    neuralese is a term first used in neuroscience to describe the internal coding or communication system within neural systems.

    it originally referred to the idea that neural signals might form an intrinsic "language" representing aspects of the world, though these signals gain meaning only through interpretation in context.

    in artificial intelligence, the term now has a more concrete role, referring to the deep communication protocols used by multiagent systems.

  • spwa4 a day ago

    There's 2 things called neuralese:

    1) internally, in latent space, LLMs use what is effectively a language, but all the words are written on top of each other instead of separately, and if you decode it as letters, it sounds like gibberish, even though it isn't. It's just a much denser language than any human language. This makes them unreadable ... and thus "hides the intentions of the LLM", if you want to make it sound dramatic and evil. But yeah, we don't know what the intermediate thoughts of an LLM sound like.

    The decoded version is often referred to as "neuralese".

    2) if 2 LLMs with sufficiently similar latent space communicate with each other (same model), it has often been observed that they switch to "gibberish" BUT when tested they are clearly still passing meaningful information to one another. One assumes they are using tokens more efficiently to get the latent space information to a specific point, rather than bothering with words (think of it like this: the thoughts of an LLM are a 3d point (in reality 2000d, but ...). Every token/letter is a 3d vector (meaning you add them), chosen so words add up to the thought that is their meaning. But when outputting text why bother with words? You can reach any thought/meaning by combining vectors, just find the letter moving the most in the right direction. Much faster)

    Btw: some specific humans (usually toddlers or children that are related) when talking to each other switch to talking gibberish to each other as well while communicating. This is especially often observed in children that initially learn language together. Might be the same thing.

    These languages are called "neuralese".

ma2rten a day ago

Presumably the model is trained in post-training to produce a response to a prompt, but not to reproduce the prompt itself. So if you prompt it with an empty prompt it's going to be out of distribution.

puttycat a day ago

> OpenAI has figured out RL. the models no longer speak english

What does this mean?

  • orbital-decay a day ago

    The model learns to reason on its own. If you only reward correct results but not readable reasoning, it will find its own way to reason that is not necessarily readable by a human. The chain may look like English, but the meaning of those words might be completely different (or even the opposite) for the model. Or it might look like a mix of languages, or just some gibberish - for you, but not for the model. Many models write one thing in the reasoning chain and a completely different in the reply.

    That's the nature of reinforcement learning and any evolutionary processes. That's why the chain of thought in reasoning models is much less useful for debugging than it seems, even if the chain was guided by the reward model or finetuning.

  • Hard_Space a day ago

    Interesting. This happens in Colossus: The Forbin Project (1970), where the rogue AI escapes the semantic drudgery of English and invents its own compressed language with which to talk to its Russian counterpart.

    • Mistletoe a day ago

      It also happens in Ex Machina at the end when the two androids whisper and talk to each other in their special faster language. I always found this to be one of the most believable, real things from that movie and one of my favorite parts.

revskill a day ago

What does that mean ?

pinoy420 a day ago

5 seems to do a better job with copyrighted content. I got it to spit out the entirely of ep IV (but you have to redact the character names)