• stingpie@lemmy.world
    link
    fedilink
    arrow-up
    69
    arrow-down
    2
    ·
    5 months ago

    This might be happening because of the ‘elegant’ (incredibly hacky) way openai encodes multiple languages into their models. Instead of using all character sets, they use a modulo operator on each character, to make all Unicode characters represented by a small range of values. On the back end, it somehow detects which language is being spoken, and uses that character set for the response. Seeing as the last line seems to be the same mathematical expression as what you asked, my guess is that your equation just happened to perfectly match some sentence that would make sense in the weird language.

        • crispy_kilt@feddit.de
          link
          fedilink
          arrow-up
          2
          ·
          5 months ago

          Seriously? Python for massive amounts of data? It’s a nice scripting language, but it’s excruciatingly slow

          • stingpie@lemmy.world
            link
            fedilink
            arrow-up
            5
            ·
            5 months ago

            There are bindings in java and c++, but python is the industry standard for AI. The libraries for machine learning are actually written in c++, but use python language bindings. Python doesn’t tend to slow things down since machine learning is gpu-bound anyway. There are also library specific programming languages which urges the user to make pythonic code that can be compiled into c++.

    • NeatNit@discuss.tchncs.de
      link
      fedilink
      arrow-up
      18
      arrow-down
      1
      ·
      5 months ago

      I suppose it’s conceivable that there’s a bug in converting between different representations of Unicode, but I’m not buying and of this “detected which language is being spoken” nonsense or the use of character sets. It would just use Unicode.

      The modulo idea makes absolutely no sense, as LLMs use tokens, not characters, and there’s soooooo many tokens. It would make no sense to make those tokens ambiguous.

      • stingpie@lemmy.world
        link
        fedilink
        arrow-up
        8
        ·
        5 months ago

        I completely agree that it’s a stupid way of doing things, but it is how openai reduced the vocab size of gpt-2 & gpt-3. As far as I know–I have only read the comments in the source code– the conversion is done as a preprocessing step. Here’s the code to gpt-2: https://github.com/openai/gpt-2/blob/master/src/encoder.py I did apparently make a mistake, as the vocab reduction is done through a lut instead of a simple mod.