• LughOPMA
    link
    English
    2211 days ago

    Current LLM models tend to extract “best practice” responses a lot. They can statistically guess the correct responses to things, because it’s what experts cite the most. I wonder if that is what is behind this? As the authors of the research point out, the significance here is not the AI’s appearance of superior intelligence, it’s that it’s yet another example of how people may be influenced by AI.

    • DarkThoughts
      link
      fedilink
      711 days ago

      That and the fact that I’m guessing they hand picked the results too instead of using just the first response given. Ultimately LLMs aren’t AI, it’s not forming its own thoughts, it’s generating text based on input that was produced by humans. So saying they rated “AI” responses better than humans is already disingenuous.

    • @CanadaPlus@lemmy.sdf.org
      link
      fedilink
      English
      1
      edit-2
      11 days ago

      Probably. They’ve mastered the art of corporate-speak; another natural language task which doesn’t require precise abstract reasoning.

      I’m kind of convinced that the set of possible moral philosophies most people would agree with in practice is the empty set, at this point, so I’m not surprised those kinds of answers do better.

    • @credo@lemmy.world
      link
      fedilink
      English
      1
      edit-2
      10 days ago

      I took one of the more complicated questions from an expert help column and fed it into Chat GPT. This was before it could perform live searches and the answer it gave was pretty close to the expert’s own answer.

    • @xor@infosec.pub
      link
      fedilink
      English
      -111 days ago

      “A representative sample of 299 U.S. adults first rated the quality of moral evaluations when blinded to their source.”

      a representative sample is probably 299 absolute idiots… i’d also question what people they had actually write the human essays…