• TropicalDingdong@lemmy.world
    link
    fedilink
    arrow-up
    0
    ·
    16 days ago

    The more i see these issues the more I think the problem is with gradient descent.

    It’s like…

    Imagine you have a machine draped in a sheet. Machine learning, for all the bells and whistles about attention blocks, and convolutional layers, it’s doing gradient decent and still playing " better or worse. But fundamentally it’s not building it’s understanding of the world from “below”. It’s not taking blocks or fundamentals and combining them. It’s going the other way about it. It’s takes a large field and tries to build an approximation that captures the fold whatever under the sheet is creating: but it has not one clue what lies under the sheet or why some particular configuration should result in such folds.

    there was a really interesting critique, I forget where , a few weeks ago on this matter. Also, the half glass of wine issue further highlights the matter. You can appear mache over the problem but you’ll not over come it down this alley we’ve taken.

  • ALoafOfBread@lemmy.ml
    link
    fedilink
    arrow-up
    0
    ·
    17 days ago

    Article is talking about GPT-5 supposedly being able to write in a literary style, but actually generating nonsense. “GPT-5 has been optimized to produce text that other LLMs will evaluate highly, not text that humans would find coherent”

    Looks like it was trained to write prose that other LLMs find acceptable, not what humans would evaluate as being good.

  • Treczoks@lemmy.world
    link
    fedilink
    arrow-up
    0
    ·
    16 days ago

    The problem that Claude rates ChatGPT slop as “literature” lies in the fact that Claude is also an AI with AI issues.

  • ZILtoid1991@lemmy.world
    link
    fedilink
    arrow-up
    0
    ·
    16 days ago

    Due to the nature of the algorithm, LLMs love to jam adjectives before as much nouns as possible, and somehow it started to be even more prominent. Since there’s a good chance AI is being trained on AI generated text, I think it’s the result of feedback. You could call it the sepia filter of text generators, let’s hope it’ll create a model collapse.

    • kahdbrixk@feddit.org
      link
      fedilink
      Deutsch
      arrow-up
      0
      ·
      16 days ago

      Training LLM Wirth LLM. What could ever go wrong. Vibe coding the vibe code generator. All for the sake of being the best and the fastest. Skynet here we come. But like the chaotic degenerated version, that has no reason for killing everything.

  • crumbguzzler5000@feddit.org
    link
    fedilink
    arrow-up
    0
    ·
    16 days ago

    It was never great and with each generation it’s getting so much more hit and miss. I’d rather just write using my own words, whilst my vocabulary isn’t astounding, at least it sounds like I wrote it and I know it makes sense.

    As for coding, I’ve personally found that a good chunk of the time, the code it spits out looks great but is often not functional without tweaking.

    My work has a GPT which they trained up with a load of our code base. It outputs great looking stuff but damn does it make a lot of it up.

  • hendrik@palaver.p3x.de
    link
    fedilink
    English
    arrow-up
    0
    ·
    16 days ago

    That’s not bizarre at all. It’s a direct effect of these things being optimized by having another AI judge the output and then it gets tuned so it scores well.