• staph@sopuli.xyz
    link
    fedilink
    English
    arrow-up
    5
    ·
    edit-2
    7 hours ago

    There is a lot more that goes into it than just being correct. 18000 waters may have been the actual order, because somebody decided to screw with the machine. A human who isn’t terminally autistic would reliably interpret that as a joke and would simply refuse to punch that in. The LLM will likely do what a human tells it to do, since it has no contextual awareness, it only has the system prompt and whatever interaction with the user it had so far.

    • Khanzarate@lemmy.world
      link
      fedilink
      English
      arrow-up
      4
      ·
      6 hours ago

      Thats part of correctness to me, delivering an order that taco bell actually would make is important.

      Semantics aside, though, we agree. That’s very important.

    • tomiant@programming.dev
      link
      fedilink
      English
      arrow-up
      1
      ·
      edit-2
      4 hours ago

      So they just trim the instructions so it doesn’t take joke orders, so it can make more reasonable decisions, like:

      “May I take your order?”

      “Two double whoppers with extra mayo and a chocolate cherry banana sundae”

      “Oh you’ve GOTTA be joking!”