• Zaktor@sopuli.xyz
    link
    fedilink
    English
    arrow-up
    1
    ·
    1 year ago

    does not have a model of the objects to which the words refer

    I’m not even sure what this is supposed to be saying. Sounds kind of like a bullshit generator.

    Words are encodings of knowledge and their expression and use represent that knowledge, and these machines ingest a repository containing a significant percent of written human communication. It encodes that the words “dog” and “bark” are often used together, but it also encodes that “dog” and “cat” are things that are both “mammals” and “mammals” are “animals”, and that the pair of them are much more likely to appear in a human household than a “porpoise”. What is this other kind of model of objects that hasn’t been in some way represented in all of the internet?

    • Veraticus@lib.lgbtOP
      link
      fedilink
      English
      arrow-up
      0
      ·
      1 year ago

      It is not a model of objects. It’s a model of words. It doesn’t know what those words themselves mean or what they refer to; it doesn’t know how they relate together, except that some words are more likely to follow other words. (It doesn’t even know what an object is!)

      When we say “cat,” we think of a cat. If we then talk about a cat, it’s because we love cats, or hate them, or want to communicate something about them.

      When an LLM says “cat,” it has done so because a tokenization process selected it from a chain of word weights.

      That’s the difference. It doesn’t think or reason or feel at all, and that does actually matter.

      • Zaktor@sopuli.xyz
        link
        fedilink
        English
        arrow-up
        1
        ·
        1 year ago

        This is just the same hand-waving repeated. What does it mean to “know what a word means”? How is a word, indexed into a complex network of word embeddings, meaningfully different as a token from this desired “object model”? Because the indexing and encoding very much does relate words together separately from their likelihood to appear in a sentence together. These embeddings may be learned from language, but language is simply a method of communicating meaning, and notably humans also learn meaning through consuming it.

        What do things like “love” or “want” or “feeling” have to do with a model of objects? How would you even recognize a system that does that and why would it be any more capable than a LLM at producing good and trustable information? Does feeling love for a concept help you explain what a random blogger does? Do you need to want something to produce meaningful output?

        This just all seems like poorly defined techno-spiritualism.

        • Veraticus@lib.lgbtOP
          link
          fedilink
          English
          arrow-up
          0
          ·
          1 year ago

          It is not hand-waving; it is the difference between an LLM, which, again, has no cognizance, no agency, and no thought – and humans, which do. Do you truly believe humans are simply mechanistic processes that when you ask them a question, a cascade of mathematics occurs and they spit out an output? People actually have an internal reality. For example, they could refuse to answer your question! Can an LLM do even something that simple?

          I find it absolutely mystifying you claim you’ve studied this when you so confidently analogize humans and LLMs when they truly are nothing alike.

          • barsoap@lemm.ee
            link
            fedilink
            arrow-up
            1
            ·
            1 year ago

            Do you truly believe humans are simply mechanistic processes that when you ask them a question, a cascade of mathematics occurs and they spit out an output? People actually have an internal reality.

            Those two things can be true at the same time.

            I find it absolutely mystifying you claim you’ve studied this when you so confidently analogize humans and LLMs when they truly are nothing alike.

            “Nothing alike” is kinda harsh, we do have about as much in common with ChatGPT as we have with flies purpose-bred to fly left or right when exposed to certain stimuli.

          • Zaktor@sopuli.xyz
            link
            fedilink
            English
            arrow-up
            1
            ·
            1 year ago

            no cognizance, no agency, and no thought

            Define your terms. And explain why any of them matter for producing valid and “intelligent” responses to questions.

            Do you truly believe humans are simply mechanistic processes that when you ask them a question, a cascade of mathematics occurs and they spit out an output?

            Why are you so confident they aren’t? Do you believe in a soul or some other ephemeral entity that wouldn’t leave us as a biological machine?

            People actually have an internal reality. For example, they could refuse to answer your question! Can an LLM do even something that simple?

            Define your terms. And again, why is that a requirement for intelligence? Most of the things we do each day don’t involve conscious internal planning and reasoning. We simply act and if asked will generate justifications and reasoning after the fact.

            It’s not that I’m claiming LLMs = humans, I’m saying you’re throwing out all these fuzzy concepts as if they’re essential features lacking in LLMs to explain their failures in some question answering as something other than just a data problem. Many people want to believe in human intellectual specialness, and more recently people are scared of losing their jobs to AI, so there’s always a kneejerk reaction to redefine intelligence whenever an animal or machine is discovered to have surpassed the previous threshold. Your thresholds are facets of the mind that you both don’t define, have no means to recognize (I assume your consciousness, but I cannot test it), and have not explained why they’re important for fact rather than BS generation.

            How the brain works and what’s important for various capabilities is not a well understood subject, and many of these seemingly essential features are not really testable or comparable between people and sometimes just don’t exist in people, either due to brain damage or a simple quirk in their development. The people with these conditions (and a host of other psychological anomalies) seem to function just fine and would not be considered unthinking. They can certainly answer (and get wrong) questions.

          • lloram239@feddit.de
            link
            fedilink
            arrow-up
            1
            ·
            1 year ago

            People actually have an internal reality.

            So do LLMs.

            Can an LLM do even something that simple?

            Ask it about any NSFW topic and it will refuse.

            analogize humans and LLMs when they truly are nothing alike.

            They seem way more similar than different. The part were they are different trivially follow from the LLMs architecture (e.g. LLMs are static, tokenizing makes character-based problems difficult, memory is limited to the prompt, no interaction with the external world, no vision, no hearing, …) and most of that can be overcome by extending the model, e.g. multi-model models with vision and hearing are on their way, DeepMind is working on models that interact with the real world, etc. This is all coming and coming fast.

        • barsoap@lemm.ee
          link
          fedilink
          arrow-up
          0
          ·
          edit-2
          1 year ago

          What does it mean to “know what a word means”?

          For one, ChatGPT has no idea what a cat or dog looks like. It has no understanding of their differences in character of movement. Lacking that kind of non-verbal understanding, when analysing art that’s actually in its domain, that is, poetry, it couldn’t even begin to make sense of the question “has this poem feline or canine qualities” – best it can do is recognise that there’s neither cats nor dogs in it and, being stumped, make up some utter nonsense. Maybe it has heard of catty and that dogs are loyal and will be looking for those themes, but feline and canine as in elegance? Forget it, unless it has read a large corpus of poet analysis that uses those terms: It can parrot that pattern matching, but it can’t do the pattern matching itself, it cannot transfer knowledge from one domain to another when it has no access to one of those domains.

          And that’s the tip of the iceberg. As humans we’re not really capable of purely symbolic thought so it’s practically impossible to appreciate just how limited those systems are because they’re not embodied.

          (And, yes, Stable Diffusion has some understanding of feline vs. canine as in elegance – but it’s an utter moron in other areas. It can’t even count to one).


          Then, that all said, and even more fundamentally, ChatGPT (as all other current AI algos we have) is a T2 system, not a T3 system. It comes with rules how to learn, it doesn’t come with rules enabling it to learn how to learn. As such it never thinks – it cannot think, as in “mull over”. It reacts with what passes as a gut in AI land, and never with “oh I’m not sure about this so let me mull it over”. It is in principle capable of not being sure but that doesn’t mean it can rectify the situation.

          • lloram239@feddit.de
            link
            fedilink
            arrow-up
            1
            ·
            1 year ago

            it couldn’t even begin to make sense of the question “has this poem feline or canine qualities”

            Which is obviously false, as a quick try will show. Poems are just language and LLMs understand that very well. That LLMs don’t have any idea how cats actually look like or move, beyond what they can gather from text books, is irrelevant here, they aren’t tasked with painting a picture (which the upcoming multi-modal models can do anyway).

            Now there can of course be problems that can be expressed in language, but not solve in the realm of language. But I find those to be incredible rare, rare enough that I never really seen a good example. ChatGPT captures an enormous amount of knowledge about the world, and humans have written about a lot of stuff. Coming up with questions that would be trivial to answer for any human, but impossible for ChatGPT is quite tricky.

            And that’s the tip of the iceberg.

            Have you actually ever actually seen an iceberg or just read about them?

            It comes with rules how to learn, it doesn’t come with rules enabling it to learn how to learn

            ChatGPT doesn’t learn. It’s a completely static model that doesn’t change. All the learning happened in a separate step back when it was created, it doesn’t happen when you interact with it. That illusion comes from the text prompt, which includes both your text as well as its output, getting feed into the model as input. But outside that text prompt, it’s just static.

            “oh I’m not sure about this so let me mull it over”.

            That’s because it fundamentally can’t mull it over. It’s a feed forward neural network, meaning everything that goes in on one side comes out on the other in a fixed amount of time. It can’t do loops by itself. It has no hidden internal monologue. The only dynamic part is the prompt, which is also why its ability to problem solve improves quite a bit when you require it to do the steps individually instead of just presenting the answer, as that allows the prompt to be it’s “internal monologue”.