Lvxferre [he/him]

I have two chimps within, called Laziness and Hyperactivity. They smoke cigs, drink yerba, fling shit at each other, and devour the faces of anyone who comes close to them.

They also devour my dreams.

  • 0 Posts
  • 1.48K Comments
Joined 2 years ago
cake
Cake day: January 12th, 2024

help-circle



  • I’ve interacted with k0e3 in the past, they’re no LLM. Even then, a quick profile check shows it. But you didn’t check it, right? Of course you didn’t, it’s easier to vomit assumptions and re-eat your own vomit, right?

    And the comment’s “tone” isn’t even remotely close to typical LLM output dammit. LLMs avoid words like “bullshit”, contracting “it is not” into “it’s not” (instead of “it isn’t”), or writing in first person. The only thing resembling LLM output is the em dash usage—but there are a thousand potential reasons for that.

    (inb4 assumer claims I’m also an LLM because I just used an em dash and listed three items.)





  • You don’t get it.

    I do get it. And that’s why I’m disdainful towards all this “simulated reasoning” babble.

    In the past, the brick throwing machine was always failing its target and nowadays it is almost always hitting near its target.

    Emphasis mine: that “near” is a sleight of hand.

    It doesn’t really matter if it’s hitting “near” or “far”; in both cases someone will need to stop the brick-throwing machine, get into the construction site (as if building a house manually), place the brick in the correct location (as if building a house manually), and then redo operations as usual.

    In other words, “hitting near the target” = “failure to hit the target”.

    And it’s obvious why it’s wrong; the idea that an auto-builder should throw bricks is silly. It should detect where the brick should be placed, and lay it down gently.

    The same thing applies to those large token* models; they won’t reach anywhere close to reasoning, just like a brick-throwing machine won’t reach anywhere close to an automatic house builder.

    *I’m calling it “large token model” instead of “large language model” to highlight another thing: those models don’t even model language fully, except in the brain of functionally illiterate tech bros who think language is just a bunch of words. Semantics and pragmatics are core parts of a language; you don’t have language if utterances don’t have meaning or purpose. The nearest of that LLMs do is to plop some mislabelled “semantic supplement” - because it’s a great red herring (if you mislabel something, you’re bound to get suckers confusing it with the real thing, and saying “I dun unrurrstand, they have semantics! Y u say they don’t? I is so confusion… lol lmao”).

    It depends on how good you are asking the machine to throw bricks (you need to assume some will miss and correct accordingly).

    If the machine relies on you to be an assumer (i.e. to make shit up, like a muppet), there’s already something wrong with it.

    Eventually, brick throwing machines will get so good that they will rely on gravitational forces to place the bricks perfectly and auto-build houses.

    To be blunt that stinks “wishful thinking” from a distance.

    As I implied in the other comment (“Can house construction be partially automated? Certainly. Perhaps even fully. But not through a brick-throwing machine.”), I don’t think reasoning algorithms are impossible; but it’s clear LLMs are not the way to go.


  • You don’t say.

    Imagine for a moment you had a machine that allows you to throw bricks at a certain distance. This shit is useful, specially if you’re a griefer; but even if you aren’t, there are some corner cases for that, like transporting construction material at a distance.

    And yet whoever sold you the machine calls it a “house auto-builder”. He tells you that it can help you to build your house. Mmmh.

    Can house construction be partially automated? Certainly. Perhaps even fully. But not through a brick-throwing machine.

    Of course trying to use the machine for its advertised purpose will go poorly, even if you only delegate brick placement to it (and still build the foundation, add cement etc. manually). You might economise a bit of time when the machine happens to throw a brick in the right place, but you’ll waste a lot of time cleaning broken bricks, or replacing them. But it’s still being sold as a house auto-builder.

    But the seller is really, really, really invested on this auto-construction babble. Because his investors gave him money to create auto-construction tools. And he keeps babbling on how “soon” we’re going to get fully auto house building, and how it’s an existential threat to builders and all that babble. So he tweaks the machines to include “simulated building”. All it does is to tweak the force and aim of the machine, so it’s slightly less worse at throwing bricks.

    It still does not solve the main problem: you don’t build a house by throwing bricks. You need to place them. But you still have some suckers saying “haha, but it’s a building machine lmao, can you prove it doesn’t build? lol”.

    That’s all what “reasoning” LLMs are about.



  • It’s completely off-topic, but:

    We used to have a rather large sisal fibre mat/rug at home, that Siegfrieda (my cat) used to scratch. However my mum got some hate boner against that mat, and replaced it with an actual rug. That’s when Frieda decided she’d hop onto the sofa and chairs and scratch them.

    We bought her a scratching post - and she simply ignored it. I solved the issue by buying two smaller sisal mats, and placing them strategically in places Frieda hangs around. And then slapping her butt every time she used them, for positive behaviour reinforcement (“I’m pet when I scratch it! I should scratch it more!”)

    I’m sharing this to highlight it’s also important to recognise each individual cat has preferences, that might not apply to other cats. She wanted a horizontal surface to scratch; so no amount of scratching posts would solve it.





  • I worded it in a dumb/certain/silly way but, unless drastic changes happen, I do find it likely to happen.

    Look at how often Nintendo is surfacing negatively on the news:

    • harassing a small dev studio over patents,
    • trying to kill emulation while profiting off it,
    • bricking hardware already sold to customers,
    • demanding unreasonable prices for new games,
    • dictating if you shall be allowed to feature one of its games in a speedrunning event…

    Nintendo stopped being seen as a company that enables your fun, to become one that gatekeeps it. That’s brand damage - and really bad for Nintendo’s console sales; people are only willing to invest in a console if they’re reasonably certain they can have fun with it.

    And at the same time, there are voices within and around Nintendo pushing the company towards the mobile market. Remember Pokémon Go? Or Ishihara saying the Switch 1 would flop, because of smartphones? If Nintendo console sales decline meaningfully, those voices will become louder and louder. Eventually Nintendo will focus primarily on the mobile market.

    However people don’t typically buy mobile games; the monetisation strategy is completely different - microtransactions, gacha, lootboxes, all that crap. Most players (the “minnows”) won’t drop a penny on the game, but huge spenders (the “whales”) compensate for that, so it works.

    The minnows aren’t just freeloaders, mind you; they’re required to keep the game alive. So mobile game companies need to fine-tune the pressure in their games - it should be just enough to encourage people to spend some money on the game, but not enough to shoo the minnows away.

    But we’re talking about Nintendo here. A company willing to damage its own brand for a few additional pennies. Nintendo would not be able to see all those minnows and say “hey, that’s cool”, it would go full “ARE THOSE FREELOADERS STUPID? DON’T THEY KNOW THEY’RE SUPPOSED TO BUY STUFF?”. It would tune the pressure way up, and ruin its mobile market, after it ruined its console market.

    …perhaps it should go back to selling playing cards.




  • [special pleading] Those are all the smallest models

    [sarcasm] Yeah, because if you randomly throw more bricks in a construction site, the bigger pile of debris will look more like a house, right. [/sarcasm]

    and you don’t seem to have reasoning [SIC] mode, or external tooling, enabled?

    Those are the chatbots available through DDG. I just found it amusing enough to share, given

    1. The logic procedure to be followed (multiplication) is rather simple, and well documented across the internet, thus certainly present in their corpora.
    2. The result is easy to judge: it’s either correct or incorrect.
    3. All answers are incorrect and different from each other.

    Small note regarding “reasoning”: just like “hallucination” and anything they say about semantics, it’s a red herring that obfuscates what is really happening.

    At the end of the day it’s simply weighting the next token based on the previous tokens + prompt, and optionally calling some external tool. It is not really reasoning; what’s doing is not too different in spirit from Markov chains, except more complex.

    [no true Scotsman] LLM ≠ AI system

    If large “language” models don’t count as “AI systems”, then what you shared in the OP does not either. You can’t eat your cake and have it too.

    It’s been known for fome time, that LLMs do “vibe math”.

    I.e. they’re unable to perform actual maths.

    [moving goalposts] Internally, they try to come up with an answer that “feels” right…

    It doesn’t matter if the answer “feels” right (whatever this means). The answer is incorrect.

    which makes it pretty impressive for them to come anywhere close, within a ±10% error margin.

    No, the fact they are unable to perform a simple logical procedure is not “impressive”. Specially not when outputting the “approximation” as if it was the true value; note how none of the models outputted anything remotely similar to “the result is close to $number” or “the result is approximately $number”.

    [arbitrary restriction + whataboutism] Ask people to tell you what a right answer could be, give them 1 second to answer… see how many come that close to the right one.

    None of the prompts had a time limit. You’re making shit up.

    Also. Sure, humans brainfart all the time; that does not magically mean that those systems are smart or doing some 4D chess as your OP implies.

    A chatbot/AI system on the other hand, will come up with some Python code to do the calculation, then run it. Still can go wrong, but it’s way less likely.

    I.e. it would need to use some external tool, since it’s unable to handle logic by itself, as exemplified by maths.

    all explanation past the «are you counting the “rr” as a single r?» is babble

    Not so sure about that. It treats r as a word, since it wasn’t specified as “r” or single letter. Then it interpretes it as… whatever. Is it the letter, phoneme,

    The output is clearly handling it as letters. It hyphenates the letters to highlight them, it mentions “digram” (i.e. a sequence of two graphemes), so goes on. And in no moment is referring to anything that can be understood as associated with sounds, phonemes. And it’s claiming there’s an ⟨r⟩ «in the middle of the “rr” combination».

    font, the programming language R…

    There’s no context whatsoever to justify any of those interpretations.

    since it wasn’t specified, it assumes “whatever, or a mix of”.

    If this was a human being, it would not be an assumption. Assumption is that sort of shit you make up from nowhere; here context dictates the reading of “r” as “the letter ⟨r⟩”.

    However since this is a bot it isn’t even assuming. Just like a boulder doesn’t “assume” you want it to roll down; it simply reacts to an external stimulus.

    It failed at detecting the ambiguity and communicating it spontaneously, but corrected once that became part of the conversation.

    There’s no ambiguity in the initial prompt. And no, it did not correct what it says; the last reply is still babble, you don’t count ⟨rr⟩ in English as a single letter.

    It’s like, in your examples… what do you mean by “by”? “3 by 6” is 36… you meant to “multiply 36”? That’s nonsense… 🤷

    I’d rather not answer this one because, if I did, I’d be pissing on Beehaw’s core values.


  • Wrong maths, you say?

    Anyway. You didn’t ask the number of times the phoneme /ɹ/ appears in the spoken word, so by context you’re talking about the written word, and the letter ⟨r⟩. And the bot interpreted it as such, note it answers

    here, let me show you: s-t-r-a-w-b-e-r-r-y

    instead of specifying the phonemes.

    By the way, all explanation past the «are you counting the “rr” as a single r?» is babble.