• 0 Posts
  • 5.99K Comments
Joined 2 years ago
cake
Cake day: August 27th, 2023

help-circle











  • Please don’t mistake vindication for a lack of ambiguity. When this took off, we had no goddamn idea what the limit was. The fact it works half this well is still absurd.

    Simple examples like addition were routinely wrong, but they were wrong in a way that indicated - the model might actually infer the rules of addition. That’s a compact way to predict a lot of arbitrary symbols. Seeing that abstraction emerge would be huge, even if it was limited to cases with a zillion examples. And it was basically impossible to reason about whether that was pessimistic or optimistic.

    A consensus for “that doesn’t happen” required all of this scholarship. If we had not reached this point, the question would still be open. Remove all the hype from grifters insisting AGI is gonna happen now, oops I mean now, oops nnnow, and you’re still left with a series of advances previously thought impossible. Backpropagation doesn’t work… okay now it does. Training only plateaus… okay it gets better. Diffusion’s cute, avocado chairs and all, but… okay that’s photoreal video. It really took people asking weird questions on high-end models to distinguish actual reasoning capability from extremely similar sentence construction.

    And if we’re there, can we please have models ask a question besides ‘what’s the next word?’






  • Charles Babbage was once asked, ‘But if someone puts in the numbers wrong, how will your calculator get the right answer?’

    Using a chatbot to code is useful if you don’t know how to code. You still need to know how to chatbot. You can’t grunt at the machine and expect it to read your mind.

    Have you never edited a Google search, because the first try didn’t work?


  • This kind of assertion wildly overestimates how well we understand intelligence.

    Higher levels of bullshitting require more abstraction and self-reference. Meaning must be inferred from observation, to make certain decisions, even when picking words from a list.

    Current models are abstract enough to see a chessboard in an Atari screenshot, figure out which pieces each jumble of pixels represents, and provide a valid move. Scoffing because it’s not actually good at chess is a bizarre line to draw, to say there’s zero understanding involved.

    Current models might be abstract enough to teach them a new game by explaining the rules.

    Current models are not abstract enough to explain why they’re bad at a game and expect them to improve.