Need to let loose a primal scream without collecting footnotes first? Have a sneer percolating in your system but not enough time/energy to make a whole post about it? Go forth and be mid: Welcome to the Stubsack, your first port of call for learning fresh Awful you’ll near-instantly regret.

Any awful.systems sub may be subsneered in this subthread, techtakes or no.

If your sneer seems higher quality than you thought, feel free to cut’n’paste it into its own post — there’s no quota for posting and the bar really isn’t that high.

The post Xitter web has spawned soo many “esoteric” right wing freaks, but there’s no appropriate sneer-space for them. I’m talking redscare-ish, reality challenged “culture critics” who write about everything but understand nothing. I’m talking about reply-guys who make the same 6 tweets about the same 3 subjects. They’re inescapable at this point, yet I don’t see them mocked (as much as they should be)

Like, there was one dude a while back who insisted that women couldn’t be surgeons because they didn’t believe in the moon or in stars? I think each and every one of these guys is uniquely fucked up and if I can’t escape them, I would love to sneer at them.

(Credit and/or blame to David Gerard for starting this.)

  • aio@awful.systems
    link
    fedilink
    English
    arrow-up
    0
    ·
    edit-2
    1 day ago

    If the growth is superexponential, we make it so that each successive doubling takes 10% less time.

    (From AI 2027, as quoted by titotal.)

    This is an incredibly silly sentence and is certainly enough to determine the output of the entire model on its own. It necessarily implies that the predicted value becomes infinite in a finite amount of time, disregarding almost all other features of how it is calculated.

    To elaborate, suppose we take as our “base model” any function f which has the property that lim_{t → ∞} f(t) = ∞. Now I define the concept of “super-f” function by saying that each subsequent block of “virtual time” as seen by f, takes 10% less “real time” than the last. This will give us a function like g(t) = f(-log(1 - t)), obtained by inverting the exponential rate of convergence of a geometric series. Then g has a vertical asymptote to infinity regardless of what the function f is, simply because we have compressed an infinite amount of “virtual time” into a finite amount of “real time”.

    • scruiser@awful.systems
      link
      fedilink
      English
      arrow-up
      0
      ·
      9 hours ago

      Yeah AI 2027’s model fails back of the envelope sketches as soon as you try working out any features of it, which really draws into question the competency of it’s authors and everyone that has signal boosted it. Like they could have easily generated the same crit-hype bullshit with “just” an exponential model, but for whatever reason they went with this model. (They had a target date they wanted to hit? They correctly realized adding in extraneous details would wow more of their audience? They are incapable of translating their intuitions into math? All three?)