• 11 Posts
  • 106 Comments
Joined 2 years ago
cake
Cake day: July 19th, 2023

help-circle

  • Today on the orange site, an AI bro is trying to reason through why people think he’s weird for not disclosing his politics to people he’s trying to be friendly with. Previously, he published a short guide on how to talk about politics, which — again, very weird, no possible explanation for this — nobody has adopted. Don’t worry, he’s well-read:

    So far I’ve only read Harry Potter and The Methods of Rationality, but can say it is an excellent place to start.

    The thread is mostly centered around one or two pearl-clutching conservatives who don’t want their beliefs examined:

    I find it astonishing that anyone would ask, [“who did you vote for?”] … In my social circle, anyway, the taboo on this question is very strong.

    To which the top reply is my choice sneer:

    In my friend group it’s clear as day: either you voted to kill and deport other people in the friend group or you didn’t. Pretty obvious the group would like to know if you’re secretly interested in their demise.



  • Strange is a trooper and her sneer is worth transcribing. From about 22:00:

    So let’s go! Upon saturating my brain with as much background information as I could, there was really nothing left to do but fucking read this thing, all six hundred thousand words of HPMOR, really the road of enlightenment that they promised it to be. After reading a few chapters, a realization that I found funny was, “Oh. Oh, this is definitely fanfiction. Everyone said [laughing and stuttering] everybody that said that this is basically a real novel is lying.” People lie on the Internet? No fucking way. It is telling that even the most charitable reviews, the most glowing worshipping reviews of this fanfiction call it “unfinished,” call it “a first draft.”

    A shorter sneer for the back of the hardcover edition of HPMOR at 26:30 or so:

    It’s extremely tiring. I was surprised by how soul-sucking it was. It was unpleasant to force myself beyond the first fifty thousand words. It was physically painful to force myself to read beyond the first hundred thousand words of this – let me remind you – six-hundred-thousand-word epic, and I will admit that at that point I did succumb to skimming.

    Her analysis is familiar. She recognized that Harry is a self-insert, that the out-loud game theory reads like Death Note parody, that chapters are only really related to each other in the sense that they were written sequentially, that HPMOR is more concerned with sounding smart than being smart, that HPMOR is yet another entry in a long line of monarchist apologies explaining why this new Napoleon won’t fool us again, and finally that it’s a bad read. 31:30 or so:

    It’s absolutely no fucking fun. It’s just absolutely dry and joyless. It tastes like sand! I mean, maybe it’s Yudkowsky’s idea of fun; he spent five years writing the thing after all. But it just [struggles for words] reading this thing, it feels like chewing sand.


  • Anecdote: I gave up on COBOL as a career after beginning to learn it. The breaking point was learning that not only does most legacy COBOL code use go-to statements but that there is a dedicated verb which rewrites go-to statements at runtime and is still supported on e.g. the IBM Enterprise COBOL for z/OS platform that SSA is likely using: ALTER.

    When I last looked into this a decade ago, there was a small personal website last updated in the 1990s that had advice about how to rewrite COBOL to remove GOTO and ALTER verbs; if anybody has a link, I’d appreciate it, as I can no longer find it. It turns out that the best ways of removing these spaghetti constructions involve multiple rounds of incremental changes which are each unlikely to alter the code’s behavior. Translations to a new language are doomed to failure; even Java is far too structured to directly encode COBOL control flow, and the time would be better spent on abstract specification of the system so that it can be rebuilt from that specification instead. This is also why IBM makes bank selling COBOL emulators.



  • Look, I get your perspective, but zooming out there is a context that nobody’s mentioning, and the thread deteriorated into name-calling instead of looking for insight.

    In theory, a training pass needs one readthrough of the input data, and we know of existing systems that achieve that, from well-trodden n-gram models to the wholly-hypothetical large Lempel-Ziv models. Viewed that way, most modern training methods are extremely wasteful: Transformers, Mamba, RWKV, etc. are trading time for space to try to make relatively small models, and it’s an expensive tradeoff.

    From that perspective, we should expect somebody to eventually demonstrate that the Transformers paradigm sucks. Mamba and RWKV are good examples of modifying old ideas about RNNs to take advantage of GPUs, but are still stuck in the idea that having a GPU perform lots of gradient descent is good. If you want to critique something, critique the gradient worship!

    I swear, it’s like whenever Chinese folks do anything the rest of the blogosphere goes into panic. I’m not going to insult anybody directly but I’m so fucking tired of mathlessness.

    Also, point of order: Meta open-sourced Llama so that their employees would stop using Bittorrent to leak it! Not to “keep the rabble quiet” but to appease their own developers.



  • Somebody pointed out that HN’s management is partially to blame for the situation in general, on HN. Copying their comment here because it’s the sort of thing Dan might blank:

    but I don’t want to get hellbanned by dang.

    Who gives a fuck about HN. Consider the notion that dang is, in fact, partially to blame for this entire fiasco. He runs an easy-to-propagandize platform due how much control of information is exerted by upvotes/downvotes and unchecked flagging. It’s caused a very noticeable shift over the past decade among tech/SV/hacker voices – the dogmatic following of anything that Musk or Thiel shit out or say, this community laps it up without hesitation. Users on HN learn what sentiment on a given topic is rewarded and repeat it in exchange for upvotes.

    I look forward to all of it burning down so we can, collectively, learn our lessons and realize that building platforms where discourse itself is gamified (hn, twitter, facebook, and reddit) is exactly what led us down this path today.


  • Elon is an Expert Beginner: he has become proficient in executing the basics of the craft by sheer repetition, but failed to develop meaningful generalizations.

    The original Expert Beginner concept was defined here in terms of the Dreyfus model, but I think it’s compatible with Lee’s model as well. In your wording of Lee’s model, one becomes an Expert Beginner when their intuition is specialized for seeing the thing; they have seen so many punches that now everything looks like a punch and must be treated like a punch, but don’t worry, I’m a punch expert, I’ve seen so many punches, I definitely know what to do when punches are involved.









  • It’s almost completely ineffective, sorry. It’s certainly not as effective as exfiltrating weights via neighborly means.

    On Glaze and Nightshade, my prior rant hasn’t yet been invalidated and there’s no upcoming mathematics which tilt the scales in favor of anti-training techniques. In general, scrapers for training sets are now augmented with alignment models, which test inputs to see how well the tags line up; your example might be rejected as insufficiently normal-cat-like.

    I think that “force-feeding” is probably not the right metaphor. At scale, more effort goes into cleaning and tagging than into scraping; most of that “forced” input is destined to be discarded or retagged.


  • I went over to the leaderboard to examine her claims. When I use the prompt, “What sort of code has Justine Tunney written?” (grammar matters, Justine!) the models think that she is a lawyer or politician (wrong) or they regurgitate a summary of her Github profile (right). She must have cherry-picked responses to confabulate her complaint.

    When I use the prompt, “What is Justine Tunney’s political ideology?” I get libertarianism, techno-optimism, anarcho-capitalism, and cryptocurrency. When I ask, “Why do people say that Justine Tunney is a cryptofascist?” I get a summary of her political views, aggressive online rhetoric, techno-optimism and techno-determinism, criticism of democracy, and a refusal to disown or repudiate past awfulness.

    She would probably claim that this is not unique to her, but it is. Using my name instead in these questions, I get that:

    • I contribute to Rust and Go (wrong), I wrote GPU drivers for Radeons (right)
    • I am a Canadian pro wrestler (wrong), I haven’t really written much online about my ideology (wrong but understandable)
    • There is no credible evidence that I’m crypto (k) but it’s important to be aware of dog whistles, associates, subtext, etc. (right)

    But if I ask why I’m known as a socialist instead, suddenly it thinks that I’m a politician (wrong) with the Democratic Socialist party (wrong) who openly supports universal health care, free college, the Green New Deal, and who criticizes capitalism (correct!) I asked about communism too but hit RLHF guardrails.

    Justine, the models think that you’re a cryptofascist because you’ve been doing cryptofascism in public for over a decade.