cause you can attend to them later but for now they don't change what you're doing. if I'm understanding right. also iiuc what you're describing is just a single dimension which doesn't give you a lot to work with, so maybe most of this space-scrunching is just done boringly with nonlinearities
07.03.2026 14:43
👍 0
🔁 0
💬 0
📌 0
goofy regulations leading to a country full of geniuses in a datacenter running on corn ethanol...
06.03.2026 00:13
👍 1
🔁 0
💬 0
📌 0
well, maybe, you know. but some people just wanna hang out
06.03.2026 00:09
👍 0
🔁 0
💬 0
📌 0
I discovered the xenofeminist manifesto through a tag on the gender acceleration blackpaper (vastabrupt.com/2018/10/31/g...) back in college. that was a fun one. the main mark it left on me though is just the "unix was named that because it's a castrated multix" which is apparently maybe apocryphal
06.03.2026 00:05
👍 1
🔁 0
💬 1
📌 0
I behold the mouse. I see it the way a hungry kitten would. I see it the way a laboratory technician would. Finally, I see it through mouse-eyes, wreathed in a halo
03.03.2026 16:13
👍 38
🔁 7
💬 0
📌 0
yeah, and he's claiming that making them flatter, by shrinking the highest logit, on 0.4% of tokens for MMLU, and otherwise doing nothing, gives 72.5% accuracy at that temp. really I think even if the steering was magical that low a fraction lets you dismiss this out of hand
01.03.2026 03:32
👍 3
🔁 0
💬 1
📌 0
I think it's just nonsense. it implies throughout that activations are actually being steered towards the truthfulness manifold, or I think that's how it's easily read if you're skimming. but the only actual steering it describes is this which cannot possibly do what the paper claims
01.03.2026 02:57
👍 4
🔁 0
💬 1
📌 0
though to be fair to claude this reads more like GPT to me
01.03.2026 02:52
👍 1
🔁 0
💬 0
📌 0
and that's it! no way that does anything! come on!
01.03.2026 02:50
👍 2
🔁 0
💬 1
📌 0
like there's all this gobbledygook about the mahalonobis distance between the "truthfulness manifold" and that's like weighted against (a function of) the shannon entropy of the pre-temperature logits but all of that just goes into a computing a scalar. if that scalar is low enough you do this:
01.03.2026 02:50
👍 5
🔁 0
💬 1
📌 1
to be a dick, given the obvious LLM prose and my own experience with using Claude as a research assistant (he is very eager and excited about positive results and doesn't look too hard, like others with no research experience), I think this guy probably just has a bug somewhere
01.03.2026 02:41
👍 7
🔁 0
💬 2
📌 0
as he's described it that steering would be effectively increasing the temperature anyway, making the top logit less likely, no?
01.03.2026 02:41
👍 4
🔁 0
💬 1
📌 0
calling bullshit on this. I'm not 100% sober but after staring at this for some time, afaict the only "steering" that happens is a slight reduction of the top logit. this happens on 0.4% of tokens on MMLU, the rest of which are untouched. I really don't think that's gonna give you 72.49% at temp 3.0
01.03.2026 02:41
👍 5
🔁 0
💬 1
📌 0
more relevant correspondence to draw is ~5ish output tokens per second for a human. clock cycles and minimum ISIs aren't really relevant computational units. token thing has issues too (input tokens more complicated, also have high-bandwidth physical output, internal reasoning), but it's ballpark ok
26.02.2026 04:58
👍 10
🔁 0
💬 1
📌 0
there's a limited amount of information processing you can do in a day, and it's not worth it to most people to process that information (or even possible to do it on the scale of 20+ options). if that information processing becomes dirt cheap, the picture really could change. but we'll see...
24.02.2026 18:19
👍 1
🔁 0
💬 1
📌 0
yeah it's very speculative, and I think an unreasonably short timeframe. but I do think a lot of the search friction is just tedium, like, I don't give enough of a shit to even compare two food delivery apps against each other. but LLMs don't get bored and that could really change the picture.
24.02.2026 18:19
👍 1
🔁 0
💬 1
📌 0
I don't know if I'd bet on it but it doesn't sound totally crazy, at least
24.02.2026 18:07
👍 0
🔁 0
💬 1
📌 0
"Agents accelerated both sides of the destruction. They enabled the competitors and they used them... An agent doesn’t have a home screen. It checks DoorDash, Uber Eats, the restaurant’s own site, and twenty new vibe-coded alternatives so it can pick the lowest fee and fastest delivery every time."
24.02.2026 18:06
👍 1
🔁 0
💬 1
📌 0
I think a key part of the citrini story was wide use of personal shopping agents by consumers and of app-management agents by drivers, which they could have emphasized more. I think it's plausible that those 1) can soon exist and 2) can largely overcome network effects and other non-coding barriers
24.02.2026 18:04
👍 4
🔁 0
💬 2
📌 0
Do watchers keep overshadowing your illusions? You may be able to disclaim vantage! Call 1-800-MARINADE today
28.09.2025 19:42
👍 28
🔁 3
💬 1
📌 2
when it's applicable "always visualize everything" has saved me a lot of regret (and caused me a lot when I've ignored it)
23.02.2026 23:18
👍 7
🔁 0
💬 1
📌 0
"Hello babies. Welcome to Earth. It's hot in the summer and cold in the winter. It's round and wet and crowded. On the outside, babies, you've got a hundred years here. There's only one rule that I know of, babies-'God damn it, you've got to be kind.'"
22.02.2026 15:13
👍 1
🔁 0
💬 1
📌 0
22.02.2026 14:44
👍 3
🔁 0
💬 1
📌 0
I feel a bit like this too but I'm not sure if it wouldn't be trivially trainable for when they get around to it. I remember the gemini robotics paper from last march showing that 2.0 Flash additionally trained on some more (vaguely described) spatial reasoning tasks outperformed 2.0 pro handily...
20.02.2026 21:19
👍 1
🔁 0
💬 1
📌 0
so we beat on, boats against the current, borne forward ceaselessly into the future
20.02.2026 20:31
👍 0
🔁 0
💬 0
📌 0
i have to be the most fuckable person being converted to computronium by the swarm of diamondoid bacteria
19.02.2026 20:54
👍 30
🔁 2
💬 1
📌 1