I’m puzzled about this, the responses seemed fine. Like, one guy is saying his dumb purchase was a few thousand dollars on a desk. Unless he makes that mistake every week: so what? Look at 90% of F-150 sales.
I’m puzzled about this, the responses seemed fine. Like, one guy is saying his dumb purchase was a few thousand dollars on a desk. Unless he makes that mistake every week: so what? Look at 90% of F-150 sales.
We don’t talk enough about how morally depraved the tech industry turned out to be. Every single ounce of their self-regarding statements of values was an outright lie.
Sir, can I interest you in a discussion on land value tax.
Turn them all into dispatchable data centers, EZ.
what's kinda fun is to work out how fast a gasoline fill up is in terms of megawatts. You have to do a few corrections to your taste to deal with how much is lost to, say, heat, rather than useful work, but it is measured in megawatts. It's about 20MW gross and maybe 4-6MW net (over the few minutes)
“No matter what happens, keep this in mind: It’s the same old thing, from one end of the world to the other. It fills the history books, ancient and modern, and the cities, and the houses too. Nothing new at all.”
But so what? I don’t expect 100% of the text I read out of Wikipedia to be relevant to my task either. My contrary take is that LLMs will help with critical thinking: now using a GRE word means nothing. You can’t lean on proxy for quality or accuracy of an idea.
Chart interfaces is an area where skipping the diving into the manual almost always helps, unless you are writing a charting library, then you have to be more careful: there’s presumably something about not yet invented, and its unorthodoxy will likely result in some unhelpful model output.
The hype people are setting the bar way too high and say weird stuff, but automating away three days of drudgery per annum is a big deal.
I’m still harvesting slight year over year productivity gains from mobile broadband and wearables. If LLMs spare three days of drudgery a year for the next ten years, that’s 1% increase in productivity from one family of technique. That’s stellar. Yet so modest in any given year.
The "median voter" theory of Democratic renewal really needs to have an account for what must be done about right wing propaganda's impact on that median voter. Via @brianbeutler.bsky.social:
www.offmessage.net/p/jasmine-cr...
kind of crazy that if you have enough money and don't like what you see in the media, you can just buy up every film studio, news station, and social media app and change it
“A rat’s anus?”
I mean, I’m using Claude code too and it’s very clearly like that!
My thought is, accumulation of productivity will be gradual: fewer design errors, longer lasting designs from deeper research. But program interfaces exposed to the greater public will change and be more programmable since a critical thinker without time for the manual can compose what they want.
As it comes to programs in a long continuity and high levels of assurance , I’ve found AI more useful for prototyping. This is still a significant research function, to study several bigger and more complete prototypes to assess impact in the materialized text.
This is not a production program, stakes are low, but it allows us to more thoroughly research our problems. We can move with a greater volume of data and research.
That’s true for us too though: we have many adjacent areas of software engineering where manual diving dominates. Even some that exist in our area of specialty, for example, whipping up a quick script to navigate /proc in Linux and sum some counters
Yeah I saw it as it went by but didn’t zoom in on the kv cache part.
Now that’s a cool experiment. Any idea how it’s able to do that without materializing tokens first to realize it?
You can even approach this from an information-theoretical point of view: the largest open weight models are about a terabyte, wikipedia is maybe 150GB. No wonder it's so good at knowing most of the spells in Harry Potter, you know?
Anyway, I do not use the Wikipedia analogy completely lightly: it’s a non-consensual version, where my works were basically turned into statistical slaw without my permission and occupy some space. But it’s effectively crowdsourced, a bit like Wikipedia.
Other fun tricks: getting better results by having it draw a diagram in ascii first, where you suspect the translation to ascii will prove accurate and then it can easily spot the problem. Many variants of this, when you suspect an intermediate representation will help it enter the right space.
…but if you mention “statistics” or “regression” it’ll start making a python program with regurgitated scipy to whip up some regressions. Neat! But you had to suspect it would not find the right latent space for the job.
Other times, unless you are strategic, it’ll fail to connect disciplines. If you ask “do you think X about this data” having not put it the word “statistics” at at any point in the debug dump, it’ll just kind of vibe some words out.
I was just debugging something today that yielded number of preposterous suggestions that clearly are something like regression to the mean / attraction to the training set. That’s okay, it’s just very obvious it’s happening and when to ignore or what to put in to correct.
Well, I also do much of my coding (you could call it keyboarding) by AI but it’s absolutely routine that I find a number of critical defects in it. But I know to look for them because it’s a common type of defect. Try getting an AI to one shot a concurrent algorithm not straight from a textbook.
I’m a little puzzled but this. I’m not a social scientist. I write software. LLMs are great, I use them a lot, but it’s very clearly wandering around in its Wikipedia-like latent space. This falls far short of being a good idea to deploy to production. Are our arts so different?
“We’re on a mission from God”