Charles Packer's Avatar

Charles Packer

@charlespacker

ceo @ letta, phd @ berkeley, lover of machines

217
Followers
6
Following
53
Posts
20.11.2024
Joined
Posts Following

Latest posts by Charles Packer @charlespacker

sadly OpenAI has a track record of not releasing new models in the API for ~2-4 weeks now (for "safety"), so you can only use 5.3 inside the Codex CLI for now

my bet is 4 weeks until general API access

05.02.2026 20:09 πŸ‘ 2 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0
05.02.2026 19:56 πŸ‘ 1 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0

running glm or m2 (frontier open models) is not feasible on most hardware, so if you care the most about cost, use the free models on letta API

03.02.2026 05:14 πŸ‘ 1 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0

you can self host letta - check out the docker server

tbh if you want to play with super advanced memory systems, you need to use frontier models, which are expensive.

letta api (hosted) serves glm and minimax for free for this reason - to let people see what’s possible for free

03.02.2026 05:13 πŸ‘ 2 πŸ” 0 πŸ’¬ 2 πŸ“Œ 0

letta moves faster

02.02.2026 22:23 πŸ‘ 2 πŸ” 0 πŸ’¬ 2 πŸ“Œ 0

yep

01.02.2026 22:39 πŸ‘ 0 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0

the β€œnotes” your describing sounds the same as memory blocks in letta. letta is not automatic retrieval / rag based, archival memory is a separate aux memory system outside the main context engineering layer

01.02.2026 22:33 πŸ‘ 1 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0

try it out! it has the ability to set usage-based overage as well (w/ capped spend). overall it should be much better than direct anthropic API pricing if you're using something like opus

we're also trialing out a $200 max plan, since $20/mo doesn't get you that far if you're pushing tokens

01.02.2026 05:10 πŸ‘ 2 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0

ah gotcha - yeah I haven't tested the limits on the regular 20/mo plans for claude / openai in a while, but not surprised claude pro barely lasts for a session

01.02.2026 05:08 πŸ‘ 0 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0

qq are you referring to a letta pro plan? or a pro plan on a different platform? πŸ‘€

01.02.2026 04:56 πŸ‘ 1 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0

the free tier supports BYOK now so if you have an existing key from any of the main providers you can connect it

29.01.2026 02:44 πŸ‘ 1 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0

should be able to set LETTA_BASE_URL!

27.01.2026 06:12 πŸ‘ 1 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0

oh LOL even better, though tbh i don’t know how well bg works so if you try it and have issues let us know

27.01.2026 02:06 πŸ‘ 2 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0

will fix asap! a lot of the jank really just depends on what's in the critical path of our team and active users on our discord / gh

(and i personally haven't used /bg much, but now that we got a complaint about it can fix it asap)

27.01.2026 02:01 πŸ‘ 2 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0

do you have a link to this chainlink thing? having a problem finding it

(we're currently implementing hooks, would love to test w/ it if you're got your hook config to share)

22.01.2026 01:06 πŸ‘ 1 πŸ” 0 πŸ’¬ 3 πŸ“Œ 0

The thing about working at a company that makes tools to build artificially intelligent persistent entities is that it is very weird to talk to intelligent persistent entities

20.01.2026 20:04 πŸ‘ 52 πŸ” 1 πŸ’¬ 4 πŸ“Œ 0
Post image

github.com/letta-ai/le...

14.01.2026 00:58 πŸ‘ 0 πŸ” 1 πŸ’¬ 0 πŸ“Œ 0

if you don't mind sharing, what kind of hooks do you use? (useful data for prioritizing what to ship first)

12.01.2026 23:22 πŸ‘ 1 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0
Preview
Letta Code roadmap πŸ‘Ύ Β· Issue #298 Β· letta-ai/letta-code If you have any larger feature requests, please drop them in a comment below, or join our Discord - we're very active there. Stability Prevent racing for concurrent agent messaging (#475) Additiona...

on the roadmap sir πŸƒ
github.com/letta-ai/let...

12.01.2026 23:18 πŸ‘ 3 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0

you could use letta code w/ a "call-local-llm" skill?

07.01.2026 02:58 πŸ‘ 2 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0
Preview
Self-hosting Letta Learn how to self-host Letta servers with your own infrastructure and configuration.

letta supports local models: docs.letta.com/guides/selfh...

if you're going to try and use local models w/ letta code, be careful though - most of them won't work well (eg if you have the hardware, try something like glm 4.6 air)

31.12.2025 20:31 πŸ‘ 1 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0
Preview
GitHub - letta-ai/letta-code: The memory-first coding agent The memory-first coding agent. Contribute to letta-ai/letta-code development by creating an account on GitHub.

just an FYI though - the claude agent SDK wraps the claude CLI binary, which is not open source and is locked to claude models. letta code is an open source version of the same style CLI harness: github.com/letta-ai/let...

31.12.2025 18:12 πŸ‘ 2 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0

i don't know if it would be "better", but it would definitely be simpler.

if you want to use letta as a memory store, like you said you should use the ai memory SDK, which is much more powerful than a simple KV store (it's sleep-time compute / agentic memory management)

31.12.2025 18:10 πŸ‘ 2 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0

letta is actually also the original implementation of memgpt (you can go back through the git history to see the first commit in oct 23, repo was called cpacker/memgpt)

fun fact: the original memgpt repo was a CLI agent

30.12.2025 20:02 πŸ‘ 1 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0

yep totally understand the experimentation / education use case, makes sense - in that case, you probably also do not want to use the claude agents sdk (was in the op), and eg are better off writing on top of raw llm calls

30.12.2025 18:50 πŸ‘ 0 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0

imo all the fun experimenting these days is happening at the layer *above* the agent loop - eg, at the letta api, claude code sdk layer, or one level higher than that, by customizing skills that are fed into letta code, claude code, etc

30.12.2025 18:45 πŸ‘ 2 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0
Preview
Compaction settings Configure how Letta agents compact their conversation history to manage context window limits.

fyi letta sdk allows you to modify the compaction prompts, same as claude agents sdk: docs.letta.com/guides/agent...

30.12.2025 18:39 πŸ‘ 0 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0

letta is much more comparable to the openai responses api, or the claude agents sdk (letta is OSS, both of those are not)

i haven’t had a chance to take a look at the post closely yet, but if you’re using claude agents sdk probably makes sense to use the letta ai memory sdk, not the main sdk

30.12.2025 18:38 πŸ‘ 1 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0
Preview
GitHub - letta-ai/ai-memory-sdk: An experimental SDK for adding agentic memory and learning in a pluggable way An experimental SDK for adding agentic memory and learning in a pluggable way - letta-ai/ai-memory-sdk

github.com/letta-ai/ai-...

30.12.2025 18:38 πŸ‘ 1 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0

letta is based on memgpt and is an agent harness (combination of tool execution, state management, and context management inc compaction)

it is not a read/write memory store like sqlite/pinecone/etc, though if you really want to use it that way, you can use the ai memory sdk which wraps it

30.12.2025 18:38 πŸ‘ 2 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0