About us

Industry
Software Development
Company size
2-10 employees
Type
Privately Held

Employees at Letta

Updates

  • Letta reposted this

    View profile for Charles Packer, graphic

    Co-Founder & CEO at Letta

    Excited to see Letta in TechCrunch's most disruptive startups of 2024 list 👀 Memory is the most important problem in AI today. LLMs are the new building block for AI systems, but LLMs are inherently stateless compute units - we need fundamental advancements in memory systems to get to anything resembling AGI. 👾 If you're interested in advancing the frontier of memory for AI, reach out! We're hiring across product, infra, and research! (link in comments) Note: our office is in SF, not Berkeley ;)

    • Letta featured in TechCrunch's "most disruptive startups of 2024"
    • No alternative text description for this image
  • View organization page for Letta, graphic

    1,569 followers

    The AI agents stack in late 2024, organized into three key layers: agent hosting/serving, agent frameworks, and LLM models & storage. Read more at https://lnkd.in/dP7JAzFr

    View profile for Charles Packer, graphic

    Co-Founder & CEO at Letta

    Introducing the AI agents stack: breaking down today’s tech stack for building AI agents into three key layers: (1) agent hosting/serving, (2) agent frameworks, (3) LLM models & storage. Sarah Wooders and I got so tired of seeing bad “market maps” for LLM / AI agents shoved in our feeds that either had layouts that made no sense or were littered with random companies on them (ie that don’t have serious community adoption), or both. As researchers / engineers actually working in the agents space, we decided to make a serious attempt at making one of these market map diagrams that actually reflects real world usage by today’s developers building AI agents. Basically - if you’re starting a vertical agents company today (November 2024), what software are you most likely to use to build out your “agents stack”? In our opinion, the AI/LLM agents stack is a significant departure from the standard LLM stack. The key difference between the two lies in managing state: LLM serving platforms are generally stateless, whereas agent serving platforms need to be stateful (retain the state of the agent server-side). Building stateful services is a lot harder of an engineering challenge compared to building developer SDKs, so unsurprisingly very few agent serving platforms actually exist today (Letta being one of them). For a full breakdown of the stack, check out our full post (link in comments).

    • The AI agents stack in late 2024, organized into three key layers: agent hosting/serving, agent frameworks, and LLM models & storage.
  • Letta reposted this

    View profile for Charles Packer, graphic

    Co-Founder & CEO at Letta

    The death of scaling laws 📈 ☠ ⁉ Over the past week there’s been a lot of chatter around “the death of scaling laws”: we have leaks from various sources coming out saying that efforts at frontier labs to scale base models are stalling, meanwhile Sam and Dario are making the rounds on podcasts "claiming AGI by 2025/2026". My take on this as someone that's been working on "agents" research for 7+ years (previously RL + LLM agents at Berkeley / now exclusively LLM agents at Letta): it’s been clear that traditional scaling laws (more parameters + more data = 🧠 📈 ) have been dead for a while now. Look at the trajectory of public model releases from OpenAI - whereas GPT-3.5 to GPT-4 was a clear increase in capabilities (yay scaling laws 🎉 ), all model releases beyond GPT-4 have existing along some tradeoff curve: - GPT-3.5 → GPT-4 (huge bump in reasoning performance*) - GPT-4 → GPT-4-Turbo (worse reasoning, better speed) - GPT-4-Turbo → GPT-4o (back to ~GPT-4 reasoning though slightly worse in many tasks, but better multi-modal support) - GPT-4o → o1 (”better” than GPT-4 in some tasks, though trading off w/ >10s-1m+ response times - in many cases GPT-4 is a better user experience) *note: “performance” here is my own personal take as someone who spends their day-to-day getting LLMs to work as agents (not referring to just public benchmarks), YMMV That being said, I strongly disagree that the death of (traditional) scaling laws implies that AI has “hit a wall” or that AI progress is about to stall. As OpenAI showed with o1, there’s a lot more you can do to improve AI beyond just scaling LLM parameters + data. To date the majority of progress in LLM-based intelligence has come from the base models - as this progress slows, the focus of researchers and engineers will shift to the layer directly above it (the LLM OS / context management layer). Relatively little focus went into the system building over the past ~2 years because of lack of stability - it’s incredibly hard to build systems around LLMs (in particular choosing the right abstractions) when the capabilities of the base models is changing so quickly. Now that LLMs are becoming more of a known quantity (we understand their limitations and their strengths as auto-regressive token predictors) - it’s off to the races to build the full agentic system around the LLM. I would love to be wrong - if GPT-5 exists it would be an amazing early Christmas gift (please drop the weights via magnet link -- thank you in advance Sam 🙏🎄)

  • Letta reposted this

    View profile for Charles Packer, graphic

    Co-Founder & CEO at Letta

    Q: How do we get from LLMs-as-chatbots to LLMs-as-agents? A: Programatic memory (context) management through an LLM OS. We're super excited that our (free!) course on LLMs as OS with Andrew Ng and DeepLearning.AI is finally out!! Andrew is one of the “godfathers of modern AI” and has turned DLAI into the de facto platform for learning about the state-of-the-art LLM technology (everything spanning research to new frameworks). Sarah Wooders and I strongly believe the next wave of advancements in AI/agents will come from memory management, but despite our conviction there’s not much material online to help developers and builders understand (1) the importance of memory and (2) how to add memory management to their LLM systems to create better agents. The biggest problem today with LLM agents today is memory. Not just “memory” in a semantic sense (how can we get LLM agents to remember facts+preferences over time, similar to how a human does?), but also memory management - what tokens do we put into the context window at each LLM inference step, and why? The memory management (or “context management”) problem is the fundamental problem in programming LLMs to turn them from autocomplete engines into compound agentic systems that can interact with the world and learn from experience. JSON mode, tool use / function calling, RAG, chain-of-thought: these are all early forays into building the “LLM OS” for context management. We’re still extremely early in the history of LLM OS development. In our DLAI course, we distill the main ideas behind what it even means to do “LLM memory management” into clean and concise example and guide you through building a version of MemGPT (one of the early examples of an LLM OS) yourself, entirely from scratch.

    View profile for Andrew Ng, graphic
    Andrew Ng Andrew Ng is an Influencer

    Founder of DeepLearning.AI; Managing General Partner of AI Fund; Exec Chairman of Landing AI

    New short course: LLMs as Operating Systems: Agent Memory, created with Letta, and taught by its founders Charles Packer and Sarah Wooders. An LLM's input context window has limited space. Using a longer input context also costs more and results in slower processing. So, managing what's stored in this context window is important. In the innovative paper "MemGPT: Towards LLMs as Operating Systems," its authors (which include Charles and Sarah) proposed using an LLM agent to manage this context window. Their system uses a large persistent memory that stores everything that could be included in the input context, and an agent decides what is actually included. Take the example of building a chatbot that needs to remember what's been said earlier in a conversation (perhaps over many days of interaction with a user). As the conversation's length grows, the memory management agent will move information from the input context to a persistent searchable database, summarize information to keep relevant facts in the input context; and restore relevant conversation elements from further back in time. This allows a chatbot to keep what's currently most relevant in its input context memory to generate the next response. When I read the original MemGPT paper, I thought it was an innovative technique for handling memory for LLMs. The open-source Letta framework, which we'll use in this course, makes MemGPT easy to implement. It adds memory to your LLM agents and gives them transparent long-term memory. In detail, you’ll learn: - How to build an agent that can edit its own limited input context memory, using tools and multi-step reasoning  - What is a memory hierarchy (an idea from computer operating systems, which use a cache to speed up memory access), and how these ideas apply to managing the LLM input context (where the input context window is a "cache" storing the most relevant information; and an agent decides what to move in and out of this to/from a larger persistent storage system)  - How to implement multi-agent collaboration by letting different agents share blocks of memory This course will give you a sophisticated understanding of memory management for LLMs, which is important for chatbots having long conversations, and for complex agentic workflows. Please sign up here!  https://lnkd.in/gu2BgqJE

  • View organization page for Letta, graphic

    1,569 followers

    Welcome Kevin!! 👾

    View profile for Charles Packer, graphic

    Co-Founder & CEO at Letta

    Kevin Lin is joining our research team from Berkeley NLP! Kevin was previously a PhD student at University of California, Berkeley, and prior to the PhD worked as a researcher at the Allen Institute for AI (Ai2). Kevin has pioneered context management for language agents, with research spanning retrieval, long-context, and evaluation. Kevin was also the third author on the MemGPT paper and shared Joseph Gonzalez as an advisor with me and Sarah Wooders (please see attached photo of the three of us enjoying a pint at the famous Joey Lab Social 🍻 ). We are incredibly excited to have him back on the team at Letta! 👾

    • Kevin Lin is joining Letta as a Member of Technical Staff
    • Charles Packer, Kevin Lin, Sarah Wooders enjoying beers together
  • View organization page for Letta, graphic

    1,569 followers

    View profile for Charles Packer, graphic

    Co-Founder & CEO at Letta

    shubham I̶̛̛̒̒̇̀̈́̐̾́̔̓̐̋͌͗̃͑̿̄͑̾̓̌͒̊̀͐̈͆̿̈́̐̊͛͛̐̕͠͝͝ naik is joining our engineering team from Snorkel AI / Ancestry / Klaviyo, and previously took multiple LLM playgrounds from 0→1 including at Together AI. Shubham is leading development on the ADE (Agent Development Environment), a new kind of interface to develop, debug, and deploy advanced LLM agents. We can’t wait to show you what he’s been working on at Letta. 👀 Welcome Shubham!! 👾

    • Welcoming Shubham to Letta as a Member of Technical Staff

Similar pages

Browse jobs

Funding

Letta 1 total round

Last Round

Seed

US$ 10.0M

See more info on crunchbase