MemGPT — Unlimited Context (Memory) for LLMs

How can you overcome the context window size limit of LLMs?

Venelin Valkov
10 min readOct 27, 2023
Photo by Raj Rana on Unsplash

One of the largest (no pun intended) limitations of Large Language Models (LLMs) is their context window size. Here’s an overview of the most popular LLMs and their context window size:

How can you overcome the limited token context window? MemGPT offers a solution inspired by traditional Operating Systems (OS) — hierarchical memory. Let’s take a look at how it works.

--

--