This is still several orders of magnitude more items than the entire training corpus for all GPT models combined. I guess if you were to index individual codepoints in the training corpus, we'd start to see those volumes.
You don't index the training data, but other data. It gives LLMs the medium-term memory that they're missing.
Think of it like an accountant. Weights are all of their experience. Prompt is the form in front of them. A vector database makes it easier to find the appropriate tax law and have that open (in the prompt) as well.
This is useful for people as well, like literally this example. But the LLM + vector combination is looking really powerful because of the tight loops.