-- Matt Mahoney, mattmahone...@gmail.com On Sun, May 11, 2025, 9:25 PM Rob Freeman <chaotic.langu...@gmail.com> wrote:
> Matt, > > What do you mean "each session creates a private copy"? > I mean that your prompts don't update the model. If it did, then information would leak between unrelated users. After the model is trained, every user sees the same fixed set of 10 to 100 billion parameters. Training and prediction costs about the same. Training costs a few million dollars per trillion tokens. Prediction costs a few dollars per million tokens. My models don't work that way. It predicts the next bit of text, then updates the model in proportion to the prediction error. It is more accurate because it has the most up to date information. For online LLM services to do this, it would have to make a private copy of the parameters. Some services charge a lower rate for cached input, so maybe they are doing this instead of using a large context window. This would be closer to the way the brain works and more economical IMHO. ------------------------------------------ Artificial General Intelligence List: AGI Permalink: https://agi.topicbox.com/groups/agi/Tdc5c19d0f38aacd6-Md357d2bb532c783380060925 Delivery options: https://agi.topicbox.com/groups/agi/subscription