-- Matt Mahoney, mattmahone...@gmail.com

On Sun, May 11, 2025, 9:25 PM Rob Freeman <chaotic.langu...@gmail.com>
wrote:

> Matt,
>
> What do you mean "each session creates a private copy"?
>

I mean that your prompts don't update the model. If it did, then
information would leak between unrelated users. After the model is trained,
every user sees the same fixed set of 10 to 100 billion parameters.

Training and prediction costs about the same. Training costs a few million
dollars per trillion tokens. Prediction costs a few dollars per million
tokens.

My models don't work that way. It predicts the next bit of text, then
updates the model in proportion to the prediction error. It is more
accurate because it has the most up to date information. For online LLM
services to do this, it would have to make a private copy of the
parameters. Some services charge a lower rate for cached input, so maybe
they are doing this instead of using a large context window. This would be
closer to the way the brain works and more economical IMHO.


------------------------------------------
Artificial General Intelligence List: AGI
Permalink: 
https://agi.topicbox.com/groups/agi/Tdc5c19d0f38aacd6-Md357d2bb532c783380060925
Delivery options: https://agi.topicbox.com/groups/agi/subscription

Reply via email to