Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

How would you prove that it's not a "stochastic parrot", in general?

I don't see why it matters if the "time signal", whatever it is - and you surely need one for an internal clock either way - is text or something else. The models that we have only have text inputs, so naturally it would be a token (but it could easily be a specialized non-text token like BOS/EOS if we trained the model that way). And the model can abstain from generating anything given any input - this is actually not uncommon for smaller models. GPT-3.5 and GPT-4 never seem to do it, but then again it's specifically fine-tuned for chat, i.e. always producing an output.

Long-term memory is a general problem with these things, but its short-term memory is its context window, so why would it have problem correlating events there? And for long-term memory, if it is implemented as an API under the hood that the model uses to store and query data, it would be trivial for it to timestamp everything according to the clock, no?



Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: