Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

How does MC warm-up fit with LLMs? With LLMs you start with a prompt, so I don't see how "warm up" applies.

You're not just sampling from them like some MC cases.

> If you let the model run a bit longer it enters a region close to the typical set and when it's ready to answer you have a high probability of getting a good answer.

What does "let the model run a bit longer" even mean in this context?





The way almost all modern LLMs work is by starting with a "thinking" phase which is explicitly not part of the output. You let that process run longer.



Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: