Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

This was part of the premise of o1 though, no? By encouraging the model to output shorter/longer chains of thought, you can scale model performance (and costs) down/up at inference time.


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: