Hacker News new | past | comments | ask | show | jobs | submit login

Im saying temperature is a hack to make the models actually produce real answers.



But they can get also real answers even if you have the temperature setting as 0, where it will always pick the highest scoring token?


I think he's saying if you set temp to 0 and answers become deterministic, it will appear that the model is just memorising and reciting. The randomness is a hack that 'forces' the model to generalise by deliberately knocking it off the track of the most probable next token




Consider applying for YC's Summer 2025 batch! Applications are open till May 13

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: