Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

Chain of thought is basically reasoning as humans do it, the only difference is that unlike humans the model can't see that its output is wrong, abandon a line of reasoning and re-prompt itself (yet).


Various attempts at feeding their output back in to check itself have shown marked improvements in accuracy.


Multi agent LLMs talking to each other can already do this. It's just not cost feasible yet because it can lead to infinite loops and no solutions




Consider applying for YC's Winter 2026 batch! Applications are open till Nov 10

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: