But note that just because a model says it's using chain-of-thought or tools to come to a certain conclusion, doesn't necessarily mean that it is: https://vgel.me/posts/tools-not-needed/
Yes, I agree. But note that the same logic applies to human beings too. Just because people say they are using chain-of-thought or tools to come to a certain conclusion, doesn't necessarily mean they are. Philosophers have been grappling with this issue for many centuries :-)
The analogy is meant to show that, while it’s possible to raise deep philosophical questions based on superficial or trivial observations, it can also also be quite silly to do that.
Yes. I have been using ChatGPT quite a bit with programming tasks. One of the things I've been trying to do is using chain-of-thought prompting to ask the model to review its own code, line by line, evaluating it for the presence of a certain type of bug or some other criterion.
This has been illuminating: as ChatGPT steps through the lines of code, its "analysis" discusses material that _is not present in the line of code_. It then reaches a "conclusion" that is either correct or incorrect, but having no real relationship to the actual code.
Yes.
There's evidence that you can get these models to write chain-of-thought explanations that are consistent with the instructions in the given text.
For example, take a look at the ReAct paper: https://arxiv.org/abs/2210.03629
and some of the LangChain tutorials that use it:
https://langchain.readthedocs.io/en/latest/modules/agents/ge...
https://langchain.readthedocs.io/en/latest/modules/agents/im...