Hacker News new | past | comments | ask | show | jobs | submit login

> Is there any evidence or reason to suspect that this would result in the desired effect?

Yes.

There's evidence that you can get these models to write chain-of-thought explanations that are consistent with the instructions in the given text.

For example, take a look at the ReAct paper: https://arxiv.org/abs/2210.03629

and some of the LangChain tutorials that use it:

https://langchain.readthedocs.io/en/latest/modules/agents/ge...

https://langchain.readthedocs.io/en/latest/modules/agents/im...




But note that just because a model says it's using chain-of-thought or tools to come to a certain conclusion, doesn't necessarily mean that it is: https://vgel.me/posts/tools-not-needed/


Yes, I agree. But note that the same logic applies to human beings too. Just because people say they are using chain-of-thought or tools to come to a certain conclusion, doesn't necessarily mean they are. Philosophers have been grappling with this issue for many centuries :-)


A car can explode, and so can a piñata.

Therefore deciding whether they are interchangeable is a deep question that may take centuries to resolve.


I don't understand this analogy.


The analogy is meant to show that, while it’s possible to raise deep philosophical questions based on superficial or trivial observations, it can also also be quite silly to do that.


Yes. I have been using ChatGPT quite a bit with programming tasks. One of the things I've been trying to do is using chain-of-thought prompting to ask the model to review its own code, line by line, evaluating it for the presence of a certain type of bug or some other criterion.

This has been illuminating: as ChatGPT steps through the lines of code, its "analysis" discusses material that _is not present in the line of code_. It then reaches a "conclusion" that is either correct or incorrect, but having no real relationship to the actual code.


I try to ask chatGPT add comments to my code, everything is good until I found it modify code, too.


But in the case of chain-of-thought and that ReAct paper, the results did have a measured increase in accuracy.


Oh yeah, absolutely. Just not something I'd use for, e.g., mortgages where denying someone unfairly could lead to a lawsuit.


Thanks! I was genuinely asking, will need some time to read and digest. LangChain looks interesting.


Thank you. Also search the web for posts on "prompt engineering," to get a sense of how people are using LLMs today. It's pretty eye-opening.




Join us for AI Startup School this June 16-17 in San Francisco!

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: