Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

What would we actually consider an AGI? Wikipedia lists the following tests

> The Turing Test (Turing): A machine and a human both converse unseen with a second human, who must evaluate which of the two is the machine, which passes the test if it can fool the evaluator a significant fraction of the time. Note: Turing does not prescribe what should qualify as intelligence, only that knowing that it is a machine should disqualify it.

> The Coffee Test (Wozniak): A machine is required to enter an average American home and figure out how to make coffee: find the coffee machine, find the coffee, add water, find a mug, and brew the coffee by pushing the proper buttons.

> The Robot College Student Test (Goertzel): A machine enrolls in a university, taking and passing the same classes that humans would, and obtaining a degree.

> The Employment Test (Nilsson): A machine performs an economically important job at least as well as humans in the same job.

LLMs don't seem very far from passing 1), 3) and 4). I wouldn't be surprised if "GPT5" passed those 3.



I think GTP is very far from solving 1), at least if you happen to know anything about LLMs.

I think the easiest one of these would be 4) actually.


Is it? I remember reading [1] back in the day, which showed some limitations of GPT3, but ChatGPT actually answers all of those perfectly.

[1] https://lacker.io/ai/2020/07/06/giving-gpt-3-a-turing-test.h...




Consider applying for YC's Winter 2026 batch! Applications are open till Nov 10

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: