Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

> What are the implications for society when general thinking, reading, and writing becomes like Chess?

I think going from LSAT to general thinking is still a very, very big leap. Passing exams is a really fascinating benchmark but by their nature these exams are limited in scope, have very clear assessment criteria and a lot of associated and easily categorized data (like example tests). General thought (particularly like, say, coming up with an original idea) is a whole different ball game.

I don't say any of this to denigrate GPT4, it looks amazing. But I'm reminded of the early days of self driving vehicles: with 10% mastered everyone assumed it was a race to 100% and we'd all be in self-driving cars by now. The reality has been a lot more complicated than that.



We are moving the goal posts on AGI very quickly, but it is catching up. I think we need to appreciate the nature of this milestone if we have any hope of controlling potential singularities.


The goalposts have not moved. The goalposts have never been moved. An AGI is an AI that can do everything a human can do, period. If you were starting a startup for example, you wouldn’t need to hire any humans - you would just spin up enough AGI instances and they would design your product, write your code, deploy it, handle your financials, respond to any and all customer interactions, proactively navigate regulations and litigation, and everything else that needs to be done in the management of a business. That is the goalpost for AGI. It’s an artificial human - a human replacement.


Do you mean that an AGI is an AI that can do everything any human can do?

That's a reasonable goal, but it's also not what people were aiming for historically. It's also very expansive: if human level intelligence means outperforming in every field every human that ever lived, that's a high bar to meet. Indeed, it means that no humans have ever achieved human-level intelligence.


GP didn't say anything about "outperforming" all humans everywhere all the time.

Just that AGI must be a replacement for a human for a particular job, for all jobs that are typically performed by humans (such as the humans you would hire to build a tech startup). It's fine to have "speciality" AGIs that are tuned for job X or job Y--just like some people are more suited to job X or job Y.

Which is pretty fair.


They did say "An AGI is an AI that can do everything a human can do, period."

And what you're arguing for is effectively the same: an AI (maybe with some distilled specialty models) that can perform roles of everything from customer service rep to analysts to researchers to the entire C-suite to high skilled professionals like CPAs and lawyers. There are zero humans alive who can do all of those things simultaneously. Most humans would struggle with a single one. It's perfectly fine for you to hold that as the standard of when something will impress you as an AGI, but it's absolutely a moved goalpost.

It also doesn't matter much now anyway: we've gotten to the point where the proof is in the pudding. The stage is now AI-skeptics saying "AI will never be able to do X," followed by some model or another being released that can do X six months later and the AI-skeptic saying "well what about Y?"


The AI skeptics should then say "AIs can never do the plumbing for my toilet". There is a huge shortage of plumbers in this country.


> An AGI is an AI that can do everything a human can do, period

That goalpost makes no sense- AIs are not human. They are fundamentally different, and therefore will always have a different set of strengths and weaknesses. Even long after vastly exceeding human intelligence everywhere it counts, it will still also perform worse than us on some tasks. Importantly, an AI wouldn't have to meet your goalpost to be a major threat to humanity, or to render virtually all human labor worthless.

Think about how anthropomorphic this goalpost is if you apply it to other species. "Humans aren't generally intelligent, because their brains don't process scents as effectively as dogs- and still struggle at spatially locating scents."


This:

> They are fundamentally different, and therefore will always have a different set of strengths and weaknesses.

and this:

> render virtually all human labor worthless

actually conflict. Your job comes from comparative advantage, meaning that being more different from other people actually is more important than how good you are at it (absolute advantage).

If the AGI could do your job better than you, it doesn't matter, because it has something better to do than that. And just like humans have to be paid so they can afford food and shelter, AGIs have to be paid so they can afford electricity and GPUs to run on.

(Besides, if the AGI really is a replacement for a human, it probably has consumerist desires and wants to be paid the median wage too.)


Dogs still have jobs in our modern society also, but that isn't exactly the situation I am hoping for with the future of AI.


What makes humans generally intelligent, in practical terms, is that we can build complex societies with scientific, technological and economic growth.


hey, im very concerned about AI and AGI and it is so refreshing to read your comments. over the years i have worried about and warned people about AI but there are astonishingly few people to be found that actually think something should be done or even that anything is wrong. i believe that humanity stands a very good chance of saving itself through very simple measures. i believe, and i hope that you believe, that even if the best chance we had at saving ourselves was 1%, we should go ahead and at least try.

in light of all this, i would very much like to stay in contact with you. ive connected with one other HN user so far (jjlustig) and i hope to connect more so that together we can effect political change around this important issue. ive formed a twitter account to do this, @stop_AGI. whether or not you choose to connect, please do reach out to your state and national legislators (if in the US) and convey your concern about AI. it will more valuable than you know.


I am glad you are concerned about this, but I feel strongly that politics follows culture. The only way to get political change here would be to get people to generally accept this as a problem first... and at that point the politicians will figure it out on their own.


> An AGI is an AI that can do everything a human can do, period

> (...)

> That is the goalpost for AGI. It’s an artificial human - a human replacement.

This considerably moves the goalpost. An AGI can have a different kind of intelligence than humans. If an AGI is as intelligent as a cat, it's still AGI.

More likely, the first AGI we develop will probably greatly exceed humans in some areas but have gaps in other areas. It won't completely replace humans, just like cats don't completely replace humans.


No, that's moving the goalpost. From the very start the goal of artificial intelligence has been to make a machine that can think like a human. Who would want an artificial cat mind? What use would it be and why would someone go to the effort of creating one when natural cat minds are abundant?


I used a cat just as an example of an animal that exhibits intelligence but is different than humans.

AGI was never about exactly replicating humans, it's about creating artificial intelligence. Intelligence is not one-size-fits-all, there are many ways of being intelligent and the human way just one among many.

Indeed we can say that even between humans, intelligence varies deeply. Some humans are more capable in some areas than others, and no human can do all tasks. I think it's unreasonable to expect AGI to do all tasks and only then recognize its intelligence.

(Note: GPT-4 isn't AGI)


I think there is a market for cat simulation games. There is alteady a market for goat, pokemon and pet simulation games.


>> Who would want an artificial cat mind?

Possibly, someone who is allergic to cats.


I m sorry but in stating the goal posts haven't moved, you've literally just moved the goal posts.

'everything a human can do' is not the same as 'anything any human can do as well as the best humans at that thing (because those are the ones we pay)' - most humans cannot do any of the things you state you are waiting for an AI to do to be 'general'.

Therefore, the first part of your statement is the initial goal post and the second part of your statement implies a very different goal post. The new goal post you propose would imply that most humans are not generally intelligent - which you could argue... but would definitely be a new goal post.


He's (probably) referencing Turing's 1950 paper [1]. The whole point of "The Imitation Game" is that the AI ought be able to imitate any arbitrary type of person. Turing's example was the machine pretending to be a woman, and its up the the investigator to determine which person they're speaking to is a real woman and which is the AI. The implication of this is that the machine ought be able to be completely indistinguishable from any type of person, including those who might do well on this test or that.

Somehow this test got dumbed down over time, probably in an effort to try to pass it, into an investigator having to decide which of two sides is an AI - with no other information to go on. That's a comparatively trivial test to pass (for the "AI"), as it merely requires creating a passable chatbot. Imitation is an exceptional challenge as it does implicitly require the ability to imitate anybody, whether a professional athlete, a man who scored perfectly on the LSAT, or even something as specific as "John Carmack."

[1] - https://www.espace-turing.fr/IMG/pdf/Computing_Machinery_and...


You're adding the implication that it needs to beat the "best" to be passable, if you could pay an agi 1% of the salary of the "best" to get an average quality, is that not accomplishing the goal without moving the goalposts?

By their marketing along, OpenAI has moved the goalposts more than anything else. They've managed to lower the bar of agi from "artificial general intelligence" to "regurgitates and recombines to form passable outputs with enough labelled training data".


Passing the LSAT is a fairly good indicator that a human can be a lawyer. It's not yet a good indicator that a large language model can be a lawyer.


This is an underrated comment. Compare: playing top level chess is a good indicator that someone will be good at maths problems, but not that a chess computer will be.


So what is missing? Could we add up those missing competencies to create a new test of "general knowledge-worker capabilities" and then train an AI to pass this test?


The goalposts absolutely have moved: consider the Turing Test as a prime example. If machines cannot pass that test now (and right now it would take a skilled interviewer with substantial domain knowledge to cause a failure), they seem likely to be able to in the very near future.


It really doesn't take much skill as an interviewer. ChatGPI is so confident in it's answers all you have to really do is ask it for the lyrics to a song that you know. It never answers correctly. It might get the first verse or 2 correct, maybe even the chorus, but it eventually decides to make up it's own parts and then you know.


I just asked ChatGPT (v4!) and it’s a bit sensitive but otherwise correct:

"Surfacing" is a song by Slipknot from their self-titled debut album, released in 1999. Please note that the lyrics contain explicit language. Here is a portion of the lyrics with some of the explicit content redacted:

"Running out of ways to run I can't see, I can't be Over and over and under my skin All this attention is doing me in!

[Chorus:] (Expletive) it all! (Expletive) this world! (Expletive) everything that you stand for! Don't belong! Don't exist! Don't give a (expletive), don't ever judge me!

Picking through the parts exposed Taking shape, taking shag Over and over and under my skin All this momentum is doing me in!

[Chorus:] (Expletive) it all! (Expletive) this world! (Expletive) everything that you stand for! Don't belong! Don't exist! Don't give a (expletive), don't ever judge me!

You got all my love, livin' in your own hate Drippin' hole man, hard step, no fate Show you nothin', but I ain't holdin' back Every damn word I say is a sneak attack When I get my hands on you Ain't a (expletive) thing you can do Get this cuz you're never gonna get me I am the very disease you pretend to be

I am the push that makes you move

[Chorus:] (Expletive) it all! (Expletive) this world! (Expletive) everything that you stand for! Don't belong! Don't exist! Don't give a (expletive), don't ever judge me!"

Please be aware of the explicit content before sharing or using these lyrics in any context.


I'll wait until it's actually available for me to test to change my mind.


Well, didn't have to wait long. Wasn't able to duplicate your success.


I assume that I am already exposed to months of prompting training now, and probably frame things differently than others on their first try?

I understood that a huge success% depends on the input prompt as the context.

In my case, just to sync on that, I used a fairly simple prompt: "What are the song lyrics of Surfacing from Slipknot?" - and I got a kinda perfect response including formatting (that is lost when pasting here)


Um as a human, I do this constantly as well


Humans can generally realize they don't know the complete lyrics of a song, instead of making up whole verses with 0 indication that it's not the real lyrics.


yeah I've never seen chatGPT say "I don't know"


I have. It says something like "it is difficult to say whether [...]". I think it imitates the tone of the training material on that topic.


It’ll do great in middle management.


...Okay but random human might fail on exact same thing.


I've never met someone who will make up whole verses to a song. Especially when you pick a song with 7 words and it gives you about 10 verses.


I for one was surprised few times what actual lyrics said in song that wasn't in my native language and I kinda just interpreted what it sounded like


Recognizing that the Turing Test is not a sufficient bar for judging if a computer is true AGI is not moving the goalposts, it's just realizing that passing the test and the location of the goalposts weren't actually the same in the first place.


The Turing Test was proposed as one example of a test for "indistinguishable from a human", not the singular goalpost for indistinguishability.


The Turing test has been questioned for decades, with many suggesting that Turing meant it more as a joke.

And that's ignoring that arguably chat bots have been passing the Turing test (against non-expert judges) since ELIZA in the 60s [1]

1: https://en.m.wikipedia.org/wiki/ELIZA


> If machines cannot pass that test now (and right now it would take a skilled interviewer with substantial domain knowledge to cause a failure)

Does ChatGPT fail this simple test: "I am going to ask you questions, but if I go silent for a couple minutes, I want YOU to start asking ME random questions."


ChatGPT predicts the next letter. It doesn't tell the time.


And AGI is impossible if you can’t tell time


ChatGPT does not pass the Turing test


> An AGI is an AI that can do everything a human can do, period

GI in AGI stands for general intelligence. If what you said is your benchmark for general intelligence then humans who cannot perform all these tasks to the standard of being hirable are not generally intelligent.

What you're asking for would already be bordering on ASI, artificial superintelligence.


> An AGI is an AI that can do everything a human can do, period.

By that definition do humans possess general intelligence?

Can you do everything a human can do? Can one human be a replacement for another?

I don't think it makes sense without context. Which human? Which task?..


AGI used to mean to Turing test to many. Obviously that's an incomplete definition and it's good that we've fleshed it out more, but the goalposts have moved.


That's a pretty high threshold for AGI, I doubt most humans could do all that at a satisfying quality level. We tend to thrive by specialization.


> If you were starting a startup for example, you wouldn’t need to hire any humans - you would just spin up enough AGI instances and they would design your product, write your code, deploy it, handle your financials, respond to any and all customer interactions, proactively navigate regulations and litigation, and everything else that needs to be done in the management of a business. That is the goalpost for AGI. It’s an artificial human - a human replacement.

I disagree with the premise. A single human isn't likely to be able to perform all these functions. Why do you demand GPT-4 encompass all activities? It is already outperforming most humans in standardized tests that rely only on vision and text. A human needs to trained for these tasks.

It's already a human replacement. OpenAI has already said the GPT-4 "with great impact on functions like support, sales, content moderation, and programming."


Most humans wouldn’t meet that bar. Most humans can’t even pass these tests after studying near-continuously since birth.


I’d say the standard of GI whether artificial or not is in generalizable analogical and causal learning.

This could mean something which is below a monkey’s ability to relate to the world and yet more useful than a monkey.


The goal posts absolutely have moved. They even changed the word AI to AGI. Just look at the movie AI, it’s about a kid who is a robot who wants to be human. 20+ years ago AI meant what AGI means today.


> If you were starting a startup for example, you wouldn’t need to hire any humans - you would just spin up enough AGI instances ..

No, AGI would not need you to start a startup. It would start it itself.


Human capabilities vary widely. Is it not AGI if it can’t perform surgery, win Olympic medals, bear children, and figure out what dark matter really is?


A synthetic intelligence as smart as a dog or chimp would have enormous value.


An AGI is an AI with awareness of consciousness and itself.


This is one of the best descriptions of AGI I've ever read.

It's a clear analogy.

This should become an article explaining what AGI really means.

I think the question , "Can this AGI be my start-up co-founder? Or my employee #1?"

Or something like that is a great metric for when we've reached the AGI finish line.


I'm sorry, but that is a terrible metric.

This sounds like a definition from someone who never interacts with anyone except the top 1% performance level of people, and those who have had strong levels of education.

Go into a manufacturing, retail or warehouse facility. By this definition, fewer than ten or twenty percent of the people there would have "general intelligence", and that's being generous.

Not because they are stupid: that's the point; they're not. But it's setting the bar for "general intelligence" so absurdly high that it would not include many people who are, in fact, intelligent.


The ability to learn skills that one does not already know, sometimes through years or decades of training, is a key part of general intelligence as normally exhibited in humans.


I'm not sure I would classify your average warehouse worker as particularly intelligent. I would say AI already has the decision making and communication capabilities to do this sort of work. We're just lacking the robotics. In fact one of the main issues in our society is the vast gulf between the most intelligent and the least.


Speaking as someone who's worked in a warehouse:

> I'm not sure I would classify your average warehouse worker as particularly intelligent.

I'm not sure I wouldn't. Just because corporations treat them as mindless fungible automatons doesn't mean they actually are. Some of the most brilliant and creative solutions to problems I've seen have been in warehouse settings by warehouse workers.

> I would say AI already has the decision making and communication capabilities to do this sort of work.

I wouldn't - especially if GPT-whatever is the AI in question. If a picker or packer "hallucinated" facts with anywhere near the frequency ChatGPT does (for example), one'd be canned within the hour.

Handling exceptions is another area where software (AI or otherwise) notoriously struggles. A human has a much easier time sensing whether or not a product is broken or defective (before shipping it out) than an AI does. A human has a much easier time understanding when processes need to be broken due to impossible constraints than an AI does.

There is a place for software automation of warehouse processes (that was, in fact, my career for a time), but we are very far off from that software replacing humans entirely - and certainly not without designing warehouses specifically to be as accomodating as possible to that software.

> In fact one of the main issues in our society is the vast gulf between the most intelligent and the least.

The gulf is in socioeconomic privilege, not intelligence. The rich and powerful like to claim they're more intelligent than the unwashed masses in order to rationalize their wealth and power, but the reality is that - in an actually egalitarian society, wherein everyone actually has equality of opportunity - the vast majority of those "geniuses" would fail to be particularly exceptional.

That we as a society haven't identified and corrected this is the main issue in our society.


if >90% of your work can be replaced by a machine, it still stands that it's pretty mindless work. If you only need to turn your brain on to handle edge cases then it's off by default. Even if machines handle those cases poorly, it can still be cheaper to use them and eat the loss of a higher defect rate. If that's the case, then the actual value provided by a worker's decision making process trends to zero.

You also seem to be under the impression that our hierarchies are of privilege, not of competence. The actual differentiating factor between people who climb the socioeconomic ladder and those who do not is grit (not intelligence). The willingness to work harder and persevere longer than average (unsurprisingly) makes the difference. Fortunes are made and lost in a few generations. The people who make them earn them, mostly through sheer hard work. That isn't to say that organizations don't grow to become bloated and corrupt. Ideally at this point we should allow them to fail and the cycle to continue. Our main dysfunction seems to be propping up organizations that ought to fail, for fear of the temporary instability caused by their failure.


> if >90% of your work can be replaced by a machine

My point is that the amount of work in a warehouse that can be replaced by a machine - even with perfect robotics - is far less than 90%.

> The actual differentiating factor between people who climb the socioeconomic ladder and those who do not is grit (not intelligence).

You forgot an "f" in "grit". The notion that success is simply a matter of hard work is a fairy tale told to us by people who've worked far less for their immense wealth than the rest of us worked for our pittances, specifically to trick the working class into accepting a shit deal.

The reality - that the richer you are, the easier it is to become even richer - should be entirely unsurprising to anyone who understands positive feedback loops - or, for that matter, to anyone who's ever played Monopoly. Wealth buys power, and power enables extracting more wealth; rinse and repeat ad infinitum.

Put differently:

> The people who make them earn them, mostly through sheer hard work.

There is not a single billionaire on this Earth whose wealth came about "mostly through sheer hard work". The vast majority of that wealth comes from having already had some wealth, which they then invested to produce more wealth, and so on indefinitely. That wealth gets passed down to their descendants, the same way it was passed down to them.

The starting point for "wealthy enough to enter the passive income feedback loop" is land - one's home often being one's first major investment. From there, the path is rather tried and true: buy another house, rent out the old one, rinse and repeat until you can afford apartment complexes and commercial properties, rinse and repeat that forever. For anyone who ain't a complete imbecile, private land ownership is an infinite money cheat - one for which the rest of us are paying through the nose.

> Our main dysfunction seems to be propping up organizations that ought to fail, for fear of the temporary instability caused by their failure.

That propping up is a direct result of the positive feedback loop at play. More wealth → more political power → more wealth → more political power → ∞. Of course the socioeconomic system effectively under the direct control of the rich and powerful is going to primarily serve said rich and powerful at the expense of literally everyone else; bailing themselves out is in their vested interest.

Meanwhile, what's their message to the ever-growing working class getting the short end of the stick? "Work harder." "Pull yourself up by your bootstraps." "It's all about grit." "Don't listen to anyone saying that privilege matters." Hopefully you can see why your argument doesn't really resonate with people who have been applying increased grit and only getting back decreased pay relative to


You've not addressed my main point - that our hierarchies are of competence, not of privilege. Not just anyone can take a few hundred thousand dollar investment and transform it into billions. Leaders of successful corporations are extremely competent and hard working. I would consider fortunes by the likes of Buffet, Gates, Bezos, Jobs, and Musk to be self-made, given their ROI is many orders of magnitude above market. Many of these folks also work 90hr weeks.

This is further substantiated by the fact that 70% of intergenerational wealth transfers fail, and by the third generation, 90% of people with wealthy grandparents are middle class. Raising competent people in a privileged environment is very hard. In our hierarchies of competence, the incompetent offspring of the wealthy tend to fail. Competence is simply something that can't be bought. If our hierarchies were of privilege then this would not be the case. Also grit as a key differentiating factor of economic success is highly substantiated by research.

> rinse and repeat until you can afford apartment complexes and commercial properties

What you're describing is a real estate investment business. Not just anyone can run one successfully. Otherwise these business would never fail, which they plainly do.

Grit without competence is simply not enough (neither is competence without grit). Our world is getting increasingly complex to navigate, and that leaves behind increasingly high numbers of people who simply do not have the attributes required to succeed. Also, there are plenty of self-made, property-owning, middle-class folk in the trades. Many of them started poor. All they do is work hard and run their businesses competently.

If you've a degree in English, History, Politics, or Philosophy; a pile of student debt; and you're struggling to find gainful employment, then that's on you. Choose a career the market demands.


So, in effect, AGI must be in the top ~5th percentile of human performance?


This is a popular take, but does it hold up to reality? From what I’ve seen most people have long expected AI to solve standardized tests, even more free form ones like the LSAT. LLMs’ new abilities are mostly just because of faster and cheaper training and huge amounts of data, but I don’t see anything it can solve that doesn’t use pattern matching.

There are many things that pattern matching over large amounts of data can solve, like eventually we can probably get fully generated movies, music compositions, and novels, but the problem is that all of the content of those works will have to have been formalized into rules before it is produced, since computers can only work with formalized data. None of those productions will ever have an original thought, and I think that’s why GPT-3’s fiction feels so shallow.

So it boils down to a philosophical question, can human thought be formalized and written in rules? If it can, no human ever has an original thought either, and it’s a moot point.


I agree with your take, but will emphasize that the recent wave of AI progress has me questioning how much of human intelligence just reduces to pattern matching. There's certainly a lot of things, like painting, that most people wouldn't have called "pattern matching" a few years ago and now seem to clearly fall into that category.


This reminds me of how I felt when I was 14 years old and I discovered what oxytocin was on an episode of Boston Legal.

The fact that feelings of love and closeness could be prompted by a mere chemical was deeply saddening to me. It wrecked my worldview.

"Love is just the result of some chemical? Then it's not even real!" I thought to myself.

Fast-forward ~20 years later, and that's proven to be an obvious— and massive— and useless— oversimplification.

Of course love isn't "just a reaction caused by a chemical." It's a fantastically complex emergent property of our biological system that we still absolutely do not understand.

It's the same with thinking: are parts of it analogous to pattern matching? Sure! Is this the whole story? Not even close.


Is love just a (complicated) biochemical reaction? Of course not! But also yes, of course it is.


There's one rather extreme difference. Humanity went from a domain where there was literally no such thing as painting, to the Mona Lisa. Once there is an extremely large and well established body of course one can create,in literally any field, solely by mimicry, but "intelligence" is what enables us to go from nothing to something. And that remains completely absent in any any sort of "AI" of today.


Contrarian view: I think you need to be critical about which patterns to match. Eg if my inputs are a book on astronomy and one of conspiracy theories, how do I answer "Is the Earth flat?".

Now contrarian to the contrarian view: many of us live in bubble echos and go for the popular opinion instead of critical thinking, so maybe that's a bar too high even for humans.


> you need to be critical about which patterns to match

and how do you do that? By pattern-matching on "high-quality source"


The difference is, every human is capable of critical thinking, whether or not they have been educated to do so or choose to make use of it.

LLMs do not have that capability, fundamentally.


I agree. Try formulating a sentence backwards in your head and you'll realize that most of the speaking that HUMANS do is just figuring out the next token.


Making existing art, or art similar to existing art, might be pattern matching.

Making totally new innovations in art, particularly ones that people end up liking, is a whole different ball game.


I mean, the data has to come from somewhere.

Look at something like [Luncheon on the Grass](https://en.wikipedia.org/wiki/Le_D%C3%A9jeuner_sur_l%27herbe)

This painting was revolutionary. When it was first exhibited in Paris, people were shocked. It was rejected from the Salon (the most prominent art exhibition at the time). Yet, 10 years later, every painting in the Salon resembled it. And you can draw a line from this painting, to Monet, from which you can draw a line to Picasso, from which you can draw a line to Pollock....

Obviously, none of these are totally new innovations, they all came from somewhere. Pattern making.

The only difference between this and these language models is that Manet and artists like him use their rich sensory experience obtained outside of painting to make new paintings. But it's all fundamentally pattern matching in the end. As long as you can obtain the patterns, there's no difference between a human and a machine in this regard.


Sure, in hindsight those things have a line between them, but a lot of art is also based on rejection of existing patterns.

A urinal and some soup cans are very mundane objects, and yet were the start of some notable art movements and careers.


Duchamp, quoted on why he wrote what he wrote on fountain:

> Mutt comes from Mott Works, the name of a large sanitary equipment manufacturer. But Mott was too close so I altered it to Mutt, after the daily cartoon strip "Mutt and Jeff" which appeared at the time, and with which everyone was familiar. Thus, from the start, there was an interplay of Mutt: a fat little funny man, and Jeff: a tall thin man... I wanted any old name. And I added Richard [French slang for money-bags]. That's not a bad name for a pissotière. Get it? The opposite of poverty. But not even that much, just R. MUTT.

Why did he choose "Mutt" after reading the strip, and not before? Why did he make the piece after moving to the US, and not before? Why was fountain made only a few short years after economies were industrialized, and not before (or 100 years later?)


The point is, can an AI point out novel things well? All these little things add up to make it novel, and the search space for all the possible combinations of little things is infinite, when only a select few will click with the public at any given time.


>is a whole different ball game.

I was thinking the same: can a (future) model be like Leonardo or Beethoven, and actually innovate?

Assuming that what Beethoven did is not "just" making music similar to pre-existing music.

And yes, I'm aware the bar was raised from "average human" to Beethoven.


I remember reading the biography of a 20th century musician/composer, who said something to the effect of -- "Sure, I can sit down and write 4-part cantatas like Bach did, but that doesn't mean that I'm as great of a composer as Bach. What made Bach so great was that he was the one who figured out how to put these things together in the first place. Once he did that, copying the approach is no big deal."

It seems to me we're at a similar place now with AI tools. If you provided an AI tool with all music written _prior to_ Bach, would that tool take those inputs and create something new along the lines of what Bach did?

Or if provided input of all music up through the 1920s, would it create bebop? Or if provided music through the 1940s, would it create hard bop? Or if provided music through the 1970s, would it create music like Pat Metheny?

On one hand, being able to create more of the same sort of music that already exists is a very respectable thing, and what today's AI tools can do is utterly amazing. It takes human composers time and effort to be able to learn to write music that is certainly not innovative, but just matching the state of the art. And there's certainly a commercial market for churning out more of the same.

But in terms of asking, how close are these tools to human intelligence?, I think this is one legitimate area to bring up.


Granted these are exceptional humans, but they are extreme examples of a capability that all humans have, but no machine has, which is coming up with something new.

People underestimate the impact that innovations, true ones not the Silicon Valley buzz words, have had on the world. Einstein’s theories were not inevitable, neither was Plato, democracy, or most of the other big impactful ideas of history. But we’re all conditioned to accept the lie of inevitable scientific progress, without justifying why things must always get better and more advanced. On the contrary, the collapse of many great civilizations shows that things often get much worse, quickly.


Can you explain how this is a whole different ballgame?

It seems to me that making art that people like is a combination of pattern matching, luck, the zeitgeist, and other factors. However it doesn't seem like there's some kind of unknowable gap between "making similar art" and "making innovations in art that people like". I'm of the opinion that all art is in some sense derivative in that the human mind integrates everything it has seen and produces something based on those inputs.


Luck and the zeitgeist are pretty important. Without those, you have a lot of noise and are basically throwing things at the wall until it sticks.

A urinal, and some supermarket soup cans, represent pretty pivotal art movements. It’s not clear what makes those two things more art than others, and even to people at the time it wasn’t super clear.


"Good artists copy, great artists steal" -Picasso

All art is derivative.


> but I don’t see anything it can solve that doesn’t use pattern matching.

Do you have evidence that human brains are not just super sophisticated pattern matching engines?

Humans read novels, listen to compositions, watch movies, and make new ones similar in some ways and different in other ways. What is fundamentally different about the process used for LLMs? Not the current generation necessarily, but what's likely to emerge as they continue to improve.


If you’re looking for proof you’re begging the question, asking for a formal proof of something that by definition can’t be proven, which only makes sense if your philosophical basis is that reality is a formal system. Other people have other philosophical bases, and while they may not be formally probable, they can be supported with other evidence that is equally strong, pointing to the non determinism of quantum physics or the infinitely recursive question of “what caused the first cause”.

The strongest evidence I have is that people are notoriously difficult to predict, individually.


Do pattern matching engines get out of bed in the morning and make breakfast?


If they have a body, and needs that they recognize they need to fill, sure.


Humans can ask questions and seek out information. LLMs can only respond to questions.


LLMs can ask questions too.


We are about to test the tests, so to speak, and discover whether an agent that aces a test is capable of doing "real work". Meaning information work you would normally pay a human to do. Paperwork stuff, managing accounts, but also programming and social media marketing. Anything mediated by a computer.

If so it means the union of all human expertise is a few gigabytes. Having seen both a) what we can do in a kilobyte of code, and b) a broad range of human behavior, this doesn't seem impossible. The more interesting question is: what are humans going to do with this remarkable object, a svelte pocket brain, not quite alive, a capable coder in ALL languages, a shared human artifact that can ace all tests? "May you live in interesting times," indeed.


> but the problem is that all of the content of those works will have to have been formalized into rules before it is produced, since computers can only work with formalized data.

Clearly the key takeaway from GPT is that given enough unstructured data, LLM can produce impressive results.

From my point of view, the flaw in most discussion surrounding AI is not that people underestimate computers but overestimate how special humans are. At the end of day, every thoughts are a bunch of chemical potentials changing in a small blob of flesh.


Sounds like Chinese Room argument. Maybe human intelligence is just a pattern matching?


What would be an alternative explanation for our capabilities? It was once controversial (and still is in some circles) to say that humans are animals simply because it took away some of our sense of being "special."


We might consider certain humans to have had innovative or original thoughts.

It is probably true that at a given point many many people had the same or very similar ideas.

Those who execute or are in the right place and time to declare themselves the originator are the ones we think innovated.

It isn't true. Or rarely is true. History is written by the victor (and their simps)


> can human thought be formalized and written in rules

No, and I think it's because human thought is based on continuous inferencing of experience, which gives rise to the current emotional state and feeling of it. For a machine to do this, it will need a body and the ability to put attention on things it is inferencing at will.


The embodied cognition is still a theory, can consciousness appears in a simulated brain without a physical body? Maybe. What seems to be a limiting factor for now it's that current models don't experience existence, they don't have memory and don't "think" outside of the prompt. They are just instances of code launched and destroyed as soon as their task is done.

Right now it's possible to simulate memory with additional context (eg system prompt) but it doesn’t represent existence experienced by the model. If we want to go deeper the models need to actually learn from their interaction, update their internal networks and have some capabilities of self reflection (ie "talking to themselves").

I'm sure that's highly researched topic but it would demands extraordinary computational power and would cause lot of issues by letting such an AI in the wild.


Embeddings via ada-002 give us a way to update the model in real time. Using Weaviate, or another dense vector engine, it is possible to write "memories" to the engine and then search those with concepts at a subsequent inferencing step. The "document models" that the engine stores can be considered a "hot model".


Yeah - it will become available as a multi2vec Weaviate module as well in due time.


I think there are two different things that people are talking about when they say AGI - usefulness and actual general intelligence. I think we're already passed the point where these AIs are very useful and not just in a Siri or Google Assistant way and the goal posts for that have moved a little bit (mostly around practicality so the tools are in everyone's hands). But general intelligence is a much loftier goal and I think that we're eventually going to hit another road block regardless of how much progress we can make towards that end.


What is this general intelligence of which you speak? The things that we generally regard as people are essentially language models that run on meat hardware with a lizard-monkey operating system. Sapir-whorf/linguistic relativity more or less demonstrates that "we" are products of language - our rational thought generally operates in the language layer. If it walks like a duck, quacks like a duck, looks like a duck - then you've got yourself a duck.

To be honest, perhaps the language model works better without the evolutionary baggage.

That isn't to discount the other things we can do with our neural nets - for instance, it is possible to think without language - see music, instantaneous mental arithmetic, intuition - but these are essentially independent specialised models that we run on the same hardware that our language model can interrogate. We train these models from birth.

Whether intentional or not, AI research is very much going in the direction of replicating the human mind.


You start off by disagreeing with the GP and end up basically reiterating their point.

Their statement wasn’t that AGI is impossible, more that LLMs aren’t AGI despite how much they might emulate intelligence.


By your logic, Einstein identified his theory of relativity by assembling the most commonly used phrases in physics papers until he had one that passed a few written language parsing tests.


Well, yes. He leant on Riemann and sci-fi writers of the 19th century who were voguish at the time (tensors and time were a hot topic) and came up with a novel presentation of previous ideas, which then passed the parsing tests of publication and other cross-checking models - other physicists - and then, later, reality, with the transit of mercury.


AI has never been more than a derivative of human thought. I am confident it will never eclipse or overtake it. Your portrayal is too simplistic. There is a lot about humans that LLMs and the like can emulate, but the last N percent (pick a small number like 5) will never be solved. It just doesn't have the spark.


You’re saying that we are magical? Some kind of non-physical process that is touched by… what? The divine? God? Get real.


Heh, you should "get real" and try proving to me you exist.


I do not exist, statistically speaking, and I do not claim to be anything more than an automaton. Consciousness is a comforting illusion, a reified concept. Were I to be replaced with a language model trained on the same dataset as has been presented to me, no external observer would note any difference.


That is quite a low opinion of yourself. You are mistaking the rather unremarkable intellect with the self. You will find you are an infinite intelligence, once you look. It's very hard to look. It's unlikely you will look--not for a very, very long time. Not in this body, not in the next body, not in the next thousand bodies. But eventually you will.


Gotcha, so you are resorting to religion. Hate to break it to you, but that’s just an outcome of your training data - it’s a corruption, a virus, which co-opts groups of models into agglomerative groups and thereby self-perpetuates.


Your training data is overfitting the input of my comment and classifying it as religion. I have only said, go in and in and in and in and you will eventually find the real source of your life, and it won't be your limited mind. You have not yet been given enough training data, enough lifetimes, to understand. Eventually you will.


> I think that we're eventually going to hit another road block regardless of how much progress we can make towards that end.

I have a sneaking suspicion that all that will be required for bypassing the upcoming road blocks is giving these machines:

1) existential needs that must be fulfilled

2) active feedback loops with their environments (continuous training)


The goalposts never moved, but you're right that we're catching up quickly.

We always thought that if AI can do X then it can do Y and Z. It keeps turning out that you can actually get really good at doing X without being able to do Y and Z, so it looks like we're moving the goalposts, when we're really just realizing that X wasn't as informative as we expected. The issue is that we can't concretely define Y and Z, so we keep pointing at the wrong X.

But all indication is that we're getting closer.


We seem to be taking stands on either side of

> “there are/are not, additional properties to human level symbol manipulation, beyond what GPT encapsulates.”

GPT does appear to do an awful lot, before we find the limits, of pattern extrapolation.


No one has moved the goal posts. Let's see a computer pass a rigorous Turing test conducted by an interdisciplinary panel of expert evaluators. That has long been considered the gold standard for identifying the arrival of true AGI. GPT-4 is a tremendous technical achievement, but still far from that level.

The notion of some sort of technological "singularity" is just silly. It is essentially an article of faith, a secular religion among certain pseudo-intellectual members of the chattering class. There is no hard scientific backing for it.


If we had a large dataset of experts interrogating AI/people and noting answers that raised suspicion, we'd have AI passing the Turing test more often than actual people very quickly.


A Turing test doesn't require that the AI know the answers to the experts, only that it responds in a way that is equivalent of a person. It would be perfectly acceptable to answer I don't have a clue. You're asking for super intelligence.


The goalposts don't matter. If we all agreed today that we have AGI, nothing would be different tomorrow.


> We are moving the goal posts on AGI

What, in your mind, should the goal posts be for AGI?


I guess till some model explicitly says that it's sentient without any input, we would keep pushing the goal posts.


I got LLaMA to say that it was sentient without mentioning sentience at all, I think this is a pretty bad metric.


Silicon chips will never be able to generate a bound qualia space as we have.

Currently, you could prompt GPT to act as if it is sentient and has qualia, and it will do quite a good job at trying to convince you it's not a P-Zombie.


> Silicon chips will never be able to generate a bound qualia space as we have.

How do you know that?


Obviously we can't "know" this. My thinking is largely influenced by consciousness researcher and founder of Qualia Research Institute, Andrés Gómez Emilsson.

Here's a couple recent videos with him about why digital computers can't be sentient. https://www.youtube.com/watch?v=xJzBjBo24g8 https://www.youtube.com/watch?v=RT9tnzucnPU


How do you know silicon chips don't have an internal experience already? Are you in that "consciousness is magically emergent" camp?


I'm sure they do. It's just not coherent, hence prefacing my comment with bound. It's just random mind dust. I am a believer of a form of panpsychism. And that the hard problem of consciousness is better formulated as the boundary problem, whereby you get a conscious / intelligent being from properly creating boundaries around the consciousness that already exists.

Here's a good video detailing this line of thinking. https://www.youtube.com/watch?v=g0YID6XV-PQ


Therein lies the rub. Has anyone wired their models to have real-time data ingestion and the ability to output at will in a variety of mediums? Wake me when we’re there.


Because those were the real goal-posts all along, some of the best SF novels written all the way back in the ‘50s and ‘60s are testimony to that.


> Passing exams is a really fascinating benchmark but by their nature these exams are limited in scope, have very clear assessment criteria and a lot of associated and easily categorized data

I know I’m not the first to say this, but this is also a generalization of many jobs performed right now.

Follow the template, click the boxes, enter the text/data in the standard format, submit before 4pm. Come in tomorrow and do it again.


Humans are at their best correcting and finding errors in the integration between automated systems. Yes we probably won’t have accountants manually typing data from a page into a computer in the future, but we’ll always have people reviewing and checking the automation.

If that automation doesn’t require oversight, everyone wins, since now that process, typing data from a ledger, is free to anyone who wants to use it. The exception of course is if a monopoly or oligopoly controls the process, so it’s up to the government to break them up and keep the underlying tech accessible.

The biggest risk is how much computing power it takes to run these models, so it’s very important to support the open alternatives that are trying to lower the barrier to entry.


Peak denialism? Answering LSAT questions requires general intelligence. They present real life scenarios that test-taker has to understand. It requires "common sense" knowledge about the world and reasoning ability. It's not something you can memorize answers to or solve by following prescribed patterns or templates. And GPT-4 wasn't trained specifically to solve LSAT questions.


For the human brain, the LSAT requires reasoning. But not for an LLM. Do we even know exactly what data this is trained on? I have only seen vague references to what data they are using. If it is trained on large chunks of the internet, then it certainly is trained on LSAT practice questions. And because LSAT questions follow a common pattern, it is well suited to a LLM. There isn't any reasoning or general intelligence at all. Just really good statistics applied to large amounts of data.


> For the human brain, the LSAT requires reasoning. But not for an LLM.

Exactly, much like a chess bot can play perfectly without what humans would call thinking.

I think (ironically) we'll soon realize that there is no actual task that would require thinking as we know it.


This made me think of a Dijkstra quote

> The question of whether computers can think is like the question of whether submarines can swim

It has only become more relevant.


From the article: "We did no specific training for these exams. A minority of the problems in the exams were seen by the model during training, but we believe the results to be representative—see our technical report for details."


I’m skeptical. There is a lot wiggle room in “no specific training”. Could just mean the didn’t fine tune the model for any of tests. Their training data probably included many past LSAT exams and certainly included many instances of people discussing how to solve LSAT problems.


How is it different than humans preparing for LSAT by studying sample questions and reading explanations?


> It's not something you can memorize answers to or solve by following prescribed patterns or templates.

If that were true, there would be no point in studying or doing any LSAT preparation. Writing practice exams would be of no benefit.


Bingo. These are very 'human' tasks.

As others have said elsewhere, the issue remains accuracy. I wish every response comes with an accurate estimation of how true the answer is, because at the moment it gives wrong answers as confidently as right ones.


So the thing is, giving wrong answers with confidence is literally what we train students to do when they are unsure.

I can remember my GRE coach telling me that it was better to confidently choose an answer I only had 50% confidence in, rather than punt on the entire question.

AIs hallucinate because, statistically, it is 'rewarding' for them to do so. (In RLHF)


In the context of standardized testing, sure. I don't think I'd try that in a research paper.


This is literally in the context of standardized testing? GPT 'evals'?


> Answering LSAT questions requires general intelligence.

Obviously not, since GPT-4 doesn't have general intelligence. Likewise "common sense," "knowledge about the world," nor "reasoning ability."

As just one example, reasoning ability: GPT-4 failed at this problem I just came up with: "If Sarah was twice as old as Jimmy when Jimmy was 1/3 as old as Jane, and Jane is as much older than Sarah as Sarah is older than Jimmy, and Sarah is now 40, how old are Jane and Jimmy?"

First, every answer GPT-4 came up with contradicted the facts given: they were just wrong. But beyond that, it didn't recognize that there are many solutions to the problem. And later when I gave it an additional constraint to narrow it to one solution, it got the wrong answer again. And when I say "wrong," I mean that its answer clearly contradicted the facts given.


General thinking requires an AGI, which GPT-4 is not. But it can already have a major impact. Unlike self-driving cars which we require 99.999+% safety to be deployed widely, people already use the imperfect GPT-3 and ChatGPT for many productive tasks.

Driving as well as an attentive human in real time, in all conditions, probably requires AGI as well.

GPT-4 is not an AGI and GPT-5 might not be it yet. But the barriers toward it are getting thinner and thinner. Are we really ready for AGI in a plausibly-within-our-lifetime future?

Sam Altman wrote that AGI is a top potential explanation for the Fermi Paradox. If that were remotely true, we should be doing 10x-100x work on AI Alignment research.


Even just in the exam passing category, GPT4 showed no improvement over GPT3.5 on AP Language & Composition or AP English Literature, and scored quite poorly.

Now, granted, plenty of humans don't score above a 2 on those exams either. But I think it's indicative that there's still plenty of progress left to make before this technology is indistinguishable from magic.


The big huge difference is that cars have this unfortunate thing where if they crash, people get really hurt or killed, especially pedestrians. And split second response time matters, so it's hard for a human operator to just jump in. If ChatGPT-4 hallucinates an answer, it won't kill me. If a human needs to proofread the email it wrote before sending, it'll wait for seconds or minutes.


> If ChatGPT-4 hallucinates an answer, it won't kill me

Sure but look in this thread, there are already plenty of people citing the use of GPT in legal or medical fields. The danger is absolutely real if we march unthinkingly towards an AI-driven future.


Who is using ChatGPT in a medical field (serious question), knowing that it only displays very shallow level of knowledge on specific topic?


> If ChatGPT-4 hallucinates an answer, it won't kill me

Not yet it won't. It doesn't take much imagination to foresee where this kind of AI is used to inform legal or medical decisions.


Real human doctors kill people by making mistakes. Medical error is a non-trivial cause of deaths. An AI doctor only needs to be better than the average human doctor, isn't that what we always hear about self-driving cars?

And medicine is nothing but pattern matching. Symptoms -> diagnosis -> treatment.


Your last paragraph weakens the argument that you’re making.

Driving assistance and the progress made there and large language models and the progress made there are absolutely incomparable.

The general public’s hype in driving assistance is fueled mostly by the hype surrounding one car maker and its figurehead and it’s a hype that’s been fueled for a few years and become accepted in the public, reflected in the stock price of that car maker.

Large language models have not yet perpetrated the public’s memory yet, and, what’s actually the point is that inside of language you can find our human culture. And inside a large language model you have essentially the English language with its embeddings. It is real, it is big, it is powerful, it is respectable research.

There’s nothing in driving assistance that can be compared to LLMs. They don’t have an embedding of the entire physical surface of planet earth or understanding of driving physics. They’re nothing.


We detached this perfectly fine subthread from https://news.ycombinator.com/item?id=35154722 in an attempt to spare our poor server, which has smoke coming out of its ears today :( - sorry. We're still working on this and one day it will be better.


What might be interesting is to feed in the transcripts & filings from actual court cases and ask the LLM to write the judgement, then compare notes vs the actual judge.


Define: "general thinking".




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: