Hacker Newsnew | past | comments | ask | show | jobs | submit | necroforest's commentslogin

yeah, and as far as I'm aware there isn't even a definition of "civilization" or "intelligence" that doesn't boil down to "sufficiently like me".


I'm with you on intelligence, but the hallmark of civilization is right in the etymology -- the existence of cities.


> hallmark of civilization is right in the etymology -- the existence of cities.

That is not really the etymology of "civilisation" though. City and civilisation share etymological roots, but city is not the etymological origin of the world civilisation.

And of course then we just ask the question: what are cities? Do gopher, or prairie dog colonies count? (or are those just towns? :)) How about ant colonies or bee hives?

Clearly all of the above share some similarities with some human settlements. They also have important differences of course. So if we want to decide if there are other "civilisations" on Earth parallel with us, we have to be more precise with our definitions.


> city is not the etymological origin of the world civilisation

Civilisations are a subset of societies [1]. Urbanisation is commonly held as a divider between complex societies and full-blown civilisations.

> then we just ask the question: what are cities?

This is valid. I’d say the defining attribute is economies of scale. Ant colonies and bee hives demonstrate elements of this; the sum is greater than the whole.

Whether ants and bees form complex societies is less debatable, unless we reduce the terms to mean intricate where we begin enveloping colonies of trees and every social animal, potentially even just multicellular life, which while poetically pleasing isn’t useful.

[1] https://en.m.wikipedia.org/wiki/Civilization


Well, all right, do ants form cities? Yes, they do, at least of a kind.

You could even argue that lichen is a city, inhabited by multiple species.

I mean, I'm not sure I'd go so far as to argue than either of those really are a civilization. But if "city" is your sole criterion...


Or a beehive. Bees moved into the basement window of my house and watching them go in and out it seemed to me that this single beehive had more departures and arrivals than all the commercial airports in the world put together.

If you look at the problem of "bee decline" from the viewpoint of the beekeeper where you are responsible for it you are responsible for a "city" of 50,000 insects that faces all kinds of threats from the inside and outside.


Ant nests and beehives aren't cities, they are households, consisting of a single family.


I don't entirely disagree with you, but "what products do people want" is overly conservative. Pre-ChatGPT, very few people wanted a (more or less) general purpose chatbot.


> Pre-ChatGPT, very few people wanted a (more or less) general purpose chatbot.

And post-ChatGPT, very few people want to have to deal with "a (more or less) general purpose chatbot."


One of my local car dealerships is using an chat system of some kind (probably an LLM?).

It's awful and a complete waste of time. I'm not sure if LLMs are getting good use yet / general chatbots are good or ready for business use.


But this is like companies forcing self-checkout at retail. Companies get to cheap out on their customer support by paying a monthly fee instead of paying a person to do it. It doesn't matter if it's worse for you, as every business is getting sold on using this technology.

You misery and wasted time is their improved stock price and bonus package.


This works if you have a dominant market position, like Meta or Amazon, but a business like a local car dealership might be very ill served with this model since their customers can just go elsewhere.


Some of these are just text matchers with hardcoded options, no ML involved. Essentially phone trees, except worse because they don't tell you your options.

There are some more advanced ones using ChatGPT now. I'm guessing they simply pre-prompt it. Can lead to funny results like a customer making the Chevy bot implement an algo in Python.


Who cares. I literally use ChatGPT 30 times a day. It answers incredibly complex queries along with citations I can verify. Isn’t “this not good enough yet” line getting old? There nothing else that can estimate the number of cinder blocks I need to use for a project and account for the volume of concrete required for it (while taking into consideration the actual available volume available in a cinder block and settling) with a few quick sentences I speak to it. I can think of literally thousands of things I have asked that would have taken hours of googling that I can get an answer for in minutes.

I think the problem is you haven’t shifted your mindset to using AI correctly yet.

Edit: More everyday examples from just the last 3 days

- Use carbide bits to drill into rocks. Googling “best bits for drilling rocks” doesn’t bring up anything obvious about carbide but it was the main thing chatGPT suggested.

- gave it dimensions for a barn I’m building and asked it how many gallons of paint I would need of a particular type. I could probably work that out myself but it’s a bunch of lookups (what’s the total sq footage, how many sq ft per gallon, what type of paint stands up to a lot of scuffing etc.)

- coarse threaded inserts for softwood when I asked it for threaded insert recommendations. I would have probably ended up not caring and fine threaded slips right out of pine.

- lookup ingredients in a face cream and list out any harms (with citations) for any of them.

- speeds and feeds for acrylic cutting for my particular CNC. Don’t use a downcut bit because it might cause a fire, something I didn’t consider.

- an explanation of relevant NEMA outlets. Something that’s very hard to figure out if you’re just dropped into it via googling.


>Who cares.

clearly anyone trying to buy a car, which is already an ordeal with a human as is.

>I literally use ChatGPT 30 times a day

good for you? I use Google. mos of my queries aren't complex.

>Isn’t “this not good enough yet” line getting old?

as long as companies pretend 2024 AI can replace skilled labor, no. It's getting old how many more snake oil salesmen keep pretending that I can just use ChatGPT to refactor this very hot loop of performance sensitive code. And no ChatGPT, I do not have the time budget (real time) to hook into some distributed load for that function.

I'm sure in a decade it will wow me. But I prefer to for it to stay in its lane and I stay in mine for that decade.

>There nothing else that can estimate the number of cinder blocks I need to use for a project

is Calculus really this insurmontable feat to be defending big tech over? I'm not a great mathmatican, but give them excel/sheets and they can do the same in minutes.

>I can think of literally thousands of things I have asked that would have taken hours of googling that I can get an answer for in minutes.

I'm glad it works out for you. I'm more scrutinous in my searches and I see that about half the time its sources are a bit off at best, and dangerously wrong at worst. 50/50 isn't worth any potential time saved for what I research.

>I think the problem is you haven’t shifted your mindset to using AI correctly yet.

perhaps. But for my line of work that's probably for the best.


"I think the problem is you haven’t shifted your mindset to using AI correctly yet"

There is an indictment of AI "products" if I ever heard one


That’s a glib, low effort dismissal but it makes sense if you consider it.

It’s like people that kept going to the library even with Google around. You’re not playing to the strengths of AI and relying on whatever suboptimal previous method you used to find the answers. It does really, really well with very specific queries with a lot of looks ups and dependencies that nothing else can really answer without a lot of work on your end.


How come this spoon won't hold my soup? Don't tell me I'm holding it wrong!


I mean if my dentist adds a Helpful Super GenAI Infused Chatbot that can't book appointments or answer any questions about their services no amount of "you're holding it wrong" insistence about LLMs in general will actually make it useful to me.

The point is ChatGPT's wild success doesn't automatically mean consumers want and possibly will never want a chatbot as their primary interface for your specific app or service.


I feel like I'm being too obvious but maybe try using it for something it's good at.


> with citations I can verify.

And do you? Every time someone tried to show me examples of “how amazing ChatGPT is at reasoning”, the answers had glaring mistakes. It would be funny if it weren’t so sad how it shows people turning off their critical thinking when using LLMs, to the point they won’t even verify answers when trying to make a point.

Here’s a small recent example of failure: I asked the “state of the art” ChatGPT model which Monty Python members have been knighted (it wasn’t a trick question, I really wanted to know). It answered Michael Palin and Terry Gilliam, and that they had been knighted for X, Y, and Z (I don’t recall the exact reasons). Then I verified the answer on the BBC, Wikipedia, and a few others, and determined only Michael Palin has been knighted, and those weren’t even the reasons.

Just for kicks, I then said I didn’t think Michael Palin had been knighted. It promptly apologised, told me I was right, and that only Terry Gilliam had been knighted. Worse than useless.


I do. It’s not complex to click on the citation, skim the abstract and results and check the reputation of the publication. It’s built into how I have always searched for information.

I also usually follow most prompts with “look it up I want accurate information”


> I also usually follow most prompts with “look it up I want accurate information”

That didn’t work so hot for two lawyers in the news a while back.


A while back when the lawyers used it, chatGPT didn’t do lookups and citation links.


Please don't take this as a glib, low effort answer, but... I am glad you're not an engineer. Taking advice from an LLM on things like outlets, ingredient safety, and construction measurements seems like a mistake.


Did you do the cinder block project? Was its estimate close? From everything I’ve seen LLMs are not that great at math.


Yes, I finished the footings for the barn. I had to get two extra bags on an estimate of 68 bags. Not bad in my opinion considering the wastage from mixing, spilling etc. Also I would have had to do a bunch of tedious math that I didn’t have to.

I had about 5-10 cinder blocks left over, not bad for an order of ~150


It works for me, therefore it should work for everyone?


It’s pretty much the vein of the comment of the GP I’m responding to.


>I'm not sure if LLMs are getting good use yet / general chatbots are good or ready for business use.

They left room for the idea that the technology could evolve to be useful. You're simply dismissing anyone who cannot use he technology as is as "using it wrong".

As someone who did a tad of UX, that's pretty much the worst thing you can say to a tester. it doesn't help them understand your POV, it builds animosity towards you and the tester, and you're ruining the idea of the test because you are not going to be there to say "you're doing it wrong" when the UX releases. There's 0 upsides to making such a response.


This isn't what the other comment you're replying to was talking about.


Not sure if this is a sarcasm or "you are holding it wrong" moment.


There’s skill involved in using these. Of course there is.

That doesn’t make them useless.


depends on the tool and purpose. There was skill in navigating a file system, and now the next generation (obbscured from folder systems by mobile) seem to be losing that ability.

You can look at it in two ways, neither are particularly wrong unless your job is in fact to navigate file systems.


Of course, LLMs would be more useful to many more people if they could be used without skill, and were as "useful" as a human mentor.

That's true, and they lack that capability. Many people seem to react as though this means they're missing all value, however. I find them incredibly useful; it just isn't possible to get much value out without investing effort myself.


>LLMs would be more useful to many more people if they could be used without skill, and were as "useful" as a human mentor.

That is partially marketing's fault, so I say that confusion is self inflicted. Because marketing isn't focusing on "make yourself more productive!"


If you’re unaware of how something can help you, learning how can only improve your outcome.


I don't think so, people have been wanting a general chatbot for a long time. It's useful for plenty of things, just not useful when embedded in random places.


I kind of remember the Turing Test was a big deal for some 70+ years.

We should have known that once we pass the Turing Test it would almost instantly become as passe as Deep Blue beating Kasparov on the road to general intelligence.

I am taking a break from my LLM subscriptions right now for the first time to gain some perspective and all I miss it for is as a code assistant. I would also miss it for learning another human language. It seems unsurprising that large language models use cases are with automated language. What is really surprising is how very limited the use cases for automated language seems to be.


Stanford researchers said that ChatGPT passes the Turing Test. Honestly I don't understand how, since it's pretty easy to tell that you're talking to it, but yeah I don't think it really matters.

Far more useful than simulating a person, OpenAI managed to index so much information and train their models to present it in a compact way, making ChatGPT better than Google search for some purposes. Also, code generation.


To be fair, the overwhelming feedback appears to be that people dont want a general purpose chatbot in every product and website, especially when it's labelled 'AI'.

So... certainly there's a space for new products.

...but perhaps for existing products, it's not as simple as 'slap some random AI on it and hope you ride the wave of AI'.


Chatbots were HUGE in the late 2010s!


And they hugely sucked, and basically were a sign you were dealing with either a fly by night company or a corporation so faceless and shitty you'd never willingly do business with them.

It was literally replacing a hierarchical link tree and that almost always was easier to use.


Or you were told this one chatbot was “haunted” by a creepypasta character and was amazed that it would actually reference that story (not knowing it was just a silly program trained on its interactions, leading to a feedback loop).


>very few people wanted a (more or less) general purpose chatbot.

I mean, I still don't. But from a cynical business point of view, cutting customer servce costs (something virtually every company of scale has) of 99% of customer calls is a very obvious application of a genera purpose chatbot.

expand that to "better search engine" and "better autocomplete" and you already have very efficient, practical, and valuable tools to sell. but of course companies took the angle of "this can replace all labor" instead of offering these as assistive productivity tools.


It's branding (see: TensorFlow); also, pretty much anything (linear) you would do with an arbitrarily ranked tensor can be expressed in terms of vector ops and matmuls


"Fixed-Function Matrix Accelerator" just doesn't have the same buzzy ring to it.


It’s the perfect name for my next EDM SoundCloud mix, though.


> "Fixed-Function Matrix Accelerator" just doesn't have the same buzzy ring to it.

FixMax™ or Maxxelerator™ would be good brands.


cosine similarity is (isomorphic to) "distances to nearby objects". and not all embeddings are word embeddings.


It's isomorphic when vectors are normalized, otherwise it's angle distance, not position distance


It’s a mistake to think of vectors as coordinates of objects in space, though. You can visualize them like that, but that’s not what they are. The vectors are the objects.

A vector is just a list of n numbers. Embedded into a n dimensional space, a vector is a distance in a direction. It isn’t ’the point you get to by going that distance in that direction from the origin of that space’. You don’t need as space to have an origin for the embedding to make sense - for ‘cosine similarity’ to make sense.

Cosine similarity is just ‘how similar is the direction these vectors point in’.

The geometric intuition of ‘angle between’ actually does a disservice here when we are talking about high dimensional vectors. We’re talking about things that are much more similar to functions than spatial vectors, and while you can readily talk about the ‘normalized dot product’ of two functions it’s much less reasonable to talk about the ‘cosine similarity’ between them - it just turns out that mathematically those are equivalent.


Fair enough.

I think people skip over that the vectors are the result of the minimization of the objective.

That objective is roughly the same since word2vec. GLoVe is mathematically equivalent. LLMs are also equivalent.

For a LM, the objective function is still roughly the same. Maximizing probability of the next token conditional on previous tokens.

This means the embedding vector of a token minimizes distance to tokens that come before it often, and maximizes distance to those that don't.


If you need to relearn your iPhone/Mac interfaces every year, you might want to get screened for Alzheimers.


And you might want to avoid reading comments too literally :)


> Who doesn’t think China has a ten year AI algorithm to takeover Taiwan?

anybody who works in either AI or natsec


> We arent going to see more progress until we have a way to generalize the compute graph as a learnable parameter

That's a bold statement since a ton of progress has been made without learning the compute graph.


From my naive perspective, there seems to be a plateau, that everyone is converging on, somewhere between ChatGPT 3.5 and 4 level of performance, with some suspecting that the implementation of 4 might involve several expert models, which would already be extra sauce, external to the LLM. This, combined with the observation that generative models converge to the same output, given the same training data, regardless of architecture (having trouble finding the link, it was posted here some weeks ago), external secret sauce, outside the model, might be where the near term gains are.

I suppose we'll see in the next year!


We already have competitors to Transformers

https://arxiv.org/abs/2312.00752


Where do I enter in my credit card info?


You hire people to implement a product based on this?


A ton of progress can be made climbing a tree, but if your goal is reaching the moon it becomes clear pretty quickly that climbing taller trees will never get you there.


True, but it is the process of climbing trees that gives the insight whether taller trees help or not and if not, what to do next.


Not true. Climbing trees for millions of years taught us nothing about orbits, or rockets, or literally incomprehensible to human distances, or the vacuum of space, or any possible way to get higher than a tree.

We eventually moved on to lighter than air flight, which once again did not teach us any of those things and also was a dead end from the "get to the sky/moon" perspective, so then we invented heavier than air flight, which once again could not teach us about orbits, rockets, distances, or the vacuum of space.

What got us to the moon was rigorous analysis of reality with math to discover Newton's laws of motion, from which you can derive rockets, orbits, the insane scale of space, etc. No amount of further progress in planes, airships, kites, birds, anything on earth would ever have taught us the techniques to get to the moon. We had to analyze the form and nature of reality itself and derive an internally consistent model of that physical reality in order to understand anything about doing space.


> Climbing trees for millions of years taught us nothing about

Considering the chasm in the number of neurons between the apes and most other animals, I think one could claim that climbing those trees had some contribution to the ability to understand those things. ;) Navigating trees, at weight and speed, has a minimum intelligence reqiurement.


With enough thrust, even p̵i̵g̵s̵ trees can fly.


We have made progress in efficiency, not functionality. Instead of searching google or stack overflow or any particular documentation, we just go to Chatgpt.

Information compression is cool, but I want actual AI.


The idea that there has been no progress in functionality is silly.

Your whole brain might just be doing "information compression" by that analogy. An LLM is sort of learning concepts. Even Word2Vec "learned" than king - male + female = queen and that's a small model that's really just one part (not exact, but similar) of a transformer.


Let me rephrase that.

One level deep information compression is cool, but I want actual AI.

Its true that our brains compress information, but we compress it in a much more complex manner, in the sense that we can not only recall stuff, but also execute a decision tree that often involves physical actions to find the answer we are looking for.


An LLM isn't just recalling stuff. Brand new stuff, which it never saw in it's training, can come out.

The minute you take a token and turn it into an embedding, then start changing the numbers in that embedding based on other embeddings and learned weights, you are playing around with concepts.

As for executing a decision tree, ReAct or Tree of Thought or Graph of Thought is doing that. It might not be doing it as well as a human does, on certain tasks, but it's pretty darn amazing.


>Brand new stuff, which it never saw in it's training, can come out.

Sort of. You can get LLMs to produce some new things, but these are statistical averages of existing information. Its kinda like a static "knowledge tree", where it can do some interpolation, but even then, its interpolation based on statistically occurring text.


The interpolation isn't really based on statistically occurring text. It's based on statistically occurring concepts. A single token can have many meanings depending on context and many tokens can represent a concept depending on context. A (good) LLM is capturing that.


Neither just text or just concepts, but text-concepts — LLMs can only manipulate concepts as they can be conveyed via text. But I think wordlessly, in pure concepts and sense-images, and serialize my thoughts to text. That I have thoughts that I am incapable of verbalizing is what makes me different from an LLM - and, I would argue, actually capable of conceptual synthesis. I have been told some people think “in words” though.


Nope, you could shove in an embedding that didn't represent an existing token. It would work just fine.

(if not obvious.. you'd shove it in right after the embedding layer...)


Fascinating. What’s “actual AI”?


> What’s “actual AI”

Is Ibn Sina (Avicenna, year ~1000) fine?

> [the higher faculty proper of humans is] the primary function of a natural body possessing organs in so far as it commits acts of rational choice and deduction through opinion; and in so far as it perceives universal matters

Or, "Intelligence is the ability to reason, determining concepts".

(And a proper artificial such thing is something that does it well.)


It is a tool that has the ability to craft a prompt that will break current state of the art model.

It is a tool that can be given a project in language X and produce an idomatic port in language Y.

It is a tool that given a 20 pages paper spec will ask the questions needed to clarify the specs.


Something that can reason and figure things out without having ever been exposed to the information during training.


This either includes GPT-4 or excludes people


It’s whatever computers can’t do, dummy! :P


which laws are broken exactly? it's not remotely settled law that "training an NN = copyright infringement"


The training isn't the issue per se, it's the regurgitation of verbatim text (or close enough to be immediately identifiable) within a for-profit product. Worse still that the regurgitation is done without attribution.


The legal argument, which I'm sure you are very well aware of, is that training a model on data, reorganizing, and then presenting that data as your own is copyright infringement.


I don't think OP is arguing in bad faith.The fact is it's unclear what laws this legal argument is supported by.


Agreed, it is unclear. It's also a very commonly discussed issue with generative AI and there's been a significant amount of buzz around this. Is the NYT testing the legal waters? Maybe. Will this case set precedent? Yes. Is this a silly, random, completely unhinged case to bring?

No.


Can you elaborate a bit more? That’s actually just a claim, not a legal argument.

Copyright law allows for transformative uses that add something new, with a further purpose or different character, and do not substitute for the original use of the work. Are LLM’s not transformative?


Right, hence the lawsuit. They allege that the Copyright Act is the law that was broken.


No, we’re seeing the first steps of it (maybe) becoming settled law.


What's an order of magnitude between friends?


Maybe "best word processor for mac" is a better search for you?


Consider applying for YC's Winter 2026 batch! Applications are open till Nov 10

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: