Its not completely true. Check their popular gas models like telluride etc, you still have to pay above MSRP or some crap accessories to get the cars. They are selling gas cars extremely well.
I agree with the comment. Especially if you index these quantized vectors with HNSW you will get even less precision. You need to retrieve lot more and do a kNN with the FP32 vectors is the way to go !
Its IMHO too. His contribution to educational content is incredible, and very few individuals have the ability to explain things the way he does. However, I am also unsure about his contribution to the field itself. It is a side effect of working in the industry on the product side. You don't have a chance to publish papers, and you don't want to reveal your secrets or bugs to everyone.
Some how my experience doesn't reflect this. I keep going back to GPT if i need an answer. I even say most of the time I prefer 3.5 better than Google Bard. GPT 4 is definitely way better than Bard for sure.
My boss had a KYC experience that involved uploading a short selfie video in which he had to turn his head/blink, smile, etc. This was recently, and in addition to the "photo of yourself holding a document with your photo in it" check.
I don't know enough about GenAI or the KYC process to say whether or not that could be faked, but I'm confident saying that with only traditional tools it's a barrier that would stop all but the most sophisticated actors.
Recently, I noticed that Amazon removed the listed ingredients for regular items like toothpaste, shampoos, and mouthwashes. Initially, I thought it was a mistake for a couple of items I regularly buy, but it seems they have removed this information for many products. I'm not sure what the intention behind this change is. My suspicion now (after this article) is that they are doing it deliberately, allowing them to claim that the page doesn't list ingredients for any item.
The issue of hallucinations won't be solved with the RAG approach. It requires a fundamentally different architecture. These aren't my words but Yann LeCun's. You could easily understand if you spend some time playing around. The autoregressive nature won't allow the LLMs to create an internally consistent model before answering the question. We have approaches like Chain of Thought and others, but they are merely band-aids and superficially address the issue.
If you build a complex Chain if Thought style Agent and then train/finetune further by reinforcement learning with this architecture then it is not a band-aid anymore, it is an integral part of the model and the weights will optimize to make use of this CoT ability.
It's been 3.5 years since GPT-3 was released, and just over a year since ChatGPT was released to the public.
If it was possible to solve LLM hallucinations with simple Chain-of-Thought style agents, someone would have done that and released a product by now.
The fact that nobody has released such a product, is pretty strong evidence that you can't fix hallucinations via Chain-of-Thought or Retrieval-Augmented Generation, or any other band-aid approaches.
I agree: but I just wanted to say that there are specific subdomains where you can mitigate some of these issues.
For example, generating json.
You can explicitly follow a defined grammar to get what will always be a valid json output.
Similarly, structured output such as code can be passed to other tools such as compilers, type checkers and test suites to ensure that at a minimum the output you selected passes some minimum threshold of “isn’t total rubbish”.
For unstructured output this a much harder problem, and bluntly, it doesn’t seem like there’s any kind of meaningful solution to it.
…but the current generation of LLMs are driven by probabilistic sampling functions.
Over the probability curve you’ll always get some rubbish, but if you sample many times for structure and verifiable output you can, to a reasonable degree, mitigate the impact that hallucinations have.
Currently that’s computationally expensive, to drive the chance of error down to a useful level, but compute scales.
We may seem some quite reasonable outputs from similar architectures wrapped in validation frameworks in the future, I guess.
I agree that the "forcing valid json output" is super cool.
But it's unrelated to the problem of LLM hallucinations. A hallucination that's been validated as correct json is still a hallucination.
And if your problem space is simple enough that you can validate the output of an LLM well enough to prove it's free of hallucinations, then your problem space doesn't need an LLM to solve it.
> your problem space doesn’t need an LLM to solve it
Hmmm… kinda opinion right?
I’m saying; in specific situations, you can validate the output and aggregate solutions based on deterministic criteria to mitigate hallucinations.
You can use statistical methods (eg. There’s a project out there that generates tests and uses “on average tests pass” as a validation criteria) to reduce the chance of an output hallucination to probability threshold that you’re prepared to accept… for certain types of problems.
That the problem space is trivial or not … that’s your opinion, right?
It has no bearing on the correctness of what I said.
There’s no specific reason to expect that just like you can validate output against a grammar to require output that is structurally correct, you can’t validate output against some logical criteria (eg. unit tests) to require output that is logically correct against the specified criteria.
It’s not particularly controversial.
Maybe the output isn’t perfectly correct if you don’t have good verification steps for your task, maybe the effort required to build those validators is high, I’m just saying: it is possible.
>>>The constant use of the battery pack is unprecedented and is causing reliability problems for Tesla and other EV truck makers [similar claims coming from Uber drivers who use their Teslas over 300 miles per day, i.e., their batteries are dying quickly.
Meaning the batteries are not meant to use more than few hours per day because they heat up ?
>>>PepsiCo did a 500-mile trip with the Tesla Semi from California to Phoenix, but it was just for PR purposes. The batteries completely burned out, which is why on PR trips like this, they bring three Tesla Semis, with two being towed on a diesel Semi truck, only to be swapped out when the battery dies the other two Semis on the 500-mile drive
> Meaning the batteries are not meant to use more than few hours per day because they heat up ?
TBH, it sounds like it might be temp related... but it could be as simple as overly conservative software. These are in such early stages of development, that I'm sure they are being very conservative about going into limp mode at the first sign of a potential problem.