Hacker News
new
|
past
|
comments
|
ask
|
show
|
jobs
|
submit
login
jasonjmcghee
1 day ago
|
parent
|
context
|
favorite
| on:
Lossless LLM compression for efficient GPU inferen...
I read it similarly - that this is a specific attribute of bfloat16, so the quants folks tend to run on local hardware don't have the same inefficiency to exploit
Join us for
AI Startup School
this June 16-17 in San Francisco!
Guidelines
|
FAQ
|
Lists
|
API
|
Security
|
Legal
|
Apply to YC
|
Contact
Search: