Hacker News
new
|
past
|
comments
|
ask
|
show
|
jobs
|
submit
login
mountainriver
3 days ago
|
parent
|
context
|
favorite
| on:
Lossless LLM compression for efficient GPU inferen...
Is it possible to run this on new models? It seem like the code is only for inference, unless I’m misunderstanding
Join us for
AI Startup School
this June 16-17 in San Francisco!
Guidelines
|
FAQ
|
Lists
|
API
|
Security
|
Legal
|
Apply to YC
|
Contact
Search: