A bag of tricks to increase either training or inference latency or memory and storage requirements for large language models.
undefined subscriptions will be displayed on your profile (edit)
Skip for now
For your security, we need to re-authenticate you.
Click the link we sent to , or click here to sign in.
Share this post
A Hacker's Guide to LLM Optimization
Share this post
A bag of tricks to increase either training or inference latency or memory and storage requirements for large language models.