Hacker News
new
|
past
|
comments
|
ask
|
show
|
jobs
|
submit
login
regularfry
8 days ago
|
parent
|
context
|
favorite
| on:
Qwen3.6-27B: Flagship-Level Coding in a 27B Dense ...
To add more complexity to the picture, you can run MoE models at a
higher
quant than you might think, because CPU expert offload is less impactful than full layer offload for dense models.
help
Consider applying for YC's Summer 2026 batch! Applications are open till May 4
Guidelines
|
FAQ
|
Lists
|
API
|
Security
|
Legal
|
Apply to YC
|
Contact
Search: